Why the multiple ECM Repository/Silo model is not a good idea

by Frank 15. November 2016 06:00

“43 Reasons why Managing Records in-Place may not be good enough”

Enterprise Content Management is a moving target, constantly evolving with new challenges and new paradigms. For example, how do we filter out only relevant information from social media? How do we avoid capturing personal data and being culpable under privacy laws? How do we capture all emails containing sexism, racism and bullying without being guilty of an invasion of privacy of the individual? How do we meet all of our compliance obligations when our staff are spread across multiple states/counties/provinces and multiple countries with different legislation and compliance requirements? All weighty challenges for the modern Knowledge Manager or CIO.

Another interesting challenge for Knowledge Managers and CIOs is the newer document management paradigm of being asked to manage all content without a single central repository. That is, to be responsible for all content across a myriad of locations controlled by a myriad of applications and a myriad of departments/organizations and people. Back when I was an employee and not an employer, my tough (ex-military) manager in Blue Bell, PA would just bang his fist on his desk and say, “Goddam Frank, just do it!” That was always a signal for me to get creative.

However, try as I may, I am finding it nigh on impossible to get creative enough to work out how I could effectively and reliably manage all content across an enterprise without a single central repository.

In multiple-repository systems we find multiple document stores; local files, network file shares, local data bases, multiple file servers, multiple copies of SharePoint and multiple Cloud repositories like Dropbox, Box, iCloud, Google Cloud Storage and other hosted document storage. The CIO may proudly claim to manage multiple information silos but what he or she really has is a laissez faire document management ecosystem that may well be centrally monitored (hopefully) but is most certainly not centrally managed.

In the multiple silo model the documents in our multiple locations are ‘managed’ by multiple people and multiple applications (e.g., SharePoint, Google Docs, etc.). We may have implemented another layer of software above all these diverse applications trying to keep up with what is happening but If I am just ‘watching’ then I don’t have an inviolate copy and I don’t have any control over what happens to the document. I am unable to enforce any standards. There is no ‘standard’ central control over versioning or retention and no control over the document life cycle or chain of evidence.

For example, you wouldn’t know if the document had since been moved to a different location that you are not monitoring. You wouldn’t know if it had been deleted. You wouldn’t know its relationship to other documents and processes in other silos. You wouldn’t know its context in your enterprise and therefore you wouldn’t know how relevant this document was. The important distinction is that under the multiple silo model you are ‘watching’ not managing; other software is managing the life-cycle and disposition of the document.

All you really know is that at a certain point in time a document existed and what its properties were at that time (e.g., historical ‘natural’ Metadata such as original filename, author, date created, etc.). However, you have no contextual Metadata, no transactional Metadata, no common indexing and no common Business Classification System. In this case, you don’t have a document management system, you have a laissez faire document management ecosystem, an assortment of independently ‘managed’ information silos. Most importantly, you are not able to link documents to business processes that transcend organizational structures and silos.

Sure, SharePoint and Cloud silos make collaboration easier but at what cost? What can’t we do with this multi-silo ecosystem? Why doesn’t this solution meet the best-practice objectives of a document management system? What are the major areas where it falls short? How does the proliferation of multiple silos and content repositories affect us? What are our risks? Here is my assessment of the major shortfalls of this paradigm.

 We are unable to:

1.    extract the critical insights that enterprise information should provide

2.    define all the relationships that link documents to enterprise business processes

3.    find the right information at the right time

4.    provide a single access point for all content

5.    Implement an effective, consistent enterprise-wide document security system

6.    effectively protect against natural or man-made disasters

7.    produce evidence-standard documents

8.    minimize document handling costs

9.    guarantee the integrity of a document

10.guarantee that a document is in fact the most recent version

11.guarantee that a document is not an older copy

12.minimize duplicate and redundant information

13.meet critical compliance targets like Sarbanes-Oxley Act (SOX) and the HIPAA

14.create secure, searchable archives for digital content

15.effectively secure all documents against loss

16.implement common enterprise version control

17.facilitate enterprise collaboration

18.Improve timeliness

19.manage enterprise document security and control

20.manage smaller and more reliable backups

21.achieve the lowest possible document management and archiving costs

22.deliver the best possible knowledge management access and search

23.guarantee consistent content

24.optimize management and executive time

25.standardize the types of documents and other content can be created within an organization.

26.define common use template to use for each type of document.

27.standardize the Metadata required for each type of document.

28.standardize where to store a document at each stage of its life cycle.

29.control access to a document at each stage of its life cycle.

30.move documents within the organization as team members contribute to the documents' creation, review, approval, publication, and disposition.

31.implement a common set of policies that apply to documents so that document-related actions are audited, documents are retained or disposed of properly, and content that is important to the organization is protected.

32.manage when and if a document has to be converted from one format to another as it moves through the stages of its life cycle.

33.guarantee that all documents are treated as corporate records, that common retention policies are applied determining which documents must be retained according to legal requirements and corporate guidelines.

34.guarantee enterprise-wide Regulatory compliance

35.produce an enterprise-wide audit trail

36.share information across departmental and/or silo boundaries

37.centrally manage the security access to documents/information across different areas of the organization.

38.consistently classify documents as each repository may be used by a different department and be classified differently.  

39.identify duplicates based on document name.

40.easily find things based on metadata, as it wouldn’t be common across repositories.

41.control access via AD single sign on

42.access all enterprise documents using a single license.

          43.centrally audit access and changes to metadata.

What are your risks?  Your risks are huge!

 

 

 

 

 

Totally Automatic, Rules-Driven Email Management & Archiving

by Frank 15. December 2015 06:00

 

More than thirty years after the advent of email as a convenient and fast means of business to business communication most organizations still don’t have an effective way to analyze, monitor, select, capture and classify emails. If your organization does, then you are the exception.

This means most businesses don’t come even close to meeting the requirements of any compliance legislation that applies to their business. Nor do they even come close to managing real corporate risk.

It also means that most businesses don’t effectively guard against sexism, racism, obscenity, theft and bullying in their email system.

Is this a case of ‘heads in the sand’ or is the problem seen as just too hard? Maybe, senior management doesn’t really see unmanaged and unmonitored emails as a problem. Well, at least until the first court case.

In my experience, many organizations think they have a solution but in reality, they don’t; at least not a one hundred-percent solution. In the case of email management, ‘good enough’ is certainly not good enough. It only takes one bad email to slip through the cracks to bring the whole house down.

In many examples, organizations rely on end users to monitor, manage and police email. The problem with this model is that end users are human and typically exhibit all the strengths and weaknesses of humans. In this case, we are more concerned with the failings. As humans we are not always on top of our game; we have good days and we have bad days. We get distracted, we are prejudiced, we are sometimes lazy, we are sometimes careless and for a small number, we are sometimes outright dishonest.

Human beings will never produce a one-hundred-percent consistent result; that is not in our nature. Maybe when AI gets to the stage that we can all be ‘upgraded’ to cyborgs this will change, but I seriously doubt it. As long as there is any trace of humanity we will still be lovingly unreliable and inconsistent entities.

You don’t have a one hundred-percent reliable system if you don’t control and standardize all the inputs. Instead, you have a form of ‘managed chaos’ and inconsistent and unreliable results. You will also probably have a false sense of security, “Sure, we are managing all emails (well, kind of).”

I have been a proponent of the fully-automatic, server-centric paradigm for email management for many years and still promote it as the only one hundred-percent reliable way to effectively and consistently manage all incoming and outgoing emails. It is also the only way to manage risk effectively.

To be one hundred-percent sure you must have a one hundred-percent consistent paradigm. That is, a common set of rules that all emails are judged against plus a common set of processes to apply after an email has been ‘judged’.

Following are some example of what our fully-automatic, rules-driven email management system should be doing 24/7:

  • Is the email of a personal nature and harmless? If so, it can be ignored, there is no reason to capture and classify it.
  • Is the email all about business? If so, it needs to be captured and correctly classified within our corporate store.
  • Does the email contain expletives or sexual references? If so, it needs to be captured, quarantined by our security system and referred to a responsible officer for further examination and possible action.
  • Does the email contain references to corporate IP or classified material? If so, it needs to be captured, quarantined by our security system and referred to a responsible officer for further examination and possible action.
  • Is the email about business and does it require some action or response? If so, it needs to be captured, correctly classified within our corporate store and appropriate workflow initiated.
  • Is the email from a senior executive, about business and does it require some action or response with appropriate access controls applied? If so, it needs to be captured, correctly classified within our corporate store with appropriate security and access rights assigned and appropriate workflow initiated.

We produced our first fully-automatic, rules-driven email managements system in 1994. By today’s standards the technology was primitive but it worked and we used it within our business to demonstrate to our customers and partners how it could function in the real world. We called that product GEM for ‘GMB’s Email Management’ system. 

GEM has been redesigned and rewritten multiple times since so as to utilize the latest technology and tools. We still call it GEM even though our company is now called Knowledgeone Corporation, not GMB; the name works for us and our customers and we see no need to change it. It is after all, a ‘gem’ of a product.

We have used each and every version of GEM since 1994 within our company (we are the primary Beta test site) to automatically analyze all incoming and outgoing emails and to store and classify captured emails in our corporate store based on the RecFind 6 relational database. I can’t imagine running my business without GEM and I don’t understand how other organizations can exist without GEM, but they do albeit, taking huge risks.

The latest version of GEM, 2.7.1, is a major upgrade and involves a significant change in the way we connect to email servers of any type (e.g., Exchange, Office 365, GroupWise, Notes, etc.). We have standardized and simplified the interface to the email server using IMAP and converted our Agents to Windows Services to make the installation and management of GEM as easy as possible for your IT staff.

We have also improved the Rules engine to make it as easy as possible to define all the rules you need to manage your emails.

Because we have 21 years’ experience installing, configuring and using GEM in a real-world production example we also have the world’s most experienced GEM consultants to assist our customers.

How GEM integrates with any other EDRMS

GEM is designed to use the RecFind 6 relational database as its image & data repository. However, we provide several options for integrating to any other EDRMS such that the other EDRMS can search for and access emails (and the associated Metadata) captured by GEM. The four main methods, in order of ease-of-use are:

  1. Capturing encapsulated XML records produced by GEM;
  2. Using the RecFind 6 Mini API;
  3. Using the RecFind 6 SharePoint Integration Module (for customers using SharePoint as their EDRMS); and
  4. Using the RecFind 6 SDK

Please contact Support at Knowledgeone Corp for more information on the above methods.

The differences between a Classification System & an Information Management System

by Frank 5. November 2015 06:00

 

We have a large number of records and document management customers using our product RecFind 6 and with new customers the question always arises about how to best organize information in the RecFind 6 database. As the Metadata and business processes in RecFind 6 are 100% configurable, every customer ends up with a unique configuration.

Some records managers want the shared drives structure replicated in the database. Some want everything filed under a strict hierarchical classification system or Taxonomy. Some customers want the whole process simplified so end users clearly know where to file stuff and where to find stuff. Different managers in a single customer site will often disagree about how the information should be managed. Usually, the IT manager disagrees with the records manager and it is up to us to come up with an agreed and workable compromise; no easy task! There is no “one size fits all” paradigm here. We have grown to accept these discussions as part of every new installation.

Whereas I fully support the principles behind most EDRMS standards as espoused and recommended or even mandated by records management consultants I also find myself agreeing with most end users who just want the whole process simplified and expressed in natural language, not as an arcane, complex, inconsistent and difficult to navigate hierarchical classification system.

To wit, the way you classify information should not dictate how you store, manage and retrieve information.

I have written a paper of this exact subject and although it was in 2009 it is still 100% relevant. Please see this link Do You Really Need a Taxonomy? You don’t have to agree with me but please try to understand the message. End users want easy, fast access, not time-consuming complexity.

Maybe I should begin by telling you how we solve the problem at Knowledgeone Corporation and manage our emails, electronic documents and shared drives with a hybrid system. That is, a combination of RecFind 6 and shared drives. This is also a model we regularly recommend to our customers as an acceptable compromise; one that is simple to implement and one that always works.

I am obviously a big fan of making information as easy as possible to capture and as easy as possible to retrieve. This is especially important to the long-suffering end-user class who have no interest in becoming part-time records managers and who simply won’t use a system if it is too difficult to use and too time-consuming.

End users want direct access to information in the easiest and most timely fashion possible, they do not want to go through a third party or ‘information broker’. This means we need to have both a simple search system as well as a security system that ensures people only see what they are supposed to see.

And of course, the biggest problem with complex, hierarchical classification systems is that no two people file the same way and even a single user will often file things differently over time. This in itself makes the act of finding something by browsing through a classification hierarchy a hit and miss affair.

At Knowledgeone Corporation, we implemented a hybrid model that uses a simply structured shared drive resource plus automated tools to ensure everything that should be captured is captured. This approach is also all about separating the functionality of the Authoring packages (e.g., Word, Excel, Outlook, etc.) from the functionality of the EDRMS. They have different roles to play.

Let’s dispense with the notion that shared drives are evil just as we should dispense with the notion that paper is evil. Each has a part to play in a well management information management system

We use our product GEM to automatically capture all work related emails and we use our product RecCapture to automatically capture all work-related electronic documents from our shared drives. We all use a common shared drive structure to write and store our original electronic documents. Note that we do not use the feature in the RecFind 6 Button to force all ‘Saves’ into RecFind 6. We have this feature because the industry dictates it should be there but it is not popular and most customers never turn it on. Not everything you write should go into RecFind 6 and not everything you write is ready to go into RecFind 6 (though we do have a special ‘draft’ type for those customers that want drafts stored in RecFind 6).

We don’t use what you would call a formal taxonomy, we use what I call a ‘natural’ classification system. For us this means a classification system that perfectly reflects our business practices, processes and vocabulary. In our case, we are customer-centric so everything (apart from a little administrative and supplier stuff) is organized in customer or prospect folders and the lower levels are minimal, being things like Correspondence, Quotes and Orders.

Our RecFind 6 database is mostly based on customer and prospect files; it is our CRM. Customers and prospective customers are our core business just as members and cases are the core business of unions. Every industry has a core business and in my mind this should always be reflected in the classification system used so that it perfectly aligns with the work practices and processes and ‘language’ of most staff. Whenever I consult to a new organization I always try to first determine its core business and its natural language and then design the implementation around these.

We also use RecFind 6 to run our business so it is also our asset management system, our help desk and incident system, our project management system and our R&D development system. For these applications and others that we have implemented in RecFind 6, we have nothing outside of RecFind 6 to capture because all relevant information (e.g., customer support calls, details of meetings, phone calls, quotes, orders, annual leave request, etc.) are entered directly into RecFind 6 by our staff or captured automatically. RecFind 6 is our company repository and the source of all knowledge for my staff.

Because we are customer centric I need to be able to see everything about any customer or prospect in one place. For us this means centralizing on the Entity record (the Entity table is where we store the basic information on each customer or prospect). As RecFind 6 is a relational database we then store all related information in linked tables, all linked to and accessible from the Entity record with a single click.

In our RecFind 6 system, every piece of information I need to refer to is just one-click away once I view the entity record. I can also find any customer’s record instantly in RecFind 6 just by entering the customer number or a part of the organization name. Once I select the customer record, everything thing else I need to know is just one-click away and all links are viewable in a single screen. We are a customer-centric business and our RecFind 6 database is therefore organized as customer centric.

In practice, if someone at Knowledgeone Corporation wants to find something they always look first in RecFind 6 because it is a lot easier and faster than trying to search the shared drives or Outlook. Because we use automated tools (GEM and RecCapture) we are confident that everything that should be captured is captured. We don’t rely on our already too busy staff to remember to capture every important email or electronic document; it is done for them. All they have to do is search and create. Plus most of our information is stored behind customer/prospect/partner numbers in the Entity table so all information is both easy to browse and search (Text, Metadata, BOOLEAN, Saved Searches, etc.).

As a backup, every staff member has the Button installed (in Word, Excel, PowerPoint and Adobe Professional) but they rarely use it.

We have a security system configured around our management structure that works fine for us. As a Director for example, most of the stuff I save is with a basic security code (e.g., a letter to a customer) because everyone needs to be able to see it. However, as a Director I also have the right to save things with higher levels of security, e.g., Manager, Director, where appropriate with restricted access. Like all good security systems, it is simple but effective. I am not a fan of overcomplicating anything.

Our searching is also structured the same way. We have configured RecFind 6 to add the objects we need to search on as menu items in the search function just as we would do for any customer. We therefore have a Metadata search menu of Attachments (electronic documents, emails and images), Entities (Customers, Prospects, Partners and Suppliers), People, Incidents, Bugs, Quotes, Invoices, Timesheets, Support agreements, etc. We repeat this with Boolean searches. We make it as easy as possible and as logical as possible so our staff can find anything as fast as possible. After all, I am paying their salaries so I want them to be as productive as possible.

Most importantly, we provide multiple entry-points for searches. I can for example search directly for emails with a Metadata search, searching by a combination of Sender, Recipient, Date, Subject, etc. Alternatively, I can search for customer emails from within the Entity record just by clicking on a single link for all attachments for that customer. We give our staff multiple options just as we give our customers multiple options.

You can search on any field and different classes of users can have different Metadata to both view and search on. The security system determines what each class of user (security group) can both see and then do with the information they can see. That is, the security system determines what tables and fields (and electronic documents and emails) you can see and then what methods (Add, Modify, Clone, Delete, Search, Print, etc.) you can use. Each security group sees only what it needs to see and has only the functionality it needs to get the job done

Because the system is flexible, the records manager for example could choose to search for emails on the way they were classified (say a 3 level hierarchy) but end users could choose to search using a natural selection of Metadata fields such as Sender, Recipient, Subject, Content, Date or ranges of these fields combined in either a Metadata or BOOLEAN or (making it easy for end users) Saved search.

Its horses for courses!

Following the above hybrid approach also means that you can still implement and manage all the recordkeeping principles such as retention and disposal schedules, location tracking, auditing, etc.

My point is that it is possible to meet the needs of all classes of users without frustrating any group.  It just requires a hybrid approach and the configuration of the system to suit each class of user.

Making everyone happy is a lot better than making some people happy and some people unhappy. Why would you do this if you had a choice?

 

 

The absolute easiest & lowest cost way to meet all Electronic Document & Records Management (EDRMS) requirements?

by Frank 19. May 2015 06:00

Because we are a software vendor that builds and markets a range of Enterprise Content Management tools under the RecFind 6 banner I have often been asked, “What is the absolute easiest and lowest cost way to meet all compliance requirements?”

I usually respond with a well-considered and ‘traditional’ response that includes information about Business Classification Systems, UI design, Retention Schedules, etc., etc. The solution proposed would also require a significant degree of consulting so that we aware entirely conversant with the customer’s requirements and business practices and also involve a significant amount of end user training.

This is what the customer expects and it falls in line with the traditional, professional approach.

However, the final solution is rarely ‘easy’ or ‘low-cost’ primarily because it has followed the traditional approach. The more we ask questions and consult and the more people we speak to the more complex the solution becomes. This is normal because we end up trying to configure a solution to meet hundreds or thousands of variables.

There is an easier and lower cost way but I fear that very few of my customers would ever consider it because it requires them to disregard everything they have ever learned about rolling out an EDRMS. We have tried proposing it a few times but never with success. It usually gets shot down by the external consultant or the internal records management professional or both.

It doesn’t require a BCS or a Taxonomy and it doesn’t require a complex Retention Schedule and it doesn’t require significant consulting or significant end-user training. Records Management professionals will surely hate it as a ‘career-ending’ trend. It does require an open mind, the ability to think laterally and a willingness to redefine both the problem and the solution.

It only has three requirements:

  1. Know what electronic documents and emails you don’t want to capture;
  2. Provide a powerful but easy-to-use search that allows anyone to find anything instantly; and
  3. Employ a risk-management approach to retention and select a single retention date (e.g., 7 or 20 years).

Fundamental to the success of this non-conformist solution is the acceptance that computers and storage are dirt-cheap compared to human time. If your IT manager or CIO still agonizes and complains about how much disk space you use up for emails and electronic documents then this is definitely not the solution for you. Your IT hierarchy is still living in the long-gone past when computers and disk were expensive and people were cheap (by comparison).

However, if you have practical, sensible IT people then the approach is worth considering especially if your organization has a long history of failing to digitize its records and automate its processes. That is, you have tried at least once to roll out an organization-wide EDRMS and have failed and/or blown the budget. The word ‘pilot’ probably appears often in your company history usually prefixed by the adjective ‘failed’. Don’t feel too bad, most pilots are initiated because management lacks conviction. They are therefore destined to fail.

We have the tools required to implement such a solution but I won’t go into detail about them now. This is a concept paper, not a detailed instruction manual. If you are interested in the concept please contact me and I can then elaborate.

So, if you really do want to rollout a successful EDRMS and do it in the fastest and least disruptive and lowest cost way possible then please write to me and pose your questions.

For the doubters, this is the same way we manage our electronic documents and emails at Knowledgeone Corporation and we have done so for many years. We use our own software; apart from a couple of accounting packages we run our whole company with the RecFind 6 Product Suite and totally automate the capture of all electronic documents and emails. All my staff have to know is how to search and yes, they can find anything in seconds even after 31 years of operation and a very, very large database.

It is not difficult, it is not ‘expensive’, it does not require a huge amount of management or maintenance time and it runs largely in the background. As I said above, all your staff have to learn is how to search.

It does however, require an open mind and a desire to finally solve the problem in the most expeditious manner possible. But, please don’t tell me you want to run a pilot. Test my solution by all means and put it through the most vigorous change control procedures but don’t damn the end result by beginning with a “we are not really sure it will work so are not really committed and won’t allocate much of a budget but let’s try a pilot anyway because that limits our exposure and risk” approach.

I don’t want to waste your time or mine.

How to simplify electronic document and email management

by Frank 17. September 2014 06:00

I have written about this topic many times in the past (see links at the end of this post) but the lesson is always the same. There are two key rules:

1.     If your system relies on people being 100% consistent and reliable it won’t work; and

2.     If you system places an additional workload on already busy workers it won’t work.

The message is, if you simplify and automate your system you give it the best possible chance of working.

If your system works as automatically as possible and doesn’t require much effort from your workforce then it has the best possible chance of being successful.

With today’s technology and tools there is simply no need to burden your workforce with capture and classification tasks. Do you still see people still using typewriters, rotary phones or Morse code? No you don’t because there is much better technology available. So why do you persist with an old, outdated and unsuccessful model? Why do you ask your staff to manually capture and classify electronic documents and emails when there are much better, much faster, much more consistent and much more reliable ways to achieve a better result? It is after all 2014, not 1914; we all use computers and smart phones now, not typewriters, wind-up rotary phones and Morse code.

Emails are managed by email servers, (yes, even Google). Email servers allow plug-ins and add-ons and are ‘open’ so you can automatically monitor and capture incoming and outgoing emails.

Electronic documents are always saved somewhere, for example on your shared drives or directly into your DMS. As such they can be captured and interrogated programmatically.

It is entirely possible to ‘parse’ any electronic document or email and its associated attributes and Metadata and make consistent decisions about whether or not to capture it and how to classify it when captured. It isn’t rocket science any more, it is just analysis, design and programming. We can go even further and determine who should be notified and what action(s) need to be initiated in response to each new email or electronic document.  

We can easily implement an end-to-end business process whereby every electronic document and email is managed from creation to destruction and we can do this with minimal human involvement. Where human involvement is required, for example making a decision or deciding upon an appropriate response, we can also automate and manage the business processes required and simply ‘present’ staff with all the required information when required.

Isn’t this was the Knowledge Management revolution was supposed to be about?

“A system that provides the user with the explicit information required, in exactly the form required at precisely the time the user needs it.”

The new model is all about automation and processing at the server rather than at the user’s workstation; a fully automatic, server-centric paradigm. A system that is all about the ‘Push’ rather than the ‘Pull’ model. A model whereby the computer services the end user, where the end user is not a slave to the computer.

We could also call it management by exception. “Please only give me what I need to see when I need to see it.”

None of the above is new or revolutionary thinking, it is all just common sense. None of the above requires yet-to-be invented technology or products, it only requires existing and proven technology and products.

The fully-automatic, server-centric approach should be the default choice and it should be a no-brainer for any organization that needs to implement an email and document management regime. Unfortunately, too often it isn’t.

If you have the responsibility of rolling out an email and document management system and the fully-automatic, server-centric approach isn’t on your agenda then your boss should be asking you why not.

References:

White papers

Posts

How to clean up your shared drives, Frank’s approach

by Frank 22. August 2014 06:00

In my time in this business (enterprise content management, records management, document management, etc.) I have been asked to help with a ‘shared drive problem’ more times than I can remember. This particular issue is analogous with the paperless office problem. Thirty years ago when I started my company I naively thought that both problems would be long gone by now but they are not.

I still get requests for purely physical records management solutions and I still get requests to assist customers in sorting out their shared drives problems.

The tools and procedures to solve both problems have been around for a long time but for whatever reason (I suspect lack of management focus) the problems still persist and could be described as systemic across most industry segments.

Yes, I know that you can implement an electronic document and records management system (we have one called RecFind 6) and take away the need for shared drives and physical records management systems completely but most organizations don’t and most organizations still struggle with shared drives and physical records. This post addresses the reality.

Unfortunately, the most important ingredient in any solution is ‘ownership’ and that is as hard to find as it ever was. Someone with authority, or someone who is prepared to assume authority, needs to take ownership of the problem in a benevolent dictator way and just steam-roll a solution through the enterprise. It isn’t solvable by committees and it requires a committed, driven person to make it happen. These kind of people are in short supply so if you don’t have one, bring one in.

In a nutshell there are three basic problems apart from ownership of the problem.

1.     How to delete all redundant information;

2.     How to structure the ‘new’ shared drives; and

3.     How to make the new system work to most people’s satisfaction.

Deleting redundant Information

Rule number one is don’t ever ask staff to delete the information they regard as redundant. It will never happen. Instead, tell staff that you will delete all documents in your shared drives with a created or last updated date greater than a nominated date (say one-year into the past) unless they tell you specifically which ‘older’ documents they need to retain. Just saying “all of them” is not an acceptable response. Give staff advance notice of a month and then delete everything that has not been nominated as important enough to retain.  Of course, take a backup of everything before you delete, just in case. This is tough love, not stupidity.

Structuring the new shared drives

If your records manager insists on using your already overly complex, hierarchical corporate classification scheme or taxonomy as the model for the new shared drive structure politely ask them to look for another job. Do you want this to work or not?

Records managers and archivists and librarians (and scientists) understand and love complex classification systems. However, end users don’t understand them, don’t like them and won’t use them. End users have no wish to become part-time records managers, they have their own work to do thank you.

By all means make the new structure a subset of the classification system, major headings only and no more than two levels if possible. If it takes longer than a few seconds to decide where to save something or to find something then it is too complex. If three people save the same document in three different places then it is too complex. If a senior manager can’t find something instantly then it is too complex. The staff aren’t to blame, you are.

I have written about this issue previously and you can reference a white paper at this link, “Do you really need a Taxonomy?”

The shared drives aren’t where we classify documents, it is where we make it as easy and as fast as possible to save, retrieve and work on documents; no more, no less. Proper classification (if I can use that term) happens later when you use intelligent software to automatically capture, analyse and store documents in your document management system.

Please note, shared drives are not a document management system and a document management system should never just be a copy of your shared drives. They have different jobs to do.

Making the new system work

Let’s fall back on one of the oldest acronyms in business, KISS, “Keep It Simple Stupid!” Simple is good and elegant, complex is bad and unfathomable.

Testing is a good example of where the KISS principle must be applied. Asking all staff to participate in the testing process may be diplomatic but it is also suicidal. You need to select your testers. You need to pick a small number of smart people from all levels of your organization. Don’t ask for volunteers, you will get the wrong people applying. Do you want participants who are committed to the system working, or those who are committed to it failing? Do you want this to succeed or not?

If I am pressed for time I use what I call the straight-line-method. Imagine all staff in a straight line from the most junior to the most senior. Select from both ends, the most junior and the most senior. Chances are that if the system works for this subset that it will also work for all the staff in between.

Make it clear to all that the shared drives are not your document management system. The shared drives are there for ease of access and to work on documents. The document management system has business rules to ensure that you have inviolate copies of important documents plus all relevant contextual information. The document management system is where you apply business rules and workflow. The document management system is all about business process management and compliance. The shared drives and the document management system are related and integrated but they have different jobs to do.

We have shared drives so staff don’t work on documents on ‘private’ drives, inaccessible and invisible to others. We provide a shared drive resource so staff can collaborate and share information and easily work on documents. We have shared drives so that when someone leaves we still have all their documents and work-in-process.

Please do all the complex processes required in your document management system using intelligent software, automate as much as possible. Productivity gains come about when you take work off staff, not when you load them up with more work. Give your staff as much time as possible so they can use their expertise to do the core job they were hired for.

If you don’t force extra work on your staff and if you make it as easy and as fast as possible to use the shared drives then your system will work. Do the opposite and I guarantee it will not work.

A simple guide to using shared drives to capture & classify electronic documents and emails

by Frank 18. July 2014 06:00

I have written previously about ways to solve the shared drives problem (click here) and I have written numerous articles (and a book) about ways to manage emails and electronic/digital records. However, we still receive multiple requests from customers and prospective customers about the best, and simplest, way to effectively manage these problems.

The biggest stumbling block and impediment to progress in most cases is the issue of a suitable taxonomy or classification system. Time and time again I see people putting off the solution while they spend years and tens of thousands or hundreds of thousands of dollars grappling with the construction of a suitable taxonomy. I have written about this topic previously as well and if you want my recommendations please click on this link.

If you really want the simplest, easiest to understand, easiest to use and lowest cost way to solve all of the above problems then please forget about spending the next twelve to eighteen months grappling with the nuances of your classification system. It isn’t necessary.

What you need instead is a natural classification structure that reflects your business processes. Please give your long-suffering end users something they will instantly recognize and can easily work with because it is familiar from their day to day work. Give them something to work with that doesn’t require them to become amateur records managers battling to decipher a complex, hierarchical classification system that requires an intricate knowledge of classification theory to interpret correctly. Give them something that makes it as easy as possible to file everything in the right place first time with absolutely minimal effort. Give them something that makes it as easy as possible to find something.

What I am proposing isn’t a hundred-percent solution and it won’t suit every organization but I guarantee that it will turn chaos into order in any organization that implements it. You may well see it as an eighty-five-percent solution but that is a hell of a lot better than no solution. It is also easy and fast to implement and relatively low cost (you will need some form of RM software).

First up you need to make decisions about what kind of business you are.  Notice that I said “what kind of business you are” not “what kind of records you manage” or “how your business is structured”.  Most importantly, strongly resist the temptation to base your classification structure on your existing business structure or organization’s departments/agencies and instead base it on your most common business processes. Please refer to the following extract from:

Overview of Classification Tools for Records Management by the National Archives of Australia, ISBN 0 642 34499 X (an excellent reference document if you need to understand classification systems).

“Classifying records and business information by functions and activities moves away from traditional classification based on organisational structure or subject. Functions and activities provide a more stable framework for classification than organisational structures that are often subject to change through amalgamation, devolution and decentralisation. The structure of an organisation may change many times, but the functions an organisation carries out usually remain much the same over time.”

I would also strongly resist the temptation to build your classification structure on content; it is way too difficult. Instead, as I have said above, base it on your common business processes.

When I say classification structure I mean the way you name and organize folders in your shared drives. I can’t give you a generic solution because I am not that clever; I don’t know enough about your business. I can however, give you an example.

Please also remember that for the most part, we are dealing with unstructured source information; Word, Excel, PowerPoint, Emails, etc. Emails are a little easier to deal with because they have a limited but common structure, e.g., Date Received, Sender, Recipient, CC and Subject. With other electronic documents we are have far less information and are  usually limited to Author (not reliable), Date Created, Date Modified and Filename. Ergo, as I said earlier, trying to base a classification system on the content of unstructured documents is both difficult and inexact. It is certainly doable but you will have to spend a lot more money on consulting and sophisticated software to achieve your ends.

In my simple example of my simple system I am going to assume that your business is customer (or client) centric, i.e., as opposed to being case-centric or project-centric, etc. The top level of your classification structure therefore will be the client name and/or number. To make it as simple as possible I am going to propose only two levels. The second level represents your most common business processes, that is, what you do with each customer. So for example, I have:

Customer Name

     Correspondence

     Contracts

     Quotes & Proposals.

     Orders

     Incidents

I am also not going to differentiate between emails and other types of electronic documents, I am going to treat them all the same.

Now how does this simple system work?

  1. Staff producing electronic documents don’t have their ‘own’ shared drive, all staff use the common classification structure. This is very important, let one or more people be exceptions and you no longer have a system you can rely on to meet your needs for reliable retrieval and any compliance legislation you are subject to.
  2. Staff drag and drop or ‘save-as’ emails from their email client to the correct sub-folder.
  3. Similarly, staff save (or drag and drop) electronic documents into the correct sub-folder. You can control access if required by applying security to electronic documents.
  4. You purchase or build a document repository (based on any common database such as SQL Server, MySQL, etc.) and within this repository you replicate the folder structure of your shared drives with logical folders and subfolders.
  5. You purchase or build a tool that constantly monitors the shared drives (e.g., using .NET Watcher technology) and that instantly captures a copy of any new or modified document (you do need to configure your repository to automatically version modified documents). You may also decide to automatically delete the original source document after it has been captured.
  6. You build or purchase a records and document management software package that allows you to index, search and report on all the information in your repository.
  7. You train your staff in how to save and search for information (shouldn’t take more than a half to one day) and then you go live.

I would also recommend applying a retention schedule based on sub folder (e.g., contracts) and date created and have the records management system automatically apply it to manage the lifecycle of captured documents. There is no sense in retaining information longer than you have to; it is also a dangerous practice.

Please note that the above is just an example and a very simple one at that. You need to determine the most appropriate folder structure for your organization.

WARNING

Do not let the folder structure become overly complex and unwieldy. If you do, it won’t work and you will end up with lots of stuff either not captured or captured to the wrong place. The basic rules are that if it takes more than few second to decide where to file something then it is too complex and that any structure more than 3 levels deep is too complex.

And finally, this isn’t just a theory, it is something we do in our organization and it is something many of our customers do. If you would like to read more on this approach there are some white papers and more explanations at this link. Alternatively, you can contact us and ask questions at this link.

Good luck.

 

What is the future of RecFind? - The Product Road Map

by Frank 19. May 2014 06:00

First a little history. We began in 1984 with our first document management application called DocFind marketed by the then Burroughs Corporation (now called Unisys). In June 1986 we sold the first version of RecFind, a fully-featured electronic records management system and a vast improvement on the DocFind product. Then we progressively added document imaging then electronic document management and workflow and then with RecFind 6 a brand new paradigm and an amalgam of all previous functionality; an Information management system able to run multiple applications concurrently with a complete set of enterprise content management functionality. RecFind 6 is the eighth completely new iteration of the iconic RecFind brand.

RecFind 6 was and is unique in our industry because it was designed to be what was previously called a Rapid Application Development system (RAD) but unlike previous examples, we provided the high level toolset so new applications could be inexpensively ‘configured’ (by using the DRM) not expensively programmed and new application tables and fields easily populated using Xchange. It immediately provided every customer with the ability to change almost anything they needed changed without needing to deal with the vendor (us).  Each customer had the same tools we used to configure multiple applications within a single copy of RecFind 6. RecFind 6 was the first ECM product to truly empower the customer and to release them from the expensive and time consuming process of having to negotiate with the vendor to “make changes and get things done.”

In essence, the future of the RecFind brand can be summarised as more of the same but as an even easier to use and more powerful product. Architecturally, we are moving away from the fat-client model (in our case based on the .NET smart-client paradigm) to the zero-footprint, thin-client model to reduce installation and maintenance costs and to support far more operating system platforms than just Microsoft Windows. The new version 2.6 web-client for instance happily runs on my iPad within the Safari browser and provides me with all the information I need on my customers when I travel or work from home (we use RecFind 6 as our Customer Relationship Management system or CRM). I no longer need a PC at home and nor do I need to carry a heavy laptop through airports.

One of my goals for the remainder of 2014 and 2015 following is to convince my customer base to move to the RecFind 6 web-client from the standard .NET smart-client. This is because the web-client provides tangible, measurable cost benefits and will be the basis for a host of new features as we gradually deprecate the .NET smart-client and expand the functionality of the web-client. We do not believe there is a future for the fat/smart-client paradigm; it has seen its day. Customers are rightfully demanding a zero footprint and the support of an extensive range of operating environments and devices including mobile devices such as smartphones and tablets. Our web-client provides the functionality, mobile device support and convenience they are demanding.

Of course the back-end of the product, the image and data repository, also comes in for major upgrades and improvements. We are sticking with MS SQL Server as our database but will incorporate a host of new features and improvements to better facilitate the handling of ‘big data’. We will continue to research and make improvements to the way we capture, store and retrieve data and because our customer’s databases are now so large (measured in hundreds of Gigabytes), we are making it easier and faster to both backup and audit the repository. The objectives as always are scalability, speed, security and robustness.

We are also adding new functionality to allow the customer to bypass our standard user interface (e.g., the .NET smart-client or web-client) and create their own user interface or presentation layer. The objective is to make it as easy as possible for the customer to create tailored interfaces for each operating unit within their organization. A simple way to think of this functionality is to imagine a single high level tool that lets you quickly and easily create your own screens and dashboards and program to our SDK.

On the add-in product front we will continue to invest in our add-in products such as the Button, the MINI API, the SDK, GEM, RecCapture, the High Speed Scanning Module and the SharePoint Integration Module. Even though the base product RecFind 6 has a full complement of enterprise content management functionality these add-on products provide options requested by our customers. They are generally a way to do things faster and more automatically.

We will continue to provide two approaches for document management; the end-user paradigm (RecFind 6 plus the Button) and the fully automatic capture and classification paradigm (RecFind 6 plus GEM and RecCapture). As has been the case, we also fully expect a lot of our customers to combine both paradigms in a hybrid solution.

The major architectural change is away from the .NET smart-client (fat-client) paradigm to the browser-based thin-client or web-client paradigm. We see this as the future for all application software, unconstrained by the strictures of proprietary operating systems like Microsoft Windows.

As always, our approach, our credo, is that we do all the hard work so you don’t have to. We provide the feature rich, scalable and robust image and data repository and we also provide all of the high level tools so you can configure your applications that access our repository. We also continue to invest in supporting and enhancing all of our products making sure that they have the feature set you require and run in the operating environments you require them to. We invest in the ongoing development of our products to protect your investment in our products. This is our responsibility and our contribution to our ongoing partnership.

 

Is this Microsoft’s worst mistake ever?

by Frank 30. November 2013 06:00

I run a software company called the Knowledgeone Corporation that has been developing application solutions for the Microsoft Windows platform since the very first release of Windows. As always, our latest product offering RecFind 6 version 2.6 has to be tested and certified against the latest release of windows. In this case that means Windows 8.1.

Like most organizations, we waited for the Windows 8.1 release before upgrading our workstations from Windows 7. The only exceptions were our developers workstations because we bought them new PCs with Windows 8 pre-installed.

We are now testing the final builds of RecFind 6 version 2.6 and have found a major problem. The problem is that Microsoft in its infinite wisdom has decided that you can’t install Windows 8.1 over a Windows 7 system and retain your already installed applications.

The only solution is to install Windows 8 first and then upgrade Windows 8 to Windows 8.1. However, if you are running Windows 7 Enterprise this won’t work either and you will be told that you will have reinstall all of your applications.

I am struggling to understand Microsoft’s logic.

Surely Microsoft wants all its customers to upgrade to Windows 8.1? If so, why has it ‘engineered’ the Windows 8.1 upgrade so customers will be discouraged from using it? Does anyone at Microsoft understand how much work and pain is involved in re-installing all your applications?

No, I am not kidding. If you have a PC or many PCs with Windows 7 installed you are going to have to install Windows 8 first in order to maintain all of your currently installed applications. Then, after spending many hours installing Windows 8 (it is not a trivial process) spend more precious time installing Windows 8.1. Microsoft has ensured that you cannot go direct from Windows 7 to Windows 8.1.

Of course, if you are unlucky, you could be living in a country where Microsoft has blocked the downloading of Windows 8, like Australia. Now you are between a rock and a hard place. Microsoft won’t let you install Windows 8 and if you install Windows 8.1 you face days or weeks of frustrating effort trying to re-install all of your existing applications.

 

Here are some quotes from Microsoft:

“You can decide what you want to keep on your PC. You won't be able to keep programs and settings when you upgrade. Be sure to locate your original program installation discs or purchase confirmation emails if you bought programs online. You'll need these to reinstall your programs after you upgrade to Windows 8.1—this includes, for example, Microsoft Office, Apache OpenOffice, and Adobe programs. It's also a good idea to back up your files at this time, too.”

If you're running Windows 7, Windows Vista, or Windows XP, all of your apps will need to be reinstalled using the original installation discs, or purchase confirmation emails if you bought the apps online.”

If the management at Microsoft wanted to ensure the failure of Windows 8.1 they couldn’t have come up with a better plan than the one they have implemented. By making Windows 8.1 so difficult to install they have ensured that its customers will stick with the tried and proven Windows 7 for as long as possible.

Can anyone at Microsoft explain why they thought this was a good idea?

Records Management in the 21st century; you have computers now, do it differently

by Frank 1. June 2013 06:32

I own and run a computer software company called the Knowledgeone Corporation and we have specialised in what is now known as enterprise content management software since 1984 when we released our first product DocFind. We are now into the 8th iteration of our core and iconic product RecFind and have sold and installed thousands of RecFind sites where we manage corporate records and electronic documents.

I have personally worked with hundreds of customers to ensure that we understand and meet their requirements and I have also designed and specified every product we have delivered over the last 29 years so while I have never been a practicing records manager, I do know a great deal about records and document management and the vagaries of the practise all around the world.

My major lament is that many records managers today still want to run their ‘business’ in exactly the same way it was run 30 or 50 or even a hundred years ago. That is, as a physical model even when using computers and automated solutions like our product RecFind 6. This means we still see overly complicated classification systems and overcomplicated file numbering systems and overcomplicated manual processes for the capture and classification of paper, document images, electronic documents and emails.

It is a mindset that is locked in the past and can’t see beyond the confines of the file room.

I also still meet records managers that believe each and every employee has a responsibility to ‘become’ a junior records manager and both fully comprehend and religiously follow all of the old-fashioned and hopelessly overcomplicated and time-consuming processes laid out for the orderly capture of corporate documents.

I have news for all those locked-in-the-past records managers. Your approach hasn’t worked in the last 30 years and it certainly will not work in the future.

Smart people don’t buy sophisticated computer hardware and application software and then try to replicate the physical model for little or no benefit. Smart people look at what a computer system can do as opposed to 20,000 linear feet of filing shelves or 40 Compactuses and 30 boxes of filing cards and immediately realize that they have the power to do everything differently, faster, most efficiently and infinitely smarter.  They also realize that there is no need to overburden already busy end users by a forcing them to become very bad and very inconsistent junior records managers. End users are not hired to be records managers they are hired to be engineers, sales people, accountants, PAs, etc., and most already have 8 hours of work a day without you imposing more on them.

There is always a better way and the best way is to roll out a records and document and email management system that does not require your end users to become very bad and inconsistent junior records managers. This way it may even have a chance of actually working.

Please throw that old physical model away. It has never worked well when applied to computerised records, document and email management and it never will. Remember that famous adage, “The definition of insanity is to keep doing the same thing and to expect the results to be different”?

I guarantee two things:

1.     Your software vendor’s consultant is more than happy to offer advice and guidance; and

2.     He/she has probably worked in significantly more records management environments than you have and has a much broader range of experience than you do.

It doesn’t hurt to ask for advice and it doesn’t hurt to listen.

Month List