Encryption Options for WebCenter Content

September 20, 2016

By: Raoul Miller – Enterprise Architect

With the increased focus on security in the workplace, TEAM is seeing that more and more of our clients have the requirement to encrypt important data. The content that is managed in their WebCenter Content instance is already an important asset, so the business may see a need to encrypt some or all of that content.

Because there are different areas that can be encrypted, there has been some confusion as to how to go about this process. The following lists the three major options for data encryption within WCC and some of the pros and cons associated with each:

Transport Layer Security (TLS) for Traffic

The easiest and quickest level of encryption to deploy is SSL (Secure Sockets Layer) configuration for web traffic, JDBC connection, and LDAP queries. All of these can be configured from the WebLogic Server (WLS) console interface and require only certificate procurement and management from the client.


  • Encryption of traffic prevents sniffing of credentials or data from the wired or wireless networks
  • Relatively quick to configure and simple to deploy


  • Does not address encryption of content or metadata
  • Requires certificate procurement and management


Metadata Encryption

The next step in encryption is to encrypt some or all of the tables or columns in the database. While encryption is now available in Microsoft SQL Server 2016, this is not officially supported by Oracle and has not been tested by TEAM. For those using Oracle Database Enterprise Edition, this path requires licensing of the Advanced Security option and deployment of TDE (transparent data encryption) within the JDBC client.

While it is possible to encrypt only some of the metadata, the overhead involved with this would be quite substantial and there would be a risk of exposing newly created custom metadata. All of TEAM’s clients that use encryption have chosen to encrypt the entire metadata schema.


  • Flexibility as to which tables or columns within the schema should be encrypted
  • Fully supported by Oracle
  • Provides additional security from internal threats and escalation of privileges


  • There is an encryption / decryption overhead at the server and client level (client in this case is WLS Server)
  • Management of database and certificates places additional burden on DBAs
  • Requires Enterprise Edition Oracle Database and Advanced Security licenses


Content Encryption

The ultimate step in security is to encrypt the content as well as the metadata. The only supported method for this is to store the content in the Oracle database using SecureFiles.  While in theory this could be done without metadata security, to do so would be very poor practice, so this assumes that both metadata and content are to be encrypted.

The FileStore Provider within WebCenter Content (WCC) manages file storage and when content is created / submitted to the content management system, it must be tagged with a metadata field (xStorageRule) indicating where it is to be stored. The system can manage multiple file system storage rules, but only a single JDBC rule. Assignment of the storage rule is normally done either through profiles or workflow.

Clients have 3 options for storage of content:

  • Encrypted in a database
  • Unencrypted in a database
  • Unencrypted on a file system (the default content storage option).

Clients may also choose to store some (or most) content unencrypted on the file system, and another portion encrypted within the database, but WCC does not (currently) support storage of some content unencrypted in the database while other content is also encrypted in the database. Combining unencrypted storage rules for content on the file system and an encrypted storage rule for the database will allow for a “mixed” system where only that content that is required to be encrypted has the overhead.


  • Encryption of content is an added layer of security
  • Performance of SecureFiles is comparable to network attached storage
  • Content is now fully encrypted in transport and at rest to the highest standards (TDE supports AES256, AES192 – default for TDE column encryption, AES128 – default for TDE tablespace encryption, and 3DES168).
  • Fully supported by Oracle


  • There is an encryption / decryption overhead at the server and client level (client in this case is WLS Server)
  • Management of database and certificates places additional burden on DBAs
  • Requires Enterprise Edition Oracle Database, Advanced Security, and Partitioning licenses
  • Setup of tablespace storage can be complex and is poorly documented

All of the above options assume that the deployment is on-premises, or deployed on infrastructure as a service (IaaS). You can still encrypt content on hosted systems, and I will follow up on your choices for hosted systems in another post in the near future.

Please feel free to contact TEAM for all of your WebCenter Content questions, particularly those around content security, encryption, and redaction.

Want to talk at OpenWorld? Email sales@teaminformatics.com

More Information

SSL Setup



TDE (Transparent Data Encryption)


Advanced Security on Oracle Database







Sending E-mail from WebCenter Content using Office 365

September 8, 2016

By: Joe Hutchings – Director, Managed Services & Dwayne Parkinson – Solution Architect

It’s a familiar tale. The whole company migrated to Office 365 over the weekend BUT… someone forgot to test WebCenter Content. Now, that critical workflow that sends e-mail message notifications has suddenly stopped working! Since your system isn’t working, we won’t waste any more words. Here’s the fix.

  1. From WebCenter Content, go to the Administration menu and select Admin Server and then Internet Configuration as shown below.


2. Change the mail server address on that screen so it points to the Mail Exchange record (aka MX record) on your DNS server that is used by Office 365. You should be able to find that information by using this link: http://mxtoolbox.com/SuperTool.aspx# and entering the domain of your company.  NOTE: the domain does not include “www” so it’s just something like myCompany.com.

3. Add an SPF entry to your DNS server with the IP address of your WebCenter Content server to prevent the WebCenter Content e-mails from being flagged as SPAM. Details about setting up the SPF record for Office 365 can be found here:

4. Restart the content server.

That’s it!  You’re off and running.  Now you can enjoy the rest of your day.

Automatic Redaction in WebCenter – A TEAM Webinar

September 7, 2016

redaction-logoProtecting sensitive data is extremely important for your organization. The release of public records, data breaches, and simple sharing between parties can cause unnecessary distress and costs when names, addresses, credit card numbers, etc. become public. Personally Identifiable Information, or PII, is any data that could potentially enable identification of a specific individual. Identifying this data is invariably important to protect your company, your employees, and your clients.

With a single disclosure breach reaching an estimated cost as high as seven million dollars ($7m) and affecting all manner of agencies, corporations, and hospitals across the globe every year, why not instead spend your resources on preemptive and preventative measures?

TEAM’s Redaction Engine utilizes multiple technologies to scan for and identify sensitive data already in your WebCenter Content repository, and newly scanned or checked-in documents. The engine then automatically redacts that information in subsequently generated PDFs.

Watch this video to learn more and join us for the webinar for a live demonstration for latest features and to answer any questions you may have.

redactionBenefits & Latest Features:

  • Integrated into Oracle WebCenter Content 11g & 12c
  • Pattern Recognition (regex and basic) identifies information such as:
    • SS Numbers
    • Date of Birth
    • Identification Numbers/Medical Forms
    • Credit Card Numbers
    • ACH Routing and Bank Account Numbers
  • Now allows multi-word phrases
  • Improved RegEx validation tester to provide better feedback for patterns
  • Redaction Templates
    • The ability to draw redactions in the UI and save them as a “template” that can be later utilized when checking in new documents
  • Create Redaction Sets – unique sets of criteria and patterns to be redacted
  • UI now includes comprehensive predefined Regular Expressions to match common use cases such as Social Security Numbers, telephone numbers, and financial amounts


TEAM’s 2nd Redaction Engine Webinar
September 29th, 2016 – 1:00PM CST 
Register Now!

Elasticsearch – The Smarter Way to Search

August 24, 2016

By: Marlene Laus – Marketing Specialist

No matter what size your business is or industry you’re engaged in, the necessity to efficiently locate content on demand is paramount. To locate an item, one must search, and to search there must be a pool of data that can be queried – an index.  The act of indexing creates searchable information that users can query to locate their target. Unfortunately, as businesses grow and content volumes rise, it becomes increasingly difficult for some search systems to keep up with the demands of indexing and querying, let alone return quality results.

elasticsearch logo

This is where Elasticsearch comes in. Elasticsearch changes the way your documents are both indexed and queried, making it more efficient and faster than ever to locate your content. TEAM has taken this technology to the next level by integrating it with Oracle WebCenter Content, an industry-leading Enterprise Content Management (ECM) system. By integrating Elasticsearch with WebCenter Content, you are able to get the latest, fastest and most advanced functionality.


Unmatched Speed

Traditionally, when data is entered into WebCenter Content it goes through Oracle Text Search (OTS) which indexes the data so that your document can be searched by any words within the document and on metadata applied to the document. Elasticsearch takes it even further by providing a data indexing system that is able to index your documents in real time at a remarkably fast speed, resulting in almost instant search results. Operating at an unmatched speed, the advanced indexing system is able to support big repositories with high volumes of search.


One of the most impressive features of Elasticsearch is its ability to be easily scaled both horizontally and vertically. Expand vertically by adding more memory and processors on individual servers and the service will attempt to utilize everything you give it. Expand horizontally when you add additional servers to create clusters and  they’ll operate in tandem to power your search efforts.

Image Source

As repositories continue to grow, it can become difficult for typical indexing systems to keep up with the high demand. Indexes start to break down and it can take days or even weeks to re-build them, possibly rendering your search capabilities inactive. This is where Elasticsearch changes the game for organizations with large repositories; in what would normally take hours, Elasticsearch can efficiently re-build your indexes within minutes.

Custom and Customizable

Elasticsearch enables you get the most out of your search indexes by allowing you to customize them! At a basic level, you have the ability to choose what metadata fields you want to use for searching your documents and the option of using friendly metadata names of your choosing. For example, if you want to find a document that contains the word “Marketing” in the title, you can type “title: Marketing” into the search bar and Elasticsearch will pull all of the documents in your repository with “Marketing” included in the title metadata field. At a more complex level, the Elasticsearch index allows dozens of intricate index configuration options, all of which are available to use with the integration. The different ways you choose to index and search for documents within your repository are limitless. Elasticsearch puts the power in hands of the user.

Screen Shot 2016-08-24 at 11.59.35 AM
Image Source


Finally, Elasticsearch enables you to utilize a practice called “stemming”. For example, if you query for the word “run,” Elasticsearch will compile your results and include documents that contain words where “run” is a stem, such as “runner” and “running”. The non-exact results will be returned with lower confidence than the exact matches but you can rest easy knowing your results will be as complete as possible.

Take Control of Your Searches

In a nutshell, Elasticsearch enables faster indexing and quicker, more accurate search results while making it easier than ever to utilize customized indexes all while staying fault- and load-tolerant. The possibilities are endless with Elasticsearch supercharging the search capabilities of your WebCenter Content repository.

How to Synchronize WebCenter Capture Configurations between Environments

July 29, 2016

By: Dwayne Parkinson – Solution Architect

Anyone who has used WebCenter Enterprise Capture appreciates the power and flexibility the tool provides for scanning, importing, converting, processing and exporting content. Its power and flexibility come from its wide range of configuration options.  However, in some situations this flexibility also creates problems as system administrators try to keep Capture configurations synchronized between environments.  Thankfully Oracle has provided a relatively easy way to synchronize Capture configurations between environments.

The process of migrating WebCenter Enterprise Capture configurations involves the following steps:

  1. Start the WebLogic scripting tool on the server running the WebCenter Capture instance that you want to export. This is done by running the wlst.cmd or wlst.sh depending on whether you’re running Windows or Linux.  Sample locations for the files are:

2. Connect to your Capture server using the connect WLST command. A sample may look like this:

connect(‘weblogic’, ‘myPassword’, ‘localhost:16400’)

3. Find the Capture workspace ID you want to export. Capture workspaces are assigned a numeric ID that you can obtain by using the listWorkspaces command from within the WebLogic Scripting Tool.  This command will list the ID followed by the name of the workspace.


4. Using the ID of the workspace you want to export, issue the WebLogic Scripting Tool command to exportWorkspaces with the ID of the workspace and a destination path and file name.


5. Copy the file that was created in step 3 to the destination server that will be updated with the configurations.

6. Start the WebLogic Scripting Tool on the WebCenter Capture instance that you want to import the configurations into (see #1).

7. Connect to the destination Capture server (see #2)

8. Perform a listWorkspaces command (see #3) to insure that the name of the workspace you’re importing does not already exist. If there is an existing workspace with the same name as the workspace you exported in step 3, you must go into the Capture Console and either rename it or delete it before proceeding.

9. Issue the importWorkspace command to load the workspace from the configuration file from step 4 and create the workspace.


10. VERY IMPORTANT: be sure to verify the configuration and change Import Source settings such as the folder locations and e-mail addresses on Capture jobs so they point to the correct location for this environment.

sync blogHINT: By using similar paths and just changing from Development to Test, QA, Production, etc. the change should be relatively small and easy to manage.

Using the steps above, you will be able to quickly and easily keep your WebCenter Enterprise Capture configurations synchronized between environments, even in the most complex installations.

For more information about WebCenter Capture and how it can be leveraged to help your business, please contact TEAM Informatics.

Don’t Regret – Redact

July 6, 2016

By: Jon Chartrand – Director of Product Management

The concept of sensitive information management is germane to pretty much every business, organization, and public sector outfit in the world. Typically, this sensitive information is classified as “PII” or Personally Identifiable Information – this would be any data which could lead to someone being personally identified and includes things like social security numbers, date of birth, and phone numbers. Other data, often revolving around financial information, includes credit card numbers, bank account numbers, and account balances. All of these data points must be carefully monitored and masked before documents can potentially be made available for distribution – externally or internally. Failure to do so can lead to devastating legal and financial consequences, bankrupting corporations and governments alike. As experts in the field of content management and in bringing order to unstructured data, we felt an obligation to assist our clients with this often expensive and time-consuming effort.

Examples of PII, according to the National Institute of Standards and Technology (NIST)1:

Name Street Address State Zip Code
Telephone Number Email Address Social Security Number Medical Record Number
Health Plan Number Account Number Account Balance ACH Number
Bank Account Routing Number Credit Card Number CCV Code
Driver’s License Number Passport Number Taxpayer ID Date of Birth


Just these example values represent a staggering amount of data across potentially every piece of content your organization creates, updates, manages, stores, distributes, and archives. The compliance costs required to scour content for this data can be monumental in terms of both dollars and hours. However, these costs can pale in comparison to the costs associated with a data breach. A recent study found that the average total cost of a data breach in the US can exceed $7 million, with an average per-record cost of more than $2002. These are some frightening numbers. So how do we help strengthen your compliance efforts while also reducing your compliance costs? That is the question we asked ourselves several months ago and the answer, we believe, is the TEAM Redaction Engine.

We built the Engine to meet three specific needs:

  • textual pattern matching in digital documents
  • integration with scanning solutions for paper documents
  • redaction of identified data in both PDFs and images

The Redaction Engine is a plugin, or component, for Oracle’s WebCenter Content (WCC) platform. This was done because WCC is a leader in the Enterprise Content Management space and it has direct integrations with powerful scanning solutions, Oracle’s cloud-based platforms, and powerful search options such as Elasticsearch. Other than enabling scanning, the component requires no additional software or hardware to perform its functions against the content in your repository – which is a revolution in the sensitive information arena.

Pattern Matching

When it comes to assisting with sensitive information compliance, the primary challenge comes in the form of identifying the data in question. Between our efforts with WebCenter Content and with Elasticsearch in the enterprise content management space, we realized that we already have access to every character of every piece of digital content that’s been indexed. What it boils down to is identifying patterns and developing a method for seeking those patterns in the available data. Look again at the table of examples above. Of the 20 data points described, 18 of them (90%!) can easily be identified based on a likely pattern.  This is where we started on our efforts.

The Redaction Engine is focused around a primary core – the Pattern Matching Engine. We allow you to craft a series of patterns using both Regular Expressions and Simple Patterns. To identify Social Security Numbers, for example, you’ll need to take into account the common variation which lacks the dashes. You could choose to use two simple patterns if you weren’t interested in specifics of SSN rules:

  • (with dashes) ###-##-####
  • (without dashes) #########

These would pick up Social Security Numbers but would also incorrectly identify any numeric value which fits this form but doesn’t actually meet certain rules for SSN’s such as that no group of digits can be all zeroes. We could instead craft a regular expression which is much more robust and is designed to meet the rules of SSN’s laid out by the Social Security Administration3:

  • ^(?!219-09-9999|078-05-1120)(?!666|000|9\d{2})\d{3}-(?!00)\d{2}-(?!0{4})\d{4}$
  • ^(?!219099999|078051120)(?!666|000|9\d{2})\d{3}(?!00)\d{2}(?!0{4})\d{4}$

This is an example of how simple and also how robust the pattern matching can be. These same tactics patternmatchingcan be applied to matching pretty much any other predictably-formatted value. The only question is the depth of complexity you want to apply to the efforts. Given that Regular Expression experts are fairly rare, we also included an expression evaluator in the interface. This provides feedback on your expressions and confirms whether each pattern makes sense to the engine or not.

Now that the patterns are configured, WebCenter Content does the heavy lifting during the check-in process of opening the document and extracting the text within so that the document’s contents can be indexed. This indexing means you can search for a word inside the document instead of just the title or metadata. It also means we have a readily available block of extracted text that we can quickly parse against our patterns and identify desired information. Once identified, we simply hand the PDF to an editing library which adds the redaction, burns it into the document, and saves a new copy as a “Redacted Rendition”. The new PDF even remains full-text searchable – it just has the redacted text removed! This is the simplest – and most common – scenario.

Scanning Integration

Less common but no less important are image-based, or scanned, documents.  As paper documents are still a fact of life, we always want to keep an eye on our methods for digitizing that physical content to bring it into the repository. Whether that’s a simple WebCenter Capture setup or some other scanning platform, the important piece is that we get this now-digital item into a managed structure such as WebCenter Content. If your choice is to stick with the WebCenter family, the Redaction Engine is specifically enhanced to work intimately with both Capture and Oracle Forms Recognition (OFR). One of the best examples of this partnership is with content that contains non-digital text, that is, handwriting.

After the paper item is digitized via the scanner and Capture, it’s passed to OFR for processing. This is where we set up “markers” and instruct OFR where to look for characters in a specific location. Even if Oracle Forms cannot interpret the handwriting (via Optical Character Recognition or OCR) it can identify the precise coordinates for the location of the handwriting. Now we simply pass the digitized document and the coordinates to WebCenter Content and the Redaction Engine.

redaction1                                     redaction2

In the end we have a perfectly redacted entry even though the text wasn’t readable by a character recognition engine. This means that as long as we can find digital “landmarks” in our document, we can train Oracle Forms Recognition to look for and identify illegible entries and pass those for redaction.

If, however, your solution for scanning physical documents does not include WebCenter Capture or Oracle Forms Recognition, the Redaction Engine is happy to work with those items as well.


A Bad Fax

In fact, any image-based content can be passed through the Redaction Engine as we’ve included an OCR library with the product. This means not only image-based PDFs but native TIFF, JPEG, or GIF files can be processed as well. The Redaction Engine OCR library will process the content item and scan for any machine-readable English text that it can find. Of course, like with any OCR process, there are limitations in terms of language, fonts, and file resolution however the vast majority of modern scanned documents will have no problems being read. If you’re submitting documents sent via fax machine in 1997 and then digitized with a consumer-grade scanner a year later, you could very well run into issues.

Something extra on this front comes from the fact that we’re finding text in these images – search. While WebCenter Content would not ordinarily be able to include these content items in the full text search index, we’ve joined the Redaction Engine with TEAM’s Elasticsearch Integration to make this happen. That mean’s any text found when an image or image-based item is passed through the engine is submitted to the Elasticsearch index, making it fully searchable. This means, for example, that a scanned invoice could possibly be found by searching for the vendor name, or the invoice ID, or the invoice total and not just by the metadata that was associated to the item at check-in.

Responsible Redaction

We’ve now covered three specific cases where content can be redacted:

  1. via full-text matching of the document contents
  2. via sets of coordinates passed to the Engine
  3. via pattern and location matching of OCR text in an image or image-based item

In all cases the Redaction Engine creates a new, specifically-redacted content item that is separate and unique from the original file. The redactions are also “burned in” to the new file ensuring that the underlying text is permanently removed. Both of these steps are taken to first ensure that no data is lost for the redaction process and, second, to simultaneously ensure that redacted items are secure in terms of information removal.

The last piece of what we have come to call “responsible redaction” is the auditing capability of the Redaction Engine. The product keeps a record of every redaction performed – not just at a document level but at the redaction level. A single content item with several redactions has every individual redaction logged, including the specific pattern that was matched in each case. Redaction Reports can be generated for any date range desired and can be exported as a Microsoft Excel document. This exported document can now be stored as a managed record in WebCenter Content or maintained elsewhere for legal purposes. The goal in all cases is simply to provide as much transparency as possible into a process that is built to, well, do the exact opposite!


The Redaction Engine is not only about helping to lessen the burden on businesses that have to manually parse, identify, and redact sensitive information but to also bolster those on-going information compliance efforts and keep trouble from finding the front door. As we’ve worked on this effort, I’ve come to find a much greater appreciation for the efforts that must be undertaken to try and keep our information safe and secure. As a group, we’re incredibly pleased to be able to offer a solution that could very well save you and your business time, money, and headaches.


1 “Guide to Protecting the Confidentiality of Personally Identifiable Information (PII)”, NIST, April 2010

2 “2016 Cost of Data Breach Study: United States”, Ponemon Institute, June 2016

3 “Validating Social Security Numbers through Regular Expressions”, Rion Williams, Codeproject.com, Sep 2013

TEAM Informatics’ Intelligent Content A smart solution for businesses to manage and control content

May 19, 2016

By: Jon Chartrand – Director of Product Management

Perhaps the primary conceit when it comes to content management is this: context is king. When your content or records have context, it means they can be both cataloged and discovered with much greater ease. When we talk about context, that means metadata – or data describing data. When a document is placed into your content management system it’s important to know who it came from, who it belongs to, what the data within is regarding, and every other aspect of context that can be known, implied, or assumed. This allows the system to catalog the item appropriately and other users to search for and locate the item easily. The problem is that while context is king, entering metadata can be a royal pain – and bad metadata can ruin an otherwise good system. As we all know: garbage in – garbage out.

Picture1TEAM’s been working in the content management space for over a decade so we’ve seen this issue arise repeatedly for our clients. Relying on end users for full, complete, and accurate metadata puts stress on them, slows down the contribution process, and can lead to human error or, even worse, human disinterest. So we set out to not only solve this problem but revolutionize how context is achieved for your content. We partnered with SmartLogic and combined the power of Oracle WebCenter Content with their extraordinary context classification software, Semaphore, to create a unified, smart solution.

This is Intelligent Content.

What is Intelligent Content and how does it work?

TEAM’s Intelligent Content solution alleviates the challenges and roadblocks of requiring users to navigate the metadata process by doing the work for them. This is started by the user simply saving their content to the WebCenter Content repository. The content can be contributed automatically by line-of-business systems or even ingested from network drives or cloud-based file systems. The Intelligent Content engine processes the stored material and leverages an information classification model, or “ontology”, rather than the traditional two-dimensional taxonomy. Intelligent Content drives the auto-classification process by opening each document at contribution time and parsing the content of the document. It is then able to automatically populate metadata based on the rules of the classification model. By automatically tagging your materials, it makes your content easily findable across what would have previously been multiple taxonomic pathways.



Perhaps an example can help here. Imagine an overview document that describes a land use project to build a park. The document may contain sections on project planning, soil samples, a work breakdown, price estimates, and more. In the old-school method the Project Manager checks the item into the repository and, on reflection, classifies the item as a Project Document type item with a subtype of Overview. This is helpful, but really doesn’t encompass the breadth and depth of what’s in the document. In the new-school method, Intelligent Content parses the text and applies predefined classifications; overview, soil samples, work breakdown, pricing… On and on. This means the item can be found by others who search based on what they’re looking for not necessarily solely on the structure of the item. The old-school method provides a single taxonomic pathway (Project Document > Overview). The new-school method enables a much more nuanced approach. When the Engineer looks for documents relating to soil samples, the item is returned. When the Construction Foreman looks for documents relating to Work Breakdown, the item is returned.

As I mentioned earlier, the ontology (AKA information classification model) is comprised of a set of terms and rules, which have the ability to be maintained as needed by the information or records management SME within your organization or through TEAM. By utilizing the information model on the search side of the equation, it allows the use of “semantically enhanced” search capabilities including a “search as you type” feature as well as the ability to browse through the model in an interactive graphical manner. Both methods create easier, faster, and more intelligent pathways for users to find the content they’re looking for in the system.


Why is this important for businesses?

Help Your Contributors.

There’s a lot of room for human error when a document is manually classified. TEAM’s Intelligent Content solution saves the content contributor time and effort by automatically tagging newly stored content. This ensures that every time new content is stored in any department of your business, its classification will be consistent and no longer susceptible to the vagueries of human interpretation.

Help Your Users.

Will the end-user always know what key words to search for when looking for a specific document? The auto classification system makes finding your documents faster and easier than ever. What could potentially take hours to locate within a large system can now be found in a matter of seconds due to the unique ontology model utilized by Intelligent Content.

Help Your Business.

By changing the way your content is cataloged and managed, TEAM’s Intelligent Content solution is a bottom line contributor to the overall enhancement of your business.

While this sounds like a sales pitch – and I admit it kind of is – I want you to understand that we’re also incredibly excited about the results we’re already seeing from Intelligent Content; better classification, less human error, simpler contribution experience, and far faster and more accurate searching. This is the next step in the evolution of enterprise content management. If you’re interested in learning more, you can check out our YouTube video on this topic or email us directly with your questions.