Week of June 7th

New updates: Knowledge AI, Conversation Builder, Conversation Assist, Conversational Cloud Infrastructure

Exact delivery dates may vary, and brands may therefore not have immediate access to all features on the date of publication. Please contact your LivePerson account team for the exact dates on which you will have access to the features.

The timing and scope of these features or functionalities remain at the sole discretion of LivePerson and are subject to change.

Conversation Assist


Rich answers in the On-Demand Recommendations widget are now editable

Rich answers include images and links to content, audio, and video. Rich answers are more engaging than plain answers, so for many brands they’re an important part of a Conversational AI solution. 

In this release, we’re pleased to announce that your agents can now edit the text in rich answer recommendations that are offered in the On-Demand Recommendations widget. Here’s how:


Editing the text of rich answer recommendations that are offered inline in the conversation isn’t yet available, but it’s coming soon.

If you’re using answer recommendations that are enriched via Generative AI, be aware that answers that are both rich (i.e., have multimedia) and enriched (i.e., are generated via Generative AI) are supported but not recommended. This is because the generated answer might not align well substantively with the rich content associated with the highest scoring article, which is what is used.

Conversational Cloud Infrastructure


Our LLM Gateway now offers URL post-processing

In Messaging contexts, URLs that are clickable are always handy and support an optimal experience for the end user. However, responses from the Large Language Model (LLM) service — which are formed via Generative AI  — are always returned as plain text. 

That’s where our new URL post-processing service in the LLM Gateway comes in. It takes all the URLs in the response and wraps them in HTML tags, so they become active, clickable URLs.

By default, the URL post-processing service is on for KnowledgeAI for all accounts. To turn it off, contact your LivePerson representative.

Conversation Builder


More flexibility when using answers enriched via Generative AI

Generally speaking, within a bot, you might want to use enriched answers in some interactions but not others. Previously, this wasn’t possible because the setting to turn on enriched answers was at the knowledge base level (in Settings). If you turned it on, any bot that used the knowledge base used enriched answers.

In this release, we’ve moved the Enriched answers via Generative AI setting to the interaction level. This gives you greater flexibility. Within a bot, you can turn it on for some interactions but not others. Or, you can turn it on for every applicable interaction. The choice is now yours.

Where’s the new interaction-level setting? In the KnowledgeAI interaction, you’ll find it on the Advanced tab in the interaction’s settings. (Typically, you use the KnowledgeAI interaction in Messaging bots.)


In the Integration interaction, the display of the setting is dynamic. You’ll find it on the Advanced tab in the interaction’s settings once you select a KnowledgeAI integration in the interaction. (Typically, you use the Integration interaction in Voice bots.)


The knowledge base-level setting still exists, but we plan to remove it in the next release. Until then, the interaction-level settings take precedence.

An enhanced Preview experience when using answers enriched via Generative AI

Conversation Builder’s Preview tool plays an important role as you design, develop, and test a bot. We’ve just updated it so that now, when automating answers enriched via Generative AI, the tool passes some conversation context to the LLM service that does the answer enrichment. The result? You’ll see better enriched answers while you’re previewing the bot. And you’ll have a consistent experience when using:

  • Preview
  • Conversation Tester
  • A deployed bot

With the change to Preview in this release, all of the above now pass conversation context to the LLM service.

Keep in mind that, semantically speaking, you should always get the same answer to the same question when the context is the same. But don’t expect the exact wording to be the same every time. That’s the nature of Generative AI at work creating unique content.

Also keep in mind that when using KnowledgeAI’s Answer Tester, no conversation context is involved. So these results will be slightly different.

Search multiple knowledge bases in a single interaction

Want more flexibility when automating answers in a bot? You can now search multiple knowledge bases in a single interaction.


KnowledgeAI interaction that specifies multiple knowledge bases


Integration of type “KnowledgeAI” that specifies multiple knowledge bases


How the search works

Searching multiple knowledge bases works like searching a single one:

1 - Find the article matches in the knowledge bases.

2 - Aggregate the results, and keep only those in one of the following groups (listed in order of priority): exact match, intent match, AI Search match, fallback text match.

3 - Sort the results in the group by confidence score.

4 - Return the top results based on the number of answers requested.

OAuth 2.0 credentials now support PKCE

When creating an OAuth 2.0 credential using the Authorization Code grant type, you can now extend the authorization code grant flow using Proof Key for Code Exchange (PKCE). 

PKCE provides additional security by requiring a code challenge and code verifier to be used in the authorization process. These prevent cross-site request forgeries (CSRF) and authorization code injection attacks.



Voice bots now automatically disregard consumer interruption functionality that applies to Messaging bots

Voice conversations are different from messaging ones, so Voice bots must handle consumer interruptions differently than Messaging bots. Since this is the case, previously you had to manually disable the functionality that applies to Messaging bots, so that the functionality that applies to Voice bots could be used. You did this by adding the system_handleIntermediateUserMessage environment variable to a bot environment, setting the variable to false, and linking the environment to your bot.

As of this release, we’ve fixed things so that this manual step is no longer needed. Voice bots now automatically disregard the consumer interruption functionality that’s meant for Messaging bots. Instead, Voice bots use their own consumer interruption handling functionality by default.



Populate a knowledge base by uploading a PDF from your computer

If you want to populate a knowledge base with the content in a PDF that’s stored on your computer, now you can. This option is now available. (Previously, you had to specify the URL of a PDF in a publicly accessible location: Google drive, Web site, etc.)




Enhanced Navigation Bar

After considering user feedback, LivePerson has addressed some key concerns:

  • Streamlining Access to Frequently Used Applications: Reducing the number of clicks required to access commonly used applications.
  • Mitigating Accidental Hovering in the Navbar: Preventing unintended menu expansion caused by accidental hovering.

Updates to the navigation menu will take place in Feb 2024.

What enhancements would affect the user experience?

Pin frequently used applications

The Agent Workspace and Manager Workspace were identified as the most frequently utilized applications, hence, these applications would now be pinned above all sections. 

Note: The mentioned applications will no longer be available through the original category they used to be available through (Engage). 

Labeled sections and applications

All categories and the pinned applications mentioned earlier will now be labeled, eliminating the need for users to remember which symbols correspond to them.

Menu behavioral change

The navbar won't expand when hovered over; instead, it will expand only when:

  • One of the categories is clicked.
  • The hamburger icon is clicked.

Menu structural change

The navbar categories order was changed, and now it is: Engage, Automate, Optimize, Manage.

The full details about the upcoming change can be found here.



Automatic Encryption Key Rotation

LivePerson customers can enable data-at-rest encryption of conversation transcripts and other engagement attributes. Enabling encryption ensures that data is stored encrypted across the Conversational Cloud datastores and protects the data from potential unauthorised access.

An Automatic Encryption Key Rotation feature has been introduced  that can rotate Encryption Key at the desired interval configured by your LivePerson account team.