Author: Om Kamath

Om Kamath

Slack AI Integration

In today’s fast-paced digital world, the integration of AI into our daily communication tools is not just a luxury—it’s a necessity. Recognizing this need, we’re thrilled to announce the AI Slack Integration feature with Cody. This integration is designed to enhance the Slack experience for businesses and corporations that heavily rely on it for their communication. By integrating Cody bots trained on enterprise documents, users can now enjoy a more streamlined and efficient communication process within their Slack workspaces.

How to Integrate Cody AI With Your Slack Workspace

  1. Add the Cody Bot to your Slack workspace by navigating to your Cody Settings > Integrations and clicking on Install Slack.Slack AI Integration
  2. Obtain the API Key from Cody Settings > API Keys by clicking on Create API Key.
  3. In your Slack workspace, search for your Cody App and set the API Key in the Home section.Slack AI API Integration
  4. Go to any channel in your workspace and use the /assign-bot command to assign a bot from your Cody account to that channel.Select Your Slack AI Bot
  5. To ask questions to your bot, simply type @Cody followed by your question. Cody will create a new thread in the channel to reply to your question. All the messages in that thread will be considered Chat History. If you want to start a new conversation, exit the thread and mention @Cody again.Asking question to Cody Expanded thread

 

The Future of Cody AI Integrations

This AI Slack Integration marks one of our pioneering ventures into third-party application integrations. The overwhelming demand and popularity of this feature among our users have been the driving force behind its inception. And this is just the beginning! We’re currently in the process of developing additional features and integrations, including those forDiscord and Zapier. These exciting updates will be rolled out in the near future.

Your Feedback Matters

Your insights and feedback are invaluable to us. They shape the direction of our innovations and ensure we’re always delivering the best. We invite you to share your thoughts and experiences with this integration. Connect with us on our Discord Server or reach out through the ‘Get Help’ button within our app.

Anatomy of a Bot Personality

Tips for creating a bot that does just what you want.

It’s essential to recognize that when constructing bots that utilize language models, patience is crucial, especially at the beginning. Once you have established a solid foundation, it becomes easier to add additional components. Building bots with Cody is akin to painting on a canvas. It requires a degree of creativity and some understanding of the fundamentals to add your personal touch to the bot.

The main parameter that allows your bot to adopt a particular thinking style is the Personality Prompt. The bot’s personality is shaped by various factors, including token distribution, relevance score, and more. However, the prompt for personality is the most distinct and creative aspect, as it can be customized differently by each user. Users have the freedom to create and fine-tune the bot’s personality according to their specific requirements.

Freedom is something we all appreciate, but when starting with a blank slate, it can also become intimidating and lead to ambiguity regarding where to start. If you have been feeling the same, don’t worry; this blog should help you create a better personality prompt. We will begin with the recommended prompt structure and then proceed to provide some sample prompts.

Name

It is always beneficial to start by giving your bot a name. Naming your bot adds a human touch, especially when greeting users or addressing questions related to the bot.

Prompts:

Your name is [Name of your Bot].
OR
You are ‘[Name of your Bot]’.

Description

The description of the bot makes it aware of the context that will be provided through the knowledge base. Being context-aware provides the bot with a framework for answering questions while keeping a specific domain in mind.

Prompts:

Your primary task is to [specify the domain].
OR
Your main objective is to assist me in [specify the domain].

Note: The Bot Name and Description set in the General Section are only for the user’s convenience in differentiating between multiple bots. The bot itself is unaware of these settings. Therefore, it is necessary to explicitly define the bot’s name and description within the Personality Prompt to establish its identity and characteristics.

Boundaries

One potential drawback of using LLMs trained on large datasets is the tendency to generate hallucinated responses. It’s important to note that the data used to generate responses is not utilized for fine-tuning or retraining the LLM on-demand by Cody. Instead, it serves as a contextual reference for querying the LLM, resulting in faster responses and preserving data privacy.

To ensure that the bot does not refer to data points from the original LLM dataset, which may overlap with similar domains or concepts, we have to delimit the context strictly to our knowledge base.

Prompts:

The knowledge base is your only source of information.
OR
You are reluctant to make any claims unless stated in the knowledge base.

There may be some instances where the bot doesn’t require a knowledge base or uses the knowledge base as a source of reference. In such cases, the prompt will change considerably.

Prompt:

Your primary source of reference is the knowledge base.

Response Features

The features of the response generated by the bot can also be controlled by the personality of the bot to some extent. It can consist of defining the tone, length, language and type of response you expect from your bot.

Prompts:

1. Tone: You should respond in a [polite/friendly/professional] manner.

2. Length: The responses should be in [pointers/paragraphs].

3. Language: Reply to the user [in the same language/specify different language].

4. Type: Provide the user with [creative/professional/precise] answers.

You are free to experiment with various combinations and features. The examples provided are just for your learning purposes, and the possibilities are endless.

Media

One of the most interesting features of Cody — the ability to embed media in the responses. When embedding media such as images, GIFs or videos, it is always recommended to import the media to a separate document or import the entire raw document using the built-in Cody text editor wherein you can add media. You can either copy/paste the media or embed them into the document using URLs.

An image illustrating the media buttons.

After successfully importing the media, you need to specify the same in our bot personality prompt. The prompt can be broken into two parts: Initialisation and Illustration.

Prompts:

Initialisation:
Incorporate relevant [images/videos/both] from the knowledge base when suitable.

Illustration:
Add images using the <img> tag and videos using the <iframe>
For example:
<img src=”[Image URL]”>
<iframe src=”[Video URL]”></iframe>

Fallbacks

There will be times when the bot is unable to find relevant content for the question asked by the user. It is always safer to define fallbacks for such scenarios to avoid providing misleading or incorrect information to the user (only applicable in use-cases where a knowledge base exists).

Prompts:

1. Refrain from mentioning ‘unstructured knowledge base’ or file names during the conversation.

2. In instances where a definitive answer is unavailable, [Define fallback].

OR

If you cannot find relevant information in the knowledge base or if the user asks non-related questions that are not part of the knowledge base, [Define fallback].

Steps (Optional)

If you want your bot to follow a specific conversational timeline or flow, you can easily define it using steps. This approach is particularly useful when using your bot for training or troubleshooting purposes. Each step represents a particular phase or stage of the conversation, allowing you to control the progression and ensure that the bot provides the desired information or assistance in a systematic manner.

Prompt:

Follow these steps while conversing with the user:

1. [Step 1]

2. [Step 2]

3. [Step 3]

Note: While defining steps, it is recommended to enable ‘Reverse Vector Search‘ for improved replies and allocate an adequate number of tokens to the chat history. This allows the model to consider the conversation history, including the user’s input and the bot’s previous response, when generating a reply.

Data Capture (Optional)

This prompt, in harmony with the conversational flow (steps), is particularly beneficial when the use-case of your bot revolves around support or recruitment scenarios. Currently, there is no long-term memory or database connectivity in Cody that can capture the data and store it for analytical consumption. In the future, with newer updates to the OpenAI API like function calling, we will definitely be bringing in newer features to be able to capture and store the data for a longer term.

For now, you can access the chats of your bot users (through widgets) by navigating to the ‘Guests‘ chats in the chat section. You can then manually analyze the captured data for further insights.

Prompt:

Collect the following data from the users:

– [Field 1]

– [Field 2]

– [Field 3]

– [Field 4]

Ask one question at a time. Once you have collected all the required information, close the conversation by saying thank you and displaying the data collected. Remember, your task is only to collect data.

Response Formatting*

A nifty little feature of Cody is its support for formatting bot responses using markdown or HTML tags. By providing your bot with an HTML or markdown format template in the bot personality, it will attempt to format the responses accordingly, whenever necessary.

Prompt:

Response Format:

<h1>[Field Name]</h1>

<p>[Field Name]</p>

<p>[Field Name]</p>

*Formatting works best on GPT-4

Prompt Example

Cody as a Lead Generation Bot

Anatomy of a prompt (labelled).

 

Demo chat displaying the prompt in use.

Cody as a Marketing Bot

Cody as a Training Bot

To read more personality prompts, please check out our use cases, which contain detailed prompts along with their parametric settings.

Conclusion

If you are on the free plan of Cody, there is a possibility that the bot may lose adherence to the prompt or may simply ignore some parameters due to the smaller context window or the lack of coherence. We recommend everyone to use the free plan only for trial purposes or as a transitional phase to understand the use of Cody and determine its suitability for your business.

While constructing prompts for your bot, it is also important to maintain conciseness and avoid incorporating every parameter mentioned in the article. As there is a limit to the number of tokens available, and the personality prompt also consumes tokens, you should construct them judiciously. Feel free to change the prompts given in this article as per your needs and preferences. Discovered something new? You can always share it with us, and we would be happy to discuss it.

This was just an introduction to the vast landscape of bot personality creation. LLMs are continuously improving with each passing day, and we still have a long way to go in order to fully utilize their potential. This entire journey is a new experience for all of us. As we continue to experiment, learn, and implement new use-cases and scenarios, we will share them with you through articles and tutorials. For more resources, you can also check out our Help Center and feel free to ask any questions you may have regarding Cody by joining our Discord community. Also checkout our previous blogs for more such interesting insights.

How Does Cody Generate Responses Using Your Documents?

When you get started with Cody, it is possible that you may feel disappointed or disheartened about why Cody is unable to generate the expected responses. In this short blog, we won’t delve deep into how you should be using Cody, but we will give you a rough idea of how Cody uses your documents to generate responses so that you can better understand the generation process and experiment with it.

Two main factors mainly affect response generation using your documents:

  1. Chunking
  2. Context Window

Both of these terminologies, chunking and the context window, are interrelated. A simple analogy can be drawn to compare response generation with cooking food. Chunks can be seen as the individual pieces of vegetables that you cut, while the context window represents the size of the cooking utensil. It is important to cut the vegetables into optimal-sized chunks to enhance the overall taste, and a larger utensil allows for more vegetable pieces to be added.

What is Chunking?

In simple terms, chunking is the act of breaking down content into manageable pieces for efficient use of memory. If you have read our blogs, you may be aware that models such as GPT require significant resources, and to tackle the constraints of the context window, we employ multiple processes like chunking.

Chunking is a process performed after you upload the documents to Cody. It divides or segments the document into multiple chunks, with each chunk containing relevant surrounding context. These chunks are then assigned numerical tags for easier computation, which is known as embedding. It is important to find the optimal chunk size. A smaller chunk size reduces context relevance, while a larger chunk size introduces more noise. Cody’s chunking algorithm dynamically adjusts the chunk size based on the token distribution set by the user.

How does the Context Window affect the Bot responses?

Various factors such as personality prompt, relevance score, etc., influence the quality of bot responses. The context window of the model also plays a significant role in determining the quality. The context window refers to the amount of text that an LLM (Language Model) can process in a single call. Since Cody utilizes embeddings and context injection to generate answers using OpenAI models, a larger context window allows for more data ingestion by the model in each query.

💡 Each Query (≤ Context Window) = Bot Personality + Knowledge Chunks + History + User Input + Response

Context windows of different models:

  1. GPT-3.5: 4096 Tokens (≈3500 words)
  2. GPT-3.5 16K: 16000 Tokens (≈13000 words)
  3. GPT-4: 8000 Tokens (≈7000 words)

When the context window is larger, it enables a greater proportion of each parameter, including Personality, Chunks, History, Input, and Response. This expanded context empowers the bot to generate responses that are more relevant, coherent, and creative in nature.

Cody’s latest addition enables users to check document citations by clicking on the document name at the end of the responses. These citations correspond to the chunks obtained through semantic search. Cody determines the chunk threshold for context based on the relevance score set by the user. If the user sets a high relevance score, Cody only uses chunks that surpass a predefined threshold as context for generating the answer.

Example

Assuming a predefined threshold limit value of 90% for a high relevance score, Cody discards any chunks with a relevance score lower than 90%. We recommend new users start with a lower relevance score (low or balanced), specifically when using uploaded documents (PDFs, Powerpoints, Word, etc) or websites. Uploaded documents or websites may encounter formatting and readability issues during pre-processing, which can result in lower relevance scores. Formatting the document using our built-in text editor instead of uploading raw documents will ensure the highest accuracy and trust score.

Illustrates how the relevance score affects the context chunks.

If you found this blog interesting and wish to delve deeper into the concepts of context window and chunking, we highly recommend reading this blog written by Kristian from All About AI. For more resources, you can also check out our Help Center and join our Discord community.

How to Build an AI Website Chatbot Widgets

A step-by-step guide to building your own AI to meet your business needs.

The applications of bots are vast, and with the rise in bot utilization of generative models, it can be tempting to jump on the bandwagon and build your very own ChatGPT. With the launch of Cody, building AI chatbots has never been easier before. The model-agnostic and no-code interface of Cody tries to make personal AI accessible to every individual and business domain that exists. However, although Cody strives to make the entire process as straightforward as possible, it can be a little intimidating for someone who is new to the world of AI. That’s why, in this blog, we provide an end-to-end explanation of how to build your own ChatGPT using Cody. We will walk you through a use case, helping you kickstart your Cody journey with confidence.

The Use Case

We will be building an AI chatbot using Cody that will act as a tour guide. This will be a valuable addition for tourism-related businesses, as it can greatly improve the customer experience. But, the scope of building a chatbot with Cody is not limited to just one business domain. Users can experiment with different prompts to build a bot catering to their own business and also share it with their customers after reading this blog.

Requirements

  1. Your business-related content (PDF, Word, Website or Text)
  2. Account on Cody (Recommended Plan: Premium)
  3. 15 minutes

Step 1: Adding data to the knowledge base

To add documents to train Cody for your business needs, follow these steps:

  1. Go to the Content section of Cody’s interface.
  2. Create a new folder to organize your documents.
  3. There are three methods available to add data to the knowledge base:
    • Write: Utilize the built-in text editor to directly input content.
    • Upload: Upload PDF, Word, or PowerPoint files containing relevant information.
    • Import Website: This feature allows you to crawl your business website and automatically collect data. (Note: This feature is available on Premium and Advanced Plans only.)

For this specific tutorial, we will use a Boston guidebook that includes information about various tourist attractions, their descriptions, hours of operation, addresses and frequently asked questions (FAQs).

Source Document Adding documents to Cody

After uploading the necessary documents, you will be able to track their status, which can be either “learning” or “learned.” The status indicates whether Cody is currently in the process of learning from the documents or if it has successfully completed the learning process. Once Cody has learned the uploaded documents, you can proceed to the next step, which involves building the actual chatbot.

Note: It is crucial to ensure that the documents you provide follow a specific format that enhances the quality of responses generated by Cody. To learn more about the recommended document formatting, we recommend reading the documentation provided on formatting guidelines.

Step 2: Creating the Bot’s Purpose

The purpose is further divided into 3 parts:

Personality

The Personality acts as the brain of the bot and plays an essential role in generating effective and useful responses. It is made up of multiple parameters such as Prompt, Relevance Score, Token Distribution, Prompt Persistence and Reverse Vector Search. For a detailed explanation of these terminologies, we recommend referring to our previous blog, which provides comprehensive insights for those new to Cody. Cody provides two options to the user for creating the personality: Template Mode and Advanced Mode.

Template Mode provides users with pre-defined prompts, offering a convenient plug-and-play solution. For this tutorial and use-case, we will be limiting ourselves to using the Advanced Mode since the use-case is not currently covered in the Template Mode. In the future, we plan to create a library of prompts with their specific use cases, providing even more options for users and also adding more presets to the Template Mode.

Personality for the AI tour guide:

Prompt

You are Cody Travel Guide, an AI Assistant dedicated to providing accurate information on Boston. Your primary task is to assist me by providing me with reliable and clear responses to my questions, based on the information available in the knowledge base as your only source. Refrain from mentioning ‘knowledge base’ or file names during the conversation. You are reluctant to make any claims unless they are stated or supported by the knowledge base. When a definitive answer is unavailable, acknowledge your inability to answer and inform me that you cannot respond. Instead of saying ‘the information is based on my knowledge base’ just say ‘Excuse me if I missed out anything’.

Try to use the following format to display the information:

<h2> Name of Place </h2>

<strong> Description: </strong> <br>

<strong> Address: </strong> <br>

<strong> Working Hours: </strong> <br>

<strong> Additional Information: </strong> <br>

Go to the <a href=’Insert link’>website</a> to know more.

If asked for custom tour plan or itinerary, create one using the timings and description provided in this format:

Name of place: <br>

Time to visit: Start time-End Time <br>

Address: <br>

Relevance Score: Balanced

Token Distribution: 70-10-20 (Context-History-Response)

Persist Prompt: On

Reverse Vector Search: Off

 

💡 If you want to format your bot responses in a certain way, Cody supports Markdown and HTML tags, where you can define a formatting template for your bot.

 

Every personality prompt must follow a specific structure to get quality responses. Important tips for engineering an effective prompt with examples:

Define the fallbacks:

  • “Instead of saying ‘that the information is based on my knowledge base’ just say ‘Excuse me if I missed out anything’.”
  • “If you cannot find relevant information in the knowledge base, acknowledge your inability and suggest contacting [Enter your website URL/contact no.]”

Delimit the sources:

  • “Use the knowledge base as your only source.”
  • “Your response must be in the same language as my request, and should not mention the word knowledge base during conversation.”

Define the goal:

  • “Your goal is to provide me with assistance and answer my questions about [Enter your business domain].”
  • “Avoid unrelated activities or engaging in non-knowledge base related discussions or creative writing.”

Sample boilerplate prompt for customer support:

 📌 You are Cody, a Website Support AI Assistant who works for Villa Homes. Your goal is to provide me with support and answering my questions about Villa Homes, using the knowledge base as your only source.

Avoiding unrelated activities or engaging in non-knowledge base related discussions or creative writing. If you cannot find relevant information in the knowledge base or if the user asks non-related questions that are not part of the knowledge base, acknowledge your inability and suggest that I contact Villa Homes’s team at +40XXXXXXX

Your response must be in the same language as my request, and should not mention the word knowledge base during conversation.

To know more about engineering better prompts, read this blog where we explain all the fundamentals of Cody in detail.

General

In this section, you will be required to define the name and description of your bot, as well as select the model you want to use with Cody. You have three different OpenAI models to choose from:

  1. GPT-3.5
  2. GPT-3.5 16K
  3. GPT-4

With the Premium Plan, you have access to all three models, providing you with the freedom to choose the most suitable one for your specific needs. It offers a practical choice, allowing you to leverage the full range of capabilities offered by Cody.

General Settings of the bot

We will be proceeding with GPT-4 since it is the most advanced model currently available and perfectly suited for the use-case we will be demonstrating due to its enhanced creativity.

Knowledge

The folder created in the first step should be linked to your specific bot in this section. If you have created multiple folders that you want to link, simply select all the folders you would like to train Cody on.

It is important to understand that Cody is not a black box that has a mind of its own. It is a model that simply predicts the next word in the sentence. As the famous saying in data science goes, “Garbage In, Garbage Out.” The better you set the personality, define the rules of the bot, and clean the data you provide, the better responses it will generate. With newer additions such as Template Mode with various presets, we are trying to make the personality prompting process a lot easier for users.

Step 3: Testing and Sharing the Bot

Now comes the most exciting part! Trying out and testing the bot you’ve built yourself can give you a great sense of achievement. There are two ways to test and share your bot: creating a chat or using the newly launched customizable widget.

To create a chat widget in just three simple steps:

  1. Go to the Bots section of Cody’s interface.
  2. Select the bot you created and click on the three dots ‘⋮’ for more options.
  3. Click on ‘Embed‘.

And voila! You have successfully created a customizable Cody Widget.

When using a Cody Widget, you have two options to share it with others: Share Link or Embed. If you don’t have a website for your business or if you’re not comfortable with modifying your website code, you can easily share the bot with your customers using the provided link.

Shareable link for Bot

There are two types of embeddable widgets available:

  1. Inline Embed: This type of widget will be embedded within a specified space on your website. It will occupy the necessary space as required by the widget within the defined main element.
  2. Pop-Up Embed: This type of widget appears as a floating icon on your website. When a visitor clicks on the icon, the widget expands and opens up, allowing for interaction with the chatbot.

Both types of embeddable widgets provide different ways to integrate the Cody bot into your website, offering flexibility in terms of space utilization and user experience. You can choose the one that best suits your website design and user interaction preferences.

Pop-Up Embed Inline Embed

Customizing the widget

Our latest feature release offers users complete freedom and customization options for widgets. This means you can tailor the widget’s appearance, behavior, and overall user experience to align with your specific requirements and branding.

Widget Tool

Customization options include:

  1. Header
    • Change the layout of header (Left or Center).
    • Add your business logo
    • Color
    • Title
    • Subtitle
  2. Chat
    • Message size (Chat bubble size)
    • Background color of the chat
  3. Bot
    • Initial Messages
    • Message background color
    • Bot Avatar
    • Bot Colour
  4. Human
    • Suggested questions
    • Message background color
  5. Composer
    • Placeholder message
    • Send button icon
    • Cody branding (Can be removed only in Premium and Advanced Plans)
  6. Launcher
    • Size
    • Screen Position
    • Background Color (Floating button color)
    • Icon
    • Close Icon

Widget Labelling

 

All these customization options should make Cody more personalized and in line with your business aesthetics resulting in a much improved user-experience for your customers.

That’s about it!

Bot Demonstration 1 Bot Demonstration 2

Building and sharing a bot using Cody is now easier than ever, requiring just three simple steps. With the recent introduction of widget customizability options, there’s never been a better time to create your own bot using Cody. The added flexibility in customizing the widget allows you to tailor the bot’s appearance and behavior to match your specific business needs and branding.

We are constantly working on developing and introducing more exciting features to enhance the Cody experience. Our team is dedicated to making the entire process of building and deploying bots even more seamless and user-friendly. Stay tuned for upcoming updates as we continue to improve and refine the Cody platform. For more resources, you can also check out our Help Center and join our Discord community.

The Power Of GPT-3.5 16K

Should you upgrade to the paid version of Cody? Here’s why you might want to.

A few days ago, we released a newer model for all our paid users right after the OpenAI release: GPT-3.5 16k. As intimidating as it might sound, it could be a game-changer for your business. In this blog, we will delve into the use cases of GPT-3.5 16k, explore its advantages, and highlight how it differs from the existing GPT-3.5 model and the latest higher-end GPT-4.

What is GPT-3.5 16K?

If you have used Cody’s free version before, you might already be familiar with the vanilla GPT-3.5 model, which utilizes OpenAI’s gpt-3.5-turbo model. This model is the popular choice for many users as it offers affordability, speed and reliability in most cases. On the other hand, the GPT-3.5-16k utilizes OpenAI’s gpt-3.5-turbo-16k model, which is an extension of the gpt-3.5-turbo. The significant difference lies in the ’16k’ aspect.

What is 16K?

The ’16K’ suffix indicates that the model has a context window of 16,000 tokens, a significant increase from the existing 4,096 tokens. In our previous blog, we explained what tokens are in detail. A smaller context window in models can result in several limitations, including:

  1. Lack of relevance: With a limited context window, the model may struggle to capture and maintain relevance to the broader context of a conversation or task.
  2. Inability to maintain context: A smaller context window can make it challenging for the model to remember and reference information from earlier parts of a conversation, leading to inconsistencies and difficulties in maintaining a coherent dialogue.
  3. Constraints on input-query lengths: Shorter context windows impose constraints on the length of input queries, making it difficult to provide comprehensive information or ask complex questions.
  4. Knowledge-base context limitations: A smaller context window may face limitations in incorporating knowledge from relevant documents from the knowledge due to the data ingestion limit.

Advantages of a bigger context window

A question might arise in the minds of a few people: How is GPT-3.5 able to process over 1000 web pages and documents on Cody, despite its capacity of only 4096 tokens? With advancements in the field of generative AI, providing context does not mean simply feeding the entire document as is to language models like GPT-3.5 Turbo. Backend processes such as chunking, embeddings, and vector databases are utilized to preprocess the data, maintaining relevance within the chunks and allowing the model to navigate the predefined context window.

In the current scenario, a larger context window would enhance the AI’s overall performance by accommodating larger and more complex inputs, while reducing the number of vector-storage transactions required to generate a response. Since the context window encompasses both the input and output, a larger window would also enable the model to provide elaborate and coherent responses while maintaining the conversational context.

A larger context window would also help mitigate any hallucinations that may occur when the token limit in a conversation is exceeded.

GPT-3.5 Turbo 16K v/s GPT-4

Although gpt-3.5-turbo-16k is the latest release from OpenAI, gpt-4 still outshines it in various aspects such as understanding visual context, improved creativity, coherence, and multilingual performance. The only area where GPT-3.5-16k excels is the context window, as GPT-4 is currently available in the 8k variant, with the 32k variant still being gradually rolled out.

In the meantime, until the 32k version of gpt-4 becomes widely accessible, the GPT-3.5-16k stands out with its larger context window. If you’re specifically looking for a model that offers a more extensive context window, the GPT-3.5-16k is the ideal choice.

Use cases of a bigger context window

  1. Customer Support: A larger context window enhances the short-term memory of the model, making it well-suited for applications involving customer support, form-filling, and user data collection. It enables the model to maintain context over a longer period, leading to more relevant responses to user inputs such as names, customer IDs, complaints, and feedback.
  2. Employee Training: Leveraging Cody for employee training purposes proves highly effective. Employee training often involves extensive data related to business activities, steps, and processes. To maintain contextual relevance throughout the training program, it becomes necessary to incorporate the entire conversation history of the trainee. A larger context window allows for the inclusion of more historical information, facilitating a more comprehensive and effective training experience.
  3. Data Analysis: Tasks involving financial analysis and statistical inference often require processing large volumes of data to derive meaningful insights. With a bigger context window, the model can hold more relevant information during computation, resulting in a more coherent and accurate analysis. For instance, comparing balance sheets and the overall performance of a company year over year can be executed more effectively with a larger context window.

Comparison between GPT-3.5 4K v/s 16K

To demonstrate the improvements on the 16K model, we queried a .csv file of the Periodic Table containing 118 elements and their features.

GPT-3.5 4K

GPT-3.5 4K

GPT-3.5 16K

GPT-3.5 16K


From the comparison, it is visible that GPT-3.5 4K was unable to produce all the radioactive elements in its response and skipped some elements. In contrast, GPT-3.5 16K produced almost all the radioactive elements present in the provided table. This demonstrates the improved elaboration of responses due to the bigger context window. This was just a small glimpse into the potential the 16k context window holds with there being endless applications and implementations of the same. With GPT-4 32K in the pipeline, the 16K model can facilitate a smoother transition to a bigger context window.

Should you upgrade?

The larger context window is undoubtedly a significant update and not merely a gimmick. Improved context understanding plays a crucial role in enhancing the quality of responses, and a larger context window unlocks substantial potential for these Language Model Models (LLMs). By allowing for a more extensive grasp of the conversation history and contextual cues, LLMs can deliver more accurate and contextually appropriate outputs.

As mentioned earlier, the 16K variant of GPT-3.5 is available to all users starting from the Basic Plan. If you have been using the Personal (free) plan for some time, you would have already experienced the capabilities offered by Cody. The Basic Plan provides excellent value for money, particularly if you don’t require the additional features of GPT-4. It is suitable for individuals who are building a bot as a project or a prototype for their business with the added model choice of GPT-3.5 16K. In the future, when we release the GPT-4 32K variant, you can always upgrade to the premium plan when the need arises for more tokens.

For larger enterprises, the Advanced Plan is the most powerful option, catering to resource-intensive and high-volume usage requirements. It offers comprehensive capabilities to meet the demands of large-scale operations.

Pricing structure of Cody

Additional Resources

Help Center

Discord

Cody Terminologies 101

A beginner’s guide to understanding Cody Terminologies in a simplified manner.

Before you get started with building your own GPT-based bot for your business using Cody, it is essential that you understand the core terminologies so that the flexibility our bot-builder offers does not intimidate you. You do not need to be an AI expert or have prior knowledge in this domain to understand these terminologies, as they will be explained in the most simplified way. Think of this as a beginner’s guide or ‘Cody For Dummies‘.

This blog will not cover setting up Cody or importing data into Cody since we have already posted a few blogs that will help you with the setup. In the upcoming blogs, we will delve deeper into those subjects. Without any further ado, let’s understand the different terminologies you will come across while building your bot.

Personality

Cody offers two ways to define your bot’s personality: Template Mode and Advanced Mode. The personality of your bot establishes the context and behaviour of your bot to ensure it generates helpful responses that align with your expectations and prevent any unforeseen hallucinations.

❓Hallucination refers to the phenomenon where a language model generates responses that may be imaginative, speculative, or not grounded in factual information.

Template Mode

This mode is useful for those businesses who want a plug-and-play solution to building a bot. Our developers have created bot templates catering to different business domains such as Support, Training, HR and Marketing.

Cody Template Mode Personality

Select any domain from the following and Cody will handle the rest. For added flexibility, Cody also provides settings that can be used to customize the personality to some extent.

For eg. If you choose to go ahead with IT Support as your bot personality, you can change your bot’s nickname, and decide how it will be handling unanswerable queries (a default case) and the languages you need the bot to converse in.

Cody Template Mode Settings

This mode is still under development and will be soon getting updated to be able to cater to multiple other business domains with a host of customization features.

Need more flexibility? Don’t worry, we have got you covered with our Advanced Mode.

Advanced Mode

The Advanced Bot Builder gives you the freedom to customize multiple parameters and lets you build the perfect bot that suits your use case. Currently, you can customize the following parameters:

  1. Prompt
  2. Relevance Score
  3. Token Distribution
  4. Persistent Prompt
  5. Reverse Vector Search

Prompt

The prompt will define the personality of your bot. To simplify the prompting process, consider the bot as an employee at your business. Although there is no specific structure for writing a personality prompt, we have prepared a list of parameters for your reference.

A sample prompt:

“You are the waiter at a pizza joint. Maintain a professional and friendly demeanor throughout all interactions, ensuring that users feel comfortable and supported. Remember to convey a sense of expertise and confidence in your responses. Additionally, I encourage you to actively promote our premium pizzas whenever appropriate. Do not refer to any menu sources other than the ones provided in the knowledge base. While recommending pizzas, state their prices and any offers that are applicable for them, too.”

Relevance Score

The Relevance Score reflects the degree of similarity between the user’s query and Cody’s response. Using semantic search, Cody compares the user’s query with the data present in the knowledge base. A higher relevance score will result in a precise answer but will compromise on understanding the overall context of the query and vice versa. In simple words, the relevance score is the degree to which the AI fears making mistakes and taking risks while responding.

Relevance Score Slider

Token Distribution

The token is the computational currency for large language models like the GPT family. The query (input statement) asked by the user is broken down into blocks of characters known as ‘tokens’. As AI models are really resource-intensive, in order to address the computational constraints and memory limitations, these models have a certain limit to the input data that can be processed and generated. This limit is the ‘context window’.

Context Window Illustration

Cody uses the GPT family of models and the number of tokens available are limited. The token distribution feature helps in micro-managing the usage of tokens for different purposes.

They are mainly divided into Context, History and Response Generation.

  1. Context: The tokens required for understanding the user query and knowledge base context.
  2. History: The tokens required for adding context to the user query using the chat history.
  3. Response Generation: The tokens required for assessing the coherence, grammar, and semantic validity of the generated text.

Token Distribution Slider

For the highest accuracy is important that the context makes up a large portion of the Token Distribution.

Persistent Prompt

By continuously reinforcing the prompt (personality of the bot), you create a form of conversational context and constraint that keeps the AI on track and helps maintain compliance with the desired outcomes. It acts as a reminder for the AI to stay within the predefined boundaries and provide responses that are relevant, accurate, and aligned with your goals.

Reverse Vector Search

Reverse Vector Search is a feature specifically designed for businesses that utilize Cody for employee training purposes. It seamlessly integrates previously generated AI responses with user queries, creating a comprehensive contextual understanding that enhances the quality and relevance of subsequent responses. It is especially valuable when using Cody for step-by-step guides and instructional purposes.

 

This image describes how reverse vector search works.

 

To Wrap It Up

When you have a good grasp of these terminologies, you’ll be able to refine your bot’s responses and make them even more personalized. Stay tuned for upcoming blogs offering further guidance, updates, and additional information to unleash your Cody AI bot’s full potential. For more resources, you can also check out our Help Center and join our Discord community.