AI Coder - Remark Labelling
Introduction
When evaluating products, it is important to ask panellists to provide argument-specific answers or to write down what they think about a product in their own words. However, making sense of a large dataset in which individual comments were collected may be difficult. Traditionally, remark labeling has been a manual process involving the creation and assignment of labels to corresponding remarks. While effective to some degree, this method can be time-consuming. Additionally, as data volumes grow, manually labeling remarks becomes increasingly laborious.
To address this requirement, an innovative feature has emerged: the AI Coder for Remark Labeling. This cutting-edge innovation promises to transform how notes are processed and sorted, enhancing productivity and saving time through the use of AI.
How it Works
Summary Generation: a Large Language Model(LLM) are used on a subset of remarks to provide a summary of the general themes present.
Individual Labeling: a Large Language Model(LLM), combined with a predefined set of labels (such as "sweet" or "sour"), labels each remark individually. Users can choose from our provided label sets or create their own custom sets.
- Sentiment Analysis: a Large Language Model(LLM) always assigns one of three sentiment-based labels to each remark: bad, neutral, or good.
At this moment we use GPT(Generative Pre-trained Transformer) as our default LLM, the GPT-N series is developed by OpenAI and is based on a deep learning architecture called the Transformer.
Here's a simplified explanation of how GPT works:
- Data Collection: GPT is trained on a large dataset of text from various sources on the internet. This dataset contains a wide range of topics and writing styles, providing the model with a diverse understanding of language.
- Tokenization: When you input text into GPT, it first breaks down your input into smaller chunks called tokens. These tokens are the basic units of text that the model processes. On average 1000 words will be 750 tokens.
- Contextual Understanding: GPT then uses its deep learning architecture to analyze the tokens in the context of each other. Unlike older language models, GPT models can understand and generate text based on the broader context of the conversation, not just the immediate previous message.
- Prediction: Based on its understanding of the context provided by the input tokens, GPT generates a response by predicting the most probable next tokens in the conversation. It does this by assigning probabilities to each token in its vocabulary and selecting the most likely ones based on the context.
- Generation: Finally, GPT generates a response by selecting the tokens with the highest probabilities. These tokens are then assembled into coherent text and returned as the model's response to the user's input.
- Iterative Learning: The model continually improves through a process called fine-tuning. This involves training the model on additional data and fine-tuning its parameters to improve its performance on specific tasks or domains.
Overall, GPT works by understanding the context of a conversation and generating text based on that understanding, leveraging its extensive training on a large dataset of diverse text from the internet.
NOTE:
- Remark or text question's in EyeQuestion that were flagged to be Pii information will not be sent to OpenAI.
- Product information that is setup in EyeQuestion will not be sent to OpenAI.
- All remarks selected are sent to OpenAI for processing. To protect personal information, you can exclude them from the analysis specific remarks where panellist have included personal information.
Related Articles
How to Use Remark Labeling
Introduction Open-ended comments provide valuable insights into the consumer's perception of a product, and help in identifying areas for improvement. However, analyzing such comments can be time-consuming and challenging. With remark labeling, the ...
Introduction to EyeQuestion 6
We could not be more excited to announce the release of our latest software update, version 6, packed with new features designed to enhance user experience, including: Improved Panel Management: Streamline workflows with our enhanced panel management ...
Remarks Representation
Purpose The remarks representation method can be used to display remark and comment data. Options Split Results On: Choose to split your results on either Judge, Attribute, Product or Session. Alternatively, choose None if you do not wish to split ...
Text Highlighting
Introduction The text highlighting is a novel method for explicit attitude measurement. This method leverages the natural, intuitive action of highlighting to gauge participants' attitudes. Participants are shown a piece of text and are asked to use ...
How to Display Feedback of Triangle and Tetrad Test
Introduction During a Triangle test a panellist is presented with three samples of which two are equal and one is different. A panellist must point out the odd sample. A user can add the screen condition in the triangle test to display message when a ...