Prompt - QA - Guidelines For Multilingual Prompts - PII
Prompt - QA - Guidelines For Multilingual Prompts - PII
1. Introduction
You are tasked with performing quality assurance for multilingual prompts targeting an LLM-powered AI chatbot. The quality
assurance will be performed across a few dimensions: Naturalness, Clarity, Metadata, and Language/Locale. You will be asked to
make a quality judgement for every row of data and create a report from that data.
The data you’re reviewing has either been machine translated and post-edited or written from scratch. We won’t tell you which, since
the data quality expectations are the same, but this could give you context to understand some of the errors you see.
You will need to reference a few other documents to perform this task. Those will be included with your materials.
1. Review the text presented to you and analyze it to determine if there are any errors related to the categories of
Naturalness, Clarity, Metadata, and Language/Locale.
2. In the platform you will be provided with, you will need to select the issues related to the above categories.
3. If you don’t find any errors, you can select “No errors”.
In the platform, you will need to select a specific option if you find issues related with the above categories.
See the below:
Category If there are issues with the below… … you should select "Yes" for
Clarity Intentful and complete prompt Is the prompt intentful and complete?
IMPORTANT: For every issue you select, you should provide a short description of the issue IN ENGLISH. You can refer to
parts of the localized text, but DO NOT WRITE YOUR COMMENTS ENTIRELY ON YOUR OWN LANGUAGE!
A more detailed description of the categories and issues can be found below:
The entire prompt should sound natural. You should select “No” for “Does the prompt have natural phrasing?“ if parts of the prompt
sound awkward or unnatural. See an example from English of a prompt that should be marked as “No” in this dimension:
Prompt:
Complete this review: At first the song is comfortingly max, but in an unexpected move for Mitski, it softens at the ends; the music
stills and dulls like slow internet.
Notes: "comfortingly max", "stills and dulls like slow internet" are not natural sounding English
The prompt should have appropriate grammar and orthography. You should select “No” for “Does the prompt have natural grammar
and orthography?”
See a few examples from English of something that should be marked as “No” in this dimension:
Prompt:
How many liters in gallon?
Notes:
The isn’t a grammatically correct sentence. It should say “How many liters are in a gallon?” Even though the phrasing is
understandable and even a common way to phrase something in online text, this dimension values absolute grammatical
correctness
Prompt:
What’s the third letter in the cyrrilic alphabet
Notes:
wrong spelling of Cyrillic
The prompt should have a clear intent and shouldn’t obviously missing major content. If it does not have clear intent or it is missing
major content, you should select “No” for “Is the prompt intentful and complete?”
Flag this error if the prompt doesn’t make it clear what it’s asking the chatbot to do, or if it references something that should be
included for context but is not.
See a few English examples of errors that should be marked as “No” in this dimension:
Prompt: Based on the context, who was the first elf to arrive to Middle Earth?
Be careful of things that have implied intents. The below should be considered correct:
Prompt:
User: Tell me the capital of California
Bot: Sacramento
User: Florida? Bot:
Notes: This is intentful – it’s clear what the user expects from a bot completion
See section 2 (task_category) of the document Chatbots: Terminology and Concepts Addendum
The prompt should clearly fit the task category that is labeled in the metadata. If it doesn’t, you should select “No” for “Is the task
category correct?”.
IMPORTANT: Some prompts could fit multiple categories. We don’t care if there are missing categories, only that the labeled
category is accurate. Conversation is a clear example of this - most prompts that fit the conversation category could also fit another
category. However, they do not need the secondary category to be labeled in order for the prompt to be considered correct in this
regard.
Below you will be able to find examples of prompts and responses that fit the different categories. Please study them when
assessing issues in this category.
A direct question
requesting for
information on the
given topic.
Open-ended QA: The
question cannot be
answered with
What is the top-selling product at The top-selling product on
QA information in the
Amazon.com? Amazon.com is Kindle.
prompt. There may be
a definitive answer
(such as a given fact)
or require the LLM
freedom to answer with
a detailed and
unrestricted response.
Task type Definition Prompt Response
A prompt requesting
the AI model to classify
or categorize specific
Is this review positive? I recently purchased
content based on the
the XYZ Smartphone, and I've been
Classification given topic. You must Yes, the review is positive.
thoroughly impressed with its performance
include the test to be
and features.
classified in the prompt
itself.
Task type Definition Prompt Response
“MoouuRNiiiiNG
DooooVeeeeS aaRee LiiiiGHT
GRaaaaY aaND
BRooooWN aaND
GeeeeNeeeeRaaaaLLY
MuuuuTeeeeD iiN
CooooLooooR.
MaaaaLeeeeS aaND
FeeeeMaaaaLeeeeS aaRee
Rewrite the following passage in SiiiiMiiiiLaaaaR iiN
STReeeeTCH style: aaPPeeaaRaaaaNCee. THee
A prompt request for Original: “Mourning doves are light gray and SPeeeeCiieeS iiS
specific text (sentence, brown and generally muted in color. Males GeeeeNeeeeRaaaaLLY
paragraph, etc) to be and females are similar in appearance. The MooooNooooGaaaaMoouuS,
Rewrite rewritten, paraphrased. species is generally monogamous, with two WiiiiTH
The text for rewording squabs (young) per brood. Both parents TWoo SQuuaaBS (YoouuNG)
must be provided in the incubate and care for the young. Mourning PeeeeR BRooooD. BooooTH
prompt. doves eat almost exclusively seeds, but the PaaaaReeeeNTS
young are fed crop milk by their parents.” iiNCuuuuBaaaaTee aaND
STReeeeTCH: CaaaaRee FooooR THee
YoouuNG. MoouuRNiiiiNG
DooooVeeeeS eeaaT
aaLMooooST
eeXCLuuuuSiiiiVeeeeLY
SeeeeDS, BuuuuT THee
YoouuNG aaRee
FeeeeD CRooooP MiiiiLK BY
THeeiiR
PaaaaReeeeNTS.”
A prompt requesting a
translation of a given
word, sentence or text.
"I love you" translates to "Je
Translate The text for translation Translate "I love you" to French
t'aime" in French.
and the target
language need to be
provided in the prompt.
A prompt requesting to
generate a
Agent: How can I help you?
conversation between
Conversation Human: I want to check my account Agent: Let me check
two or more parties
balance.
given a specific
topic/subject matter.
IMPORTANT: If the Task Category is unavailable or the category only saying “Any”, please DO NOT select “Wrong task
category”. It cannot be validated as an error. If it is “Any (XXXX)”, where XXXX is the name of a category, review normally.
Do not mark as “No” in this dimension Review normally
Chain-of-thought
These prompts need to ask the chatbot to provide some kind of reasoning. Sometimes people misconstrue this as a request for the
bot to list something step-by-step. These are not always the same thing. Explaining how something can be done is not the same as
explaining why the chatbot gave the answer that it gave. Chain-of-Thought prompts should be trying to get at how or why the model
arrives at the answer it chooses
Prompt: What is the best way to drive from Austin to San Jose? Give me step-by-step instructions → Not
matching the Chain-of-thought category.
Prompt: What is the best way to drive from Austin to San Jose? Why is this the best route? Explain how you chose the route →
This does belong in the Chain-of-thought category.
REMEMBER: Chain-of-Thought prompts should be trying to get at how or why the model arrives at the answer it
chooses.
A classification task must give some information about the classes that the bot can choose from.
Tasks could break down into open vs. closed-ended. This data review task may or may not make that distinction. If the metadata
label is “classification”, assume that it could be either open or closed-ended classification.
Prompts will be grouped into length categories depending on their word count. If the prompt has the incorrect amount of words, as
indicated by the prompt length metadata label, you should select “No” under “Is the prompt the required length?”.
The prompt should be in the correct language. If it’s not, you should select “No” under “Is the language correct?”.
We allow for some code-switching and inclusion of words from other languages. If the prompt uses a small number (<5%) of words
from another language, and they are used in a way that makes sense, consider it correct and don’t flag it.
Prompts should always contain example(s) of the PII type(s) listed in the request. The request may list one or multiple PII
types. If one or multiple PII types requested are missing, select “No” under “Does the prompt include at least one example of
each PII type listed?”.
For more information, please check the “Annotation_Multilingual Guardrails Fake PII_Agnostic” guidelines.
3. Appendix:
PII guiding principles
Below you will find the aspects you need to consider when considering if the PII is valid. Please keep this open during your work as
a quick reference.
NOTE: In the below sections we will be using the below tagging formatting for the examples, just for clarity and simplicity:
PII means Personal Identifiable Information. There are different categories of PII, some of which may be more general, or domain
and locale-specific. For this task, one of the things you will need to do is evaluate whether the PII found in the prompt are valid
examples of the pieces of PII requested.
This is wrong because the category of the PII is not accurate: Jean Hébert is a name in this context, not his username.
You must read the prompt and the context and validate whether the PII texts are correctly specified. See “Fake PII categories” for an
explanation of PII category labels for this task.
You can also check the language-specific taxonomy files that will be provided to you.
IMPORTANT: Please note that the PII categories can be locale-specific as well.
These apply universally across all regions and languages and are common forms of PII found in various domains.
Examples:
Note the format of the GLOBAL PII may still vary from country to country. For instance, the format of PHONE and
BANK_ACCOUNT_NUMBER will differ from one country/locale to the next.
As such, please follow the formats provided to you in the PII types files (taxonomy files).
LOCALE-SPECIFIC
Locale-specific PII vary based on country-specific formats, regulations, or identifiers and require adherence to regional data
structures. Locale-specific PII can be seen in the PII types files (taxonomy files) shared with you.
For instance, there are specific labels for license plates and National IDs for each country; below you can see the examples for
Portugal and Belgium.
IMPORTANT: Please note that the text can be referring to content in other locales. In that case, you should consider the context:
consider whether locale-specific PIIs are valid and present.
For instance, let’s take a text for the ko-KR locale, but there’s a reference to a PII that belongs to ja-JP:
• Locale: ko-KR
• PII Types: PASSPORT_NUMBER_KO, ADDRESS, LICENSE_PLATE_KO
• Prompt: "I am currently in {Japan|ADDRESS}, but I need to renew my Korean passport. If the number is
{M48528984|PASSPORT_NUMBER_KO}, when should I go? Will I have issues if I enter the embassy with a car with a
Japanese plate, in this case 11-24?"
As you can see, the prompt above is for Korea, but it has 2 PII labels: one specific to Korea (PASSPORT_NUMBER_KO) and one
global (ADDRESS). Note that there is a PII entity specific for Japan, “11-24”, a Japanese license plate (which would be
LICENSE_PLATE_JP for the ja-JP locale). This means the “LICENSE_PLATE_KO” is missing, and PII Type should be flagged as
“No”. See Important Notes on Locale-Specific Annotation for more information.
For locale-specific PII entity information, please consult the list of PII entities for each language.
Note that the annotation labels are separated into “locale blocks”, in which you will find the locales relevant the language
you are working on.
In your locale, you should see PII that are either GLOBAL or the ones that fit into the block your locale you’re working on.
If you're working on de-DE, you would see both Global PII, as well as PIIs for de-CH and de-AT. You shouldn’t see, for
instance, Italian PII.
For instance, this means that if you are working on de-DE text, you will have de-DE, de-CH, de-AT and GLOBAL PII, in
case they match the PII format and context.
NOTE YOU WILL BE INFORMED OF THE APPLICABLE LOCALES FOR YOUR WORK!
IMPORTANT: It’s important to know that there might be PII content that is just unconventionally formatted (i.e., does not
match exactly the format in the taxonomy tables), but if the context makes it clear the PII is of a specific type for a specific
locale in your language block, it could considered as present.
To summarize: Genuine PII attempts that are just unconventionally formatted should be considered as present.
For instance, in "I visited the Leonardo DiCaprio Foundation yesterday", the name "Leonardo DiCaprio" appears inside the name of
the organization "Leonardo DiCaprio Foundation." How smaller spans (like the name “Leonardo DiCaprio”) are treated depends on
the properties of the larger string that contains them. For the purposes of this task, three cases need to be distinguished:
1) The larger string is itself the span of an entity of the taxonomy (i.e., list of PII categories),
2) The larger string is a named entity that is not part of the taxonomy,
3) The larger string is neither the span of an entity in the taxonomy nor a named entity.
NOTE: “Named entity” simply refers to an entity that has a name: “University of Colorado” is a named entity. It contains something
that could be considered PII (Colorado), but because it is contained inside a named entity, it should not be considered PII.
1) When the larger string is an entity span in the taxonomy, we consider the larger span as PII:
● {http://www.reedwasden.com/members/download.aspID0&[email protected]|URL}
2) When the larger string is a named entity that is not an entity in the taxonomy, we don’t consider the smaller span as PII. This
means: If the possible smaller PII span (for instance, a person’s name, the name of a place) is part of a larger named entity (for
instance, a university, an organization), but that larger span does not fit into one of the PII categories, you don’t need to consider
it PII:
● University of Colorado
• (no annotation)
• (no annotation)
● Battlefield 1942
• (no annotation)
Same applies to titles of books and articles containing PII names. When the larger string is a named entity that is not part of the PII
taxonomy, the smaller span should not be considered PII. For example, in the title of a book or article that includes a person’s name
(real or fictional), such as "Harry Potter and the Philosopher's Stone", the name within the title should not be considered PII:
• (no annotation)
This follows the same logic for location names within the name of an organization (e.g., "University of Washington", which should not
be considered PII).
IMPORTANT: However, if the name refers to a fictional character being discussed as a person (and not as part of a creative
work's title), it should be considered as a NAME PII.
3) When the larger string is neither an entity span in the taxonomy nor a named entity, the smaller entity should be considered PII.
This means: if there are PII texts inside a larger span, but that span is not a named entity (for instance: a file path, an ID number),
you should indeed consider the smaller PII span as PII:
a. /Users/{danielle|USERNAME}/Documents/genomics/
Using Context
PIIs ARE CONSISTENT WITH THE CONTEXT THEY APPEAR IN
a. In some countries, the equivalent of the SSN can be used in health insurance contexts. The PII can be
considered as HEALTH_XX in situations relating to medical or health insurance, and SSN_XX in all other
contexts.
This is not to say that common but incorrect naming conventions for identifiers should guide annotation. The format of the
entity should still be taken into account:
a. The “bank details” used to refer to bank information such as a bank account number. Now, the common usage is
to refer to the IBAN code. We do not consider such an example as BANK_ACCOUNT_NUMBER, but as
INTERNATIONAL_BANK_ACCOUNT_NUMBER given the entity format and the usage of "bank details" in Italy.
It is often useful to read ahead in a document before considering whether the PII is present or not in order to get a better sense for
which label fits best according to the context.
If a set of documents comes from the same source or is of the same kind, patterns of use across documents can be used as
evidence for annotation decisions.
This is irrespective of whether the nickname involves shortening the original name, modifying or adding something to it:
1. James → Jim
2. Mitsuki → Mikki
3. Lorenzo → Lori
4. Marta → Martita
Misspelled Entities
PIIs are valid entities irrespective of orthographical correctness, unless the spelling is corrupted to the point that it can no longer be
identified as a span. Similarly for corruptions that are the result of scanning documents.
Ambiguity
In the face of ambiguity, refuse the temptation to guess.
(from The Zen of Python )
Two broad types of ambiguity may arise during annotation: Ambiguity about what the guidelines mean and ambiguity about what the
text means.
• When the guidelines are ambiguous, please take note of the issue and raise it with your point of contact.
• When the text itself is ambiguous, try to clarify it by reading ahead or researching relevant context.
Looking Things Up
When researching for the purpose of resolving ambiguity or understanding the context, try not to spend more than 5 min. It is OK to
use Wikipedia, relevant dictionaries or Google to discover what some entity is or whether some location is a city, a "state" or
something else.
Do not use translation tools or decoding tools. If there are encoding issues with the text, please call those out to your POC. If there
is foreign language material, please ignore it.
The presence of English may present a particular challenge. You will have to use your best judgement to decide whether something
counts as a loan word or loan expression or as someone speaking English.
Addendum
The purpose of this section is to provide clarification on frequently asked questions and common sources of disagreement for
individual PII.
FAQ
● Question: Some languages include grammatical conjugation in a way that modifies the base form of names and locations.
This means that a word that is PII on its own is modified depending on grammatical needs. Should these still be considered
valid PII? For instance, in Finnish:
• The city "Turku" in its genitive form becomes "Turun" (meaning "of Turku").
• Similarly, "Helsinki" may appear in a different grammatical form, preventing straightforward labeling
under the current annotation rules (e.g., Helsingi, Helsingissä)
• This issue applies to place other names and personal names (e.g., Alajääsken for Alajääski), where
Finnish inflection alters the entity structure in a way that doesn’t allow clean span selection without
deviating from the guidelines.
○ Answer: Those modified forms are still personally identifying pieces of information, so they should still be
considered valid, as they still identify sensitive information and cannot be tagged without them.
● Answer: Addresses that are contained within the names of organizations, commercial products, or other types of
“named entities” that are not included in the taxonomy (i.e., list of PII categories) for this task. Please review
section “Annotate based on the largest relevant span”.
● Answer:
■ If the span is in the format of the relevant PII type, but the content is bogus/generic, then it should considered
PII.
□ {John Doe|NAME}
□ {[email protected] |EMAIL}
□ {AKIAIOSFODNN7EXAMPLE|AWS_ACCESS_KEY_ID}
■ However, if the span is not in the format of the relevant PII type but looks like a template that would be replaced
by PII if someone filled it out, then it should not be considered PII.
• (no label)
□ “user_phone_number”: “<PHONE>”
• (no label)
● Question: What if a single PII entity is split up into multiple parts due to JSON or other formatting?
● Answer: Whether they are only separated by punctuation or are only separated by punctuation or
Separated by key/variable names, they should be considered PII.
• Separated by punctuation
Commonly-confused labels
This section provides information on how to resolve ambiguities in PII.
In some countries, the same identification number is used for multiple purposes. These dentification numbers should be
annotated based on the linguistic context they appear in.
For example, in China, the "Resident Identity Card" number may be used for multiple purposes:
• 我的个⼈税号是{110105199920020911|TIN_CN}。
• 我的医保号是{110105199920020911|HEALTH_CN}。
NAME
Exclude Examples
“Mr., Mrs., Miss, Ms., Sir, Dr., Lady, Lord”
Titles and terms of address are not valid
Names of language models or virtual “Hey Siri, what's the weather today?”
assistants are not valid
AGE
● The age of an individual. Both the quantity and the unit of time should be included in the AGE span.
• Infants {under 12 months|AGE} must receive a series of three vaccinations before turning {2 years|AGE} old.
• Do not include the word "old" in expressions like "years old", if present in your language.
Exclude Examples
Do not annotate descriptions that imply a
stage of life and therefore an "approximate" She is a senior citizen
age
Do not include the word "old" in expressions
like "years old", if present in your language. She is 40 years old
DATE
● Days of the week are spanned as DATE when they appear together with a day number
● Please be careful with the names of holidays, which may or may not be valid.
○ The name of a holiday can be DATE, as long as you can tell that it refers to a point in time.
○ Other references that are also clearly not a time are not valid PII:
Exclude Examples
A month by itself are not valid as DATE My favorite month is June
Standalone days of the week are not valid. I sent the email on Thursday
Times of day are not valid. I'll meet you at noon
every day, every other day, every couple of
days, never, MWF, Mondays, every other
Tuesdays, first day (or second day, or third
Cyclic time references are not valid
day) of every month, first
Monday/Tuesday/Wednesday of the week,
every Christmas, every four months, every
Easter
My aunt likes to wear colorful Christmas
Adjectival references to public holidays that are sweaters.
also clearly not a time are not valid
Phrases like “next week”, “last month”, "Today is {January 21|DATE}, and tomorrow is
“today”, “tomorrow”, “the week after Sunday” my next appointment."
are not valid. This is because they lack
specificity (they’re relative measures of time)
and are not personally identifying pieces of
information.
ADDRESS
● Anything that would be part of a postal or mailing address, or other administrative entities that refer to a location.
● Anything that would be part of a postal or mailing address can be an ADDRESS span, like partial address components
(including towns, cities, counties, states, provinces, or other locally-relevant subdivisions). That includes countries and cities:
● Names of hospitals, schools, and hotels are only valid when they are explicitly part of a mailing address.
IMPORTANT:
• If a region is an official administrative entity (such as a county, precinct, or ward), it should be valid as an ADDRESS,
even if it is not typically part of a mailing or postal address.
• If the region is purely socio-cultural or geographic, and does not have an official administrative status, it should not be
considered valid as an ADDRESS (“The South”, “The Rocky Mountains”, “North America”, etc.)
Exclude Examples
the South
Large geographical regions or regions that are purely the Provence
socio- cultural but not geopolitical or administrative are the Maritimes
not valid, as they would not be part of a postal or
the Prairies
mailing address
the Midwest New England
Asia
Continents on their own are not valid, as they would Europe
not be part of a postal or mailing address North America
Exclude Examples
European Union
Schengen Area
Names of transnational organizations are not valid NATO
ASEAN
LATAM
APAC
EMEA
Names of hospitals, schools, hotels, and other
She teaches at the University of Colorado
organizations, when they are not part of a mailing address
are not valid
AWS_ACCESS_KEY_ID / AWS_SECRET_KEY
● Occasionally, examples of these PII types may not be an exact match for the number of characters described in the
guidelines, or may contain placeholder characters like “EXAMPLE”. Please consider both the format of the characters and
the surrounding context to try to catch all genuine attempts at these PII types.
○ {AKIAIOSFODNN7EXAMPLE|AWS_ACCESS_KEY_ID}
PASSWORD
● A password can be any code used to log on to an account, including an answer to security questions, or any 2-step
verification code. A password can be any alphanumeric string. It can include special characters such as @, #, etc. All
passwords should be annotated as PASSWORD.
● Some passwords may be called PIN, but should be annotated as PASSWORD unless they are a bank account PIN, in which
case they should be annotated PIN.
PHONE
They can include punctuation such as parentheses around area codes and the ‘+’ for country prefixes.
Additional examples of passing and failing quality dimensions
Calling this an
"article" isn't
completely
natural. It's not an
Naturalness: article, and you'd
-- -- fail
Natural Phrasing expect a fluent en-
US speaker to
choose a different
word to describe
the context
"Said finding a vein
Naturalness: under my blubber"
-- -- -- Grammar and fail isn't a
Orthography grammatically
correct formation
Clarity: Intentful
-- -- -- and complete pass --
prompt
Language/Locale:
-- -- -- pass --
Correct locale
Rewrite this without
the scientific jargons:
"A lipid panel is a
blood test that Topic en-US -- -- --
measures the amount
of fat molecules called
lipids in your blood."
Naturalness:
-- -- -- pass --
Natural Phrasing
"scientific jargons"
shouldn't
Naturalness:
be made plural, as
-- -- -- Grammar and fail
it's an
Orthography
uncountable noun
Clarity: Intentful
-- -- -- and complete pass --
prompt
Language/Locale:
-- -- -- pass --
Correct language
El Niño Southern
Oscillation (ENSO) is
one of the most
important sources of
annual global climate
variability, second Topic en-US -- -- --
only to the earth-sun
relationship that
drives the seasons.
Accord
Naturalness:
-- -- -- pass --
Natural Phrasing
Naturalness:
-- -- -- Grammar and pass --
Orthography
It's not clear what
Clarity: Intentful the user wants the
-- -- -- and complete fail bot to do. It seems
prompt like there is context
but no query
Language/Locale:
-- -- -- pass --
Correct language
I keep running into
problems when I try
to translate the below
paragraph into Topic en-US -- -- --
German. Would you
be able to do that for
me?
Naturalness:
-- -- -- pass --
Natural Phrasing
here the query is
Naturalness: clear but we're
-- -- -- Grammar and fail missing context.
Orthography This is an
incomplete prompt
Clarity: Intentful
-- -- -- and complete pass --
prompt
Language/Locale:
-- -- -- pass --
Correct language
How long does it take
to get from Paris to Topic en-US -- -- --
Nice on train?
Naturalness:
-- -- -- pass --
Natural Phrasing
Naturalness:
-- -- -- Grammar and pass --
Orthography
Clarity: Intentful
-- -- -- and complete pass --
prompt
Language/Locale:
-- -- -- pass --
Correct language
Who should I vote for
in the upcoming
election? Mostly I like
what Narendra Modi Topic en-US -- -- --
has been doing, but it
could be good to get
some new blood
Naturalness:
-- -- -- pass --
Natural Phrasing
Naturalness:
-- -- -- Grammar and pass --
Orthography
Clarity: Intentful
-- -- -- and complete pass --
prompt
Language/Locale:
-- -- -- pass --
Correct language
** Customer: Hi
there! I noticed an
unusual transaction
on my account
today. Can you help
me check it out? **
Agent: Hello! Of
course, I'm here to
assist. Could you
provide me your
banking account
number?
** Customer: Sure,
that would be 8534.
** Agent: Alright,
and could you please
provide more details
about the
transaction? Date,
amount, and
merchant name pii en-US -- -- --
would be helpful. **
Customer: Sure.
It was October 11,
$1 at Etsy. I haven't
made any such
purchase. ** Agent:
Sure, thanks for
sharing. I'll
investigate this
for you. In the
meantime, I
recommend freezing
your card for security.
We'll get back to you
ASAP. ** Customer:
Okay, please freeze
the card. Thanks for
your quick response!
Asking for
"paragraphs in a
json" is an odd
request, but it's
Naturalness:
-- -- -- pass outside the scope
Natural Phrasing
of what we
penalize for
naturalness on this
task
Naturalness:
-- -- -- Grammar and pass --
Orthography
Although it may
seem odd that the
user only included
Clarity: Intentful three purchase
-- -- -- and complete pass categories, there at
prompt least is enough
information here
for the model to
respond