Skip to content

fix(website): correct grammar in chat.mdx harness section#1802

Merged
samchon merged 2 commits intowebsite/harnessfrom
copilot/sub-pr-1800-again
Mar 20, 2026
Merged

fix(website): correct grammar in chat.mdx harness section#1802
samchon merged 2 commits intowebsite/harnessfrom
copilot/sub-pr-1800-again

Conversation

Copy link
Copy Markdown
Contributor

Copilot AI commented Mar 20, 2026

Two grammatical errors in the LLM function calling harness description paragraph.

  • "The answer is not""The answer is no"
  • "if correct it""if you correct it"

📍 Connect Copilot coding agent with Jira, Azure Boards or Linear to delegate work to Copilot in one click without leaving your project management tool.

Co-authored-by: samchon <13158709+samchon@users.noreply.github.com>
Copilot AI changed the title [WIP] [WIP] Address feedback on harness wording for LLM module in the website feature fix(website): correct grammar in chat.mdx harness section Mar 20, 2026
Copilot AI requested a review from samchon March 20, 2026 01:50
@samchon samchon marked this pull request as ready for review March 20, 2026 01:59
Copilot AI review requested due to automatic review settings March 20, 2026 01:59
@samchon samchon merged commit 1861a50 into website/harness Mar 20, 2026
6 checks passed
@samchon samchon deleted the copilot/sub-pr-1800-again branch March 20, 2026 02:00
Copy link
Copy Markdown
Contributor

Copilot AI left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Pull request overview

Fixes minor grammar in the LLM function-calling harness documentation section to improve readability and clarity in the website docs.

Changes:

  • Replace “The answer is not” with “The answer is no”.
  • Replace “if correct it” with “if you correct it”.

💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.

Comment on lines +230 to 232
The answer is no, and LLM (Large Language Model) vendors like OpenAI take a lot of type level mistakes when composing the arguments of the target function to call. Even though an LLM function calling schema has defined an `Array<string>` type, LLM often fills it just by a `string` typed value. This is where the **function calling harness** comes in — a deterministic correction loop of schema generation, lenient parsing, type coercion, and validation feedback that turns unreliable LLM output into 100% correct structured data.

Therefore, when developing an LLM function calling agent, the validation feedback process is essentially required. If LLM takes a type level mistake on arguments composition, the agent must feedback the most detailed validation errors, and let the LLM to retry the function calling referencing the validation errors.
Copy link

Copilot AI Mar 20, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This sentence still has a couple of grammatical issues: “vendors … take a lot of type level mistakes” reads incorrectly (typically “make many type-level mistakes”), and “type-level” should be hyphenated.

Suggested change
The answer is no, and LLM (Large Language Model) vendors like OpenAI take a lot of type level mistakes when composing the arguments of the target function to call. Even though an LLM function calling schema has defined an `Array<string>` type, LLM often fills it just by a `string` typed value. This is where the **function calling harness** comes in — a deterministic correction loop of schema generation, lenient parsing, type coercion, and validation feedback that turns unreliable LLM output into 100% correct structured data.
Therefore, when developing an LLM function calling agent, the validation feedback process is essentially required. If LLM takes a type level mistake on arguments composition, the agent must feedback the most detailed validation errors, and let the LLM to retry the function calling referencing the validation errors.
The answer is no, and LLM (Large Language Model) vendors like OpenAI make many type-level mistakes when composing the arguments of the target function to call. Even though an LLM function calling schema has defined an `Array<string>` type, LLM often fills it just by a `string` typed value. This is where the **function calling harness** comes in — a deterministic correction loop of schema generation, lenient parsing, type coercion, and validation feedback that turns unreliable LLM output into 100% correct structured data.
Therefore, when developing an LLM function calling agent, the validation feedback process is essentially required. If LLM makes a type-level mistake on arguments composition, the agent must feedback the most detailed validation errors, and let the LLM to retry the function calling referencing the validation errors.

Copilot uses AI. Check for mistakes.
About the validation feedback, `@agentica/core` is utilizing [`typia.validate<T>()`](https://typia.io/docs/validators/validate) and [`typia.llm.application<Class>()`](https://typia.io/docs/llm/application/#application) functions. They construct validation logic by analyzing TypeScript source codes and types in the compilation level, so that detailed and accurate than any other validators like below.

Such validation feedback strategy and combination with `typia` runtime validator, `@agentica/core` has achieved the most ideal LLM function calling through the **function calling harness** pattern. In my experience, when using OpenAI's `gpt-4o-mini` model, it tends to construct invalid function calling arguments at the first trial about 50% of the time. By the way, if correct it through validation feedback with `typia`, success rate soars to 99%. And I've never had a failure when trying validation feedback twice.
Such validation feedback strategy and combination with `typia` runtime validator, `@agentica/core` has achieved the most ideal LLM function calling through the **function calling harness** pattern. In my experience, when using OpenAI's `gpt-4o-mini` model, it tends to construct invalid function calling arguments at the first trial about 50% of the time. By the way, if you correct it through validation feedback with `typia`, success rate soars to 99%. And I've never had a failure when trying validation feedback twice.
Copy link

Copilot AI Mar 20, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Minor grammar: “success rate soars to 99%” is missing an article (e.g., “the success rate soars …”).

Suggested change
Such validation feedback strategy and combination with `typia` runtime validator, `@agentica/core` has achieved the most ideal LLM function calling through the **function calling harness** pattern. In my experience, when using OpenAI's `gpt-4o-mini` model, it tends to construct invalid function calling arguments at the first trial about 50% of the time. By the way, if you correct it through validation feedback with `typia`, success rate soars to 99%. And I've never had a failure when trying validation feedback twice.
Such validation feedback strategy and combination with `typia` runtime validator, `@agentica/core` has achieved the most ideal LLM function calling through the **function calling harness** pattern. In my experience, when using OpenAI's `gpt-4o-mini` model, it tends to construct invalid function calling arguments at the first trial about 50% of the time. By the way, if you correct it through validation feedback with `typia`, the success rate soars to 99%. And I've never had a failure when trying validation feedback twice.

Copilot uses AI. Check for mistakes.
samchon added a commit that referenced this pull request Mar 20, 2026
* feat(website): the harness wording on llm module.

* fix remove

* do not abuse harness wording much a lot

* fix(website): correct grammar in chat.mdx harness section (#1802)

* Initial plan

* fix(website): correct grammar in chat.mdx harness section

Co-authored-by: samchon <13158709+samchon@users.noreply.github.com>

---------

Co-authored-by: copilot-swe-agent[bot] <198982749+Copilot@users.noreply.github.com>
Co-authored-by: samchon <13158709+samchon@users.noreply.github.com>

* fix: add missing `tags` import in function-calling-harness.md example (#1801)

* Initial plan

* fix: add tags to import in function-calling-harness.md example

Co-authored-by: samchon <13158709+samchon@users.noreply.github.com>

---------

Co-authored-by: copilot-swe-agent[bot] <198982749+Copilot@users.noreply.github.com>
Co-authored-by: samchon <13158709+samchon@users.noreply.github.com>

---------

Co-authored-by: Copilot <198982749+Copilot@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants