TapTechNews July 11th news, Anthropic company released a press release yesterday (July 10th) and announced the launch of several new tools, automating and refining prompts based on the Claude 3.5 Sonnet large language model.
TapTechNews attached the complete demonstration video as follows:
Anthropic's control panel has added an Evaluate unit, with the help of the Claude 3.5 Sonnet large language model to help developers efficiently generate, fine-tune and test prompts (Prompt).
These enhanced functions are designed to improve the language model's response to various tasks and provide valuable resources for enterprises using Claude to develop artificial intelligence products.
Developers only need to describe the task (such as diverting to handle customer support requests), and Claude will generate a high-quality prompt for you.
In the Evaluate tab, developers can test the prompts of their artificial intelligence applications according to various scenarios.
Users can upload real examples or use Claude to generate cases, thereby comparing the effectiveness between different prompts, and the answers are evaluated on a five-point scale for developers to intuitively evaluate.
In a blog example, developers only need to adjust one line to generate longer and more detailed answers in all tests, thereby saving time and improving productivity.
The new tool supports manual and automatic testing of prompts. Developers can generate input variables, see Claude's response, and manually enter test cases when needed.
Testing based on multiple real inputs helps verify the quality of prompts before production deployment. Developers can manually add or import new test cases from the CSV, or request Claude to create test cases.