~/Projects/ChatGLM3
git clone https://code.lsong.org/ChatGLM3
Commit
- Commit
- 667fe99fcba7684c5e645e8867310e01496ede70
- Author
- Longin-Yu <[email protected]>
- Date
- 2023-10-30 15:28:55 +0800 +0800
- Diffstat
README_en.md | 6 +++---
Add en doc
diff --git a/README_en.md b/README_en.md index 52deef92efd996d3dcd9743c630e914bf3d13fa0..8c18309c6aaaa7e0e913d98fbbbaaf7f5fd60434 100644 --- a/README_en.md +++ b/README_en.md @@ -51,7 +51,7 @@ > In the tests of ChatGLM3-6B-Base, BBH used a 3-shot test, GSM8K and MATH that require inference used a 0-shot CoT test, MBPP used a 0-shot generation followed by running test cases to calculate Pass@1, and other multiple-choice type datasets all used a 0-shot test. We have conducted manual evaluation tests on ChatGLM3-6B-32K in multiple long-text application scenarios. Compared with the second-generation model, its effect has improved by more than 50% on average. In applications such as paper reading, document summarization, and financial report analysis, this improvement is particularly significant. In addition, we also tested the model on the LongBench evaluation set, and the specific results are shown in the table below. -| Model | 平均 | Summary | Single-Doc QA | Multi-Doc QA | Code | Few-shot | Synthetic | +| Model | Average | Summary | Single-Doc QA | Multi-Doc QA | Code | Few-shot | Synthetic | |----------------------|:-----:|:----:|:----:|:----:|:------:|:-----:|:-----:| | ChatGLM2-6B-32K | 41.5 | 24.8 | 37.6 | 34.7 | 52.8 | 51.3 | 47.7 | | ChatGLM3-6B-32K | 50.2 | 26.6 | 45.8 | 46.1 | 56.2 | 61.2 | 65 | @@ -82,9 +82,9 @@ ![tool](resources/tool.png) - Code Interpreter: Code interpreter mode, where the model can execute code in a Jupyter environment and obtain results to complete complex tasks. ![code](resources/heart.png) -### 代码调用 +### Usage -可以通过如下代码调用 ChatGLM 模型来生成对话: +The ChatGLM model can be called to start a conversation using the following code: ```python >>> from transformers import AutoTokenizer, AutoModel