Upload README.md
Browse files
README.md
CHANGED
|
@@ -19,7 +19,7 @@ model_creator: OpenChat
|
|
| 19 |
model_name: Openchat 3.5 1210
|
| 20 |
model_type: mistral
|
| 21 |
pipeline_tag: text-generation
|
| 22 |
-
prompt_template: 'GPT4 User: {prompt}<|end_of_turn|>GPT4 Assistant:
|
| 23 |
|
| 24 |
'
|
| 25 |
quantized_by: TheBloke
|
|
@@ -71,10 +71,10 @@ These files were quantised using hardware kindly provided by [Massed Compute](ht
|
|
| 71 |
<!-- repositories-available end -->
|
| 72 |
|
| 73 |
<!-- prompt-template start -->
|
| 74 |
-
## Prompt template: OpenChat
|
| 75 |
|
| 76 |
```
|
| 77 |
-
GPT4 User: {prompt}<|end_of_turn|>GPT4 Assistant:
|
| 78 |
|
| 79 |
```
|
| 80 |
|
|
@@ -246,7 +246,7 @@ from huggingface_hub import InferenceClient
|
|
| 246 |
endpoint_url = "https://your-endpoint-url-here"
|
| 247 |
|
| 248 |
prompt = "Tell me about AI"
|
| 249 |
-
prompt_template=f'''GPT4 User: {prompt}<|end_of_turn|>GPT4 Assistant:
|
| 250 |
'''
|
| 251 |
|
| 252 |
client = InferenceClient(endpoint_url)
|
|
@@ -303,7 +303,7 @@ tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)
|
|
| 303 |
|
| 304 |
prompt = "Write a story about llamas"
|
| 305 |
system_message = "You are a story writing assistant"
|
| 306 |
-
prompt_template=f'''GPT4 User: {prompt}<|end_of_turn|>GPT4 Assistant:
|
| 307 |
'''
|
| 308 |
|
| 309 |
print("\n\n*** Generate:")
|
|
|
|
| 19 |
model_name: Openchat 3.5 1210
|
| 20 |
model_type: mistral
|
| 21 |
pipeline_tag: text-generation
|
| 22 |
+
prompt_template: 'GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:
|
| 23 |
|
| 24 |
'
|
| 25 |
quantized_by: TheBloke
|
|
|
|
| 71 |
<!-- repositories-available end -->
|
| 72 |
|
| 73 |
<!-- prompt-template start -->
|
| 74 |
+
## Prompt template: OpenChat-Correct
|
| 75 |
|
| 76 |
```
|
| 77 |
+
GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:
|
| 78 |
|
| 79 |
```
|
| 80 |
|
|
|
|
| 246 |
endpoint_url = "https://your-endpoint-url-here"
|
| 247 |
|
| 248 |
prompt = "Tell me about AI"
|
| 249 |
+
prompt_template=f'''GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:
|
| 250 |
'''
|
| 251 |
|
| 252 |
client = InferenceClient(endpoint_url)
|
|
|
|
| 303 |
|
| 304 |
prompt = "Write a story about llamas"
|
| 305 |
system_message = "You are a story writing assistant"
|
| 306 |
+
prompt_template=f'''GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:
|
| 307 |
'''
|
| 308 |
|
| 309 |
print("\n\n*** Generate:")
|