Xueqing commited on
Commit
e103c70
·
verified ·
1 Parent(s): 25bcdde

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -53,7 +53,7 @@ Users (both direct and downstream) should critically evaluate the outputs in the
53
 
54
  To load and run plutus-8B-instruct, you can use the Hugging Face Transformers library. For example:
55
 
56
- --------------------------------------------------------------------
57
  from transformers import AutoModelForCausalLM, AutoTokenizer
58
 
59
  model_name = "TheFinAI/plutus-8B-instruct"
@@ -64,7 +64,7 @@ prompt = "Παρακαλώ δώσε μου ανάλυση για την οικο
64
  inputs = tokenizer(prompt, return_tensors="pt")
65
  outputs = model.generate(**inputs)
66
  print(tokenizer.decode(outputs[0], skip_special_tokens=True))
67
- --------------------------------------------------------------------
68
 
69
  Additional details for running with mixed precision, LoRA configuration, or int4 quantization can be found in the training documentation.
70
 
 
53
 
54
  To load and run plutus-8B-instruct, you can use the Hugging Face Transformers library. For example:
55
 
56
+ ```python
57
  from transformers import AutoModelForCausalLM, AutoTokenizer
58
 
59
  model_name = "TheFinAI/plutus-8B-instruct"
 
64
  inputs = tokenizer(prompt, return_tensors="pt")
65
  outputs = model.generate(**inputs)
66
  print(tokenizer.decode(outputs[0], skip_special_tokens=True))
67
+ ```
68
 
69
  Additional details for running with mixed precision, LoRA configuration, or int4 quantization can be found in the training documentation.
70