Lord-Raven commited on
Commit
5021bd8
·
1 Parent(s): 6449289

Messing with configuration.

Browse files
Files changed (2) hide show
  1. app.py +2 -1
  2. requirements.txt +3 -3
app.py CHANGED
@@ -36,12 +36,13 @@ print(f"CUDA device: {torch.cuda.get_device_name(torch.cuda.current_device())}")
36
  # model = ORTModelForSequenceClassification.from_pretrained(model_name, export=True, provider="CUDAExecutionProvider")
37
  # tokenizer = AutoTokenizer.from_pretrained(tokenizer_name, model_max_length=512)
38
 
 
39
  model = ORTModelForSequenceClassification.from_pretrained(
40
  "philschmid/tiny-bert-sst2-distilled",
41
  export=True,
42
  provider="CUDAExecutionProvider",
43
  )
44
-
45
  tokenizer = AutoTokenizer.from_pretrained("philschmid/tiny-bert-sst2-distilled")
46
 
47
  # classifier = pipeline(task="zero-shot-classification", model=model, tokenizer=tokenizer, device="cuda:0")
 
36
  # model = ORTModelForSequenceClassification.from_pretrained(model_name, export=True, provider="CUDAExecutionProvider")
37
  # tokenizer = AutoTokenizer.from_pretrained(tokenizer_name, model_max_length=512)
38
 
39
+ print(f"ORTModelForSequenceClassification.from_pretrained")
40
  model = ORTModelForSequenceClassification.from_pretrained(
41
  "philschmid/tiny-bert-sst2-distilled",
42
  export=True,
43
  provider="CUDAExecutionProvider",
44
  )
45
+ print(f"AutoTokenizer.from_pretrained")
46
  tokenizer = AutoTokenizer.from_pretrained("philschmid/tiny-bert-sst2-distilled")
47
 
48
  # classifier = pipeline(task="zero-shot-classification", model=model, tokenizer=tokenizer, device="cuda:0")
requirements.txt CHANGED
@@ -2,8 +2,8 @@ torch==2.4.0
2
  fastapi==0.88.0
3
  huggingface_hub==0.23.5
4
  json5==0.9.25
5
- numpy<2.0
6
- optimum[exporters,onnxruntime-gpu]==1.21.4
7
- transformers
8
  sentence-transformers==3.0.1
9
  safetensors
 
2
  fastapi==0.88.0
3
  huggingface_hub==0.23.5
4
  json5==0.9.25
5
+ numpy
6
+ optimum[exporters,onnxruntime-gpu]==1.24.0
7
+ transformers==4.50.3
8
  sentence-transformers==3.0.1
9
  safetensors