Spaces:
Running
Running
login to HF
Browse files- hf_retrieval.py +3 -1
hf_retrieval.py
CHANGED
@@ -7,6 +7,7 @@ from datetime import datetime
|
|
7 |
from collections import defaultdict
|
8 |
from datasets import load_dataset
|
9 |
from qdrant_client import QdrantClient
|
|
|
10 |
|
11 |
tokenizer = open_clip.get_tokenizer("ViT-B-32")
|
12 |
|
@@ -20,6 +21,7 @@ qdrant = QdrantClient(url=os.environ["QDRANT_CLOUD_URL"],
|
|
20 |
prefer_grpc=False)
|
21 |
collection_name = "video_chunks"
|
22 |
|
|
|
23 |
|
24 |
def timestamp_to_seconds(ts):
|
25 |
h, m, s = ts.split(":")
|
@@ -90,7 +92,7 @@ def merge_chunks(chunks):
|
|
90 |
|
91 |
|
92 |
def get_video_segment(video_id, start_time, end_time):
|
93 |
-
dataset = load_dataset("aegean-ai/ai-lectures-spring-24", split="train", streaming=True
|
94 |
for sample in dataset:
|
95 |
if sample["__key__"] == video_id:
|
96 |
break
|
|
|
7 |
from collections import defaultdict
|
8 |
from datasets import load_dataset
|
9 |
from qdrant_client import QdrantClient
|
10 |
+
from huggingface_hub import login
|
11 |
|
12 |
tokenizer = open_clip.get_tokenizer("ViT-B-32")
|
13 |
|
|
|
21 |
prefer_grpc=False)
|
22 |
collection_name = "video_chunks"
|
23 |
|
24 |
+
login(token=os.environ["HF_API_TOKEN"])
|
25 |
|
26 |
def timestamp_to_seconds(ts):
|
27 |
h, m, s = ts.split(":")
|
|
|
92 |
|
93 |
|
94 |
def get_video_segment(video_id, start_time, end_time):
|
95 |
+
dataset = load_dataset("aegean-ai/ai-lectures-spring-24", split="train", streaming=True)
|
96 |
for sample in dataset:
|
97 |
if sample["__key__"] == video_id:
|
98 |
break
|