file_name
large_stringlengths 4
140
| prefix
large_stringlengths 0
12.1k
| suffix
large_stringlengths 0
12k
| middle
large_stringlengths 0
7.51k
| fim_type
large_stringclasses 4
values |
---|---|---|---|---|
test2.py | :
def __init__(self, master):
self.video = None
self.frame_rate = 0
self.video_length = 0
# The scaled image used for display. Needs to persist for display
self.display_image = None
self.display_ratio = 0
self.awaiting_corners = False
self.corners = []
#Tkinter related fields
self.master = master
self.master.title("Auto Kifu Test2")
self.window_width = root.winfo_screenwidth()
self.window_height = root.winfo_screenheight() - 100
self.master.geometry("%dx%d+0+0" % (self.window_width, self.window_height))
self.master.configure(background='grey')
self.canvas = Tkinter.Canvas(self.master)
self.canvas.place(x=0,
y=0,
width=self.window_width,
height=self.window_height)
self.canvas.bind("<Button-1>", self.mouse_clicked)
self.menubar = Tkinter.Menu(root)
root.config(menu=self.menubar)
self.fileMenu = Tkinter.Menu(self.menubar)
self.fileMenu.add_command(label="Load Image", command=self.load())
self.menubar.add_cascade(label="File", menu=self.fileMenu)
def mouse_clicked(self, event):
if self.awaiting_corners:
self.draw_x(event.x, event.y)
self.corners += [(event.x/self.display_ratio, event.y/self.display_ratio)]
if len(self.corners) == 4:
self.awaiting_corners = False
self.main()
def main(self):
board_positions, crop_window = self.find_grid(self.corners)
frames = self.parse_video(crop_window)
for x in range(len(frames)):
frames[x] = cv2.cvtColor(frames[x], cv2.COLOR_BGR2GRAY)
frames[x] = cv2.GaussianBlur(frames[x], (51, 51), 0)
thresholds = self.determine_thresholds(frames[-1], board_positions)
for x in range(len(frames)):
cv2.imwrite('output/2/frames'+str(x)+'.png', frames[x])
for x in range(len(frames)):
frames[x] = self.parse_frames(frames[x], board_positions, thresholds)
for x in range(1, len(frames)):
print "Board: "+str(x)
self.print_board(frames[x])
output = "(;GM[1]FF[4]CA[UTF-8]AP[CGoban:3]ST[2]SZ[19]"
for i in range(1, len(frames)):
moves = self.frame_difference(frames[i-1], frames[i])
for move in moves:
color = move["color"]
x = LETTERS[move["position"][0]]
y = LETTERS[move["position"][1]]
output += ";"+color+"["+x+y+"]"
output += ")"
file = open("output.txt", "w")
file.write(output)
file.close()
def find_grid(self, corners):
top_left = corners[0]
bottom_right = corners[2]
board_width = bottom_right[0] - top_left[0]
board_height = bottom_right[1] - top_left[1]
horizontal_spacing = board_width / 18
vertical_spacing = board_height / 18
crop_window = Rectangle()
crop_window.x = int(top_left[0] - horizontal_spacing)
crop_window.y = int(top_left[1] - vertical_spacing)
crop_window.w = int(board_width + (2 * horizontal_spacing))
crop_window.h = int(board_height + (2 * vertical_spacing))
board_positions = []
for x in range(0, 19):
board_positions += [[]]
for y in range(0, 19):
x_coord = int(top_left[0] + horizontal_spacing * x)
y_coord = int(top_left[1] + vertical_spacing * y)
x_coord -= crop_window.x
y_coord -= crop_window.y
board_positions[x] += [(y_coord, x_coord)]
return board_positions, crop_window
def print_board(self, frame):
print "-------------------"
for y in range(19):
string = ""
for x in range(19):
string += frame[x][y]
print string
print "-------------------"
def parse_video(self, crop_window):
fourcc = cv2.VideoWriter_fourcc(*'XVID')
out1 = cv2.VideoWriter('output.avi', fourcc, 30.0, (crop_window.w, crop_window.h))
success, current_frame = self.video.read()
current_frame = current_frame[crop_window.y:crop_window.y + crop_window.h,
crop_window.x:crop_window.x + crop_window.w]
differences = []
final_video = [current_frame]
while (self.video.isOpened() and success):
last_frame = current_frame
success, current_frame = self.video.read()
if not success: break
current_frame = current_frame[crop_window.y:crop_window.y+crop_window.h,
crop_window.x:crop_window.x+crop_window.w]
out1.write(current_frame)
s = self.mse_total(last_frame, current_frame)
#s = ssim(last_frame, current_frame) # Doesn't Work
differences += [s]
recently_still = True
still_duration = 15
for x in range(still_duration):
if x<len(differences) and differences[-x]>4:
recently_still = False
if recently_still:
#out1.write(current_frame)
s = self.mse_total(current_frame, final_video[-1])
if s>20:
final_video += [current_frame]
#plt.hist(differences, bins=400)
plt.title("Frame Difference Historgram")
plt.xlabel("Difference (mean squared error)")
plt.ylabel("Number of Frames")
#plt.show()
time = np.arange(0, self.video_length/self.frame_rate, 1.0/self.frame_rate)
time = time[:len(differences)]
#plt.plot(time, differences)
plt.xlabel('time (s)')
plt.ylabel('Difference')
plt.title('MSE over Time')
plt.grid(True)
#plt.show()
out1.release()
'''
fourcc = cv2.VideoWriter_fourcc(*'XVID')
out2 = cv2.VideoWriter('output2.avi', fourcc, 30.0,
(self.crop_w, self.crop_h))
for x in final_video:
for y in range(30):
out2.write(x)
out2.release()
'''
return final_video
def mse_total(self, imageA, imageB):
err = np.sum((imageA.astype("float") - imageB.astype("float")) ** 2)
err /= float(imageA.shape[0] * imageA.shape[1])
return err
def mse_image(self, imageA, imageB):
return (imageA - imageB) ** 2
def determine_thresholds(self, image, board_positions):
samples = []
for x in range(0, 19):
for y in range(0, 19):
position = board_positions[x][y]
samples += [float(image[position[0]][position[1]])]
plt.hist(samples, bins=255)
plt.title("Intersection Intensity Historgram")
plt.xlabel("Intensity (Greyscale)")
plt.ylabel("Number of Intersections")
# plt.show() | plt.axvline(x=centroids[0], color="red")
plt.axvline(x=centroids[1], color="red")
plt.axvline(x=centroids[2], color="red")
plt.show()
min = 0
mid = 0
max = 0
for x in range(0, 3):
if centroids[x] < centroids[min]:
min = x
if centroids[x] > centroids[max]:
max = x
for x in range(0, 3):
if x != min and x != max:
mid = x
min = centroids[min]
mid = centroids[mid]
max = centroids[max]
threshold1 = (min + mid) / 2
threshold2 = (max + mid) / 2
print "threshold 1 = "+str(threshold1)
print "threshold 2 = "+str(threshold2)
#return [threshold1, threshold2]
return [120,185]
def parse_frames(self, image, board_positions, thresholds):
return_array = []
for x in range(0, 19):
return_array += [[]]
for y in range(0, 19):
position = board_positions[x][y]
intensity = image[position[0]][position[1]]
if intensity < thresholds[0]:
return_array[x] += ["B"]
elif intensity > thresholds[1]:
return_array[x] += ["W"]
else:
return_array[x] += ["+"]
return return_array
def frame_difference(self, former_frame, later_frame):
moves = []
for x in range(19):
for y in range(19):
if (later_frame[x][ |
centroids, _ = kmeans(samples, 3) | random_line_split |
test2.py | :
def __init__(self, master):
self.video = None
self.frame_rate = 0
self.video_length = 0
# The scaled image used for display. Needs to persist for display
self.display_image = None
self.display_ratio = 0
self.awaiting_corners = False
self.corners = []
#Tkinter related fields
self.master = master
self.master.title("Auto Kifu Test2")
self.window_width = root.winfo_screenwidth()
self.window_height = root.winfo_screenheight() - 100
self.master.geometry("%dx%d+0+0" % (self.window_width, self.window_height))
self.master.configure(background='grey')
self.canvas = Tkinter.Canvas(self.master)
self.canvas.place(x=0,
y=0,
width=self.window_width,
height=self.window_height)
self.canvas.bind("<Button-1>", self.mouse_clicked)
self.menubar = Tkinter.Menu(root)
root.config(menu=self.menubar)
self.fileMenu = Tkinter.Menu(self.menubar)
self.fileMenu.add_command(label="Load Image", command=self.load())
self.menubar.add_cascade(label="File", menu=self.fileMenu)
def mouse_clicked(self, event):
if self.awaiting_corners:
self.draw_x(event.x, event.y)
self.corners += [(event.x/self.display_ratio, event.y/self.display_ratio)]
if len(self.corners) == 4:
self.awaiting_corners = False
self.main()
def main(self):
board_positions, crop_window = self.find_grid(self.corners)
frames = self.parse_video(crop_window)
for x in range(len(frames)):
frames[x] = cv2.cvtColor(frames[x], cv2.COLOR_BGR2GRAY)
frames[x] = cv2.GaussianBlur(frames[x], (51, 51), 0)
thresholds = self.determine_thresholds(frames[-1], board_positions)
for x in range(len(frames)):
cv2.imwrite('output/2/frames'+str(x)+'.png', frames[x])
for x in range(len(frames)):
frames[x] = self.parse_frames(frames[x], board_positions, thresholds)
for x in range(1, len(frames)):
print "Board: "+str(x)
self.print_board(frames[x])
output = "(;GM[1]FF[4]CA[UTF-8]AP[CGoban:3]ST[2]SZ[19]"
for i in range(1, len(frames)):
moves = self.frame_difference(frames[i-1], frames[i])
for move in moves:
color = move["color"]
x = LETTERS[move["position"][0]]
y = LETTERS[move["position"][1]]
output += ";"+color+"["+x+y+"]"
output += ")"
file = open("output.txt", "w")
file.write(output)
file.close()
def find_grid(self, corners):
top_left = corners[0]
bottom_right = corners[2]
board_width = bottom_right[0] - top_left[0]
board_height = bottom_right[1] - top_left[1]
horizontal_spacing = board_width / 18
vertical_spacing = board_height / 18
crop_window = Rectangle()
crop_window.x = int(top_left[0] - horizontal_spacing)
crop_window.y = int(top_left[1] - vertical_spacing)
crop_window.w = int(board_width + (2 * horizontal_spacing))
crop_window.h = int(board_height + (2 * vertical_spacing))
board_positions = []
for x in range(0, 19):
board_positions += [[]]
for y in range(0, 19):
x_coord = int(top_left[0] + horizontal_spacing * x)
y_coord = int(top_left[1] + vertical_spacing * y)
x_coord -= crop_window.x
y_coord -= crop_window.y
board_positions[x] += [(y_coord, x_coord)]
return board_positions, crop_window
def print_board(self, frame):
print "-------------------"
for y in range(19):
string = ""
for x in range(19):
string += frame[x][y]
print string
print "-------------------"
def parse_video(self, crop_window):
fourcc = cv2.VideoWriter_fourcc(*'XVID')
out1 = cv2.VideoWriter('output.avi', fourcc, 30.0, (crop_window.w, crop_window.h))
success, current_frame = self.video.read()
current_frame = current_frame[crop_window.y:crop_window.y + crop_window.h,
crop_window.x:crop_window.x + crop_window.w]
differences = []
final_video = [current_frame]
while (self.video.isOpened() and success):
last_frame = current_frame
success, current_frame = self.video.read()
if not success: break
current_frame = current_frame[crop_window.y:crop_window.y+crop_window.h,
crop_window.x:crop_window.x+crop_window.w]
out1.write(current_frame)
s = self.mse_total(last_frame, current_frame)
#s = ssim(last_frame, current_frame) # Doesn't Work
differences += [s]
recently_still = True
still_duration = 15
for x in range(still_duration):
if x<len(differences) and differences[-x]>4:
recently_still = False
if recently_still:
#out1.write(current_frame)
s = self.mse_total(current_frame, final_video[-1])
if s>20:
|
#plt.hist(differences, bins=400)
plt.title("Frame Difference Historgram")
plt.xlabel("Difference (mean squared error)")
plt.ylabel("Number of Frames")
#plt.show()
time = np.arange(0, self.video_length/self.frame_rate, 1.0/self.frame_rate)
time = time[:len(differences)]
#plt.plot(time, differences)
plt.xlabel('time (s)')
plt.ylabel('Difference')
plt.title('MSE over Time')
plt.grid(True)
#plt.show()
out1.release()
'''
fourcc = cv2.VideoWriter_fourcc(*'XVID')
out2 = cv2.VideoWriter('output2.avi', fourcc, 30.0,
(self.crop_w, self.crop_h))
for x in final_video:
for y in range(30):
out2.write(x)
out2.release()
'''
return final_video
def mse_total(self, imageA, imageB):
err = np.sum((imageA.astype("float") - imageB.astype("float")) ** 2)
err /= float(imageA.shape[0] * imageA.shape[1])
return err
def mse_image(self, imageA, imageB):
return (imageA - imageB) ** 2
def determine_thresholds(self, image, board_positions):
samples = []
for x in range(0, 19):
for y in range(0, 19):
position = board_positions[x][y]
samples += [float(image[position[0]][position[1]])]
plt.hist(samples, bins=255)
plt.title("Intersection Intensity Historgram")
plt.xlabel("Intensity (Greyscale)")
plt.ylabel("Number of Intersections")
# plt.show()
centroids, _ = kmeans(samples, 3)
plt.axvline(x=centroids[0], color="red")
plt.axvline(x=centroids[1], color="red")
plt.axvline(x=centroids[2], color="red")
plt.show()
min = 0
mid = 0
max = 0
for x in range(0, 3):
if centroids[x] < centroids[min]:
min = x
if centroids[x] > centroids[max]:
max = x
for x in range(0, 3):
if x != min and x != max:
mid = x
min = centroids[min]
mid = centroids[mid]
max = centroids[max]
threshold1 = (min + mid) / 2
threshold2 = (max + mid) / 2
print "threshold 1 = "+str(threshold1)
print "threshold 2 = "+str(threshold2)
#return [threshold1, threshold2]
return [120,185]
def parse_frames(self, image, board_positions, thresholds):
return_array = []
for x in range(0, 19):
return_array += [[]]
for y in range(0, 19):
position = board_positions[x][y]
intensity = image[position[0]][position[1]]
if intensity < thresholds[0]:
return_array[x] += ["B"]
elif intensity > thresholds[1]:
return_array[x] += ["W"]
else:
return_array[x] += ["+"]
return return_array
def frame_difference(self, former_frame, later_frame):
moves = []
for x in range(19):
for y in range(19):
if (later_frame[x][ | final_video += [current_frame] | conditional_block |
test2.py | :
def __init__(self, master):
self.video = None
self.frame_rate = 0
self.video_length = 0
# The scaled image used for display. Needs to persist for display
self.display_image = None
self.display_ratio = 0
self.awaiting_corners = False
self.corners = []
#Tkinter related fields
self.master = master
self.master.title("Auto Kifu Test2")
self.window_width = root.winfo_screenwidth()
self.window_height = root.winfo_screenheight() - 100
self.master.geometry("%dx%d+0+0" % (self.window_width, self.window_height))
self.master.configure(background='grey')
self.canvas = Tkinter.Canvas(self.master)
self.canvas.place(x=0,
y=0,
width=self.window_width,
height=self.window_height)
self.canvas.bind("<Button-1>", self.mouse_clicked)
self.menubar = Tkinter.Menu(root)
root.config(menu=self.menubar)
self.fileMenu = Tkinter.Menu(self.menubar)
self.fileMenu.add_command(label="Load Image", command=self.load())
self.menubar.add_cascade(label="File", menu=self.fileMenu)
def mouse_clicked(self, event):
if self.awaiting_corners:
self.draw_x(event.x, event.y)
self.corners += [(event.x/self.display_ratio, event.y/self.display_ratio)]
if len(self.corners) == 4:
self.awaiting_corners = False
self.main()
def main(self):
board_positions, crop_window = self.find_grid(self.corners)
frames = self.parse_video(crop_window)
for x in range(len(frames)):
frames[x] = cv2.cvtColor(frames[x], cv2.COLOR_BGR2GRAY)
frames[x] = cv2.GaussianBlur(frames[x], (51, 51), 0)
thresholds = self.determine_thresholds(frames[-1], board_positions)
for x in range(len(frames)):
cv2.imwrite('output/2/frames'+str(x)+'.png', frames[x])
for x in range(len(frames)):
frames[x] = self.parse_frames(frames[x], board_positions, thresholds)
for x in range(1, len(frames)):
print "Board: "+str(x)
self.print_board(frames[x])
output = "(;GM[1]FF[4]CA[UTF-8]AP[CGoban:3]ST[2]SZ[19]"
for i in range(1, len(frames)):
moves = self.frame_difference(frames[i-1], frames[i])
for move in moves:
color = move["color"]
x = LETTERS[move["position"][0]]
y = LETTERS[move["position"][1]]
output += ";"+color+"["+x+y+"]"
output += ")"
file = open("output.txt", "w")
file.write(output)
file.close()
def find_grid(self, corners):
top_left = corners[0]
bottom_right = corners[2]
board_width = bottom_right[0] - top_left[0]
board_height = bottom_right[1] - top_left[1]
horizontal_spacing = board_width / 18
vertical_spacing = board_height / 18
crop_window = Rectangle()
crop_window.x = int(top_left[0] - horizontal_spacing)
crop_window.y = int(top_left[1] - vertical_spacing)
crop_window.w = int(board_width + (2 * horizontal_spacing))
crop_window.h = int(board_height + (2 * vertical_spacing))
board_positions = []
for x in range(0, 19):
board_positions += [[]]
for y in range(0, 19):
x_coord = int(top_left[0] + horizontal_spacing * x)
y_coord = int(top_left[1] + vertical_spacing * y)
x_coord -= crop_window.x
y_coord -= crop_window.y
board_positions[x] += [(y_coord, x_coord)]
return board_positions, crop_window
def print_board(self, frame):
print "-------------------"
for y in range(19):
string = ""
for x in range(19):
string += frame[x][y]
print string
print "-------------------"
def parse_video(self, crop_window):
fourcc = cv2.VideoWriter_fourcc(*'XVID')
out1 = cv2.VideoWriter('output.avi', fourcc, 30.0, (crop_window.w, crop_window.h))
success, current_frame = self.video.read()
current_frame = current_frame[crop_window.y:crop_window.y + crop_window.h,
crop_window.x:crop_window.x + crop_window.w]
differences = []
final_video = [current_frame]
while (self.video.isOpened() and success):
last_frame = current_frame
success, current_frame = self.video.read()
if not success: break
current_frame = current_frame[crop_window.y:crop_window.y+crop_window.h,
crop_window.x:crop_window.x+crop_window.w]
out1.write(current_frame)
s = self.mse_total(last_frame, current_frame)
#s = ssim(last_frame, current_frame) # Doesn't Work
differences += [s]
recently_still = True
still_duration = 15
for x in range(still_duration):
if x<len(differences) and differences[-x]>4:
recently_still = False
if recently_still:
#out1.write(current_frame)
s = self.mse_total(current_frame, final_video[-1])
if s>20:
final_video += [current_frame]
#plt.hist(differences, bins=400)
plt.title("Frame Difference Historgram")
plt.xlabel("Difference (mean squared error)")
plt.ylabel("Number of Frames")
#plt.show()
time = np.arange(0, self.video_length/self.frame_rate, 1.0/self.frame_rate)
time = time[:len(differences)]
#plt.plot(time, differences)
plt.xlabel('time (s)')
plt.ylabel('Difference')
plt.title('MSE over Time')
plt.grid(True)
#plt.show()
out1.release()
'''
fourcc = cv2.VideoWriter_fourcc(*'XVID')
out2 = cv2.VideoWriter('output2.avi', fourcc, 30.0,
(self.crop_w, self.crop_h))
for x in final_video:
for y in range(30):
out2.write(x)
out2.release()
'''
return final_video
def mse_total(self, imageA, imageB):
err = np.sum((imageA.astype("float") - imageB.astype("float")) ** 2)
err /= float(imageA.shape[0] * imageA.shape[1])
return err
def mse_image(self, imageA, imageB):
return (imageA - imageB) ** 2
def determine_thresholds(self, image, board_positions):
samples = []
for x in range(0, 19):
for y in range(0, 19):
position = board_positions[x][y]
samples += [float(image[position[0]][position[1]])]
plt.hist(samples, bins=255)
plt.title("Intersection Intensity Historgram")
plt.xlabel("Intensity (Greyscale)")
plt.ylabel("Number of Intersections")
# plt.show()
centroids, _ = kmeans(samples, 3)
plt.axvline(x=centroids[0], color="red")
plt.axvline(x=centroids[1], color="red")
plt.axvline(x=centroids[2], color="red")
plt.show()
min = 0
mid = 0
max = 0
for x in range(0, 3):
if centroids[x] < centroids[min]:
min = x
if centroids[x] > centroids[max]:
max = x
for x in range(0, 3):
if x != min and x != max:
mid = x
min = centroids[min]
mid = centroids[mid]
max = centroids[max]
threshold1 = (min + mid) / 2
threshold2 = (max + mid) / 2
print "threshold 1 = "+str(threshold1)
print "threshold 2 = "+str(threshold2)
#return [threshold1, threshold2]
return [120,185]
def parse_frames(self, image, board_positions, thresholds):
return_array = []
for x in range(0, 19):
return_array += [[]]
for y in range(0, 19):
position = board_positions[x][y]
intensity = image[position[0]][position[1]]
if intensity < thresholds[0]:
return_array[x] += ["B"]
elif intensity > thresholds[1]:
return_array[x] += ["W"]
else:
return_array[x] += ["+"]
return return_array
def frame_difference(self, former_frame, later_frame):
moves = []
for x in range(19):
for y in range(19):
if (later_frame[x][ | MainWindow | identifier_name |
|
test_install.py |
def _get_logger(filename='test_install.log'):
"""
Convenience function to set-up output and logging.
"""
logger = logging.getLogger('test_install.py')
logger.setLevel(logging.DEBUG)
console_handler = logging.StreamHandler()
console_handler.setLevel(logging.INFO)
file_handler = logging.FileHandler(filename)
file_handler.setLevel(logging.DEBUG)
logger.addHandler(console_handler)
logger.addHandler(file_handler)
return logger
logger = _get_logger()
def check_output(cmd):
"""
Run the specified command and capture its outputs.
Returns
-------
out : tuple
The (stdout, stderr) output tuple.
"""
logger.info(cmd)
args = shlex.split(cmd)
p = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
out = [ii.decode() for ii in p.communicate()]
return out
def report(out, name, line, item, value, eps=None, return_item=False,
match_numbers=False):
"""
Check that `item` at `line` of the output string `out` is equal
to `value`. If not, print the output.
"""
try:
if match_numbers:
status = out.split('\n')[line]
else:
status = out.split('\n')[line].split()
except IndexError:
logger.error(' not enough output from command!')
ok = False
else:
try:
if match_numbers:
pat = '([-+]?(?:\d+(?:\.\d*)?|\.\d+)(?:[eE][-+]?\d+)?[jJ]?)'
matches = re.findall(pat, status)
status_item = matches[item]
else:
status_item = status[item]
logger.info(' comparing: %s %s', status_item, value)
if eps is None:
ok = (status_item == value)
else:
try:
ok = abs(float(status_item) - float(value)) < eps
except:
ok = False
except IndexError:
ok = False
logger.info(' %s: %s', name, ok)
if not ok:
logger.debug(DEBUG_FMT, out)
if return_item:
return ok, status[item]
else:
return ok
def report2(out, name, items, return_item=False):
"""
Check that `items` are in the output string `out`.
If not, print the output.
"""
ok = True
for s in items:
logger.info(' checking: %s', s)
if s not in out:
ok = False
break
logger.info(' %s: %s', name, ok)
if not ok:
logger.debug(DEBUG_FMT, out)
if return_item:
return ok, s
else:
return ok
def report_tests(out, return_item=False):
"""
Check that all tests in the output string `out` passed.
If not, print the output.
"""
from pyparsing import (Word, Combine, Suppress, Optional, OneOrMore,
delimitedList, nums, Literal)
from functools import partial
integer = Word(nums).setName('integer')
real = Combine(Word(nums) + '.' + Optional(Word(nums))).setName('real')
equals = Suppress(OneOrMore('='))
_stats = {}
def add_stat(s, loc, toks, key=None):
if key is None:
key = toks[1]
_stats[key] = toks[0]
return toks
word = ((integer + 'failed') |
(integer + 'passed') |
(integer + 'deselected') |
(integer + 'warnings')).setParseAction(add_stat)
line = (equals +
Optional(delimitedList(word)) +
'in' +
(real + (Literal('s') | 'seconds'))
.setParseAction(partial(add_stat, key='seconds')) +
equals)
line.searchString(out)
keys = ['failed', 'passed', 'deselected', 'warnings', 'seconds']
stats = {key : _stats.get(key, '0') for key in keys}
ok = stats['failed'] == '0'
logger.info(
(' {failed} failed, {passed} passed, {deselected} deselected,'
' {warnings} warnings in {seconds} seconds').format(**stats)
)
if not ok:
logger.debug(DEBUG_FMT, out)
if return_item:
return ok, stats['failed']
else:
return ok
def main():
parser = ArgumentParser(description=__doc__,
formatter_class=RawDescriptionHelpFormatter)
parser.add_argument('--version', action='version', version='%(prog)s')
parser.parse_args()
fd = open('test_install.log', 'w')
fd.close()
eok = 0
t0 = time.time()
out, err = check_output('python3 sfepy/scripts/blockgen.py')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/cylindergen.py')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/convert_mesh.py meshes/3d/cylinder.vtk out.mesh')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/convert_mesh.py --tile 2,2 meshes/elements/2_4_2.mesh out-per.mesh')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/convert_mesh.py --extract-surface --print-surface=- meshes/various_formats/octahedron.node surf_octahedron.mesh')
eok += report(out, '...', -4, 0, '1185')
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/diffusion/poisson.py')
eok += report(out, '...', -3, 5, '1.173819e-16', eps=1e-15)
out, err = check_output("""python3 sfepy/scripts/simple.py -c "ebc_2 : {'name' : 't2', 'region' : 'Gamma_Right', 'dofs' : {'t.0' : -5.0}}" sfepy/examples/diffusion/poisson.py""")
eok += report(out, '...', -3, 5, '2.308051e-16', eps=1e-15)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/diffusion/poisson_iga.py')
eok += report(out, '...', -3, 5, '3.373487e-15', eps=1e-14)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/navier_stokes/stokes.py')
eok += report(out, '...', -3, 5, '1.210678e-13', eps=1e-11)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/diffusion/poisson_parametric_study.py')
eok += report(out, '...', -3, 5, '1.606408e-14', eps=1e-13)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/linear_elasticity/its2D_3.py')
eok += report(out, '...', -24, 5, '3.964886e-12', eps=1e-11)
eok += report(out, '...', -4, 4, '2.58660e+01', eps=1e-5)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/linear_elasticity/linear_elastic.py --format h5')
eok += report(out, '...', -3, 5, '4.638192e-18', eps=1e-15)
out, err = check_output('python3 sfepy/scripts/extractor.py -d cylinder.h5')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/resview.py --off-screen -o cylinder.png cylinder.h5')
eok += report(out, '...', -2, 1, 'cylinder.png')
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/phononic/band_gaps.py')
eok += report(out, '...', -9, 0, '2.08545116e+08', match_numbers=True)
eok += report(out, '...', -8, 1, '1.16309223e+11', match_numbers=True)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/phononic/band_gaps.py | random_line_split |
||
test_install.py |
def report(out, name, line, item, value, eps=None, return_item=False,
match_numbers=False):
"""
Check that `item` at `line` of the output string `out` is equal
to `value`. If not, print the output.
"""
try:
if match_numbers:
status = out.split('\n')[line]
else:
status = out.split('\n')[line].split()
except IndexError:
logger.error(' not enough output from command!')
ok = False
else:
try:
if match_numbers:
pat = '([-+]?(?:\d+(?:\.\d*)?|\.\d+)(?:[eE][-+]?\d+)?[jJ]?)'
matches = re.findall(pat, status)
status_item = matches[item]
else:
status_item = status[item]
logger.info(' comparing: %s %s', status_item, value)
if eps is None:
ok = (status_item == value)
else:
try:
ok = abs(float(status_item) - float(value)) < eps
except:
ok = False
except IndexError:
ok = False
logger.info(' %s: %s', name, ok)
if not ok:
logger.debug(DEBUG_FMT, out)
if return_item:
return ok, status[item]
else:
return ok
def report2(out, name, items, return_item=False):
"""
Check that `items` are in the output string `out`.
If not, print the output.
"""
ok = True
for s in items:
logger.info(' checking: %s', s)
if s not in out:
ok = False
break
logger.info(' %s: %s', name, ok)
if not ok:
logger.debug(DEBUG_FMT, out)
if return_item:
return ok, s
else:
return ok
def report_tests(out, return_item=False):
"""
Check that all tests in the output string `out` passed.
If not, print the output.
"""
from pyparsing import (Word, Combine, Suppress, Optional, OneOrMore,
delimitedList, nums, Literal)
from functools import partial
integer = Word(nums).setName('integer')
real = Combine(Word(nums) + '.' + Optional(Word(nums))).setName('real')
equals = Suppress(OneOrMore('='))
_stats = {}
def add_stat(s, loc, toks, key=None):
if key is None:
key = toks[1]
_stats[key] = toks[0]
return toks
word = ((integer + 'failed') |
(integer + 'passed') |
(integer + 'deselected') |
(integer + 'warnings')).setParseAction(add_stat)
line = (equals +
Optional(delimitedList(word)) +
'in' +
(real + (Literal('s') | 'seconds'))
.setParseAction(partial(add_stat, key='seconds')) +
equals)
line.searchString(out)
keys = ['failed', 'passed', 'deselected', 'warnings', 'seconds']
stats = {key : _stats.get(key, '0') for key in keys}
ok = stats['failed'] == '0'
logger.info(
(' {failed} failed, {passed} passed, {deselected} deselected,'
' {warnings} warnings in {seconds} seconds').format(**stats)
)
if not ok:
logger.debug(DEBUG_FMT, out)
if return_item:
return ok, stats['failed']
else:
return ok
def main():
parser = ArgumentParser(description=__doc__,
formatter_class=RawDescriptionHelpFormatter)
parser.add_argument('--version', action='version', version='%(prog)s')
parser.parse_args()
fd = open('test_install.log', 'w')
fd.close()
eok = 0
t0 = time.time()
out, err = check_output('python3 sfepy/scripts/blockgen.py')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/cylindergen.py')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/convert_mesh.py meshes/3d/cylinder.vtk out.mesh')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/convert_mesh.py --tile 2,2 meshes/elements/2_4_2.mesh out-per.mesh')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/convert_mesh.py --extract-surface --print-surface=- meshes/various_formats/octahedron.node surf_octahedron.mesh')
eok += report(out, '...', -4, 0, '1185')
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/diffusion/poisson.py')
eok += report(out, '...', -3, 5, '1.173819e-16', eps=1e-15)
out, err = check_output("""python3 sfepy/scripts/simple.py -c "ebc_2 : {'name' : 't2', 'region' : 'Gamma_Right', 'dofs' : {'t.0' : -5.0}}" sfepy/examples/diffusion/poisson.py""")
eok += report(out, '...', -3, 5, '2.308051e-16', eps=1e-15)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/diffusion/poisson_iga.py')
eok += report(out, '...', -3, 5, '3.373487e-15', eps=1e-14)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/navier_stokes/stokes.py')
eok += report(out, '...', -3, 5, '1.210678e-13', eps=1e-11)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/diffusion/poisson_parametric_study.py')
eok += report(out, '...', -3, 5, '1.606408e-14', eps=1e-13)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/linear_elasticity/its2D_3.py')
eok += report(out, '...', -24, 5, '3.964886e-12', eps=1e-11)
eok += report(out, '...', -4, 4, '2.58660e+01', eps=1e-5)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/linear_elasticity/linear_elastic.py --format h5')
eok += report(out, '...', -3, 5, '4.638192e-18', eps=1e-15)
out, err = check_output('python3 sfepy/scripts/extractor.py -d cylinder.h5')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/resview.py --off-screen -o cylinder.png cylinder.h5')
eok += report(out, '...', -2, 1, 'cylinder.png')
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/phononic/band_gaps.py')
eok += report(out, '...', -9, 0, '2.08545116e+08', match_numbers=True)
eok += report(out, '...', -8, 1, '1.16309223e+11', match_numbers=True)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/phononic/band_gaps.py --phonon-phase-velocity')
eok += report(out, '...', -2, 0, '4189.41229592', match_numbers=True)
eok += report(out, '...', -2, 1, '2620.55608256', match_numbers=True)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/phononic/band_gaps | """
Run the specified command and capture its outputs.
Returns
-------
out : tuple
The (stdout, stderr) output tuple.
"""
logger.info(cmd)
args = shlex.split(cmd)
p = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
out = [ii.decode() for ii in p.communicate()]
return out | identifier_body |
|
test_install.py | = out.split('\n')[line].split()
except IndexError:
logger.error(' not enough output from command!')
ok = False
else:
try:
if match_numbers:
pat = '([-+]?(?:\d+(?:\.\d*)?|\.\d+)(?:[eE][-+]?\d+)?[jJ]?)'
matches = re.findall(pat, status)
status_item = matches[item]
else:
status_item = status[item]
logger.info(' comparing: %s %s', status_item, value)
if eps is None:
ok = (status_item == value)
else:
try:
ok = abs(float(status_item) - float(value)) < eps
except:
ok = False
except IndexError:
ok = False
logger.info(' %s: %s', name, ok)
if not ok:
logger.debug(DEBUG_FMT, out)
if return_item:
return ok, status[item]
else:
return ok
def report2(out, name, items, return_item=False):
"""
Check that `items` are in the output string `out`.
If not, print the output.
"""
ok = True
for s in items:
logger.info(' checking: %s', s)
if s not in out:
ok = False
break
logger.info(' %s: %s', name, ok)
if not ok:
logger.debug(DEBUG_FMT, out)
if return_item:
return ok, s
else:
return ok
def report_tests(out, return_item=False):
"""
Check that all tests in the output string `out` passed.
If not, print the output.
"""
from pyparsing import (Word, Combine, Suppress, Optional, OneOrMore,
delimitedList, nums, Literal)
from functools import partial
integer = Word(nums).setName('integer')
real = Combine(Word(nums) + '.' + Optional(Word(nums))).setName('real')
equals = Suppress(OneOrMore('='))
_stats = {}
def add_stat(s, loc, toks, key=None):
if key is None:
key = toks[1]
_stats[key] = toks[0]
return toks
word = ((integer + 'failed') |
(integer + 'passed') |
(integer + 'deselected') |
(integer + 'warnings')).setParseAction(add_stat)
line = (equals +
Optional(delimitedList(word)) +
'in' +
(real + (Literal('s') | 'seconds'))
.setParseAction(partial(add_stat, key='seconds')) +
equals)
line.searchString(out)
keys = ['failed', 'passed', 'deselected', 'warnings', 'seconds']
stats = {key : _stats.get(key, '0') for key in keys}
ok = stats['failed'] == '0'
logger.info(
(' {failed} failed, {passed} passed, {deselected} deselected,'
' {warnings} warnings in {seconds} seconds').format(**stats)
)
if not ok:
logger.debug(DEBUG_FMT, out)
if return_item:
return ok, stats['failed']
else:
return ok
def | ():
parser = ArgumentParser(description=__doc__,
formatter_class=RawDescriptionHelpFormatter)
parser.add_argument('--version', action='version', version='%(prog)s')
parser.parse_args()
fd = open('test_install.log', 'w')
fd.close()
eok = 0
t0 = time.time()
out, err = check_output('python3 sfepy/scripts/blockgen.py')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/cylindergen.py')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/convert_mesh.py meshes/3d/cylinder.vtk out.mesh')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/convert_mesh.py --tile 2,2 meshes/elements/2_4_2.mesh out-per.mesh')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/convert_mesh.py --extract-surface --print-surface=- meshes/various_formats/octahedron.node surf_octahedron.mesh')
eok += report(out, '...', -4, 0, '1185')
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/diffusion/poisson.py')
eok += report(out, '...', -3, 5, '1.173819e-16', eps=1e-15)
out, err = check_output("""python3 sfepy/scripts/simple.py -c "ebc_2 : {'name' : 't2', 'region' : 'Gamma_Right', 'dofs' : {'t.0' : -5.0}}" sfepy/examples/diffusion/poisson.py""")
eok += report(out, '...', -3, 5, '2.308051e-16', eps=1e-15)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/diffusion/poisson_iga.py')
eok += report(out, '...', -3, 5, '3.373487e-15', eps=1e-14)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/navier_stokes/stokes.py')
eok += report(out, '...', -3, 5, '1.210678e-13', eps=1e-11)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/diffusion/poisson_parametric_study.py')
eok += report(out, '...', -3, 5, '1.606408e-14', eps=1e-13)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/linear_elasticity/its2D_3.py')
eok += report(out, '...', -24, 5, '3.964886e-12', eps=1e-11)
eok += report(out, '...', -4, 4, '2.58660e+01', eps=1e-5)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/linear_elasticity/linear_elastic.py --format h5')
eok += report(out, '...', -3, 5, '4.638192e-18', eps=1e-15)
out, err = check_output('python3 sfepy/scripts/extractor.py -d cylinder.h5')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/resview.py --off-screen -o cylinder.png cylinder.h5')
eok += report(out, '...', -2, 1, 'cylinder.png')
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/phononic/band_gaps.py')
eok += report(out, '...', -9, 0, '2.08545116e+08', match_numbers=True)
eok += report(out, '...', -8, 1, '1.16309223e+11', match_numbers=True)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/phononic/band_gaps.py --phonon-phase-velocity')
eok += report(out, '...', -2, 0, '4189.41229592', match_numbers=True)
eok += report(out, '...', -2, 1, '2620.55608256', match_numbers=True)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/phononic/band_gaps.py --phonon-dispersion')
eok += report(out, '...', -6, 1, '[0,')
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/phononic/band_gaps_rigid.py')
eok += report(out, '...', -9, 0, '4.58709531e+07', match_numbers=True)
eok += report(out, '...', -8, 1, '1.13929200e+11', match_numbers=True)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/quantum/hydrogen.py')
eok += report(out, '...', | main | identifier_name |
test_install.py | = out.split('\n')[line].split()
except IndexError:
logger.error(' not enough output from command!')
ok = False
else:
try:
if match_numbers:
pat = '([-+]?(?:\d+(?:\.\d*)?|\.\d+)(?:[eE][-+]?\d+)?[jJ]?)'
matches = re.findall(pat, status)
status_item = matches[item]
else:
status_item = status[item]
logger.info(' comparing: %s %s', status_item, value)
if eps is None:
|
else:
try:
ok = abs(float(status_item) - float(value)) < eps
except:
ok = False
except IndexError:
ok = False
logger.info(' %s: %s', name, ok)
if not ok:
logger.debug(DEBUG_FMT, out)
if return_item:
return ok, status[item]
else:
return ok
def report2(out, name, items, return_item=False):
"""
Check that `items` are in the output string `out`.
If not, print the output.
"""
ok = True
for s in items:
logger.info(' checking: %s', s)
if s not in out:
ok = False
break
logger.info(' %s: %s', name, ok)
if not ok:
logger.debug(DEBUG_FMT, out)
if return_item:
return ok, s
else:
return ok
def report_tests(out, return_item=False):
"""
Check that all tests in the output string `out` passed.
If not, print the output.
"""
from pyparsing import (Word, Combine, Suppress, Optional, OneOrMore,
delimitedList, nums, Literal)
from functools import partial
integer = Word(nums).setName('integer')
real = Combine(Word(nums) + '.' + Optional(Word(nums))).setName('real')
equals = Suppress(OneOrMore('='))
_stats = {}
def add_stat(s, loc, toks, key=None):
if key is None:
key = toks[1]
_stats[key] = toks[0]
return toks
word = ((integer + 'failed') |
(integer + 'passed') |
(integer + 'deselected') |
(integer + 'warnings')).setParseAction(add_stat)
line = (equals +
Optional(delimitedList(word)) +
'in' +
(real + (Literal('s') | 'seconds'))
.setParseAction(partial(add_stat, key='seconds')) +
equals)
line.searchString(out)
keys = ['failed', 'passed', 'deselected', 'warnings', 'seconds']
stats = {key : _stats.get(key, '0') for key in keys}
ok = stats['failed'] == '0'
logger.info(
(' {failed} failed, {passed} passed, {deselected} deselected,'
' {warnings} warnings in {seconds} seconds').format(**stats)
)
if not ok:
logger.debug(DEBUG_FMT, out)
if return_item:
return ok, stats['failed']
else:
return ok
def main():
parser = ArgumentParser(description=__doc__,
formatter_class=RawDescriptionHelpFormatter)
parser.add_argument('--version', action='version', version='%(prog)s')
parser.parse_args()
fd = open('test_install.log', 'w')
fd.close()
eok = 0
t0 = time.time()
out, err = check_output('python3 sfepy/scripts/blockgen.py')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/cylindergen.py')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/convert_mesh.py meshes/3d/cylinder.vtk out.mesh')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/convert_mesh.py --tile 2,2 meshes/elements/2_4_2.mesh out-per.mesh')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/convert_mesh.py --extract-surface --print-surface=- meshes/various_formats/octahedron.node surf_octahedron.mesh')
eok += report(out, '...', -4, 0, '1185')
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/diffusion/poisson.py')
eok += report(out, '...', -3, 5, '1.173819e-16', eps=1e-15)
out, err = check_output("""python3 sfepy/scripts/simple.py -c "ebc_2 : {'name' : 't2', 'region' : 'Gamma_Right', 'dofs' : {'t.0' : -5.0}}" sfepy/examples/diffusion/poisson.py""")
eok += report(out, '...', -3, 5, '2.308051e-16', eps=1e-15)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/diffusion/poisson_iga.py')
eok += report(out, '...', -3, 5, '3.373487e-15', eps=1e-14)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/navier_stokes/stokes.py')
eok += report(out, '...', -3, 5, '1.210678e-13', eps=1e-11)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/diffusion/poisson_parametric_study.py')
eok += report(out, '...', -3, 5, '1.606408e-14', eps=1e-13)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/linear_elasticity/its2D_3.py')
eok += report(out, '...', -24, 5, '3.964886e-12', eps=1e-11)
eok += report(out, '...', -4, 4, '2.58660e+01', eps=1e-5)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/linear_elasticity/linear_elastic.py --format h5')
eok += report(out, '...', -3, 5, '4.638192e-18', eps=1e-15)
out, err = check_output('python3 sfepy/scripts/extractor.py -d cylinder.h5')
eok += report(out, '...', -2, 1, '...done')
out, err = check_output('python3 sfepy/scripts/resview.py --off-screen -o cylinder.png cylinder.h5')
eok += report(out, '...', -2, 1, 'cylinder.png')
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/phononic/band_gaps.py')
eok += report(out, '...', -9, 0, '2.08545116e+08', match_numbers=True)
eok += report(out, '...', -8, 1, '1.16309223e+11', match_numbers=True)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/phononic/band_gaps.py --phonon-phase-velocity')
eok += report(out, '...', -2, 0, '4189.41229592', match_numbers=True)
eok += report(out, '...', -2, 1, '2620.55608256', match_numbers=True)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/phononic/band_gaps.py --phonon-dispersion')
eok += report(out, '...', -6, 1, '[0,')
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/phononic/band_gaps_rigid.py')
eok += report(out, '...', -9, 0, '4.58709531e+07', match_numbers=True)
eok += report(out, '...', -8, 1, '1.13929200e+11', match_numbers=True)
out, err = check_output('python3 sfepy/scripts/simple.py sfepy/examples/quantum/hydrogen.py')
eok += report(out, '...', | ok = (status_item == value) | conditional_block |
Triangle.js | this;
}
/**
* 设置三角形的顶点坐标为数组中的坐标
* @param {*} points 顶点数组
* @param {*} i0 第一个点的索引
* @param {*} i1 第二个点的索引
* @param {*} i2 第三个点的索引
* @returns
*/
setFromPointsAndIndices( points, i0, i1, i2 ) {
this.a.copy( points[ i0 ] );
this.b.copy( points[ i1 ] );
this.c.copy( points[ i2 ] );
return this;
}
/**
* 返回一个该平面三角形的拷贝
* @returns
*/
clone() {
return new this.constructor().copy( this );
}
/**
* 将自己设置为一个指定三角形的复制
* @param {*} triangle
* @returns
*/
copy( triangle ) {
this.a.copy( triangle.a );
this.b.copy( triangle.b );
this.c.copy( triangle.c );
return this;
}
/**
* 计算平面三角形的面积
* @returns
*/
getArea() {
_v0.subVectors( this.c, this.b );
_v1.subVectors( this.a, this.b );
return _v0.cross( _v1 ).length() * 0.5;
}
/**
* 计算三角形的中点。
*/
getMidpoint( target ) {
if ( target === undefined ) {
console.warn( 'THREE.Triangle: .getMidpoint() target is now required' );
target = new Vector3();
}
return target.addVectors( this.a, this.b ).add( this.c ).multiplyScalar( 1 / 3 );
}
/**
* 获取三角面的normal
* @param {*} target
* @returns
*/
getNormal( target ) {
return Triangle.getNormal( this.a, this.b, this.c, target );
}
/**
* 获取平面三角形坐在的平面
* @param {*} target
* @returns
*/
getPlane( target ) {
if ( target === undefined ) {
console.warn( 'THREE.Triangle: .getPlane() target is now required' );
target = new Plane();
}
// 通过空间中三个不共线的三个点确定一个平面
return target.setFromCoplanarPoints( this.a, this.b, this.c );
}
/**
* 计算指定点的重心坐标
* @param {*} point 指定的点
* @param {*} target 结果将会被拷贝到这个Vector3中
* @returns
*/
getBarycoord( point, target ) {
return Triangle.getBarycoord( point, this.a, this.b, this.c, target );
}
/**
* uv插值计算,通过本平面三角形三个顶点的uv,插值计算出任意点的uv。
* @param {*} point 要进行uv插值的点的位置
* @param {*} uv1 顶点1的uv
* @param {*} uv2 顶点2的uv
* @param {*} uv3 顶点3的uv
* @param {*} target 结果将会被拷贝到这个Vector2中
* @returns
*/
getUV( point, uv1, uv2, uv3, target ) {
return Triangle.getUV( point, this.a, this.b, this.c, uv1, uv2, uv3, target );
}
/**
* 判断指定点是否在本平面三角形中
* @param {*} point
* @returns
*/
containsPoint( point ) {
return Triangle.containsPoint( point, this.a, this.b, this.c );
}
/**
* 判断一个给定向量是否朝向平面三角形法向量
* @param {*} direction
* @returns
*/
isFrontFacing( direction ) {
return Triangle.isFrontFacing( this.a, this.b, this.c, direction );
}
/**
* 判定三角形与传入的box是否相交
* @param {*} box
* @returns
*/
intersectsBox( box ) {
return box.intersectsTriangle( this );
}
/**
* 返回三角形上最靠近所给定的point的点
* @param {*} p 指定的点
* @param {*} target 结果将会被拷贝到这个Vector3中
* @returns
*/
closestPointToPoint( p, target ) {
if ( target === undefined ) {
console.warn( 'THREE.Triangle: .closestPointToPoint() target is now required' );
target = new Vector3();
}
const a = this.a, b = this.b, c = this.c;
let v, w;
// 算法来自christer Ericsion所著的《Real-Time Collision Detection》
// Morgan Kaufmann 出版社, (c) 2005 Elsevier Inc.,
// 版权许可。请查看5.1.5章节来了解详情。
// 基本上来说,我们希望通过最少的冗余计算来得到指定点p位于哪个voronoi区域
_vab.subVectors( b, a );
_vac.subVectors( c, a );
_vap.subVectors( p, a );
const d1 = _vab.dot( _vap );
const d2 = _vac.dot( _vap );
if ( d1 <= 0 && d2 <= 0 ) {
// vertex region of A; barycentric coords (1, 0, 0)
// 顶点A的区域,重心坐标(1,0,0)
return target.copy( a );
}
_vbp.subVectors( p, b );
const d3 = _vab.dot( _vbp );
const d4 = _vac.dot( _vbp );
if ( d3 >= 0 && d4 <= d3 ) {
// vertex region of B; barycentric coords (0, 1, 0)
// 顶点B的区域,重心坐标(0,1,0)
return target.copy( b );
}
const vc = d1 * d4 - d3 * d2;
if ( vc <= 0 && d1 >= 0 && d3 <= 0 ) {
v = d1 / ( d1 - d3 );
// edge region of AB; barycentric coords (1-v, v, 0)
// AB边的区域; 重心坐标 (1-v, v, 0)
return target.copy( a ).addScaledVector( _vab, v );
}
_vcp.subVectors( p, c );
const d5 = _vab.dot( _vcp );
const d6 = _vac.dot( _vcp );
if ( d6 >= 0 && d5 <= d6 ) {
// vertex region of C; barycentric coords (0, 0, 1)
// 顶点C的区域,重心坐标(0, 0, 1)
return target.copy( c );
}
const vb = d5 * d2 - d1 * d6;
if ( vb <= 0 && d2 >= 0 && d6 <= 0 ) {
w = d2 / ( d2 - d6 );
// edge region of AC; barycentric coords (1-w, 0, w)
// AC边的区域; 重心坐标 (1-w, 0, w)
return target.copy( a ).addScaledVector( _vac, w );
}
const va = d3 * d6 - d5 * d4;
if ( va <= 0 && ( d4 - d3 ) >= 0 && ( d5 - d6 ) >= 0 ) {
_vbc.subVectors( c, b );
w = ( d4 - d3 ) / ( ( d4 - d3 ) + ( d5 - d6 ) );
// edge region of BC; barycentric coords (0, 1-w, w)
// BC边的区域; 重心坐标 (0, 1-w, w)
return target.copy( b ).addScaledVector( _vbc, w ); // edge region of BC
}
// face region
// 平面三角形面上的区域
const denom = 1 / ( va + vb + vc );
// u = va * denom
v = vb * denom;
w = vc * denom;
return target.copy( a ).addScaledVector( _vab, v ).addScaledVector( _vac, w );
}
/**
* 判断本平面三角形是否和传入的平面三角形相同
* @param {*} triangle
* @returns
*/
equals( triangle ) {
return triangle.a.equals( this.a ) && triangle.b.equals( this.b ) && triangle.c.equals( this.c );
}
}
export { Triangle };
| identifier_body |
||
Triangle.js | , b );
_v0.subVectors( a, b );
target.cross( _v0 );
const targetLengthSq = target.lengthSq();
if ( targetLengthSq > 0 ) {
// 归一化处理
return target.multiplyScalar( 1 / Math.sqrt( targetLengthSq ) );
}
return target.set( 0, 0, 0 );
}
/**
* 用来计算重心坐标的静态/实例方法
* @param {*} point 指定的点位
* @param {*} a 平面三角形的顶点1
* @param {*} b 平面三角形的顶点2
* @param {*} c 平面三角形的顶点3
* @param {*} target 结果将会被拷贝到这个Vector3中
* @returns
*
* 函数算法参考:http://www.blackpawn.com/texts/pointinpoly/default.html
* 注意这个函数并不是用来计算一个三角形的重心,而是通过类似计算重心坐标的方式来表示平面上任意一点的位置
* 这样可以简单判断一个点是否在三角形中
*/
static getBarycoord( point, a, b, c, target ) {
const ca = new Vector3();
const ba = new Vector3();
const pa = new Vector3();
ca.subVectors( c, a );
ba.subVectors( b, a );
pa.subVectors( point, a );
const dot00 = ca.dot( ca );
const dot01 = ca.dot( ba );
const dot02 = ca.dot( pa );
const dot11 = ba.dot( ba );
const dot12 = ba.dot( pa );
const denom = ( dot00 * dot11 - dot01 * dot01 );
if ( target === undefined ) {
console.warn( 'THREE.Triangle: .getBarycoord() target is now required' );
target = new Vector3();
}
// 三个点共线 or 奇异三角形
// TODO: 为什么奇异三角形会返回0?
if ( denom === 0 ) {
// 任意点都在三角形外
// 不确定这是不是最好的返回值,或许应该返回undefined
return target.set( - 2, - 1, - 1 );
}
const invDenom = 1 / denom;
const u = ( dot11 * dot02 - dot01 * dot12 ) * invDenom;
const v = ( dot00 * dot12 - dot01 * dot02 ) * invDenom;
// 重心坐标的x,y,z必须加起来等于1
return target.set( 1 - u - v, v, u );
}
/**
* 计算三角形中是否包含某个点
* @param {*} point 指定的点位
* @param {*} a 平面三角形的顶点1
* @param {*} b 平面三角形的顶点2
* @param {*} c 平面三角形的顶点3
* @returns
*/
static containsPoint( point, a, b, c ) {
this.getBarycoord( point, a, b, c, _v3 );
return ( _v3.x >= 0 ) && ( _v3.y >= 0 ) && ( ( _v3.x + _v3.y ) <= 1 );
}
/**
* uv插值计算,通过三个点的uv,插值计算出任意点的uv。
* @param {*} point 要进行uv插值的点的位置
* @param {*} p1 顶点1的位置
* @param {*} p2 顶点2的位置
* @param {*} p3 顶点3的位置
* @param {*} uv1 顶点1的uv
* @param {*} uv2 顶点2的uv
* @param {*} uv3 顶点3的uv
* @param {*} target 结果将会被拷贝到这个Vector2中
* @returns
*/
static getUV( point, p1, p2, p3, uv1, uv2, uv3, target ) {
this.getBarycoord( point, p1, p2, p3, _v3 );
target.set( 0, 0 );
target.addScaledVector( uv1, _v3.x );
target.addScaledVector( uv2, _v3.y );
target.addScaledVector( uv3, _v3.z );
return target;
}
/**
* 判断一个给定向量是否朝向给定三点组成的平面三角形法向量
* @param {*} a 平面三角形的顶点1
* @param {*} b 平面三角形的顶点2
* @param {*} c 平面三角形的顶点3
* @param {*} direction
* @returns
*/
static isFrontFacing( a, b, c, direction ) {
_v0.subVectors( c, b );
_v1.subVectors( a, b );
// strictly front facing
return ( _v0.cross( _v1 ).dot( direction ) < 0 ) ? true : false;
}
/**
* 设置平米三角形的三个点
* @param {*} a
* @param {*} b
* @param {*} c
* @returns
*/
set( a, b, c ) {
this.a.copy( a );
this.b.copy( b );
this.c.copy( c );
return this;
}
/**
* 设置三角形的顶点坐标为数组中的坐标
* @param {*} points 顶点数组
* @param {*} i0 第一个点的索引
* @param {*} i1 第二个点的索引
* @param {*} i2 第三个点的索引
* @returns
*/
setFromPointsAndIndices( points, i0, i1, i2 ) {
this.a.copy( points[ i0 ] );
this.b.copy( points[ i1 ] );
this.c.copy( points[ i2 ] );
return this;
}
/**
* 返回一个该平面三角形的拷贝
* @returns
*/
clone() {
return new this.constructor().copy( this );
}
/**
* 将自己设置为一个指定三角形的复制
* @param {*} triangle
* @returns
*/
copy( triangle ) {
this.a.copy( triangle.a );
this.b.copy( triangle.b );
this.c.copy( triangle.c );
return this;
}
/**
* 计算平面三角形的面积
* @returns
*/
getArea() {
_v0.subVectors( this.c, this.b );
_v1.subVectors( this.a, this.b );
return _v0.cross( _v1 ).length() * 0.5;
}
/**
* 计算三角形的中点。
*/
getMidpoint( target ) {
if ( target === undefined ) {
console.warn( 'THREE.Triangle: .getMidpoint() target is now required' );
target = new Vector3();
}
return target.addVectors( this.a, this.b ).add( this.c ).multiplyScalar( 1 / 3 );
}
/**
* 获取三角面的normal
* @param {*} target
* @returns
*/
getNormal( target ) {
return Triangle.getNormal( this.a, this.b, this.c, target );
}
/** | * @returns
*/
getPlane( target ) {
if ( target === undefined ) {
console.warn( 'THREE.Triangle: .getPlane() target is now required' );
target = new Plane();
}
// 通过空间中三个不共线的三个点确定一个平面
return target.setFromCoplanarPoints( this.a, this.b, this.c );
}
/**
* 计算指定点的重心坐标
* @param {*} point 指定的点
* @param {*} target 结果将会被拷贝到这个Vector3中
* @returns
*/
getBarycoord( point, target ) {
return Triangle.getBarycoord( point, this.a, this.b, this.c, target );
}
/**
* uv插值计算,通过本平面三角形三个顶点的uv,插值计算出任意点的uv。
* @param {*} point 要进行uv插值的点的位置
* @param {*} uv1 顶点1的uv
* @param {*} uv2 顶点2的uv
* @param {*} uv3 顶点3的uv
* @param {*} target 结果将会被拷贝到这个Vector2中
* @returns
*/
getUV( point, uv1, uv2, uv3, target ) {
return Triangle.getUV( point, this.a, this.b, this.c, | * 获取平面三角形坐在的平面
* @param {*} target | random_line_split |
Triangle.js | b );
_v0.subVectors( a, b );
target.cross( _v0 );
const targetLengthSq = target.lengthSq();
if ( targetLengthSq > 0 ) {
// 归一化处理
return target.multiplyScalar( 1 / Math.sqrt( targetLengthSq ) );
}
return target.set( 0, 0, 0 );
}
/**
* 用来计算重心坐标的静态/实例方法
* @param {*} point 指定的点位
* @param {*} a 平面三角形的顶点1
* @param {*} b 平面三角形的顶点2
* @param {*} c 平面三角形的顶点3
* @param {*} target 结果将会被拷贝到这个Vector3中
* @returns
*
* 函数算法参考:http://www.blackpawn.com/texts/pointinpoly/default.html
* 注意这个函数并不是用来计算一个三角形的重心,而是通过类似计算重心坐标的方式来表示平面上任意一点的位置
* 这样可以简单判断一个点是否在三角形中
*/
static getBarycoord( point, a, b, c, target ) {
const ca = new Vector3();
const ba = new Vector3();
const pa = new Vector3();
ca.subVectors( c, a );
ba.subVectors( b, a );
pa.subVectors( point, a );
const dot00 = ca.dot( ca );
const dot01 = ca.dot( ba );
const dot02 = ca.dot( pa );
const dot11 = ba.dot( ba );
const dot12 = ba.dot( pa );
const denom = ( dot00 * dot11 - dot01 * dot01 );
if ( target === undefined ) {
console.warn( 'THREE.Triangle: .getBarycoord() target is now required' );
target = new Vector3();
}
// 三个点共线 or 奇异三角形
// TODO: 为什么奇异三角形会返回0?
if ( denom === 0 ) {
// 任意点都在三角形外
// 不确定这是不是最好的返回值,或许应该返回undefined
return target.set( - 2, - 1, - 1 );
}
const invDenom = 1 / denom;
const u = ( dot11 * dot02 - dot01 * dot12 ) * invDenom;
const v = ( dot00 * dot12 - dot01 * dot02 ) * invDenom;
// 重心坐标的x,y,z必须加起来等于1
return target.set( 1 - u - v, v, u );
}
/**
* 计算三角形中是否包含某个点
* @param {*} point 指定的点位
* @param {*} a 平面三角形的顶点1
* @param {*} b 平面三角形的顶点2
* @param {*} c 平面三角形的顶点3
* @returns
*/
static containsPoint( point, a, b, c ) {
this.getBarycoord( point, a, b, c, _v3 );
return ( _v3.x >= 0 ) && ( _v3.y >= 0 ) && ( ( _v3.x + _v3.y ) <= 1 );
}
/**
* uv插值计算,通过三个点的uv,插值计算出任意点的uv。
* @param {*} point 要进行uv插值的点的位置
* @param {*} p1 顶点1的位置
* @param {*} p2 顶点2的位置
* @param {*} p3 顶点3的位置
* @param {*} uv1 顶点1的uv
* @param {*} uv2 顶点2的uv
* @param {*} uv3 顶点3的uv
* @param {*} target 结果将会被拷贝到这个Vector2中
* @returns
*/
static getUV( point, p1, p2, p3, uv1, uv2, uv3, target ) {
this.getBarycoord( point, p1, p2, p3, _v3 );
target.set( 0, 0 );
target.addScaledVector( uv1, _v3.x );
target.addScaledVector( uv2, _v3.y );
target.addScaledVector( uv3, _v3.z );
return target;
}
/**
* 判断一个给定向量是否朝向给定三点组成的平面三角形法向量
* @param {*} a 平面三角形的顶点1
* @param {*} b 平面三角形的顶点2
* @param {*} c 平面三角形的顶点3
* @param {*} direction
* @returns
*/
static isFrontFacing( a, b, c, direction ) {
_v0.subVectors( c, b );
_v1.subVectors( a, b );
// strictly front facing
return ( _v0.cross( _v1 ).dot( direction ) < 0 ) ? true : false;
}
/**
* 设置平米三角形的三个点
* @param {*} a
* @param {*} b
* @param {*} c
* @returns
*/
set( a, b, c ) {
this.a.copy( a );
this.b.copy( b );
this.c.copy( c );
return this;
}
/**
* 设置三角形的顶点坐标为数组中的坐标
* @param {*} points 顶点数组
* @param {*} i0 第一个点的索引
* @param {*} i1 第二个点的索引
* @param {*} i2 第三个点的索引
* @returns
*/
setFromPointsAndIndices( points, i0, i1, i2 ) {
this.a.copy( points[ i0 ] );
this.b.copy( points[ i1 ] );
this.c.copy( points[ i2 ] );
return this;
}
/**
* 返回一个该平面三角形的拷贝
* @returns
*/
clone() {
return new this.constructor().copy( this );
}
/**
* 将自己设置为一个指定三角形的复制
* @param {*} triangle
* @returns
*/
copy( triangle ) {
this.a.copy( triangle.a );
this.b.copy( triangle.b );
this.c.copy( triangle.c );
return this;
}
/**
* 计算平面三角形的面积
* @returns
*/
getArea() {
_v0.subVectors( this.c, this.b );
_v1.subVectors( this.a, this.b );
return _v0.cross( _v1 ).length() * 0.5;
}
/**
* 计算三角形的中点。
*/
getMidpoint( target ) {
if ( target === undefined ) {
console.warn( 'THREE.Triangle: .getMidpoint() target is now required' );
target = new Vector3();
}
return targ | addVectors( this.a, this.b ).add( this.c ).multiplyScalar( 1 / 3 );
}
/**
* 获取三角面的normal
* @param {*} target
* @returns
*/
getNormal( target ) {
return Triangle.getNormal( this.a, this.b, this.c, target );
}
/**
* 获取平面三角形坐在的平面
* @param {*} target
* @returns
*/
getPlane( target ) {
if ( target === undefined ) {
console.warn( 'THREE.Triangle: .getPlane() target is now required' );
target = new Plane();
}
// 通过空间中三个不共线的三个点确定一个平面
return target.setFromCoplanarPoints( this.a, this.b, this.c );
}
/**
* 计算指定点的重心坐标
* @param {*} point 指定的点
* @param {*} target 结果将会被拷贝到这个Vector3中
* @returns
*/
getBarycoord( point, target ) {
return Triangle.getBarycoord( point, this.a, this.b, this.c, target );
}
/**
* uv插值计算,通过本平面三角形三个顶点的uv,插值计算出任意点的uv。
* @param {*} point 要进行uv插值的点的位置
* @param {*} uv1 顶点1的uv
* @param {*} uv2 顶点2的uv
* @param {*} uv3 顶点3的uv
* @param {*} target 结果将会被拷贝到这个Vector2中
* @returns
*/
getUV( point, uv1, uv2, uv3, target ) {
return Triangle.getUV( point, this.a, this.b, this.c | et. | identifier_name |
Triangle.js | this;
}
/**
* 设置三角形的顶点坐标为数组中的坐标
* @param {*} points 顶点数组
* @param {*} i0 第一个点的索引
* @param {*} i1 第二个点的索引
* @param {*} i2 第三个点的索引
* @returns
*/
setFromPointsAndIndices( points, i0, i1, i2 ) {
this.a.copy( points[ i0 ] );
this.b.copy( points[ i1 ] );
this.c.copy( points[ i2 ] );
return this;
}
/**
* 返回一个该平面三角形的拷贝
* @returns
*/
clone() {
return new this.constructor().copy( this );
}
/**
* 将自己设置为一个指定三角形的复制
* @param {*} triangle
* @returns
*/
copy( triangle ) {
this.a.copy( triangle.a );
this.b.copy( triangle.b );
this.c.copy( triangle.c );
return this;
}
/**
* 计算平面三角形的面积
* @returns
*/
getArea() {
_v0.subVectors( this.c, this.b );
_v1.subVectors( this.a, this.b );
return _v0.cross( _v1 ).length() * 0.5;
}
/**
* 计算三角形的中点。
*/
getMidpoint( target ) {
if ( target === undefined ) {
console.warn( 'THREE.Triangle: .getMidpoint() target is now required' );
target = new Vector3();
}
return target.addVectors( this.a, this.b ).add( this.c ).multiplyScalar( 1 / 3 );
}
/**
* 获取三角面的normal
* @param {*} target
* @returns
*/
getNormal( target ) {
return Triangle.getNormal( this.a, this.b, this.c, target );
}
/**
* 获取平面三角形坐在的平面
* @param {*} target
* @returns
*/
getPlane( target ) {
if ( target === undefined ) {
console.warn( 'THREE.Triangle: .getPlane() target is now required' );
target = new Plane();
}
// 通过空间中三个不共线的三个点确定一个平面
return target.setFromCoplanarPoints( this.a, this.b, this.c );
}
/**
* 计算指定点的重心坐标
* @param {*} point 指定的点
* @param {*} target 结果将会被拷贝到这个Vector3中
* @returns
*/
getBarycoord( point, target ) {
return Triangle.getBarycoord( point, this.a, this.b, this.c, target );
}
/**
* uv插值计算,通过本平面三角形三个顶点的uv,插值计算出任意点的uv。
* @param {*} point 要进行uv插值的点的位置
* @param {*} uv1 顶点1的uv
* @param {*} uv2 顶点2的uv
* @param {*} uv3 顶点3的uv
* @param {*} target 结果将会被拷贝到这个Vector2中
* @returns
*/
getUV( point, uv1, uv2, uv3, target ) {
return Triangle.getUV( point, this.a, this.b, this.c, uv1, uv2, uv3, target );
}
/**
* 判断指定点是否在本平面三角形中
* @param {*} point
* @returns
*/
containsPoint( point ) {
return Triangle.containsPoint( point, this.a, this.b, this.c );
}
/**
* 判断一个给定向量是否朝向平面三角形法向量
* @param {*} direction
* @returns
*/
isFrontFacing( direction ) {
return Triangle.isFrontFacing( this.a, this.b, this.c, direction );
}
/**
* 判定三角形与传入的box是否相交
* @param {*} box
* @returns
*/
intersectsBox( box ) {
return box.intersectsTriangle( this );
}
/**
* 返回三角形上最靠近所给定的point的点
* @param {*} p 指定的点
* @param {*} target 结果将会被拷贝到这个Vector3中
* @returns
*/
closestPointToPoint( p, target ) {
if ( target === undefined ) {
console.warn( 'THREE.Triangle: .closestPointToPoint() target is now required' );
target = new Vector3();
}
const a = this.a, b = this.b, c = this.c;
let v, w;
// 算法来自christer Ericsion所著的《Real-Time Collision Detection》
// Morgan Kaufmann 出版社, (c) 2005 Elsevier Inc.,
// 版权许可。请查看5.1.5章节来了解详情。
// 基本上来说,我们希望通过最少的冗余计算来得到指定点p位于哪个voronoi区域
_vab.subVectors( b, a );
_vac.subVectors( c, a );
_vap.subVectors( p, a );
const d1 = _vab.dot( _vap );
const d2 = _vac.dot( _vap );
if ( d1 <= 0 && d2 <= 0 ) {
// vertex region of A; barycentric coords (1, 0, 0)
// 顶点A的区域,重心坐标(1,0,0)
return target.copy( a );
}
_vbp.subVectors( p, b );
const d3 = _vab.dot( _vbp );
const d4 = _vac.dot( _vbp );
if ( d3 >= 0 && d4 <= d3 ) {
// vertex region of B; barycentric coords (0, 1, 0)
// 顶点B的区域,重心坐标(0,1,0)
return target.copy( b );
}
const vc = d1 * d4 - d3 * d2;
if ( vc <= 0 && d1 >= 0 && d3 <= 0 ) {
v = d1 / ( d1 - d3 );
// edge region of AB; barycentric coords (1-v, v, 0)
// AB边的区域; 重心坐标 (1-v, v, 0)
return target.copy( a ).addScaledVector( _vab, v );
}
_vcp.subVectors( p, c );
const d5 = _vab.dot( _vcp );
const d6 = _vac.dot( _vcp );
if ( d6 >= 0 && d5 <= d6 ) {
// vertex region of C; barycentric coords (0, 0, 1)
// 顶点C的区域,重心坐标(0, 0, 1)
return target.copy( c );
}
const vb = d5 * d2 - d1 * d6;
if ( vb <= 0 && d2 >= 0 && d6 <= 0 ) {
w = d2 / ( d2 - d6 );
// edge region of AC; barycentric coords (1-w, 0, w)
// AC边的区域; 重心坐标 (1-w, 0, w)
return target.copy( a ).addScaledVector( _vac, w );
}
const va = d3 * d6 - d5 * d4;
if ( va <= 0 && ( d4 - d3 ) >= 0 && ( d5 - d6 ) >= 0 ) {
_vbc.subVectors( c, b );
w = ( d4 - d3 ) / ( ( d4 - d3 ) + ( d5 - d6 ) );
// edge region of BC; barycentric coords (0, 1-w, w)
// BC边的区域; 重心坐标 (0, 1-w, w)
return target.copy( b ).addScaledVector( _vbc, w ); // edge region of BC
}
// face region
// 平面三角形面上的区域
const denom = 1 / ( va + vb + vc );
// u = va * denom
v = vb * denom;
w = vc * denom;
return target.copy( a ).addScaledVector( _vab, v ).addScaledVector( _vac, w );
}
/**
* 判断本平面三角形是否和传入的平面三角形相同
* @param {*} triangle
* @returns
*/
equals( triangle ) {
return triangle.a.equals( this.a ) && triangle.b.equals( this.b ) && triangle.c.equals( this.c );
}
}
export { Triangle };
| conditional_block |
||
daemon.py | 8")
handle_mode(command)
# Define command to handle callback for when MQTT command "temp" arrives.
# Tell us and confirm on MQTT, then call sendtoheatpump to set the new room value.
def on_message_temp(client, userdata, msg):
print("Received temp ")
print(msg.topic + ": " + str(msg.payload))
command = msg.payload.decode("utf-8").lower()
mqttc.publish(mqtttopic + "/status/temp", int(float(command)))
sendtoheatpump('0203', int(float(command)))
# Define command to handle callback for when MQTT command "curve" arrives.
def on_message_curve(client, userdata, msg):
print("Received curve ")
print(msg.topic + ": " + str(msg.payload))
command = msg.payload.decode("utf-8").lower()
mqttc.publish(mqtttopic + "/status/curve", int(float(command)))
sendtoheatpump('0205', int(float(command)))
# Define command to handle callback for when other commands arrive on MQTT. We ignore these.
def on_message(client, userdata, msg):
print("Received unknown command ")
print(msg.topic + ": " + str(msg.payload))
# Define command for handling heatpump mode commands.
# Respond on MQTT and then call sendtoheatpump.
def handle_mode(command):
if 'Auto' == command:
print("Set mode auto")
mqttc.publish(mqtttopic + "/status/mode", "Auto")
sendtoheatpump('2201', '1')
elif 'Heatpump' == command:
print("Set mode heatpump")
mqttc.publish(mqtttopic + "/status/mode", "Heatpump")
sendtoheatpump('2201', '2')
elif 'Electricity' == command:
print("Set mode electricity")
mqttc.publish(mqtttopic + "/status/mode", "Electricity")
sendtoheatpump('2201', '3')
elif 'Water' == command:
print("Set mode water")
mqttc.publish(mqtttopic + "/status/mode", "Water")
sendtoheatpump('2201', '4')
elif 'Off' == command:
print("Set mode off")
mqttc.publish(mqtttopic + "/status/mode", "Off")
sendtoheatpump('2201', '0')
else:
print("Unknown command!")
# Define command to parse incoming status line (XRxxxxyyyy) from H1, and send using MQTT and/or HTTP
# Example: XR010701B8 107 Heating setpoint (44.0c)
def parseandsend(line):
# We need the global dict to know about recent commands sent to heatpump.
global wantvalue
# Process the XR line to get register, value and a good readable label.
print line
splitline=line.split(' (')
labels=splitline[0].split(' ')
label=labels[4:]
label.insert(0,labels[0][2:6])
label='+'.join(label)
register=labels[0][2:6]
# Confirm there were actually a label parsed. A bit paranoid at this point.
if label:
# Clean up label and value from unnessecary characters
label=re.sub('/', '+', label)
label=re.sub(',', '', label)
value = re.sub('[hpcd\) %]', '', splitline[1])
# Make sure we actually got a value there
if value:
# Now start sending received values to MQTT and/or HTTP,
# but dont send anything if we are waiting for a specific
# register to be confirmed (if exists in dict wantvalue)
if not wantvalue.has_key(register):
if mqttservername:
mqttc.publish(mqtttopic + "/" + register, int(float(value)))
# That one there sent the raw register as an MQTT topic.
# Now if the line is temp, curve or mode, send those in a friendlier way.
if labels[0][2:6] == "0203":
mqttc.publish(mqtttopic + "/status/temp", int(float(value)))
if labels[0][2:6] == "0205":
mqttc.publish(mqtttopic + "/status/curve", int(float(value)))
if labels[0][2:6] == "2201":
if int(float(value)) == 0:
mqttc.publish(mqtttopic + "/status/mode", "Off")
if int(float(value)) == 10:
mqttc.publish(mqtttopic + "/status/mode", "Auto")
if int(float(value)) == 20:
mqttc.publish(mqtttopic + "/status/mode", "Heatpump")
if int(float(value)) == 30:
mqttc.publish(mqtttopic + "/status/mode", "Electricity")
if int(float(value)) == 40:
mqttc.publish(mqtttopic + "/status/mode", "Water")
if httpservername:
url="http://" + httpservername + "/iot/iotstore.php?id=HeatPump+" + label + "&set=" + value
urllib2.urlopen(url).read()
# Return a list of the register and data we saw.
return register, int(float(value))
else:
# Corrupted line. No value. Make sure we return something anyway.
return "error", "novalue"
else:
# Corrupted line. No label parsed. Make sure we return something.
return "error", "nolabel"
# Define command to send XW commands through H1 interface to set heatpump settings
# Uses register (4 char hex string) and data (normally integer as int or string)
def sendtoheatpump(dataregister, svalue):
# We need the global dict to put the newly set registers and values in, for later verification.
global wantvalue
# Convert value string to int, multiply 10 for one decimal and format as 2 byte HEX, then form H1 XW command.
hexvalue=format(int(float(svalue))*10, '04X')
sendcommand="XW" + dataregister + hexvalue + "\r"
# Flush buffers and send XW string.
print "Writing command: " + sendcommand
ser.flushOutput()
ser.flushInput()
ser.write(sendcommand)
# Save register and value in dict wantvalue. This is later compared to incoming values
# to make sure it was received by the heatpump.
if dataregister == "2201" and float(svalue) < 10:
# For 2201 register we multiply 10 for the compare value to match later, if we havent already multiplied (resending).
# Mode number is reported by H1 as a direct decimal from the hex data (000A = 10).
# All other values are returned with its base at one decimal (000A = 1.0).
wantvalue[dataregister] = int(float(svalue))*10
else:
wantvalue[dataregister] = int(float(svalue))
# Define command to send 2 byte (XP, XS etc.) settings to H1 interface and verify it was accepted.
# Todo: only allow 2 byte strings.
def | (h1command):
h1response="init"
# Resend command every 0.5s until answer is received
while h1response[:2] != h1command:
print "Writing command: " + h1command
ser.flushOutput()
ser.flushInput()
ser.write(h1command + "\r\n")
h1response=ser.readline()
h1response=re.sub('[\n\r]', '', h1response)
print h1response
time.sleep(0.5)
print "Confirmed " + h1command
# Define command to reset H1 interface. Unused for now, but saved for reference.
def reseth1():
print "Sending reset"
ser.write("!\r\n")
# ------------------
# Start main program
# ------------------
# Set MQTT and launch its subthread, but only if we want MQTT.
if mqttservername:
mqttc = mqtt.Client(mqtttopic)
if mqttuser:
mqttc.username_pw_set(mqttuser, mqttpassword)
mqttc.message_callback_add(mqtttopic + '/command/mode', on_message_mode)
mqttc.message_callback_add(mqtttopic + '/command/temp', on_message_temp)
mqttc.message_callback_add(mqtttopic + '/command/curve', on_message_curve)
mqttc.on_connect = on_connect
mqttc.on_message = on_message
mqttc.connect(mqttservername)
mqttc.loop_start()
# Define H1 interface serial port.
ser = serial.Serial(
port='/dev/serial0',
baudrate = 19200,
timeout=10
)
# Give MQTT time to connect, then clear the serial buffers and start the business.
time.sleep(1)
print "Start collection of data..."
ser.flushOutput()
ser.flushInput()
# The business... (Main loop)
while 1:
# Read line from H1. Strip CR/LF. Times out after 10 seconds (see above)
line=ser.readline()
line=re.sub('[\n\r]', '', line)
if line:
| sendtoh1 | identifier_name |
daemon.py | 8")
handle_mode(command)
# Define command to handle callback for when MQTT command "temp" arrives.
# Tell us and confirm on MQTT, then call sendtoheatpump to set the new room value.
def on_message_temp(client, userdata, msg):
print("Received temp ")
print(msg.topic + ": " + str(msg.payload))
command = msg.payload.decode("utf-8").lower()
mqttc.publish(mqtttopic + "/status/temp", int(float(command)))
sendtoheatpump('0203', int(float(command)))
# Define command to handle callback for when MQTT command "curve" arrives.
def on_message_curve(client, userdata, msg):
print("Received curve ")
print(msg.topic + ": " + str(msg.payload))
command = msg.payload.decode("utf-8").lower()
mqttc.publish(mqtttopic + "/status/curve", int(float(command)))
sendtoheatpump('0205', int(float(command)))
# Define command to handle callback for when other commands arrive on MQTT. We ignore these.
def on_message(client, userdata, msg):
print("Received unknown command ")
print(msg.topic + ": " + str(msg.payload))
# Define command for handling heatpump mode commands.
# Respond on MQTT and then call sendtoheatpump.
def handle_mode(command):
if 'Auto' == command:
|
elif 'Heatpump' == command:
print("Set mode heatpump")
mqttc.publish(mqtttopic + "/status/mode", "Heatpump")
sendtoheatpump('2201', '2')
elif 'Electricity' == command:
print("Set mode electricity")
mqttc.publish(mqtttopic + "/status/mode", "Electricity")
sendtoheatpump('2201', '3')
elif 'Water' == command:
print("Set mode water")
mqttc.publish(mqtttopic + "/status/mode", "Water")
sendtoheatpump('2201', '4')
elif 'Off' == command:
print("Set mode off")
mqttc.publish(mqtttopic + "/status/mode", "Off")
sendtoheatpump('2201', '0')
else:
print("Unknown command!")
# Define command to parse incoming status line (XRxxxxyyyy) from H1, and send using MQTT and/or HTTP
# Example: XR010701B8 107 Heating setpoint (44.0c)
def parseandsend(line):
# We need the global dict to know about recent commands sent to heatpump.
global wantvalue
# Process the XR line to get register, value and a good readable label.
print line
splitline=line.split(' (')
labels=splitline[0].split(' ')
label=labels[4:]
label.insert(0,labels[0][2:6])
label='+'.join(label)
register=labels[0][2:6]
# Confirm there were actually a label parsed. A bit paranoid at this point.
if label:
# Clean up label and value from unnessecary characters
label=re.sub('/', '+', label)
label=re.sub(',', '', label)
value = re.sub('[hpcd\) %]', '', splitline[1])
# Make sure we actually got a value there
if value:
# Now start sending received values to MQTT and/or HTTP,
# but dont send anything if we are waiting for a specific
# register to be confirmed (if exists in dict wantvalue)
if not wantvalue.has_key(register):
if mqttservername:
mqttc.publish(mqtttopic + "/" + register, int(float(value)))
# That one there sent the raw register as an MQTT topic.
# Now if the line is temp, curve or mode, send those in a friendlier way.
if labels[0][2:6] == "0203":
mqttc.publish(mqtttopic + "/status/temp", int(float(value)))
if labels[0][2:6] == "0205":
mqttc.publish(mqtttopic + "/status/curve", int(float(value)))
if labels[0][2:6] == "2201":
if int(float(value)) == 0:
mqttc.publish(mqtttopic + "/status/mode", "Off")
if int(float(value)) == 10:
mqttc.publish(mqtttopic + "/status/mode", "Auto")
if int(float(value)) == 20:
mqttc.publish(mqtttopic + "/status/mode", "Heatpump")
if int(float(value)) == 30:
mqttc.publish(mqtttopic + "/status/mode", "Electricity")
if int(float(value)) == 40:
mqttc.publish(mqtttopic + "/status/mode", "Water")
if httpservername:
url="http://" + httpservername + "/iot/iotstore.php?id=HeatPump+" + label + "&set=" + value
urllib2.urlopen(url).read()
# Return a list of the register and data we saw.
return register, int(float(value))
else:
# Corrupted line. No value. Make sure we return something anyway.
return "error", "novalue"
else:
# Corrupted line. No label parsed. Make sure we return something.
return "error", "nolabel"
# Define command to send XW commands through H1 interface to set heatpump settings
# Uses register (4 char hex string) and data (normally integer as int or string)
def sendtoheatpump(dataregister, svalue):
# We need the global dict to put the newly set registers and values in, for later verification.
global wantvalue
# Convert value string to int, multiply 10 for one decimal and format as 2 byte HEX, then form H1 XW command.
hexvalue=format(int(float(svalue))*10, '04X')
sendcommand="XW" + dataregister + hexvalue + "\r"
# Flush buffers and send XW string.
print "Writing command: " + sendcommand
ser.flushOutput()
ser.flushInput()
ser.write(sendcommand)
# Save register and value in dict wantvalue. This is later compared to incoming values
# to make sure it was received by the heatpump.
if dataregister == "2201" and float(svalue) < 10:
# For 2201 register we multiply 10 for the compare value to match later, if we havent already multiplied (resending).
# Mode number is reported by H1 as a direct decimal from the hex data (000A = 10).
# All other values are returned with its base at one decimal (000A = 1.0).
wantvalue[dataregister] = int(float(svalue))*10
else:
wantvalue[dataregister] = int(float(svalue))
# Define command to send 2 byte (XP, XS etc.) settings to H1 interface and verify it was accepted.
# Todo: only allow 2 byte strings.
def sendtoh1(h1command):
h1response="init"
# Resend command every 0.5s until answer is received
while h1response[:2] != h1command:
print "Writing command: " + h1command
ser.flushOutput()
ser.flushInput()
ser.write(h1command + "\r\n")
h1response=ser.readline()
h1response=re.sub('[\n\r]', '', h1response)
print h1response
time.sleep(0.5)
print "Confirmed " + h1command
# Define command to reset H1 interface. Unused for now, but saved for reference.
def reseth1():
print "Sending reset"
ser.write("!\r\n")
# ------------------
# Start main program
# ------------------
# Set MQTT and launch its subthread, but only if we want MQTT.
if mqttservername:
mqttc = mqtt.Client(mqtttopic)
if mqttuser:
mqttc.username_pw_set(mqttuser, mqttpassword)
mqttc.message_callback_add(mqtttopic + '/command/mode', on_message_mode)
mqttc.message_callback_add(mqtttopic + '/command/temp', on_message_temp)
mqttc.message_callback_add(mqtttopic + '/command/curve', on_message_curve)
mqttc.on_connect = on_connect
mqttc.on_message = on_message
mqttc.connect(mqttservername)
mqttc.loop_start()
# Define H1 interface serial port.
ser = serial.Serial(
port='/dev/serial0',
baudrate = 19200,
timeout=10
)
# Give MQTT time to connect, then clear the serial buffers and start the business.
time.sleep(1)
print "Start collection of data..."
ser.flushOutput()
ser.flushInput()
# The business... (Main loop)
while 1:
# Read line from H1. Strip CR/LF. Times out after 10 seconds (see above)
line=ser.readline()
line=re.sub('[\n\r]', '', line)
if line:
| print("Set mode auto")
mqttc.publish(mqtttopic + "/status/mode", "Auto")
sendtoheatpump('2201', '1') | conditional_block |
daemon.py | 8")
handle_mode(command)
# Define command to handle callback for when MQTT command "temp" arrives.
# Tell us and confirm on MQTT, then call sendtoheatpump to set the new room value.
def on_message_temp(client, userdata, msg):
print("Received temp ")
print(msg.topic + ": " + str(msg.payload))
command = msg.payload.decode("utf-8").lower()
mqttc.publish(mqtttopic + "/status/temp", int(float(command)))
sendtoheatpump('0203', int(float(command)))
# Define command to handle callback for when MQTT command "curve" arrives.
def on_message_curve(client, userdata, msg):
print("Received curve ")
print(msg.topic + ": " + str(msg.payload))
command = msg.payload.decode("utf-8").lower()
mqttc.publish(mqtttopic + "/status/curve", int(float(command)))
sendtoheatpump('0205', int(float(command)))
# Define command to handle callback for when other commands arrive on MQTT. We ignore these.
def on_message(client, userdata, msg):
|
# Define command for handling heatpump mode commands.
# Respond on MQTT and then call sendtoheatpump.
def handle_mode(command):
if 'Auto' == command:
print("Set mode auto")
mqttc.publish(mqtttopic + "/status/mode", "Auto")
sendtoheatpump('2201', '1')
elif 'Heatpump' == command:
print("Set mode heatpump")
mqttc.publish(mqtttopic + "/status/mode", "Heatpump")
sendtoheatpump('2201', '2')
elif 'Electricity' == command:
print("Set mode electricity")
mqttc.publish(mqtttopic + "/status/mode", "Electricity")
sendtoheatpump('2201', '3')
elif 'Water' == command:
print("Set mode water")
mqttc.publish(mqtttopic + "/status/mode", "Water")
sendtoheatpump('2201', '4')
elif 'Off' == command:
print("Set mode off")
mqttc.publish(mqtttopic + "/status/mode", "Off")
sendtoheatpump('2201', '0')
else:
print("Unknown command!")
# Define command to parse incoming status line (XRxxxxyyyy) from H1, and send using MQTT and/or HTTP
# Example: XR010701B8 107 Heating setpoint (44.0c)
def parseandsend(line):
# We need the global dict to know about recent commands sent to heatpump.
global wantvalue
# Process the XR line to get register, value and a good readable label.
print line
splitline=line.split(' (')
labels=splitline[0].split(' ')
label=labels[4:]
label.insert(0,labels[0][2:6])
label='+'.join(label)
register=labels[0][2:6]
# Confirm there were actually a label parsed. A bit paranoid at this point.
if label:
# Clean up label and value from unnessecary characters
label=re.sub('/', '+', label)
label=re.sub(',', '', label)
value = re.sub('[hpcd\) %]', '', splitline[1])
# Make sure we actually got a value there
if value:
# Now start sending received values to MQTT and/or HTTP,
# but dont send anything if we are waiting for a specific
# register to be confirmed (if exists in dict wantvalue)
if not wantvalue.has_key(register):
if mqttservername:
mqttc.publish(mqtttopic + "/" + register, int(float(value)))
# That one there sent the raw register as an MQTT topic.
# Now if the line is temp, curve or mode, send those in a friendlier way.
if labels[0][2:6] == "0203":
mqttc.publish(mqtttopic + "/status/temp", int(float(value)))
if labels[0][2:6] == "0205":
mqttc.publish(mqtttopic + "/status/curve", int(float(value)))
if labels[0][2:6] == "2201":
if int(float(value)) == 0:
mqttc.publish(mqtttopic + "/status/mode", "Off")
if int(float(value)) == 10:
mqttc.publish(mqtttopic + "/status/mode", "Auto")
if int(float(value)) == 20:
mqttc.publish(mqtttopic + "/status/mode", "Heatpump")
if int(float(value)) == 30:
mqttc.publish(mqtttopic + "/status/mode", "Electricity")
if int(float(value)) == 40:
mqttc.publish(mqtttopic + "/status/mode", "Water")
if httpservername:
url="http://" + httpservername + "/iot/iotstore.php?id=HeatPump+" + label + "&set=" + value
urllib2.urlopen(url).read()
# Return a list of the register and data we saw.
return register, int(float(value))
else:
# Corrupted line. No value. Make sure we return something anyway.
return "error", "novalue"
else:
# Corrupted line. No label parsed. Make sure we return something.
return "error", "nolabel"
# Define command to send XW commands through H1 interface to set heatpump settings
# Uses register (4 char hex string) and data (normally integer as int or string)
def sendtoheatpump(dataregister, svalue):
# We need the global dict to put the newly set registers and values in, for later verification.
global wantvalue
# Convert value string to int, multiply 10 for one decimal and format as 2 byte HEX, then form H1 XW command.
hexvalue=format(int(float(svalue))*10, '04X')
sendcommand="XW" + dataregister + hexvalue + "\r"
# Flush buffers and send XW string.
print "Writing command: " + sendcommand
ser.flushOutput()
ser.flushInput()
ser.write(sendcommand)
# Save register and value in dict wantvalue. This is later compared to incoming values
# to make sure it was received by the heatpump.
if dataregister == "2201" and float(svalue) < 10:
# For 2201 register we multiply 10 for the compare value to match later, if we havent already multiplied (resending).
# Mode number is reported by H1 as a direct decimal from the hex data (000A = 10).
# All other values are returned with its base at one decimal (000A = 1.0).
wantvalue[dataregister] = int(float(svalue))*10
else:
wantvalue[dataregister] = int(float(svalue))
# Define command to send 2 byte (XP, XS etc.) settings to H1 interface and verify it was accepted.
# Todo: only allow 2 byte strings.
def sendtoh1(h1command):
h1response="init"
# Resend command every 0.5s until answer is received
while h1response[:2] != h1command:
print "Writing command: " + h1command
ser.flushOutput()
ser.flushInput()
ser.write(h1command + "\r\n")
h1response=ser.readline()
h1response=re.sub('[\n\r]', '', h1response)
print h1response
time.sleep(0.5)
print "Confirmed " + h1command
# Define command to reset H1 interface. Unused for now, but saved for reference.
def reseth1():
print "Sending reset"
ser.write("!\r\n")
# ------------------
# Start main program
# ------------------
# Set MQTT and launch its subthread, but only if we want MQTT.
if mqttservername:
mqttc = mqtt.Client(mqtttopic)
if mqttuser:
mqttc.username_pw_set(mqttuser, mqttpassword)
mqttc.message_callback_add(mqtttopic + '/command/mode', on_message_mode)
mqttc.message_callback_add(mqtttopic + '/command/temp', on_message_temp)
mqttc.message_callback_add(mqtttopic + '/command/curve', on_message_curve)
mqttc.on_connect = on_connect
mqttc.on_message = on_message
mqttc.connect(mqttservername)
mqttc.loop_start()
# Define H1 interface serial port.
ser = serial.Serial(
port='/dev/serial0',
baudrate = 19200,
timeout=10
)
# Give MQTT time to connect, then clear the serial buffers and start the business.
time.sleep(1)
print "Start collection of data..."
ser.flushOutput()
ser.flushInput()
# The business... (Main loop)
while 1:
# Read line from H1. Strip CR/LF. Times out after 10 seconds (see above)
line=ser.readline()
line=re.sub('[\n\r]', '', line)
if line:
| print("Received unknown command ")
print(msg.topic + ": " + str(msg.payload)) | identifier_body |
daemon.py | ', int(float(command)))
# Define command to handle callback for when MQTT command "curve" arrives.
def on_message_curve(client, userdata, msg):
print("Received curve ")
print(msg.topic + ": " + str(msg.payload))
command = msg.payload.decode("utf-8").lower()
mqttc.publish(mqtttopic + "/status/curve", int(float(command)))
sendtoheatpump('0205', int(float(command)))
# Define command to handle callback for when other commands arrive on MQTT. We ignore these.
def on_message(client, userdata, msg):
print("Received unknown command ")
print(msg.topic + ": " + str(msg.payload))
# Define command for handling heatpump mode commands.
# Respond on MQTT and then call sendtoheatpump.
def handle_mode(command):
if 'Auto' == command:
print("Set mode auto")
mqttc.publish(mqtttopic + "/status/mode", "Auto")
sendtoheatpump('2201', '1')
elif 'Heatpump' == command:
print("Set mode heatpump")
mqttc.publish(mqtttopic + "/status/mode", "Heatpump")
sendtoheatpump('2201', '2')
elif 'Electricity' == command:
print("Set mode electricity")
mqttc.publish(mqtttopic + "/status/mode", "Electricity")
sendtoheatpump('2201', '3')
elif 'Water' == command:
print("Set mode water")
mqttc.publish(mqtttopic + "/status/mode", "Water")
sendtoheatpump('2201', '4')
elif 'Off' == command:
print("Set mode off")
mqttc.publish(mqtttopic + "/status/mode", "Off")
sendtoheatpump('2201', '0')
else:
print("Unknown command!")
# Define command to parse incoming status line (XRxxxxyyyy) from H1, and send using MQTT and/or HTTP
# Example: XR010701B8 107 Heating setpoint (44.0c)
def parseandsend(line):
# We need the global dict to know about recent commands sent to heatpump.
global wantvalue
# Process the XR line to get register, value and a good readable label.
print line
splitline=line.split(' (')
labels=splitline[0].split(' ')
label=labels[4:]
label.insert(0,labels[0][2:6])
label='+'.join(label)
register=labels[0][2:6]
# Confirm there were actually a label parsed. A bit paranoid at this point.
if label:
# Clean up label and value from unnessecary characters
label=re.sub('/', '+', label)
label=re.sub(',', '', label)
value = re.sub('[hpcd\) %]', '', splitline[1])
# Make sure we actually got a value there
if value:
# Now start sending received values to MQTT and/or HTTP,
# but dont send anything if we are waiting for a specific
# register to be confirmed (if exists in dict wantvalue)
if not wantvalue.has_key(register):
if mqttservername:
mqttc.publish(mqtttopic + "/" + register, int(float(value)))
# That one there sent the raw register as an MQTT topic.
# Now if the line is temp, curve or mode, send those in a friendlier way.
if labels[0][2:6] == "0203":
mqttc.publish(mqtttopic + "/status/temp", int(float(value)))
if labels[0][2:6] == "0205":
mqttc.publish(mqtttopic + "/status/curve", int(float(value)))
if labels[0][2:6] == "2201":
if int(float(value)) == 0:
mqttc.publish(mqtttopic + "/status/mode", "Off")
if int(float(value)) == 10:
mqttc.publish(mqtttopic + "/status/mode", "Auto")
if int(float(value)) == 20:
mqttc.publish(mqtttopic + "/status/mode", "Heatpump")
if int(float(value)) == 30:
mqttc.publish(mqtttopic + "/status/mode", "Electricity")
if int(float(value)) == 40:
mqttc.publish(mqtttopic + "/status/mode", "Water")
if httpservername:
url="http://" + httpservername + "/iot/iotstore.php?id=HeatPump+" + label + "&set=" + value
urllib2.urlopen(url).read()
# Return a list of the register and data we saw.
return register, int(float(value))
else:
# Corrupted line. No value. Make sure we return something anyway.
return "error", "novalue"
else:
# Corrupted line. No label parsed. Make sure we return something.
return "error", "nolabel"
# Define command to send XW commands through H1 interface to set heatpump settings
# Uses register (4 char hex string) and data (normally integer as int or string)
def sendtoheatpump(dataregister, svalue):
# We need the global dict to put the newly set registers and values in, for later verification.
global wantvalue
# Convert value string to int, multiply 10 for one decimal and format as 2 byte HEX, then form H1 XW command.
hexvalue=format(int(float(svalue))*10, '04X')
sendcommand="XW" + dataregister + hexvalue + "\r"
# Flush buffers and send XW string.
print "Writing command: " + sendcommand
ser.flushOutput()
ser.flushInput()
ser.write(sendcommand)
# Save register and value in dict wantvalue. This is later compared to incoming values
# to make sure it was received by the heatpump.
if dataregister == "2201" and float(svalue) < 10:
# For 2201 register we multiply 10 for the compare value to match later, if we havent already multiplied (resending).
# Mode number is reported by H1 as a direct decimal from the hex data (000A = 10).
# All other values are returned with its base at one decimal (000A = 1.0).
wantvalue[dataregister] = int(float(svalue))*10
else:
wantvalue[dataregister] = int(float(svalue))
# Define command to send 2 byte (XP, XS etc.) settings to H1 interface and verify it was accepted.
# Todo: only allow 2 byte strings.
def sendtoh1(h1command):
h1response="init"
# Resend command every 0.5s until answer is received
while h1response[:2] != h1command:
print "Writing command: " + h1command
ser.flushOutput()
ser.flushInput()
ser.write(h1command + "\r\n")
h1response=ser.readline()
h1response=re.sub('[\n\r]', '', h1response)
print h1response
time.sleep(0.5)
print "Confirmed " + h1command
# Define command to reset H1 interface. Unused for now, but saved for reference.
def reseth1():
print "Sending reset"
ser.write("!\r\n")
# ------------------
# Start main program
# ------------------
# Set MQTT and launch its subthread, but only if we want MQTT.
if mqttservername:
mqttc = mqtt.Client(mqtttopic)
if mqttuser:
mqttc.username_pw_set(mqttuser, mqttpassword)
mqttc.message_callback_add(mqtttopic + '/command/mode', on_message_mode)
mqttc.message_callback_add(mqtttopic + '/command/temp', on_message_temp)
mqttc.message_callback_add(mqtttopic + '/command/curve', on_message_curve)
mqttc.on_connect = on_connect
mqttc.on_message = on_message
mqttc.connect(mqttservername)
mqttc.loop_start()
# Define H1 interface serial port.
ser = serial.Serial(
port='/dev/serial0',
baudrate = 19200,
timeout=10
)
# Give MQTT time to connect, then clear the serial buffers and start the business.
time.sleep(1)
print "Start collection of data..."
ser.flushOutput()
ser.flushInput()
# The business... (Main loop)
while 1:
# Read line from H1. Strip CR/LF. Times out after 10 seconds (see above)
line=ser.readline()
line=re.sub('[\n\r]', '', line)
if line:
# If we got a line (no timeout) we do checking and parsing of it
if line[:2] == "XR":
# Only care about lines with heatpump data. (XRxxxxyyyy)
if len(line) <= 10:
# If the data line is only 10 characters, we assume the H1 reset and lost its settings. | # Make H1 send readable labels and regular full updates
sendtoh1("XP")
sendtoh1("XM")
else: | random_line_split |
|
worker_actions.rs | No such hobo id")?;
}
validate_ability(db, task.task_type, worker_id, timestamp)?;
let new_task = NewTask {
worker_id: worker_id.num(),
task_type: task.task_type,
x: task.x as i32,
y: task.y as i32,
start_time: Some(timestamp),
target_hobo_id: task.target,
};
simulate_begin_task(&new_task, &mut town, &mut worker)?;
let duration = simulate_finish_task(&new_task, &mut town, &mut worker)?;
tasks.push(new_task);
timestamp += duration;
}
Ok(ValidatedTaskList {
new_tasks: tasks,
update_tasks: vec![current_task],
village_id,
})
}
pub(crate) fn replace_worker_tasks(
db: &DB,
worker: &Addr<TownWorker>,
worker_id: WorkerKey,
tasks: &[NewTask],
village_id: VillageKey,
) {
db.flush_task_queue(worker_id);
let _inserted = db.insert_tasks(tasks);
let current_task =
execute_worker_tasks(db, worker_id, village_id).expect("Worker has no current task");
if let Some(next_task) = db.earliest_future_task(worker_id) {
let event = Event::WorkerTask {
task_id: current_task.key(),
};
worker
.send(TownWorkerEventMsg(
event,
Utc.from_utc_datetime(&next_task.start_time),
))
.wait()
.expect("Send msg to actor");
}
}
fn interrupt_task(current_task: &mut Task, worker: &Worker) -> Option<NaiveDateTime> {
match current_task.task_type {
TaskType::Idle
| TaskType::ChopTree
| TaskType::Defend
| TaskType::GatherSticks
| TaskType::CollectReward => {
let now = chrono::Utc::now().naive_utc();
Some(now)
}
TaskType::Walk => {
let speed = unit_speed_to_worker_tiles_per_second(worker.speed) as f64;
let time_so_far: Duration = Utc::now().naive_utc() - current_task.start_time;
let steps = (speed * time_so_far.num_microseconds().unwrap() as f64 / 1_000_000.0)
.ceil() as i32;
let total_time = steps as f64 / speed;
let moment = current_task.start_time
+ chrono::Duration::microseconds((total_time * 1_000_000.0) as i64);
let dx = current_task.x - worker.x;
let dy = current_task.y - worker.y;
let x = if dx == 0 {
worker.x
} else if dx < 0 {
worker.x - steps
} else {
worker.x + steps
};
let y = if dy == 0 {
worker.y
} else if dy < 0 {
worker.y - steps
} else {
worker.y + steps
};
// Walking must terminate earlier
current_task.x = x;
current_task.y = y;
Some(moment)
}
TaskType::WelcomeAbility => {
let cast_time = current_task.start_time + AbilityType::Welcome.busy_duration();
Some(cast_time)
}
}
}
/// For the given worker, executes tasks on the DB that are due
fn execute_worker_tasks(db: &DB, worker_id: WorkerKey, village: VillageKey) -> Option<Task> {
let mut tasks = db.past_worker_tasks(worker_id);
let current_task = tasks.pop();
let mut town = TownView::load_village(db, village);
for task in tasks {
if let Err(e) = finish_task(db, task.key(), Some(task), Some(&mut town)) {
println!("Executing task failed: {}", e)
}
}
current_task
}
pub(crate) fn finish_task(
db: &DB,
task_id: TaskKey,
task: Option<Task>,
town: Option<&mut TownView>,
) -> Result<Option<(Event, DateTime<Utc>)>, Box<dyn std::error::Error>> {
let task = task.or_else(|| db.task(task_id));
if let Some(task) = task {
let mut worker = db
.worker_priv(task.worker())
.ok_or("Task references non-existing worker")?;
if let Some(town) = town {
crate::worker_actions::simulate_finish_task(&task, town, &mut worker)?;
apply_task_to_db(db, &task, &mut worker)?;
} else {
let mut town = TownView::load_village(db, VillageKey(worker.home));
crate::worker_actions::simulate_finish_task(&task, &mut town, &mut worker)?;
apply_task_to_db(db, &task, &mut worker)?;
}
db.update_worker(&worker);
db.update_worker_flag_timestamp_now(worker.key(), WorkerFlagType::Work);
db.delete_task(&task);
Ok(Event::load_next_worker_task(db, task.worker()))
} else {
// Already executed.
Ok(None)
}
}
fn apply_task_to_db(db: &DB, task: &Task, worker: &mut Worker) -> Result<(), String> {
match task.task_type {
TaskType::WelcomeAbility => {
let a = AbilityType::Welcome;
let (attribute, strength) = a.apply();
let ne = NewEffect {
hobo_id: task.target().ok_or("Ability must have a target")?.num(),
attribute,
strength: Some(strength),
start_time: None, // default = now
};
db.insert_effect(&ne);
db.update_ability_used_timestamp(WorkerKey(worker.id), a);
*worker.mana.as_mut().unwrap() -= AbilityType::Welcome.mana_cost();
}
TaskType::CollectReward => {
if let Some(building) = db.find_building_by_coordinates(task.x, task.y, worker.home()) {
match building.building_type.reward_exp() {
Some(exp) => {
worker.add_exp(exp);
db.delete_building(&building);
}
None => {
return Err(format!(
"Tried to collect {} as reward",
building.building_type
));
}
}
} else {
return Err(format!("No reward to collect at {},{}", task.x, task.y));
}
}
_ => { /* NOP */ }
}
Ok(())
}
/// (Try to) apply changes to village state that happen when a worker stops doing a given task.
/// E.g. remove unit from building.
/// Returns the time it takes until the task is actually finished.
fn simulate_finish_task<T: WorkerAction>(
task: &T,
town: &mut TownView,
worker: &mut Worker,
) -> Result<Duration, String> {
match task.task_type() {
TaskType::Idle => Ok(Duration::milliseconds(0)),
TaskType::Walk => Ok(worker_walk(
town,
worker,
(task.x() as usize, task.y() as usize),
)?),
TaskType::GatherSticks | TaskType::ChopTree => {
town.state
.register_task_end(*task.task_type())
.map_err(|e| e.to_string())?;
worker_out_of_building(town, worker, (task.x() as usize, task.y() as usize))
}
TaskType::WelcomeAbility => {
let a = AbilityType::Welcome;
let duration = a.busy_duration();
Ok(duration)
}
TaskType::CollectReward => {
// Lookup object to be collected, then delete it in TownView
// Note: DB update is separate
let index = (task.x() as usize, task.y() as usize);
town.state.remove(&index);
Ok(Duration::milliseconds(0))
}
TaskType::Defend => Err("Task not implemented".to_owned()),
}
}
/// (Try to) apply changes to village state that happen when a worker starts a given task.
/// E.g. add unit to a building, or pay required price (only if it is TownView), ...
fn simulate_begin_task<T: WorkerAction>(
task: &T,
town: &mut TownView,
worker: &mut Worker,
) -> Result<(), String> {
match task.task_type() {
TaskType::Idle | TaskType::Walk | TaskType::CollectReward => Ok(()),
TaskType::GatherSticks | TaskType::ChopTree => {
town.state
.register_task_begin(*task.task_type())
.map_err(|e| e.to_string())?;
worker_into_building(town, worker, (task.x() as usize, task.y() as usize))
}
TaskType::WelcomeAbility => {
if let Some(mana) = &mut worker.mana {
let cost = AbilityType::Welcome.mana_cost();
if *mana >= cost {
*mana = *mana - cost;
Ok(())
} else {
Err("Not enough mana".to_owned())
}
} else {
Err("Worker has no mana but tries to use welcome ability".to_owned())
}
}
TaskType::Defend => Err("Task not implemented".to_owned()),
}
}
impl WorkerAction for NewTask {
fn x(&self) -> i32 {
self.x
}
fn | y | identifier_name |
|
worker_actions.rs | // check timing and effect of current task interruption
let mut current_task = db
.current_task(worker.key())
.expect("Must have a current task");
let mut timestamp =
interrupt_task(&mut current_task, &worker).ok_or("Cannot interrupt current task.")?;
worker.x = current_task.x;
worker.y = current_task.y;
// iterate tasks and match for task types
let mut tasks = vec![];
for task in tl.tasks.iter() {
// Validate target hobo exists if there is one
if let Some(target_id) = task.target {
db.hobo(HoboKey(target_id)).ok_or("No such hobo id")?;
}
validate_ability(db, task.task_type, worker_id, timestamp)?;
let new_task = NewTask {
worker_id: worker_id.num(),
task_type: task.task_type,
x: task.x as i32,
y: task.y as i32,
start_time: Some(timestamp),
target_hobo_id: task.target,
};
simulate_begin_task(&new_task, &mut town, &mut worker)?;
let duration = simulate_finish_task(&new_task, &mut town, &mut worker)?;
tasks.push(new_task);
timestamp += duration;
}
Ok(ValidatedTaskList {
new_tasks: tasks,
update_tasks: vec![current_task],
village_id,
})
}
pub(crate) fn replace_worker_tasks(
db: &DB,
worker: &Addr<TownWorker>,
worker_id: WorkerKey,
tasks: &[NewTask],
village_id: VillageKey,
) {
db.flush_task_queue(worker_id);
let _inserted = db.insert_tasks(tasks);
let current_task =
execute_worker_tasks(db, worker_id, village_id).expect("Worker has no current task");
if let Some(next_task) = db.earliest_future_task(worker_id) {
let event = Event::WorkerTask {
task_id: current_task.key(),
};
worker
.send(TownWorkerEventMsg(
event,
Utc.from_utc_datetime(&next_task.start_time),
))
.wait()
.expect("Send msg to actor");
}
}
fn interrupt_task(current_task: &mut Task, worker: &Worker) -> Option<NaiveDateTime> {
match current_task.task_type {
TaskType::Idle
| TaskType::ChopTree
| TaskType::Defend
| TaskType::GatherSticks
| TaskType::CollectReward => {
let now = chrono::Utc::now().naive_utc();
Some(now)
}
TaskType::Walk => {
let speed = unit_speed_to_worker_tiles_per_second(worker.speed) as f64;
let time_so_far: Duration = Utc::now().naive_utc() - current_task.start_time;
let steps = (speed * time_so_far.num_microseconds().unwrap() as f64 / 1_000_000.0)
.ceil() as i32;
let total_time = steps as f64 / speed;
let moment = current_task.start_time
+ chrono::Duration::microseconds((total_time * 1_000_000.0) as i64);
let dx = current_task.x - worker.x;
let dy = current_task.y - worker.y;
let x = if dx == 0 {
worker.x
} else if dx < 0 {
worker.x - steps
} else {
worker.x + steps
};
let y = if dy == 0 {
worker.y
} else if dy < 0 {
worker.y - steps
} else {
worker.y + steps
};
// Walking must terminate earlier
current_task.x = x;
current_task.y = y;
Some(moment)
}
TaskType::WelcomeAbility => {
let cast_time = current_task.start_time + AbilityType::Welcome.busy_duration();
Some(cast_time)
}
}
}
/// For the given worker, executes tasks on the DB that are due
fn execute_worker_tasks(db: &DB, worker_id: WorkerKey, village: VillageKey) -> Option<Task> {
let mut tasks = db.past_worker_tasks(worker_id);
let current_task = tasks.pop();
let mut town = TownView::load_village(db, village);
for task in tasks {
if let Err(e) = finish_task(db, task.key(), Some(task), Some(&mut town)) {
println!("Executing task failed: {}", e)
}
}
current_task
}
pub(crate) fn finish_task(
db: &DB,
task_id: TaskKey,
task: Option<Task>,
town: Option<&mut TownView>,
) -> Result<Option<(Event, DateTime<Utc>)>, Box<dyn std::error::Error>> {
let task = task.or_else(|| db.task(task_id));
if let Some(task) = task {
let mut worker = db
.worker_priv(task.worker())
.ok_or("Task references non-existing worker")?;
if let Some(town) = town {
crate::worker_actions::simulate_finish_task(&task, town, &mut worker)?;
apply_task_to_db(db, &task, &mut worker)?;
} else {
let mut town = TownView::load_village(db, VillageKey(worker.home));
crate::worker_actions::simulate_finish_task(&task, &mut town, &mut worker)?;
apply_task_to_db(db, &task, &mut worker)?;
}
db.update_worker(&worker);
db.update_worker_flag_timestamp_now(worker.key(), WorkerFlagType::Work);
db.delete_task(&task);
Ok(Event::load_next_worker_task(db, task.worker()))
} else {
// Already executed.
Ok(None)
}
}
fn apply_task_to_db(db: &DB, task: &Task, worker: &mut Worker) -> Result<(), String> {
match task.task_type {
TaskType::WelcomeAbility => {
let a = AbilityType::Welcome;
let (attribute, strength) = a.apply();
let ne = NewEffect {
hobo_id: task.target().ok_or("Ability must have a target")?.num(),
attribute,
strength: Some(strength),
start_time: None, // default = now
};
db.insert_effect(&ne);
db.update_ability_used_timestamp(WorkerKey(worker.id), a);
*worker.mana.as_mut().unwrap() -= AbilityType::Welcome.mana_cost();
}
TaskType::CollectReward => {
if let Some(building) = db.find_building_by_coordinates(task.x, task.y, worker.home()) {
match building.building_type.reward_exp() {
Some(exp) => {
worker.add_exp(exp);
db.delete_building(&building);
}
None => {
return Err(format!(
"Tried to collect {} as reward",
building.building_type
));
}
}
} else {
return Err(format!("No reward to collect at {},{}", task.x, task.y));
}
}
_ => { /* NOP */ }
}
Ok(())
}
/// (Try to) apply changes to village state that happen when a worker stops doing a given task.
/// E.g. remove unit from building.
/// Returns the time it takes until the task is actually finished.
fn simulate_finish_task<T: WorkerAction>(
task: &T,
town: &mut TownView,
worker: &mut Worker,
) -> Result<Duration, String> {
match task.task_type() {
TaskType::Idle => Ok(Duration::milliseconds(0)),
TaskType::Walk => Ok(worker_walk(
town,
worker,
(task.x() as usize, task.y() as usize),
)?),
TaskType::GatherSticks | TaskType::ChopTree => {
town.state
.register_task_end(*task.task_type())
.map_err(|e| e.to_string())?;
worker_out_of_building(town, worker, (task.x() as usize, task.y() as usize))
}
TaskType::WelcomeAbility => {
let a = AbilityType::Welcome;
let duration = a.busy_duration();
Ok(duration)
}
TaskType::CollectReward => {
// Lookup object to be collected, then delete it in TownView
// Note: DB update is separate
let index = (task.x() as usize, task.y() as usize);
town.state.remove(&index);
Ok(Duration::milliseconds(0))
}
TaskType::Defend => Err("Task not implemented".to_owned()),
}
}
/// (Try to) apply changes to village state that happen when a worker starts a given task.
/// E.g. add unit to a building, or pay required price (only if it is TownView), ...
fn simulate_begin_task<T: WorkerAction>(
task: &T,
town: &mut TownView,
worker: &mut Worker,
) -> Result<(), String> {
match task.task_type() {
TaskType::Idle | TaskType::Walk | TaskType::CollectReward => Ok(()),
TaskType::GatherSticks | TaskType::ChopTree => | {
town.state
.register_task_begin(*task.task_type())
.map_err(|e| e.to_string())?;
worker_into_building(town, worker, (task.x() as usize, task.y() as usize))
} | conditional_block |
|
worker_actions.rs | };
simulate_begin_task(&new_task, &mut town, &mut worker)?;
let duration = simulate_finish_task(&new_task, &mut town, &mut worker)?;
tasks.push(new_task);
timestamp += duration;
}
Ok(ValidatedTaskList {
new_tasks: tasks,
update_tasks: vec![current_task],
village_id,
})
}
pub(crate) fn replace_worker_tasks(
db: &DB,
worker: &Addr<TownWorker>,
worker_id: WorkerKey,
tasks: &[NewTask],
village_id: VillageKey,
) {
db.flush_task_queue(worker_id);
let _inserted = db.insert_tasks(tasks);
let current_task =
execute_worker_tasks(db, worker_id, village_id).expect("Worker has no current task");
if let Some(next_task) = db.earliest_future_task(worker_id) {
let event = Event::WorkerTask {
task_id: current_task.key(),
};
worker
.send(TownWorkerEventMsg(
event,
Utc.from_utc_datetime(&next_task.start_time),
))
.wait()
.expect("Send msg to actor");
}
}
fn interrupt_task(current_task: &mut Task, worker: &Worker) -> Option<NaiveDateTime> {
match current_task.task_type {
TaskType::Idle
| TaskType::ChopTree
| TaskType::Defend
| TaskType::GatherSticks
| TaskType::CollectReward => {
let now = chrono::Utc::now().naive_utc();
Some(now)
}
TaskType::Walk => {
let speed = unit_speed_to_worker_tiles_per_second(worker.speed) as f64;
let time_so_far: Duration = Utc::now().naive_utc() - current_task.start_time;
let steps = (speed * time_so_far.num_microseconds().unwrap() as f64 / 1_000_000.0)
.ceil() as i32;
let total_time = steps as f64 / speed;
let moment = current_task.start_time
+ chrono::Duration::microseconds((total_time * 1_000_000.0) as i64);
let dx = current_task.x - worker.x;
let dy = current_task.y - worker.y;
let x = if dx == 0 {
worker.x
} else if dx < 0 {
worker.x - steps
} else {
worker.x + steps
};
let y = if dy == 0 {
worker.y
} else if dy < 0 {
worker.y - steps
} else {
worker.y + steps
};
// Walking must terminate earlier
current_task.x = x;
current_task.y = y;
Some(moment)
}
TaskType::WelcomeAbility => {
let cast_time = current_task.start_time + AbilityType::Welcome.busy_duration();
Some(cast_time)
}
}
}
/// For the given worker, executes tasks on the DB that are due
fn execute_worker_tasks(db: &DB, worker_id: WorkerKey, village: VillageKey) -> Option<Task> {
let mut tasks = db.past_worker_tasks(worker_id);
let current_task = tasks.pop();
let mut town = TownView::load_village(db, village);
for task in tasks {
if let Err(e) = finish_task(db, task.key(), Some(task), Some(&mut town)) {
println!("Executing task failed: {}", e)
}
}
current_task
}
pub(crate) fn finish_task(
db: &DB,
task_id: TaskKey,
task: Option<Task>,
town: Option<&mut TownView>,
) -> Result<Option<(Event, DateTime<Utc>)>, Box<dyn std::error::Error>> {
let task = task.or_else(|| db.task(task_id));
if let Some(task) = task {
let mut worker = db
.worker_priv(task.worker())
.ok_or("Task references non-existing worker")?;
if let Some(town) = town {
crate::worker_actions::simulate_finish_task(&task, town, &mut worker)?;
apply_task_to_db(db, &task, &mut worker)?;
} else {
let mut town = TownView::load_village(db, VillageKey(worker.home));
crate::worker_actions::simulate_finish_task(&task, &mut town, &mut worker)?;
apply_task_to_db(db, &task, &mut worker)?;
}
db.update_worker(&worker);
db.update_worker_flag_timestamp_now(worker.key(), WorkerFlagType::Work);
db.delete_task(&task);
Ok(Event::load_next_worker_task(db, task.worker()))
} else {
// Already executed.
Ok(None)
}
}
fn apply_task_to_db(db: &DB, task: &Task, worker: &mut Worker) -> Result<(), String> {
match task.task_type {
TaskType::WelcomeAbility => {
let a = AbilityType::Welcome;
let (attribute, strength) = a.apply();
let ne = NewEffect {
hobo_id: task.target().ok_or("Ability must have a target")?.num(),
attribute,
strength: Some(strength),
start_time: None, // default = now
};
db.insert_effect(&ne);
db.update_ability_used_timestamp(WorkerKey(worker.id), a);
*worker.mana.as_mut().unwrap() -= AbilityType::Welcome.mana_cost();
}
TaskType::CollectReward => {
if let Some(building) = db.find_building_by_coordinates(task.x, task.y, worker.home()) {
match building.building_type.reward_exp() {
Some(exp) => {
worker.add_exp(exp);
db.delete_building(&building);
}
None => {
return Err(format!(
"Tried to collect {} as reward",
building.building_type
));
}
}
} else {
return Err(format!("No reward to collect at {},{}", task.x, task.y));
}
}
_ => { /* NOP */ }
}
Ok(())
}
/// (Try to) apply changes to village state that happen when a worker stops doing a given task.
/// E.g. remove unit from building.
/// Returns the time it takes until the task is actually finished.
fn simulate_finish_task<T: WorkerAction>(
task: &T,
town: &mut TownView,
worker: &mut Worker,
) -> Result<Duration, String> {
match task.task_type() {
TaskType::Idle => Ok(Duration::milliseconds(0)),
TaskType::Walk => Ok(worker_walk(
town,
worker,
(task.x() as usize, task.y() as usize),
)?),
TaskType::GatherSticks | TaskType::ChopTree => {
town.state
.register_task_end(*task.task_type())
.map_err(|e| e.to_string())?;
worker_out_of_building(town, worker, (task.x() as usize, task.y() as usize))
}
TaskType::WelcomeAbility => {
let a = AbilityType::Welcome;
let duration = a.busy_duration();
Ok(duration)
}
TaskType::CollectReward => {
// Lookup object to be collected, then delete it in TownView
// Note: DB update is separate
let index = (task.x() as usize, task.y() as usize);
town.state.remove(&index);
Ok(Duration::milliseconds(0))
}
TaskType::Defend => Err("Task not implemented".to_owned()),
}
}
/// (Try to) apply changes to village state that happen when a worker starts a given task.
/// E.g. add unit to a building, or pay required price (only if it is TownView), ...
fn simulate_begin_task<T: WorkerAction>(
task: &T,
town: &mut TownView,
worker: &mut Worker,
) -> Result<(), String> {
match task.task_type() {
TaskType::Idle | TaskType::Walk | TaskType::CollectReward => Ok(()),
TaskType::GatherSticks | TaskType::ChopTree => {
town.state
.register_task_begin(*task.task_type())
.map_err(|e| e.to_string())?;
worker_into_building(town, worker, (task.x() as usize, task.y() as usize))
}
TaskType::WelcomeAbility => {
if let Some(mana) = &mut worker.mana {
let cost = AbilityType::Welcome.mana_cost();
if *mana >= cost {
*mana = *mana - cost;
Ok(())
} else {
Err("Not enough mana".to_owned())
}
} else {
Err("Worker has no mana but tries to use welcome ability".to_owned())
}
}
TaskType::Defend => Err("Task not implemented".to_owned()),
}
}
impl WorkerAction for NewTask {
fn x(&self) -> i32 {
self.x
}
fn y(&self) -> i32 {
self.y
}
fn task_type(&self) -> &TaskType {
&self.task_type
}
fn target(&self) -> Option<HoboKey> {
self.target_hobo_id.map(HoboKey)
}
}
impl WorkerAction for Task {
fn x(&self) -> i32 | {
self.x
} | identifier_body |
|
worker_actions.rs | _type(&self) -> &TaskType;
fn target(&self) -> Option<HoboKey>;
}
pub struct ValidatedTaskList {
pub new_tasks: Vec<NewTask>,
pub update_tasks: Vec<Task>,
pub village_id: VillageKey,
}
pub(crate) fn validate_task_list(
db: &DB,
tl: &TaskList,
) -> Result<ValidatedTaskList, Box<dyn std::error::Error>> {
let worker_id = tl.worker_id;
// Load relevant data into memory
let mut worker = db.worker_priv(worker_id).ok_or("Worker does not exist")?;
let village_id = VillageKey(worker.home);
let mut town = TownView::load_village(db, village_id);
// check timing and effect of current task interruption
let mut current_task = db
.current_task(worker.key())
.expect("Must have a current task");
let mut timestamp =
interrupt_task(&mut current_task, &worker).ok_or("Cannot interrupt current task.")?;
worker.x = current_task.x;
worker.y = current_task.y;
// iterate tasks and match for task types | // Validate target hobo exists if there is one
if let Some(target_id) = task.target {
db.hobo(HoboKey(target_id)).ok_or("No such hobo id")?;
}
validate_ability(db, task.task_type, worker_id, timestamp)?;
let new_task = NewTask {
worker_id: worker_id.num(),
task_type: task.task_type,
x: task.x as i32,
y: task.y as i32,
start_time: Some(timestamp),
target_hobo_id: task.target,
};
simulate_begin_task(&new_task, &mut town, &mut worker)?;
let duration = simulate_finish_task(&new_task, &mut town, &mut worker)?;
tasks.push(new_task);
timestamp += duration;
}
Ok(ValidatedTaskList {
new_tasks: tasks,
update_tasks: vec![current_task],
village_id,
})
}
pub(crate) fn replace_worker_tasks(
db: &DB,
worker: &Addr<TownWorker>,
worker_id: WorkerKey,
tasks: &[NewTask],
village_id: VillageKey,
) {
db.flush_task_queue(worker_id);
let _inserted = db.insert_tasks(tasks);
let current_task =
execute_worker_tasks(db, worker_id, village_id).expect("Worker has no current task");
if let Some(next_task) = db.earliest_future_task(worker_id) {
let event = Event::WorkerTask {
task_id: current_task.key(),
};
worker
.send(TownWorkerEventMsg(
event,
Utc.from_utc_datetime(&next_task.start_time),
))
.wait()
.expect("Send msg to actor");
}
}
fn interrupt_task(current_task: &mut Task, worker: &Worker) -> Option<NaiveDateTime> {
match current_task.task_type {
TaskType::Idle
| TaskType::ChopTree
| TaskType::Defend
| TaskType::GatherSticks
| TaskType::CollectReward => {
let now = chrono::Utc::now().naive_utc();
Some(now)
}
TaskType::Walk => {
let speed = unit_speed_to_worker_tiles_per_second(worker.speed) as f64;
let time_so_far: Duration = Utc::now().naive_utc() - current_task.start_time;
let steps = (speed * time_so_far.num_microseconds().unwrap() as f64 / 1_000_000.0)
.ceil() as i32;
let total_time = steps as f64 / speed;
let moment = current_task.start_time
+ chrono::Duration::microseconds((total_time * 1_000_000.0) as i64);
let dx = current_task.x - worker.x;
let dy = current_task.y - worker.y;
let x = if dx == 0 {
worker.x
} else if dx < 0 {
worker.x - steps
} else {
worker.x + steps
};
let y = if dy == 0 {
worker.y
} else if dy < 0 {
worker.y - steps
} else {
worker.y + steps
};
// Walking must terminate earlier
current_task.x = x;
current_task.y = y;
Some(moment)
}
TaskType::WelcomeAbility => {
let cast_time = current_task.start_time + AbilityType::Welcome.busy_duration();
Some(cast_time)
}
}
}
/// For the given worker, executes tasks on the DB that are due
fn execute_worker_tasks(db: &DB, worker_id: WorkerKey, village: VillageKey) -> Option<Task> {
let mut tasks = db.past_worker_tasks(worker_id);
let current_task = tasks.pop();
let mut town = TownView::load_village(db, village);
for task in tasks {
if let Err(e) = finish_task(db, task.key(), Some(task), Some(&mut town)) {
println!("Executing task failed: {}", e)
}
}
current_task
}
pub(crate) fn finish_task(
db: &DB,
task_id: TaskKey,
task: Option<Task>,
town: Option<&mut TownView>,
) -> Result<Option<(Event, DateTime<Utc>)>, Box<dyn std::error::Error>> {
let task = task.or_else(|| db.task(task_id));
if let Some(task) = task {
let mut worker = db
.worker_priv(task.worker())
.ok_or("Task references non-existing worker")?;
if let Some(town) = town {
crate::worker_actions::simulate_finish_task(&task, town, &mut worker)?;
apply_task_to_db(db, &task, &mut worker)?;
} else {
let mut town = TownView::load_village(db, VillageKey(worker.home));
crate::worker_actions::simulate_finish_task(&task, &mut town, &mut worker)?;
apply_task_to_db(db, &task, &mut worker)?;
}
db.update_worker(&worker);
db.update_worker_flag_timestamp_now(worker.key(), WorkerFlagType::Work);
db.delete_task(&task);
Ok(Event::load_next_worker_task(db, task.worker()))
} else {
// Already executed.
Ok(None)
}
}
fn apply_task_to_db(db: &DB, task: &Task, worker: &mut Worker) -> Result<(), String> {
match task.task_type {
TaskType::WelcomeAbility => {
let a = AbilityType::Welcome;
let (attribute, strength) = a.apply();
let ne = NewEffect {
hobo_id: task.target().ok_or("Ability must have a target")?.num(),
attribute,
strength: Some(strength),
start_time: None, // default = now
};
db.insert_effect(&ne);
db.update_ability_used_timestamp(WorkerKey(worker.id), a);
*worker.mana.as_mut().unwrap() -= AbilityType::Welcome.mana_cost();
}
TaskType::CollectReward => {
if let Some(building) = db.find_building_by_coordinates(task.x, task.y, worker.home()) {
match building.building_type.reward_exp() {
Some(exp) => {
worker.add_exp(exp);
db.delete_building(&building);
}
None => {
return Err(format!(
"Tried to collect {} as reward",
building.building_type
));
}
}
} else {
return Err(format!("No reward to collect at {},{}", task.x, task.y));
}
}
_ => { /* NOP */ }
}
Ok(())
}
/// (Try to) apply changes to village state that happen when a worker stops doing a given task.
/// E.g. remove unit from building.
/// Returns the time it takes until the task is actually finished.
fn simulate_finish_task<T: WorkerAction>(
task: &T,
town: &mut TownView,
worker: &mut Worker,
) -> Result<Duration, String> {
match task.task_type() {
TaskType::Idle => Ok(Duration::milliseconds(0)),
TaskType::Walk => Ok(worker_walk(
town,
worker,
(task.x() as usize, task.y() as usize),
)?),
TaskType::GatherSticks | TaskType::ChopTree => {
town.state
.register_task_end(*task.task_type())
.map_err(|e| e.to_string())?;
worker_out_of_building(town, worker, (task.x() as usize, task.y() as usize))
}
TaskType::WelcomeAbility => {
let a = AbilityType::Welcome;
let duration = a.busy_duration();
Ok(duration)
}
TaskType::CollectReward => {
// Lookup object to be collected, then delete it in TownView
// Note: DB update is separate
let index = (task.x() as usize, task.y() as usize);
town.state.remove(&index);
Ok(Duration::milliseconds(0))
}
TaskType::Defend => Err("Task not implemented".to_owned()),
}
}
/// (Try to) apply changes to village state that happen when a worker starts a given | let mut tasks = vec![];
for task in tl.tasks.iter() { | random_line_split |
lockfree.rs | them, need more tests to
// see what the impact of that is
buckets: AtomicBucketList,
/// The default capacity of each bucket
///
/// Invariant: `bucket_capacity` must never be zero
bucket_capacity: AtomicUsize,
memory_usage: AtomicUsize,
max_memory_usage: AtomicUsize,
}
impl LockfreeArena {
/// Create a new Arena with the default bucket size of 4096 bytes
pub fn new(capacity: NonZeroUsize, max_memory_usage: usize) -> LassoResult<Self> {
Ok(Self {
// Allocate one bucket
buckets: AtomicBucketList::new(capacity)?,
bucket_capacity: AtomicUsize::new(capacity.get()),
// The current capacity is whatever size the bucket we just allocated is
memory_usage: AtomicUsize::new(capacity.get()),
max_memory_usage: AtomicUsize::new(max_memory_usage),
})
}
#[inline]
pub(crate) fn current_memory_usage(&self) -> usize {
self.memory_usage.load(Ordering::Relaxed)
}
#[inline]
pub(crate) fn set_max_memory_usage(&self, max_memory_usage: usize) {
self.max_memory_usage
.store(max_memory_usage, Ordering::Relaxed);
}
#[inline]
pub(crate) fn get_max_memory_usage(&self) -> usize {
self.max_memory_usage.load(Ordering::Relaxed)
}
fn set_bucket_capacity(&self, capacity: usize) {
debug_assert_ne!(capacity, 0);
self.bucket_capacity.store(capacity, Ordering::Relaxed);
}
/// Doesn't actually allocate anything, but increments `self.memory_usage` and returns `None` if
/// the attempted amount surpasses `max_memory_usage`
// TODO: Make this return a `Result`
fn | (&self, requested_mem: usize) -> LassoResult<()> {
if self.memory_usage.load(Ordering::Relaxed) + requested_mem
> self.max_memory_usage.load(Ordering::Relaxed)
{
Err(LassoError::new(LassoErrorKind::MemoryLimitReached))
} else {
self.memory_usage
.fetch_add(requested_mem, Ordering::Relaxed);
Ok(())
}
}
/// Store a slice in the Arena, returning `None` if memory is exhausted
///
/// # Safety
///
/// The reference passed back must be dropped before the arena that created it is
///
pub unsafe fn store_str(&self, string: &str) -> LassoResult<&'static str> {
// If the string is empty, simply return an empty string.
// This ensures that only strings with lengths greater
// than zero will be allocated within the arena
if string.is_empty() {
return Ok("");
}
let slice = string.as_bytes();
debug_assert_ne!(slice.len(), 0);
// Iterate over all of the buckets within the list while attempting to find one
// that has enough space to fit our string within it
//
// This is a tradeoff between allocation speed and memory usage. As-is we prioritize
// allocation speed in exchange for potentially missing possible reuse situations
// and then allocating more memory than is strictly necessary. In practice this shouldn't
// really matter, but it's worth that the opposite tradeoff can be made by adding bounded
// retries within this loop, the worst-case performance suffers in exchange for potentially
// better memory usage.
for bucket in self.buckets.iter() {
if let Ok(start) = bucket.try_inc_length(slice.len()) {
// Safety: We now have exclusive access to `bucket[start..start + slice.len()]`
let allocated = unsafe { bucket.slice_mut(start) };
// Copy the given slice into the allocation
unsafe { allocated.copy_from_nonoverlapping(slice.as_ptr(), slice.len()) };
// Return the successfully allocated string
let string = unsafe {
str::from_utf8_unchecked(slice::from_raw_parts(allocated, slice.len()))
};
return Ok(string);
}
// Otherwise the bucket doesn't have sufficient capacity for the string
// so we carry on searching through allocated buckets
}
// If we couldn't find a pre-existing bucket with enough room in it, allocate our own bucket
let next_capacity = self.bucket_capacity.load(Ordering::Relaxed) * 2;
debug_assert_ne!(next_capacity, 0);
// If the current string's length is greater than the doubled current capacity, allocate a bucket exactly the
// size of the large string and push it back in the buckets vector. This ensures that obscenely large strings will
// not permanently affect the resource consumption of the interner
if slice.len() > next_capacity {
// Check that we haven't exhausted our memory limit
self.allocate_memory(slice.len())?;
// Safety: `len` will never be zero since we explicitly handled zero-length strings
// at the beginning of the function
let non_zero_len = unsafe { NonZeroUsize::new_unchecked(slice.len()) };
debug_assert_ne!(slice.len(), 0);
let mut bucket = AtomicBucket::with_capacity(non_zero_len)?;
// Safety: The new bucket will have exactly enough room for the string and we have
// exclusive access to the bucket since we just created it
let allocated_string = unsafe { bucket.push_slice(slice) };
self.buckets.push_front(bucket.into_ref());
Ok(allocated_string)
} else {
let memory_usage = self.current_memory_usage();
let max_memory_usage = self.get_max_memory_usage();
// If trying to use the doubled capacity will surpass our memory limit, just allocate as much as we can
if memory_usage + next_capacity > max_memory_usage {
let remaining_memory = max_memory_usage.saturating_sub(memory_usage);
// Check that we haven't exhausted our memory limit
self.allocate_memory(remaining_memory)?;
// Set the capacity to twice of what it currently is to allow for fewer allocations as more strings are interned
let mut bucket = AtomicBucket::with_capacity(
NonZeroUsize::new(remaining_memory)
.ok_or_else(|| LassoError::new(LassoErrorKind::MemoryLimitReached))?,
)?;
// Safety: The new bucket will have exactly enough room for the string and we have
// exclusive access to the bucket since we just created it
let allocated_string = unsafe { bucket.push_slice(slice) };
// TODO: Push the bucket to the back or something so that we can get it somewhat out
// of the search path, reduce the `n` in the `O(n)` list traversal
self.buckets.push_front(bucket.into_ref());
Ok(allocated_string)
// Otherwise just allocate a normal doubled bucket
} else {
// Check that we haven't exhausted our memory limit
self.allocate_memory(next_capacity)?;
// Set the capacity to twice of what it currently is to allow for fewer allocations as more strings are interned
self.set_bucket_capacity(next_capacity);
// Safety: `next_capacity` will never be zero
let capacity = unsafe { NonZeroUsize::new_unchecked(next_capacity) };
debug_assert_ne!(next_capacity, 0);
let mut bucket = AtomicBucket::with_capacity(capacity)?;
// Safety: The new bucket will have enough room for the string
let allocated_string = unsafe { bucket.push_slice(slice) };
self.buckets.push_front(bucket.into_ref());
Ok(allocated_string)
}
}
}
}
impl Default for LockfreeArena {
fn default() -> Self {
Self::new(
Capacity::default().bytes,
MemoryLimits::default().max_memory_usage,
)
.expect("failed to create default arena")
}
}
impl Debug for LockfreeArena {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
struct TotalBuckets(usize);
impl Debug for TotalBuckets {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
if self.0 == 1 {
f.write_str("...1 bucket")
} else {
write!(f, "...{} buckets", self.0)
}
}
}
f.debug_struct("Arena")
.field("buckets", &TotalBuckets(self.buckets.len()))
.field(
"bucket_capacity",
&self.bucket_capacity.load(Ordering::Relaxed),
)
.field("memory_usage", &self.memory_usage.load(Ordering::Relaxed))
.field(
"max_memory_usage",
&self.max_memory_usage.load(Ordering::Relaxed),
)
.finish()
}
}
#[cfg(test)]
mod tests {
use super::*;
#[test]
fn string() {
let arena = LockfreeArena::default();
unsafe {
let idx = arena.store_str("test");
assert_eq!(idx, Ok("test"));
}
}
#[test]
fn empty_str() {
let arena = LockfreeArena::default();
unsafe {
let zst = arena.store_str("");
let zst1 = arena.store_str("");
let zst2 = arena.store_str("");
assert_eq!(zst, Ok(""));
assert_eq!(zst1, Ok(""));
assert_eq!(zst2, Ok(""));
}
}
#[test]
fn exponential_allocations() {
let arena = LockfreeArena::default();
let mut len = | allocate_memory | identifier_name |
lockfree.rs | them, need more tests to
// see what the impact of that is
buckets: AtomicBucketList,
/// The default capacity of each bucket
///
/// Invariant: `bucket_capacity` must never be zero
bucket_capacity: AtomicUsize,
memory_usage: AtomicUsize,
max_memory_usage: AtomicUsize,
}
impl LockfreeArena {
/// Create a new Arena with the default bucket size of 4096 bytes
pub fn new(capacity: NonZeroUsize, max_memory_usage: usize) -> LassoResult<Self> {
Ok(Self {
// Allocate one bucket
buckets: AtomicBucketList::new(capacity)?,
bucket_capacity: AtomicUsize::new(capacity.get()),
// The current capacity is whatever size the bucket we just allocated is
memory_usage: AtomicUsize::new(capacity.get()),
max_memory_usage: AtomicUsize::new(max_memory_usage),
})
}
#[inline]
pub(crate) fn current_memory_usage(&self) -> usize {
self.memory_usage.load(Ordering::Relaxed)
}
#[inline]
pub(crate) fn set_max_memory_usage(&self, max_memory_usage: usize) {
self.max_memory_usage
.store(max_memory_usage, Ordering::Relaxed);
}
#[inline]
pub(crate) fn get_max_memory_usage(&self) -> usize {
self.max_memory_usage.load(Ordering::Relaxed)
}
fn set_bucket_capacity(&self, capacity: usize) {
debug_assert_ne!(capacity, 0);
self.bucket_capacity.store(capacity, Ordering::Relaxed);
}
/// Doesn't actually allocate anything, but increments `self.memory_usage` and returns `None` if
/// the attempted amount surpasses `max_memory_usage`
// TODO: Make this return a `Result`
fn allocate_memory(&self, requested_mem: usize) -> LassoResult<()> {
if self.memory_usage.load(Ordering::Relaxed) + requested_mem
> self.max_memory_usage.load(Ordering::Relaxed)
{
Err(LassoError::new(LassoErrorKind::MemoryLimitReached))
} else {
self.memory_usage
.fetch_add(requested_mem, Ordering::Relaxed);
Ok(())
}
}
/// Store a slice in the Arena, returning `None` if memory is exhausted
///
/// # Safety
///
/// The reference passed back must be dropped before the arena that created it is
///
pub unsafe fn store_str(&self, string: &str) -> LassoResult<&'static str> {
// If the string is empty, simply return an empty string.
// This ensures that only strings with lengths greater
// than zero will be allocated within the arena
if string.is_empty() {
return Ok("");
}
let slice = string.as_bytes();
debug_assert_ne!(slice.len(), 0);
// Iterate over all of the buckets within the list while attempting to find one
// that has enough space to fit our string within it
//
// This is a tradeoff between allocation speed and memory usage. As-is we prioritize
// allocation speed in exchange for potentially missing possible reuse situations
// and then allocating more memory than is strictly necessary. In practice this shouldn't
// really matter, but it's worth that the opposite tradeoff can be made by adding bounded
// retries within this loop, the worst-case performance suffers in exchange for potentially
// better memory usage.
for bucket in self.buckets.iter() {
if let Ok(start) = bucket.try_inc_length(slice.len()) {
// Safety: We now have exclusive access to `bucket[start..start + slice.len()]`
let allocated = unsafe { bucket.slice_mut(start) };
// Copy the given slice into the allocation
unsafe { allocated.copy_from_nonoverlapping(slice.as_ptr(), slice.len()) };
// Return the successfully allocated string
let string = unsafe {
str::from_utf8_unchecked(slice::from_raw_parts(allocated, slice.len()))
};
return Ok(string);
}
// Otherwise the bucket doesn't have sufficient capacity for the string
// so we carry on searching through allocated buckets
}
// If we couldn't find a pre-existing bucket with enough room in it, allocate our own bucket
let next_capacity = self.bucket_capacity.load(Ordering::Relaxed) * 2;
debug_assert_ne!(next_capacity, 0);
// If the current string's length is greater than the doubled current capacity, allocate a bucket exactly the
// size of the large string and push it back in the buckets vector. This ensures that obscenely large strings will
// not permanently affect the resource consumption of the interner
if slice.len() > next_capacity {
// Check that we haven't exhausted our memory limit
self.allocate_memory(slice.len())?;
// Safety: `len` will never be zero since we explicitly handled zero-length strings
// at the beginning of the function
let non_zero_len = unsafe { NonZeroUsize::new_unchecked(slice.len()) };
debug_assert_ne!(slice.len(), 0);
let mut bucket = AtomicBucket::with_capacity(non_zero_len)?;
// Safety: The new bucket will have exactly enough room for the string and we have
// exclusive access to the bucket since we just created it
let allocated_string = unsafe { bucket.push_slice(slice) };
self.buckets.push_front(bucket.into_ref());
Ok(allocated_string)
} else | // TODO: Push the bucket to the back or something so that we can get it somewhat out
// of the search path, reduce the `n` in the `O(n)` list traversal
self.buckets.push_front(bucket.into_ref());
Ok(allocated_string)
// Otherwise just allocate a normal doubled bucket
} else {
// Check that we haven't exhausted our memory limit
self.allocate_memory(next_capacity)?;
// Set the capacity to twice of what it currently is to allow for fewer allocations as more strings are interned
self.set_bucket_capacity(next_capacity);
// Safety: `next_capacity` will never be zero
let capacity = unsafe { NonZeroUsize::new_unchecked(next_capacity) };
debug_assert_ne!(next_capacity, 0);
let mut bucket = AtomicBucket::with_capacity(capacity)?;
// Safety: The new bucket will have enough room for the string
let allocated_string = unsafe { bucket.push_slice(slice) };
self.buckets.push_front(bucket.into_ref());
Ok(allocated_string)
}
}
}
}
impl Default for LockfreeArena {
fn default() -> Self {
Self::new(
Capacity::default().bytes,
MemoryLimits::default().max_memory_usage,
)
.expect("failed to create default arena")
}
}
impl Debug for LockfreeArena {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
struct TotalBuckets(usize);
impl Debug for TotalBuckets {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
if self.0 == 1 {
f.write_str("...1 bucket")
} else {
write!(f, "...{} buckets", self.0)
}
}
}
f.debug_struct("Arena")
.field("buckets", &TotalBuckets(self.buckets.len()))
.field(
"bucket_capacity",
&self.bucket_capacity.load(Ordering::Relaxed),
)
.field("memory_usage", &self.memory_usage.load(Ordering::Relaxed))
.field(
"max_memory_usage",
&self.max_memory_usage.load(Ordering::Relaxed),
)
.finish()
}
}
#[cfg(test)]
mod tests {
use super::*;
#[test]
fn string() {
let arena = LockfreeArena::default();
unsafe {
let idx = arena.store_str("test");
assert_eq!(idx, Ok("test"));
}
}
#[test]
fn empty_str() {
let arena = LockfreeArena::default();
unsafe {
let zst = arena.store_str("");
let zst1 = arena.store_str("");
let zst2 = arena.store_str("");
assert_eq!(zst, Ok(""));
assert_eq!(zst1, Ok(""));
assert_eq!(zst2, Ok(""));
}
}
#[test]
fn exponential_allocations() {
let arena = LockfreeArena::default();
let mut len = | {
let memory_usage = self.current_memory_usage();
let max_memory_usage = self.get_max_memory_usage();
// If trying to use the doubled capacity will surpass our memory limit, just allocate as much as we can
if memory_usage + next_capacity > max_memory_usage {
let remaining_memory = max_memory_usage.saturating_sub(memory_usage);
// Check that we haven't exhausted our memory limit
self.allocate_memory(remaining_memory)?;
// Set the capacity to twice of what it currently is to allow for fewer allocations as more strings are interned
let mut bucket = AtomicBucket::with_capacity(
NonZeroUsize::new(remaining_memory)
.ok_or_else(|| LassoError::new(LassoErrorKind::MemoryLimitReached))?,
)?;
// Safety: The new bucket will have exactly enough room for the string and we have
// exclusive access to the bucket since we just created it
let allocated_string = unsafe { bucket.push_slice(slice) }; | conditional_block |
lockfree.rs | over them, need more tests to
// see what the impact of that is
buckets: AtomicBucketList,
/// The default capacity of each bucket
///
/// Invariant: `bucket_capacity` must never be zero | memory_usage: AtomicUsize,
max_memory_usage: AtomicUsize,
}
impl LockfreeArena {
/// Create a new Arena with the default bucket size of 4096 bytes
pub fn new(capacity: NonZeroUsize, max_memory_usage: usize) -> LassoResult<Self> {
Ok(Self {
// Allocate one bucket
buckets: AtomicBucketList::new(capacity)?,
bucket_capacity: AtomicUsize::new(capacity.get()),
// The current capacity is whatever size the bucket we just allocated is
memory_usage: AtomicUsize::new(capacity.get()),
max_memory_usage: AtomicUsize::new(max_memory_usage),
})
}
#[inline]
pub(crate) fn current_memory_usage(&self) -> usize {
self.memory_usage.load(Ordering::Relaxed)
}
#[inline]
pub(crate) fn set_max_memory_usage(&self, max_memory_usage: usize) {
self.max_memory_usage
.store(max_memory_usage, Ordering::Relaxed);
}
#[inline]
pub(crate) fn get_max_memory_usage(&self) -> usize {
self.max_memory_usage.load(Ordering::Relaxed)
}
fn set_bucket_capacity(&self, capacity: usize) {
debug_assert_ne!(capacity, 0);
self.bucket_capacity.store(capacity, Ordering::Relaxed);
}
/// Doesn't actually allocate anything, but increments `self.memory_usage` and returns `None` if
/// the attempted amount surpasses `max_memory_usage`
// TODO: Make this return a `Result`
fn allocate_memory(&self, requested_mem: usize) -> LassoResult<()> {
if self.memory_usage.load(Ordering::Relaxed) + requested_mem
> self.max_memory_usage.load(Ordering::Relaxed)
{
Err(LassoError::new(LassoErrorKind::MemoryLimitReached))
} else {
self.memory_usage
.fetch_add(requested_mem, Ordering::Relaxed);
Ok(())
}
}
/// Store a slice in the Arena, returning `None` if memory is exhausted
///
/// # Safety
///
/// The reference passed back must be dropped before the arena that created it is
///
pub unsafe fn store_str(&self, string: &str) -> LassoResult<&'static str> {
// If the string is empty, simply return an empty string.
// This ensures that only strings with lengths greater
// than zero will be allocated within the arena
if string.is_empty() {
return Ok("");
}
let slice = string.as_bytes();
debug_assert_ne!(slice.len(), 0);
// Iterate over all of the buckets within the list while attempting to find one
// that has enough space to fit our string within it
//
// This is a tradeoff between allocation speed and memory usage. As-is we prioritize
// allocation speed in exchange for potentially missing possible reuse situations
// and then allocating more memory than is strictly necessary. In practice this shouldn't
// really matter, but it's worth that the opposite tradeoff can be made by adding bounded
// retries within this loop, the worst-case performance suffers in exchange for potentially
// better memory usage.
for bucket in self.buckets.iter() {
if let Ok(start) = bucket.try_inc_length(slice.len()) {
// Safety: We now have exclusive access to `bucket[start..start + slice.len()]`
let allocated = unsafe { bucket.slice_mut(start) };
// Copy the given slice into the allocation
unsafe { allocated.copy_from_nonoverlapping(slice.as_ptr(), slice.len()) };
// Return the successfully allocated string
let string = unsafe {
str::from_utf8_unchecked(slice::from_raw_parts(allocated, slice.len()))
};
return Ok(string);
}
// Otherwise the bucket doesn't have sufficient capacity for the string
// so we carry on searching through allocated buckets
}
// If we couldn't find a pre-existing bucket with enough room in it, allocate our own bucket
let next_capacity = self.bucket_capacity.load(Ordering::Relaxed) * 2;
debug_assert_ne!(next_capacity, 0);
// If the current string's length is greater than the doubled current capacity, allocate a bucket exactly the
// size of the large string and push it back in the buckets vector. This ensures that obscenely large strings will
// not permanently affect the resource consumption of the interner
if slice.len() > next_capacity {
// Check that we haven't exhausted our memory limit
self.allocate_memory(slice.len())?;
// Safety: `len` will never be zero since we explicitly handled zero-length strings
// at the beginning of the function
let non_zero_len = unsafe { NonZeroUsize::new_unchecked(slice.len()) };
debug_assert_ne!(slice.len(), 0);
let mut bucket = AtomicBucket::with_capacity(non_zero_len)?;
// Safety: The new bucket will have exactly enough room for the string and we have
// exclusive access to the bucket since we just created it
let allocated_string = unsafe { bucket.push_slice(slice) };
self.buckets.push_front(bucket.into_ref());
Ok(allocated_string)
} else {
let memory_usage = self.current_memory_usage();
let max_memory_usage = self.get_max_memory_usage();
// If trying to use the doubled capacity will surpass our memory limit, just allocate as much as we can
if memory_usage + next_capacity > max_memory_usage {
let remaining_memory = max_memory_usage.saturating_sub(memory_usage);
// Check that we haven't exhausted our memory limit
self.allocate_memory(remaining_memory)?;
// Set the capacity to twice of what it currently is to allow for fewer allocations as more strings are interned
let mut bucket = AtomicBucket::with_capacity(
NonZeroUsize::new(remaining_memory)
.ok_or_else(|| LassoError::new(LassoErrorKind::MemoryLimitReached))?,
)?;
// Safety: The new bucket will have exactly enough room for the string and we have
// exclusive access to the bucket since we just created it
let allocated_string = unsafe { bucket.push_slice(slice) };
// TODO: Push the bucket to the back or something so that we can get it somewhat out
// of the search path, reduce the `n` in the `O(n)` list traversal
self.buckets.push_front(bucket.into_ref());
Ok(allocated_string)
// Otherwise just allocate a normal doubled bucket
} else {
// Check that we haven't exhausted our memory limit
self.allocate_memory(next_capacity)?;
// Set the capacity to twice of what it currently is to allow for fewer allocations as more strings are interned
self.set_bucket_capacity(next_capacity);
// Safety: `next_capacity` will never be zero
let capacity = unsafe { NonZeroUsize::new_unchecked(next_capacity) };
debug_assert_ne!(next_capacity, 0);
let mut bucket = AtomicBucket::with_capacity(capacity)?;
// Safety: The new bucket will have enough room for the string
let allocated_string = unsafe { bucket.push_slice(slice) };
self.buckets.push_front(bucket.into_ref());
Ok(allocated_string)
}
}
}
}
impl Default for LockfreeArena {
fn default() -> Self {
Self::new(
Capacity::default().bytes,
MemoryLimits::default().max_memory_usage,
)
.expect("failed to create default arena")
}
}
impl Debug for LockfreeArena {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
struct TotalBuckets(usize);
impl Debug for TotalBuckets {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
if self.0 == 1 {
f.write_str("...1 bucket")
} else {
write!(f, "...{} buckets", self.0)
}
}
}
f.debug_struct("Arena")
.field("buckets", &TotalBuckets(self.buckets.len()))
.field(
"bucket_capacity",
&self.bucket_capacity.load(Ordering::Relaxed),
)
.field("memory_usage", &self.memory_usage.load(Ordering::Relaxed))
.field(
"max_memory_usage",
&self.max_memory_usage.load(Ordering::Relaxed),
)
.finish()
}
}
#[cfg(test)]
mod tests {
use super::*;
#[test]
fn string() {
let arena = LockfreeArena::default();
unsafe {
let idx = arena.store_str("test");
assert_eq!(idx, Ok("test"));
}
}
#[test]
fn empty_str() {
let arena = LockfreeArena::default();
unsafe {
let zst = arena.store_str("");
let zst1 = arena.store_str("");
let zst2 = arena.store_str("");
assert_eq!(zst, Ok(""));
assert_eq!(zst1, Ok(""));
assert_eq!(zst2, Ok(""));
}
}
#[test]
fn exponential_allocations() {
let arena = LockfreeArena::default();
let mut len = | bucket_capacity: AtomicUsize, | random_line_split |
lockfree.rs | , 0);
self.bucket_capacity.store(capacity, Ordering::Relaxed);
}
/// Doesn't actually allocate anything, but increments `self.memory_usage` and returns `None` if
/// the attempted amount surpasses `max_memory_usage`
// TODO: Make this return a `Result`
fn allocate_memory(&self, requested_mem: usize) -> LassoResult<()> {
if self.memory_usage.load(Ordering::Relaxed) + requested_mem
> self.max_memory_usage.load(Ordering::Relaxed)
{
Err(LassoError::new(LassoErrorKind::MemoryLimitReached))
} else {
self.memory_usage
.fetch_add(requested_mem, Ordering::Relaxed);
Ok(())
}
}
/// Store a slice in the Arena, returning `None` if memory is exhausted
///
/// # Safety
///
/// The reference passed back must be dropped before the arena that created it is
///
pub unsafe fn store_str(&self, string: &str) -> LassoResult<&'static str> {
// If the string is empty, simply return an empty string.
// This ensures that only strings with lengths greater
// than zero will be allocated within the arena
if string.is_empty() {
return Ok("");
}
let slice = string.as_bytes();
debug_assert_ne!(slice.len(), 0);
// Iterate over all of the buckets within the list while attempting to find one
// that has enough space to fit our string within it
//
// This is a tradeoff between allocation speed and memory usage. As-is we prioritize
// allocation speed in exchange for potentially missing possible reuse situations
// and then allocating more memory than is strictly necessary. In practice this shouldn't
// really matter, but it's worth that the opposite tradeoff can be made by adding bounded
// retries within this loop, the worst-case performance suffers in exchange for potentially
// better memory usage.
for bucket in self.buckets.iter() {
if let Ok(start) = bucket.try_inc_length(slice.len()) {
// Safety: We now have exclusive access to `bucket[start..start + slice.len()]`
let allocated = unsafe { bucket.slice_mut(start) };
// Copy the given slice into the allocation
unsafe { allocated.copy_from_nonoverlapping(slice.as_ptr(), slice.len()) };
// Return the successfully allocated string
let string = unsafe {
str::from_utf8_unchecked(slice::from_raw_parts(allocated, slice.len()))
};
return Ok(string);
}
// Otherwise the bucket doesn't have sufficient capacity for the string
// so we carry on searching through allocated buckets
}
// If we couldn't find a pre-existing bucket with enough room in it, allocate our own bucket
let next_capacity = self.bucket_capacity.load(Ordering::Relaxed) * 2;
debug_assert_ne!(next_capacity, 0);
// If the current string's length is greater than the doubled current capacity, allocate a bucket exactly the
// size of the large string and push it back in the buckets vector. This ensures that obscenely large strings will
// not permanently affect the resource consumption of the interner
if slice.len() > next_capacity {
// Check that we haven't exhausted our memory limit
self.allocate_memory(slice.len())?;
// Safety: `len` will never be zero since we explicitly handled zero-length strings
// at the beginning of the function
let non_zero_len = unsafe { NonZeroUsize::new_unchecked(slice.len()) };
debug_assert_ne!(slice.len(), 0);
let mut bucket = AtomicBucket::with_capacity(non_zero_len)?;
// Safety: The new bucket will have exactly enough room for the string and we have
// exclusive access to the bucket since we just created it
let allocated_string = unsafe { bucket.push_slice(slice) };
self.buckets.push_front(bucket.into_ref());
Ok(allocated_string)
} else {
let memory_usage = self.current_memory_usage();
let max_memory_usage = self.get_max_memory_usage();
// If trying to use the doubled capacity will surpass our memory limit, just allocate as much as we can
if memory_usage + next_capacity > max_memory_usage {
let remaining_memory = max_memory_usage.saturating_sub(memory_usage);
// Check that we haven't exhausted our memory limit
self.allocate_memory(remaining_memory)?;
// Set the capacity to twice of what it currently is to allow for fewer allocations as more strings are interned
let mut bucket = AtomicBucket::with_capacity(
NonZeroUsize::new(remaining_memory)
.ok_or_else(|| LassoError::new(LassoErrorKind::MemoryLimitReached))?,
)?;
// Safety: The new bucket will have exactly enough room for the string and we have
// exclusive access to the bucket since we just created it
let allocated_string = unsafe { bucket.push_slice(slice) };
// TODO: Push the bucket to the back or something so that we can get it somewhat out
// of the search path, reduce the `n` in the `O(n)` list traversal
self.buckets.push_front(bucket.into_ref());
Ok(allocated_string)
// Otherwise just allocate a normal doubled bucket
} else {
// Check that we haven't exhausted our memory limit
self.allocate_memory(next_capacity)?;
// Set the capacity to twice of what it currently is to allow for fewer allocations as more strings are interned
self.set_bucket_capacity(next_capacity);
// Safety: `next_capacity` will never be zero
let capacity = unsafe { NonZeroUsize::new_unchecked(next_capacity) };
debug_assert_ne!(next_capacity, 0);
let mut bucket = AtomicBucket::with_capacity(capacity)?;
// Safety: The new bucket will have enough room for the string
let allocated_string = unsafe { bucket.push_slice(slice) };
self.buckets.push_front(bucket.into_ref());
Ok(allocated_string)
}
}
}
}
impl Default for LockfreeArena {
fn default() -> Self {
Self::new(
Capacity::default().bytes,
MemoryLimits::default().max_memory_usage,
)
.expect("failed to create default arena")
}
}
impl Debug for LockfreeArena {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
struct TotalBuckets(usize);
impl Debug for TotalBuckets {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
if self.0 == 1 {
f.write_str("...1 bucket")
} else {
write!(f, "...{} buckets", self.0)
}
}
}
f.debug_struct("Arena")
.field("buckets", &TotalBuckets(self.buckets.len()))
.field(
"bucket_capacity",
&self.bucket_capacity.load(Ordering::Relaxed),
)
.field("memory_usage", &self.memory_usage.load(Ordering::Relaxed))
.field(
"max_memory_usage",
&self.max_memory_usage.load(Ordering::Relaxed),
)
.finish()
}
}
#[cfg(test)]
mod tests {
use super::*;
#[test]
fn string() {
let arena = LockfreeArena::default();
unsafe {
let idx = arena.store_str("test");
assert_eq!(idx, Ok("test"));
}
}
#[test]
fn empty_str() {
let arena = LockfreeArena::default();
unsafe {
let zst = arena.store_str("");
let zst1 = arena.store_str("");
let zst2 = arena.store_str("");
assert_eq!(zst, Ok(""));
assert_eq!(zst1, Ok(""));
assert_eq!(zst2, Ok(""));
}
}
#[test]
fn exponential_allocations() {
let arena = LockfreeArena::default();
let mut len = 4096;
for _ in 0..10 {
let large_string = "a".repeat(len);
let arena_string = unsafe { arena.store_str(&large_string) };
assert_eq!(arena_string, Ok(large_string.as_str()));
len *= 2;
}
}
#[test]
fn memory_exhausted() {
let arena = LockfreeArena::new(NonZeroUsize::new(10).unwrap(), 10).unwrap();
unsafe {
assert!(arena.store_str("0123456789").is_ok());
// ZSTs take up zero bytes
arena.store_str("").unwrap();
let err = arena.store_str("a").unwrap_err();
assert!(err.kind().is_memory_limit());
let err = arena.store_str("dfgsagdfgsdf").unwrap_err();
assert!(err.kind().is_memory_limit());
}
}
#[test]
fn allocate_too_much() {
let arena = LockfreeArena::new(NonZeroUsize::new(1).unwrap(), 10).unwrap();
unsafe {
let err = arena.store_str("abcdefghijklmnopqrstuvwxyz").unwrap_err();
assert!(err.kind().is_memory_limit());
}
}
#[test]
fn allocate_more_than_double() | {
let arena = LockfreeArena::new(NonZeroUsize::new(1).unwrap(), 1000).unwrap();
unsafe {
assert!(arena.store_str("abcdefghijklmnopqrstuvwxyz").is_ok());
}
} | identifier_body |
|
config.rs | ,
}
}
}
#[derive(Debug, Clone, Serialize, Deserialize)]
#[serde(rename_all = "kebab-case", default, deny_unknown_fields)]
pub struct Carbon {
// TODO: will be used when multiple backends support is implemented
///// Enable sending to carbon protocol backend
//pub enabled: bool,
/// IP and port of the carbon-protocol backend to send aggregated data to
pub address: String,
/// client bind address
pub bind_address: Option<SocketAddr>,
/// How often to send metrics to this backend, ms
pub interval: u64,
/// How much to sleep when connection to backend fails, ms
pub connect_delay: u64,
/// Multiply delay to this value for each consequent connection failure
pub connect_delay_multiplier: f32,
/// Maximum retry delay, ms
pub connect_delay_max: u64,
/// How much times to retry when sending data to backend before giving up and dropping all metrics
/// note, that 0 means 1 try
pub send_retries: usize,
/// The whole metrica array can be split into smaller chunks for each chunk to be sent
/// in a separate connection. This is a workaround for go-carbon and carbon-c-relay doing
/// per-connection processing and working ineffectively when lots of metrics is sent in one
/// connection
pub chunks: usize,
}
impl Default for Carbon {
fn default() -> Self {
Self {
// enabled: true,
address: "127.0.0.1:2003".to_string(),
bind_address: None,
interval: 30000,
connect_delay: 250,
connect_delay_multiplier: 2f32,
connect_delay_max: 10000,
send_retries: 30,
chunks: 1,
}
}
}
#[derive(Debug, Clone, Serialize, Deserialize)]
#[serde(rename_all = "kebab-case", default, deny_unknown_fields)]
pub struct Network {
/// Address and UDP port to listen for statsd metrics on
pub listen: SocketAddr,
/// Address and port for replication server to listen on
pub peer_listen: SocketAddr,
/// Snapshot client bind address
pub peer_client_bind: Option<SocketAddr>,
/// Address and port for management server to listen on
pub mgmt_listen: SocketAddr,
/// UDP buffer size for single packet. Needs to be around MTU. Packet's bytes after that value
/// may be lost
pub bufsize: usize,
/// Enable multimessage(recvmmsg) mode
pub multimessage: bool,
/// Number of multimessage packets to receive at once if in multimessage mode
pub mm_packets: usize,
/// Number of multimessage packets to receive at once if in multimessage mode
pub mm_async: bool,
/// A timeout to return from multimessage mode syscall
pub mm_timeout: u64,
/// A timer to flush incoming buffer making sure metrics are not stuck there
pub buffer_flush_time: u64,
/// A length of incoming buffer to flush it making sure metrics are not stuck there
pub buffer_flush_length: usize,
/// Nmber of green threads for single-message mode
pub greens: usize,
/// Socket pool size for single-message mode
pub async_sockets: usize,
/// List of nodes to replicate metrics to
pub nodes: Vec<String>,
/// Interval to send snapshots to nodes, ms
pub snapshot_interval: usize,
}
impl Default for Network {
fn default() -> Self {
Self {
listen: "127.0.0.1:8125".parse().unwrap(),
peer_listen: "127.0.0.1:8136".parse().unwrap(),
peer_client_bind: None,
mgmt_listen: "127.0.0.1:8137".parse().unwrap(),
bufsize: 1500,
multimessage: false,
mm_packets: 100,
mm_async: false,
mm_timeout: 0,
buffer_flush_length: 0,
buffer_flush_time: 0,
greens: 4,
async_sockets: 4,
nodes: Vec::new(),
snapshot_interval: 1000,
}
}
}
#[derive(Debug, Clone, Serialize, Deserialize)]
#[serde(rename_all = "kebab-case", default, deny_unknown_fields)]
pub struct Consul {
/// Start in disabled leader finding mode
pub start_as: ConsensusState,
/// Consul agent address
pub agent: SocketAddr,
/// TTL of consul session, ms (consul cannot set it to less than 10s)
pub session_ttl: usize,
/// How often to renew consul session, ms
pub renew_time: usize,
/// Name of ke to be locked in consul
pub key_name: String,
}
impl Default for Consul {
fn default() -> Self {
Self { start_as: ConsensusState::Disabled, agent: "127.0.0.1:8500".parse().unwrap(), session_ttl: 11000, renew_time: 1000, key_name: "service/bioyino/lock".to_string() }
}
}
#[derive(Debug, Clone, Serialize, Deserialize)]
#[serde(rename_all = "kebab-case", default, deny_unknown_fields)]
pub struct Raft {
/// Delay raft after start (ms)
pub start_delay: u64,
/// Raft heartbeat timeout (ms)
pub heartbeat_timeout: u64,
/// Raft heartbeat timeout (ms)
pub election_timeout_min: u64,
/// Raft heartbeat timeout (ms)
pub election_timeout_max: u64,
/// Name of this node. By default is taken by resolving hostname in DNS.
pub this_node: Option<String>,
/// List of Raft nodes, may include this_node
pub nodes: HashMap<String, u64>,
/// Bind raft client to specific IP when connecting nodes
pub client_bind: Option<SocketAddr>,
}
impl Default for Raft {
fn default() -> Self {
Self { start_delay: 0, heartbeat_timeout: 250, election_timeout_min: 500, election_timeout_max: 750, this_node: None, nodes: HashMap::new(), client_bind: None }
}
}
impl Raft {
pub fn get_raft_options(&self) -> RaftOptions {
RaftOptions {
heartbeat_timeout: Duration::from_millis(self.heartbeat_timeout),
election_timeout: Range { start: Duration::from_millis(self.election_timeout_min), end: Duration::from_millis(self.election_timeout_max) },
}
}
}
#[derive(Debug)]
pub enum Command {
Daemon,
Query(MgmtCommand, String),
}
impl System {
pub fn load() -> (Self, Command) {
// This is a first copy of args - with the "config" option
let app = app_from_crate!()
.long_version(concat!(crate_version!(), " ", env!("VERGEN_COMMIT_DATE"), " ", env!("VERGEN_SHA_SHORT")))
.arg(Arg::with_name("config").help("configuration file path").long("config").short("c").required(true).takes_value(true).default_value("/etc/bioyino/bioyino.toml"))
.arg(Arg::with_name("verbosity").short("v").help("logging level").takes_value(true))
.subcommand(SubCommand::with_name("query").about("send a management command to running bioyino server").arg(Arg::with_name("host").short("h").default_value("127.0.0.1:8137")).subcommand(SubCommand::with_name("status").about("get server state")).subcommand(SubCommand::with_name("consensus").arg(Arg::with_name("action").index(1)).arg(Arg::with_name("leader_action").index(2).default_value("unchanged"))))
.get_matches();
let config = value_t!(app.value_of("config"), String).expect("config file must be string");
let mut file = File::open(&config).expect(&format!("opening config file at {}", &config));
let mut config_str = String::new();
file.read_to_string(&mut config_str).expect("reading config file");
let mut system: System = toml::de::from_str(&config_str).expect("parsing config");
if let Some(v) = app.value_of("verbosity") {
system.verbosity = v.into()
}
if let Some(query) = app.subcommand_matches("query") {
let server = value_t!(query.value_of("host"), String).expect("bad server");
if let Some(_) = query.subcommand_matches("status") {
(system, Command::Query(MgmtCommand::Status, server))
} else if let Some(args) = query.subcommand_matches("consensus") | {
let c_action = value_t!(args.value_of("action"), ConsensusAction).expect("bad consensus action");
let l_action = value_t!(args.value_of("leader_action"), LeaderAction).expect("bad leader action");
(system, Command::Query(MgmtCommand::ConsensusCommand(c_action, l_action), server))
} | conditional_block |
|
config.rs | all CPU cores
pub w_threads: usize,
/// queue size for single counting thread before packet is dropped
pub task_queue_size: usize,
/// Should we start as leader state enabled or not
pub start_as_leader: bool,
/// How often to gather own stats, in ms. Use 0 to disable (stats are still gathered, but not included in
/// metric dump)
pub stats_interval: u64,
/// Prefix to send own metrics with
pub stats_prefix: String,
/// Consensus kind to use
pub consensus: ConsensusKind,
}
impl Default for System {
fn default() -> Self {
Self {
verbosity: "warn".to_string(),
network: Network::default(),
raft: Raft::default(),
consul: Consul::default(),
metrics: Metrics::default(),
carbon: Carbon::default(),
n_threads: 4,
w_threads: 4,
stats_interval: 10000,
task_queue_size: 2048,
start_as_leader: false,
stats_prefix: "resources.monitoring.bioyino".to_string(),
consensus: ConsensusKind::None,
}
}
}
#[derive(Debug, Clone, Serialize, Deserialize)]
#[serde(rename_all = "kebab-case", default, deny_unknown_fields)]
pub struct Metrics {
// TODO: Maximum metric array size, 0 for unlimited
// max_metrics: usize,
/// Should we provide metrics with top update numbers
pub count_updates: bool,
/// Prefix for metric update statistics
pub update_counter_prefix: String,
/// Suffix for metric update statistics
pub update_counter_suffix: String,
/// Minimal update count to be reported
pub update_counter_threshold: u32,
/// Consistent parsing
pub consistent_parsing: bool,
/// Whether we should spam parsing errors in logs
pub log_parse_errors: bool,
/// Maximum length of data parser can keep in buffer befor considering it trash and throwing
/// away
pub max_unparsed_buffer: usize,
/// Choose the way of aggregation
pub aggregation_mode: AggregationMode,
/// Number of threads when aggregating in "multi" mode
pub aggregation_threads: Option<usize>,
}
impl Default for Metrics {
fn default() -> Self {
Self {
// max_metrics: 0,
count_updates: true,
update_counter_prefix: "resources.monitoring.bioyino.updates".to_string(),
update_counter_suffix: String::new(),
update_counter_threshold: 200,
consistent_parsing: true,
log_parse_errors: false,
max_unparsed_buffer: 10000,
aggregation_mode: AggregationMode::Single,
aggregation_threads: None,
}
}
}
#[derive(Debug, Clone, Serialize, Deserialize)]
#[serde(rename_all = "kebab-case", default, deny_unknown_fields)]
pub struct Carbon {
// TODO: will be used when multiple backends support is implemented
///// Enable sending to carbon protocol backend
//pub enabled: bool,
/// IP and port of the carbon-protocol backend to send aggregated data to
pub address: String,
/// client bind address
pub bind_address: Option<SocketAddr>,
/// How often to send metrics to this backend, ms
pub interval: u64,
/// How much to sleep when connection to backend fails, ms
pub connect_delay: u64,
/// Multiply delay to this value for each consequent connection failure
pub connect_delay_multiplier: f32,
/// Maximum retry delay, ms
pub connect_delay_max: u64,
/// How much times to retry when sending data to backend before giving up and dropping all metrics
/// note, that 0 means 1 try
pub send_retries: usize,
/// The whole metrica array can be split into smaller chunks for each chunk to be sent
/// in a separate connection. This is a workaround for go-carbon and carbon-c-relay doing
/// per-connection processing and working ineffectively when lots of metrics is sent in one
/// connection
pub chunks: usize,
}
impl Default for Carbon {
fn default() -> Self {
Self {
// enabled: true,
address: "127.0.0.1:2003".to_string(),
bind_address: None,
interval: 30000,
connect_delay: 250,
connect_delay_multiplier: 2f32,
connect_delay_max: 10000,
send_retries: 30,
chunks: 1,
}
}
}
#[derive(Debug, Clone, Serialize, Deserialize)]
#[serde(rename_all = "kebab-case", default, deny_unknown_fields)]
pub struct Network {
/// Address and UDP port to listen for statsd metrics on
pub listen: SocketAddr,
/// Address and port for replication server to listen on
pub peer_listen: SocketAddr,
/// Snapshot client bind address
pub peer_client_bind: Option<SocketAddr>,
/// Address and port for management server to listen on
pub mgmt_listen: SocketAddr,
/// UDP buffer size for single packet. Needs to be around MTU. Packet's bytes after that value
/// may be lost
pub bufsize: usize,
/// Enable multimessage(recvmmsg) mode
pub multimessage: bool,
/// Number of multimessage packets to receive at once if in multimessage mode
pub mm_packets: usize,
/// Number of multimessage packets to receive at once if in multimessage mode
pub mm_async: bool,
/// A timeout to return from multimessage mode syscall
pub mm_timeout: u64,
/// A timer to flush incoming buffer making sure metrics are not stuck there
pub buffer_flush_time: u64,
/// A length of incoming buffer to flush it making sure metrics are not stuck there
pub buffer_flush_length: usize,
/// Nmber of green threads for single-message mode
pub greens: usize,
/// Socket pool size for single-message mode
pub async_sockets: usize,
/// List of nodes to replicate metrics to
pub nodes: Vec<String>,
/// Interval to send snapshots to nodes, ms
pub snapshot_interval: usize,
}
impl Default for Network {
fn default() -> Self {
Self {
listen: "127.0.0.1:8125".parse().unwrap(),
peer_listen: "127.0.0.1:8136".parse().unwrap(),
peer_client_bind: None,
mgmt_listen: "127.0.0.1:8137".parse().unwrap(),
bufsize: 1500,
multimessage: false,
mm_packets: 100,
mm_async: false,
mm_timeout: 0,
buffer_flush_length: 0, | async_sockets: 4,
nodes: Vec::new(),
snapshot_interval: 1000,
}
}
}
#[derive(Debug, Clone, Serialize, Deserialize)]
#[serde(rename_all = "kebab-case", default, deny_unknown_fields)]
pub struct Consul {
/// Start in disabled leader finding mode
pub start_as: ConsensusState,
/// Consul agent address
pub agent: SocketAddr,
/// TTL of consul session, ms (consul cannot set it to less than 10s)
pub session_ttl: usize,
/// How often to renew consul session, ms
pub renew_time: usize,
/// Name of ke to be locked in consul
pub key_name: String,
}
impl Default for Consul {
fn default() -> Self {
Self { start_as: ConsensusState::Disabled, agent: "127.0.0.1:8500".parse().unwrap(), session_ttl: 11000, renew_time: 1000, key_name: "service/bioyino/lock".to_string() }
}
}
#[derive(Debug, Clone, Serialize, Deserialize)]
#[serde(rename_all = "kebab-case", default, deny_unknown_fields)]
pub struct Raft {
/// Delay raft after start (ms)
pub start_delay: u64,
/// Raft heartbeat timeout (ms)
pub heartbeat_timeout: u64,
/// Raft heartbeat timeout (ms)
pub election_timeout_min: u64,
/// Raft heartbeat timeout (ms)
pub election_timeout_max: u64,
/// Name of this node. By default is taken by resolving hostname in DNS.
pub this_node: Option<String>,
/// List of Raft nodes, may include this_node
pub nodes: HashMap<String, u64>,
/// Bind raft client to specific IP when connecting nodes
pub client_bind: Option<SocketAddr>,
}
impl Default for Raft {
fn default() -> Self {
Self { start_delay: 0, heartbeat_timeout: 250, election_timeout_min: 500, election_timeout_max: 750, this_node: None, nodes: HashMap::new(), client_bind: None }
}
}
impl Raft {
pub fn get_raft_options(&self) -> RaftOptions {
RaftOptions {
heartbeat_timeout: Duration::from_m | buffer_flush_time: 0,
greens: 4, | random_line_split |
config.rs | /// How often to gather own stats, in ms. Use 0 to disable (stats are still gathered, but not included in
/// metric dump)
pub stats_interval: u64,
/// Prefix to send own metrics with
pub stats_prefix: String,
/// Consensus kind to use
pub consensus: ConsensusKind,
}
impl Default for System {
fn default() -> Self {
Self {
verbosity: "warn".to_string(),
network: Network::default(),
raft: Raft::default(),
consul: Consul::default(),
metrics: Metrics::default(),
carbon: Carbon::default(),
n_threads: 4,
w_threads: 4,
stats_interval: 10000,
task_queue_size: 2048,
start_as_leader: false,
stats_prefix: "resources.monitoring.bioyino".to_string(),
consensus: ConsensusKind::None,
}
}
}
#[derive(Debug, Clone, Serialize, Deserialize)]
#[serde(rename_all = "kebab-case", default, deny_unknown_fields)]
pub struct Metrics {
// TODO: Maximum metric array size, 0 for unlimited
// max_metrics: usize,
/// Should we provide metrics with top update numbers
pub count_updates: bool,
/// Prefix for metric update statistics
pub update_counter_prefix: String,
/// Suffix for metric update statistics
pub update_counter_suffix: String,
/// Minimal update count to be reported
pub update_counter_threshold: u32,
/// Consistent parsing
pub consistent_parsing: bool,
/// Whether we should spam parsing errors in logs
pub log_parse_errors: bool,
/// Maximum length of data parser can keep in buffer befor considering it trash and throwing
/// away
pub max_unparsed_buffer: usize,
/// Choose the way of aggregation
pub aggregation_mode: AggregationMode,
/// Number of threads when aggregating in "multi" mode
pub aggregation_threads: Option<usize>,
}
impl Default for Metrics {
fn default() -> Self {
Self {
// max_metrics: 0,
count_updates: true,
update_counter_prefix: "resources.monitoring.bioyino.updates".to_string(),
update_counter_suffix: String::new(),
update_counter_threshold: 200,
consistent_parsing: true,
log_parse_errors: false,
max_unparsed_buffer: 10000,
aggregation_mode: AggregationMode::Single,
aggregation_threads: None,
}
}
}
#[derive(Debug, Clone, Serialize, Deserialize)]
#[serde(rename_all = "kebab-case", default, deny_unknown_fields)]
pub struct Carbon {
// TODO: will be used when multiple backends support is implemented
///// Enable sending to carbon protocol backend
//pub enabled: bool,
/// IP and port of the carbon-protocol backend to send aggregated data to
pub address: String,
/// client bind address
pub bind_address: Option<SocketAddr>,
/// How often to send metrics to this backend, ms
pub interval: u64,
/// How much to sleep when connection to backend fails, ms
pub connect_delay: u64,
/// Multiply delay to this value for each consequent connection failure
pub connect_delay_multiplier: f32,
/// Maximum retry delay, ms
pub connect_delay_max: u64,
/// How much times to retry when sending data to backend before giving up and dropping all metrics
/// note, that 0 means 1 try
pub send_retries: usize,
/// The whole metrica array can be split into smaller chunks for each chunk to be sent
/// in a separate connection. This is a workaround for go-carbon and carbon-c-relay doing
/// per-connection processing and working ineffectively when lots of metrics is sent in one
/// connection
pub chunks: usize,
}
impl Default for Carbon {
fn default() -> Self {
Self {
// enabled: true,
address: "127.0.0.1:2003".to_string(),
bind_address: None,
interval: 30000,
connect_delay: 250,
connect_delay_multiplier: 2f32,
connect_delay_max: 10000,
send_retries: 30,
chunks: 1,
}
}
}
#[derive(Debug, Clone, Serialize, Deserialize)]
#[serde(rename_all = "kebab-case", default, deny_unknown_fields)]
pub struct Network {
/// Address and UDP port to listen for statsd metrics on
pub listen: SocketAddr,
/// Address and port for replication server to listen on
pub peer_listen: SocketAddr,
/// Snapshot client bind address
pub peer_client_bind: Option<SocketAddr>,
/// Address and port for management server to listen on
pub mgmt_listen: SocketAddr,
/// UDP buffer size for single packet. Needs to be around MTU. Packet's bytes after that value
/// may be lost
pub bufsize: usize,
/// Enable multimessage(recvmmsg) mode
pub multimessage: bool,
/// Number of multimessage packets to receive at once if in multimessage mode
pub mm_packets: usize,
/// Number of multimessage packets to receive at once if in multimessage mode
pub mm_async: bool,
/// A timeout to return from multimessage mode syscall
pub mm_timeout: u64,
/// A timer to flush incoming buffer making sure metrics are not stuck there
pub buffer_flush_time: u64,
/// A length of incoming buffer to flush it making sure metrics are not stuck there
pub buffer_flush_length: usize,
/// Nmber of green threads for single-message mode
pub greens: usize,
/// Socket pool size for single-message mode
pub async_sockets: usize,
/// List of nodes to replicate metrics to
pub nodes: Vec<String>,
/// Interval to send snapshots to nodes, ms
pub snapshot_interval: usize,
}
impl Default for Network {
fn default() -> Self {
Self {
listen: "127.0.0.1:8125".parse().unwrap(),
peer_listen: "127.0.0.1:8136".parse().unwrap(),
peer_client_bind: None,
mgmt_listen: "127.0.0.1:8137".parse().unwrap(),
bufsize: 1500,
multimessage: false,
mm_packets: 100,
mm_async: false,
mm_timeout: 0,
buffer_flush_length: 0,
buffer_flush_time: 0,
greens: 4,
async_sockets: 4,
nodes: Vec::new(),
snapshot_interval: 1000,
}
}
}
#[derive(Debug, Clone, Serialize, Deserialize)]
#[serde(rename_all = "kebab-case", default, deny_unknown_fields)]
pub struct Consul {
/// Start in disabled leader finding mode
pub start_as: ConsensusState,
/// Consul agent address
pub agent: SocketAddr,
/// TTL of consul session, ms (consul cannot set it to less than 10s)
pub session_ttl: usize,
/// How often to renew consul session, ms
pub renew_time: usize,
/// Name of ke to be locked in consul
pub key_name: String,
}
impl Default for Consul {
fn default() -> Self {
Self { start_as: ConsensusState::Disabled, agent: "127.0.0.1:8500".parse().unwrap(), session_ttl: 11000, renew_time: 1000, key_name: "service/bioyino/lock".to_string() }
}
}
#[derive(Debug, Clone, Serialize, Deserialize)]
#[serde(rename_all = "kebab-case", default, deny_unknown_fields)]
pub struct Raft {
/// Delay raft after start (ms)
pub start_delay: u64,
/// Raft heartbeat timeout (ms)
pub heartbeat_timeout: u64,
/// Raft heartbeat timeout (ms)
pub election_timeout_min: u64,
/// Raft heartbeat timeout (ms)
pub election_timeout_max: u64,
/// Name of this node. By default is taken by resolving hostname in DNS.
pub this_node: Option<String>,
/// List of Raft nodes, may include this_node
pub nodes: HashMap<String, u64>,
/// Bind raft client to specific IP when connecting nodes
pub client_bind: Option<SocketAddr>,
}
impl Default for Raft {
fn default() -> Self {
Self { start_delay: 0, heartbeat_timeout: 250, election_timeout_min: 500, election_timeout_max: 750, this_node: None, nodes: HashMap::new(), client_bind: None }
}
}
impl Raft {
pub fn get_raft_options(&self) -> RaftOptions {
RaftOptions {
heartbeat_timeout: Duration::from_millis(self.heartbeat_timeout),
election_timeout: Range { start: Duration::from_millis(self.election_timeout_min), end: Duration::from_millis(self.election_timeout_max) },
}
}
}
#[derive(Debug)]
pub enum | Command | identifier_name |
|
building.rs | ,
pub address: String,
pub name: Option<NamePerLanguage>,
pub orig_id: osm::OsmID,
/// Where a text label should be centered to have the best chances of being contained within
/// the polygon.
pub label_center: Pt2D,
pub amenities: Vec<Amenity>,
pub bldg_type: BuildingType,
pub parking: OffstreetParking,
/// Depending on options while importing, these might be empty, to save file space.
pub osm_tags: Tags,
/// The building's connection for any agent can change based on map edits. Just store the one
/// for pedestrians and lazily calculate the others.
pub sidewalk_pos: Position,
/// Goes from building to sidewalk
pub driveway_geom: PolyLine,
}
/// A business located inside a building.
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct Amenity {
pub names: NamePerLanguage,
/// This is the specific amenity listed in OSM, not the more general `AmenityType` category.
pub amenity_type: String,
/// Depending on options while importing, these might be empty, to save file space.
pub osm_tags: Tags,
}
/// Represent no parking as Private(0, false).
#[derive(Serialize, Deserialize, Debug, PartialEq, Clone)]
pub enum OffstreetParking {
/// (Name, spots)
PublicGarage(String, usize),
/// (Spots, explicitly tagged as a garage)
Private(usize, bool),
}
#[derive(Serialize, Deserialize, Clone, Debug)]
pub enum BuildingType {
Residential {
num_residents: usize,
num_housing_units: usize,
},
/// An estimated number of residents, workers
ResidentialCommercial(usize, usize),
/// An estimated number of workers
Commercial(usize),
Empty,
}
impl BuildingType {
pub fn has_residents(&self) -> bool {
match self {
BuildingType::Residential { .. } | BuildingType::ResidentialCommercial(_, _) => true,
BuildingType::Commercial(_) | BuildingType::Empty => false,
}
}
}
/// None corresponds to the native name
#[derive(Serialize, Deserialize, Debug, PartialEq, Eq, PartialOrd, Ord, Clone)]
pub struct NamePerLanguage(
#[serde(
serialize_with = "serialize_btreemap",
deserialize_with = "deserialize_btreemap"
)]
pub(crate) BTreeMap<Option<String>, String>,
);
impl NamePerLanguage {
pub fn get(&self, lang: Option<&String>) -> &String {
// TODO Can we avoid this clone?
let lang = lang.cloned();
if let Some(name) = self.0.get(&lang) {
return name;
}
&self.0[&None]
}
pub fn new(tags: &Tags) -> Option<NamePerLanguage> {
let native_name = tags.get(osm::NAME)?;
let mut map = BTreeMap::new();
map.insert(None, native_name.to_string());
for (k, v) in tags.inner() {
if let Some(lang) = k.strip_prefix("name:") {
map.insert(Some(lang.to_string()), v.to_string());
}
}
Some(NamePerLanguage(map))
}
pub fn unnamed() -> NamePerLanguage {
let mut map = BTreeMap::new();
map.insert(None, "unnamed".to_string());
NamePerLanguage(map)
}
}
impl Building {
pub fn sidewalk(&self) -> LaneID {
self.sidewalk_pos.lane()
}
pub fn house_number(&self) -> Option<String> {
let num = self.address.split(' ').next().unwrap();
if num != "???" {
Some(num.to_string())
} else {
None
}
}
/// The polyline goes from the building to the driving position
// TODO Make this handle parking_blackhole
pub fn driving_connection(&self, map: &Map) -> Option<(Position, PolyLine)> |
/// Returns (biking position, sidewalk position). Could fail if the biking graph is
/// disconnected.
pub fn biking_connection(&self, map: &Map) -> Option<(Position, Position)> {
// Easy case: the building is directly next to a usable lane
if let Some(pair) = sidewalk_to_bike(self.sidewalk_pos, map) {
return Some(pair);
}
// Floodfill the sidewalk graph until we find a sidewalk<->bike connection.
let mut queue: VecDeque<LaneID> = VecDeque::new();
let mut visited: HashSet<LaneID> = HashSet::new();
queue.push_back(self.sidewalk());
loop {
if queue.is_empty() {
return None;
}
let l = queue.pop_front().unwrap();
if visited.contains(&l) {
continue;
}
visited.insert(l);
// TODO Could search by sidewalk endpoint
if let Some(pair) = sidewalk_to_bike(Position::new(l, map.get_l(l).length() / 2.0), map)
{
return Some(pair);
}
for t in map.get_turns_from_lane(l) {
if !visited.contains(&t.id.dst) {
queue.push_back(t.id.dst);
}
}
}
}
pub fn num_parking_spots(&self) -> usize {
match self.parking {
OffstreetParking::PublicGarage(_, n) => n,
OffstreetParking::Private(n, _) => n,
}
}
/// Does this building contain any amenity matching the category?
pub fn has_amenity(&self, category: AmenityType) -> bool {
for amenity in &self.amenities {
if AmenityType::categorize(&amenity.amenity_type) == Some(category) {
return true;
}
}
false
}
}
fn sidewalk_to_bike(sidewalk_pos: Position, map: &Map) -> Option<(Position, Position)> {
let lane = map
.get_parent(sidewalk_pos.lane())
.find_closest_lane(sidewalk_pos.lane(), |l| {
!l.biking_blackhole && PathConstraints::Bike.can_use(l, map)
})?;
// No buffer needed
Some((sidewalk_pos.equiv_pos(lane, map), sidewalk_pos))
}
/// Businesses are categorized into one of these types.
#[derive(Clone, Copy, PartialEq, Eq, PartialOrd, Ord, EnumString, Display, EnumIter)]
pub enum AmenityType {
Bank,
Bar,
Beauty,
Bike,
Cafe,
CarRepair,
CarShare,
Childcare,
ConvenienceStore,
Culture,
Exercise,
FastFood,
Food,
GreenSpace,
Hotel,
Laundry,
Library,
Medical,
Pet,
Playground,
Pool,
PostOffice,
Religious,
School,
Shopping,
Supermarket,
Tourism,
University,
}
impl AmenityType {
fn types(self) -> Vec<&'static str> {
match self {
AmenityType::Bank => vec!["bank"],
AmenityType::Bar => vec!["bar", "pub", "nightclub", "biergarten"],
AmenityType::Beauty => vec!["hairdresser", "beauty", "chemist", "cosmetics"],
AmenityType::Bike => vec!["bicycle"],
AmenityType::Cafe => vec!["cafe", "pastry", "coffee", "tea", "bakery"],
AmenityType::CarRepair => vec!["car_repair"],
AmenityType::CarShare => vec!["car_sharing"],
AmenityType::Childcare => vec!["childcare", "kindergarten"],
AmenityType::ConvenienceStore => vec!["convenience"],
AmenityType::Culture => vec!["arts_centre", "art", "cinema", "theatre"],
AmenityType::Exercise => vec!["fitness_centre", "sports_centre", "track", "pitch"],
AmenityType::FastFood => vec!["fast_food", "food_court"],
AmenityType::Food => vec![
"restaurant",
"farm",
"ice_cream",
"seafood",
"cheese",
"chocolate",
"deli",
"butcher",
"confectionery",
"beverages",
"alcohol",
],
AmenityType::GreenSpace => vec!["park", "garden", "nature_reserve"],
AmenityType::Hotel => vec!["hotel", "hostel", "guest_house", "motel"],
AmenityType::Laundry => vec!["dry_cleaning", "laundry", "tailor"],
AmenityType::Library => vec!["library"],
AmenityType::Medical => vec![
"clinic", | {
let lane = map
.get_parent(self.sidewalk())
.find_closest_lane(self.sidewalk(), |l| PathConstraints::Car.can_use(l, map))?;
// TODO Do we need to insist on this buffer, now that we can make cars gradually appear?
let pos = self
.sidewalk_pos
.equiv_pos(lane, map)
.buffer_dist(Distance::meters(7.0), map)?;
Some((pos, self.driveway_geom.clone().optionally_push(pos.pt(map))))
} | identifier_body |
building.rs | ,
pub address: String,
pub name: Option<NamePerLanguage>,
pub orig_id: osm::OsmID,
/// Where a text label should be centered to have the best chances of being contained within
/// the polygon.
pub label_center: Pt2D,
pub amenities: Vec<Amenity>,
pub bldg_type: BuildingType,
pub parking: OffstreetParking,
/// Depending on options while importing, these might be empty, to save file space.
pub osm_tags: Tags,
/// The building's connection for any agent can change based on map edits. Just store the one
/// for pedestrians and lazily calculate the others.
pub sidewalk_pos: Position,
/// Goes from building to sidewalk
pub driveway_geom: PolyLine,
}
/// A business located inside a building.
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct Amenity {
pub names: NamePerLanguage,
/// This is the specific amenity listed in OSM, not the more general `AmenityType` category.
pub amenity_type: String,
/// Depending on options while importing, these might be empty, to save file space.
pub osm_tags: Tags,
}
/// Represent no parking as Private(0, false).
#[derive(Serialize, Deserialize, Debug, PartialEq, Clone)]
pub enum OffstreetParking {
/// (Name, spots)
PublicGarage(String, usize),
/// (Spots, explicitly tagged as a garage)
Private(usize, bool),
}
#[derive(Serialize, Deserialize, Clone, Debug)]
pub enum BuildingType {
Residential {
num_residents: usize,
num_housing_units: usize,
},
/// An estimated number of residents, workers
ResidentialCommercial(usize, usize),
/// An estimated number of workers
Commercial(usize),
Empty,
}
impl BuildingType {
pub fn has_residents(&self) -> bool {
match self {
BuildingType::Residential { .. } | BuildingType::ResidentialCommercial(_, _) => true,
BuildingType::Commercial(_) | BuildingType::Empty => false,
}
}
}
/// None corresponds to the native name
#[derive(Serialize, Deserialize, Debug, PartialEq, Eq, PartialOrd, Ord, Clone)]
pub struct NamePerLanguage(
#[serde(
serialize_with = "serialize_btreemap",
deserialize_with = "deserialize_btreemap"
)]
pub(crate) BTreeMap<Option<String>, String>,
);
impl NamePerLanguage {
pub fn get(&self, lang: Option<&String>) -> &String {
// TODO Can we avoid this clone?
let lang = lang.cloned();
if let Some(name) = self.0.get(&lang) {
return name;
}
&self.0[&None]
}
pub fn new(tags: &Tags) -> Option<NamePerLanguage> {
let native_name = tags.get(osm::NAME)?;
let mut map = BTreeMap::new();
map.insert(None, native_name.to_string());
for (k, v) in tags.inner() {
if let Some(lang) = k.strip_prefix("name:") {
map.insert(Some(lang.to_string()), v.to_string());
}
}
Some(NamePerLanguage(map))
}
pub fn unnamed() -> NamePerLanguage {
let mut map = BTreeMap::new();
map.insert(None, "unnamed".to_string());
NamePerLanguage(map)
}
}
impl Building {
pub fn sidewalk(&self) -> LaneID {
self.sidewalk_pos.lane()
}
pub fn house_number(&self) -> Option<String> {
let num = self.address.split(' ').next().unwrap();
if num != "???" {
Some(num.to_string())
} else {
None
}
}
/// The polyline goes from the building to the driving position
// TODO Make this handle parking_blackhole
pub fn driving_connection(&self, map: &Map) -> Option<(Position, PolyLine)> {
let lane = map
.get_parent(self.sidewalk())
.find_closest_lane(self.sidewalk(), |l| PathConstraints::Car.can_use(l, map))?;
// TODO Do we need to insist on this buffer, now that we can make cars gradually appear?
let pos = self
.sidewalk_pos
.equiv_pos(lane, map)
.buffer_dist(Distance::meters(7.0), map)?;
Some((pos, self.driveway_geom.clone().optionally_push(pos.pt(map))))
}
/// Returns (biking position, sidewalk position). Could fail if the biking graph is
/// disconnected.
pub fn biking_connection(&self, map: &Map) -> Option<(Position, Position)> {
// Easy case: the building is directly next to a usable lane
if let Some(pair) = sidewalk_to_bike(self.sidewalk_pos, map) {
return Some(pair);
}
// Floodfill the sidewalk graph until we find a sidewalk<->bike connection.
let mut queue: VecDeque<LaneID> = VecDeque::new();
let mut visited: HashSet<LaneID> = HashSet::new();
queue.push_back(self.sidewalk());
loop {
if queue.is_empty() {
return None;
}
let l = queue.pop_front().unwrap();
if visited.contains(&l) {
continue;
}
visited.insert(l);
// TODO Could search by sidewalk endpoint
if let Some(pair) = sidewalk_to_bike(Position::new(l, map.get_l(l).length() / 2.0), map)
{
return Some(pair);
}
for t in map.get_turns_from_lane(l) {
if !visited.contains(&t.id.dst) {
queue.push_back(t.id.dst);
}
}
}
}
pub fn num_parking_spots(&self) -> usize {
match self.parking {
OffstreetParking::PublicGarage(_, n) => n,
OffstreetParking::Private(n, _) => n,
}
}
/// Does this building contain any amenity matching the category?
pub fn has_amenity(&self, category: AmenityType) -> bool {
for amenity in &self.amenities {
if AmenityType::categorize(&amenity.amenity_type) == Some(category) {
return true;
}
}
false
}
}
fn sidewalk_to_bike(sidewalk_pos: Position, map: &Map) -> Option<(Position, Position)> {
let lane = map
.get_parent(sidewalk_pos.lane())
.find_closest_lane(sidewalk_pos.lane(), |l| {
!l.biking_blackhole && PathConstraints::Bike.can_use(l, map)
})?;
// No buffer needed
Some((sidewalk_pos.equiv_pos(lane, map), sidewalk_pos))
}
/// Businesses are categorized into one of these types.
#[derive(Clone, Copy, PartialEq, Eq, PartialOrd, Ord, EnumString, Display, EnumIter)]
pub enum | {
Bank,
Bar,
Beauty,
Bike,
Cafe,
CarRepair,
CarShare,
Childcare,
ConvenienceStore,
Culture,
Exercise,
FastFood,
Food,
GreenSpace,
Hotel,
Laundry,
Library,
Medical,
Pet,
Playground,
Pool,
PostOffice,
Religious,
School,
Shopping,
Supermarket,
Tourism,
University,
}
impl AmenityType {
fn types(self) -> Vec<&'static str> {
match self {
AmenityType::Bank => vec!["bank"],
AmenityType::Bar => vec!["bar", "pub", "nightclub", "biergarten"],
AmenityType::Beauty => vec!["hairdresser", "beauty", "chemist", "cosmetics"],
AmenityType::Bike => vec!["bicycle"],
AmenityType::Cafe => vec!["cafe", "pastry", "coffee", "tea", "bakery"],
AmenityType::CarRepair => vec!["car_repair"],
AmenityType::CarShare => vec!["car_sharing"],
AmenityType::Childcare => vec!["childcare", "kindergarten"],
AmenityType::ConvenienceStore => vec!["convenience"],
AmenityType::Culture => vec!["arts_centre", "art", "cinema", "theatre"],
AmenityType::Exercise => vec!["fitness_centre", "sports_centre", "track", "pitch"],
AmenityType::FastFood => vec!["fast_food", "food_court"],
AmenityType::Food => vec![
"restaurant",
"farm",
"ice_cream",
"seafood",
"cheese",
"chocolate",
"deli",
"butcher",
"confectionery",
"beverages",
"alcohol",
],
AmenityType::GreenSpace => vec!["park", "garden", "nature_reserve"],
AmenityType::Hotel => vec!["hotel", "hostel", "guest_house", "motel"],
AmenityType::Laundry => vec!["dry_cleaning", "laundry", "tailor"],
AmenityType::Library => vec!["library"],
AmenityType::Medical => vec![
"clinic", | AmenityType | identifier_name |
building.rs | 4,
pub address: String,
pub name: Option<NamePerLanguage>,
pub orig_id: osm::OsmID,
/// Where a text label should be centered to have the best chances of being contained within
/// the polygon.
pub label_center: Pt2D,
pub amenities: Vec<Amenity>,
pub bldg_type: BuildingType,
pub parking: OffstreetParking,
/// Depending on options while importing, these might be empty, to save file space.
pub osm_tags: Tags,
/// The building's connection for any agent can change based on map edits. Just store the one
/// for pedestrians and lazily calculate the others.
pub sidewalk_pos: Position,
/// Goes from building to sidewalk
pub driveway_geom: PolyLine,
}
/// A business located inside a building.
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct Amenity {
pub names: NamePerLanguage,
/// This is the specific amenity listed in OSM, not the more general `AmenityType` category.
pub amenity_type: String,
/// Depending on options while importing, these might be empty, to save file space.
pub osm_tags: Tags,
}
/// Represent no parking as Private(0, false).
#[derive(Serialize, Deserialize, Debug, PartialEq, Clone)]
pub enum OffstreetParking {
/// (Name, spots)
PublicGarage(String, usize),
/// (Spots, explicitly tagged as a garage)
Private(usize, bool),
}
#[derive(Serialize, Deserialize, Clone, Debug)]
pub enum BuildingType {
Residential {
num_residents: usize,
num_housing_units: usize,
},
/// An estimated number of residents, workers
ResidentialCommercial(usize, usize),
/// An estimated number of workers
Commercial(usize),
Empty,
}
impl BuildingType {
pub fn has_residents(&self) -> bool {
match self {
BuildingType::Residential { .. } | BuildingType::ResidentialCommercial(_, _) => true,
BuildingType::Commercial(_) | BuildingType::Empty => false,
}
}
}
/// None corresponds to the native name
#[derive(Serialize, Deserialize, Debug, PartialEq, Eq, PartialOrd, Ord, Clone)]
pub struct NamePerLanguage(
#[serde(
serialize_with = "serialize_btreemap",
deserialize_with = "deserialize_btreemap"
)]
pub(crate) BTreeMap<Option<String>, String>,
);
impl NamePerLanguage {
pub fn get(&self, lang: Option<&String>) -> &String {
// TODO Can we avoid this clone?
let lang = lang.cloned();
if let Some(name) = self.0.get(&lang) { | return name;
}
&self.0[&None]
}
pub fn new(tags: &Tags) -> Option<NamePerLanguage> {
let native_name = tags.get(osm::NAME)?;
let mut map = BTreeMap::new();
map.insert(None, native_name.to_string());
for (k, v) in tags.inner() {
if let Some(lang) = k.strip_prefix("name:") {
map.insert(Some(lang.to_string()), v.to_string());
}
}
Some(NamePerLanguage(map))
}
pub fn unnamed() -> NamePerLanguage {
let mut map = BTreeMap::new();
map.insert(None, "unnamed".to_string());
NamePerLanguage(map)
}
}
impl Building {
pub fn sidewalk(&self) -> LaneID {
self.sidewalk_pos.lane()
}
pub fn house_number(&self) -> Option<String> {
let num = self.address.split(' ').next().unwrap();
if num != "???" {
Some(num.to_string())
} else {
None
}
}
/// The polyline goes from the building to the driving position
// TODO Make this handle parking_blackhole
pub fn driving_connection(&self, map: &Map) -> Option<(Position, PolyLine)> {
let lane = map
.get_parent(self.sidewalk())
.find_closest_lane(self.sidewalk(), |l| PathConstraints::Car.can_use(l, map))?;
// TODO Do we need to insist on this buffer, now that we can make cars gradually appear?
let pos = self
.sidewalk_pos
.equiv_pos(lane, map)
.buffer_dist(Distance::meters(7.0), map)?;
Some((pos, self.driveway_geom.clone().optionally_push(pos.pt(map))))
}
/// Returns (biking position, sidewalk position). Could fail if the biking graph is
/// disconnected.
pub fn biking_connection(&self, map: &Map) -> Option<(Position, Position)> {
// Easy case: the building is directly next to a usable lane
if let Some(pair) = sidewalk_to_bike(self.sidewalk_pos, map) {
return Some(pair);
}
// Floodfill the sidewalk graph until we find a sidewalk<->bike connection.
let mut queue: VecDeque<LaneID> = VecDeque::new();
let mut visited: HashSet<LaneID> = HashSet::new();
queue.push_back(self.sidewalk());
loop {
if queue.is_empty() {
return None;
}
let l = queue.pop_front().unwrap();
if visited.contains(&l) {
continue;
}
visited.insert(l);
// TODO Could search by sidewalk endpoint
if let Some(pair) = sidewalk_to_bike(Position::new(l, map.get_l(l).length() / 2.0), map)
{
return Some(pair);
}
for t in map.get_turns_from_lane(l) {
if !visited.contains(&t.id.dst) {
queue.push_back(t.id.dst);
}
}
}
}
pub fn num_parking_spots(&self) -> usize {
match self.parking {
OffstreetParking::PublicGarage(_, n) => n,
OffstreetParking::Private(n, _) => n,
}
}
/// Does this building contain any amenity matching the category?
pub fn has_amenity(&self, category: AmenityType) -> bool {
for amenity in &self.amenities {
if AmenityType::categorize(&amenity.amenity_type) == Some(category) {
return true;
}
}
false
}
}
fn sidewalk_to_bike(sidewalk_pos: Position, map: &Map) -> Option<(Position, Position)> {
let lane = map
.get_parent(sidewalk_pos.lane())
.find_closest_lane(sidewalk_pos.lane(), |l| {
!l.biking_blackhole && PathConstraints::Bike.can_use(l, map)
})?;
// No buffer needed
Some((sidewalk_pos.equiv_pos(lane, map), sidewalk_pos))
}
/// Businesses are categorized into one of these types.
#[derive(Clone, Copy, PartialEq, Eq, PartialOrd, Ord, EnumString, Display, EnumIter)]
pub enum AmenityType {
Bank,
Bar,
Beauty,
Bike,
Cafe,
CarRepair,
CarShare,
Childcare,
ConvenienceStore,
Culture,
Exercise,
FastFood,
Food,
GreenSpace,
Hotel,
Laundry,
Library,
Medical,
Pet,
Playground,
Pool,
PostOffice,
Religious,
School,
Shopping,
Supermarket,
Tourism,
University,
}
impl AmenityType {
fn types(self) -> Vec<&'static str> {
match self {
AmenityType::Bank => vec!["bank"],
AmenityType::Bar => vec!["bar", "pub", "nightclub", "biergarten"],
AmenityType::Beauty => vec!["hairdresser", "beauty", "chemist", "cosmetics"],
AmenityType::Bike => vec!["bicycle"],
AmenityType::Cafe => vec!["cafe", "pastry", "coffee", "tea", "bakery"],
AmenityType::CarRepair => vec!["car_repair"],
AmenityType::CarShare => vec!["car_sharing"],
AmenityType::Childcare => vec!["childcare", "kindergarten"],
AmenityType::ConvenienceStore => vec!["convenience"],
AmenityType::Culture => vec!["arts_centre", "art", "cinema", "theatre"],
AmenityType::Exercise => vec!["fitness_centre", "sports_centre", "track", "pitch"],
AmenityType::FastFood => vec!["fast_food", "food_court"],
AmenityType::Food => vec![
"restaurant",
"farm",
"ice_cream",
"seafood",
"cheese",
"chocolate",
"deli",
"butcher",
"confectionery",
"beverages",
"alcohol",
],
AmenityType::GreenSpace => vec!["park", "garden", "nature_reserve"],
AmenityType::Hotel => vec!["hotel", "hostel", "guest_house", "motel"],
AmenityType::Laundry => vec!["dry_cleaning", "laundry", "tailor"],
AmenityType::Library => vec!["library"],
AmenityType::Medical => vec![
"clinic", " | random_line_split |
|
building.rs | ,
pub address: String,
pub name: Option<NamePerLanguage>,
pub orig_id: osm::OsmID,
/// Where a text label should be centered to have the best chances of being contained within
/// the polygon.
pub label_center: Pt2D,
pub amenities: Vec<Amenity>,
pub bldg_type: BuildingType,
pub parking: OffstreetParking,
/// Depending on options while importing, these might be empty, to save file space.
pub osm_tags: Tags,
/// The building's connection for any agent can change based on map edits. Just store the one
/// for pedestrians and lazily calculate the others.
pub sidewalk_pos: Position,
/// Goes from building to sidewalk
pub driveway_geom: PolyLine,
}
/// A business located inside a building.
#[derive(Serialize, Deserialize, Debug, Clone)]
pub struct Amenity {
pub names: NamePerLanguage,
/// This is the specific amenity listed in OSM, not the more general `AmenityType` category.
pub amenity_type: String,
/// Depending on options while importing, these might be empty, to save file space.
pub osm_tags: Tags,
}
/// Represent no parking as Private(0, false).
#[derive(Serialize, Deserialize, Debug, PartialEq, Clone)]
pub enum OffstreetParking {
/// (Name, spots)
PublicGarage(String, usize),
/// (Spots, explicitly tagged as a garage)
Private(usize, bool),
}
#[derive(Serialize, Deserialize, Clone, Debug)]
pub enum BuildingType {
Residential {
num_residents: usize,
num_housing_units: usize,
},
/// An estimated number of residents, workers
ResidentialCommercial(usize, usize),
/// An estimated number of workers
Commercial(usize),
Empty,
}
impl BuildingType {
pub fn has_residents(&self) -> bool {
match self {
BuildingType::Residential { .. } | BuildingType::ResidentialCommercial(_, _) => true,
BuildingType::Commercial(_) | BuildingType::Empty => false,
}
}
}
/// None corresponds to the native name
#[derive(Serialize, Deserialize, Debug, PartialEq, Eq, PartialOrd, Ord, Clone)]
pub struct NamePerLanguage(
#[serde(
serialize_with = "serialize_btreemap",
deserialize_with = "deserialize_btreemap"
)]
pub(crate) BTreeMap<Option<String>, String>,
);
impl NamePerLanguage {
pub fn get(&self, lang: Option<&String>) -> &String {
// TODO Can we avoid this clone?
let lang = lang.cloned();
if let Some(name) = self.0.get(&lang) {
return name;
}
&self.0[&None]
}
pub fn new(tags: &Tags) -> Option<NamePerLanguage> {
let native_name = tags.get(osm::NAME)?;
let mut map = BTreeMap::new();
map.insert(None, native_name.to_string());
for (k, v) in tags.inner() {
if let Some(lang) = k.strip_prefix("name:") {
map.insert(Some(lang.to_string()), v.to_string());
}
}
Some(NamePerLanguage(map))
}
pub fn unnamed() -> NamePerLanguage {
let mut map = BTreeMap::new();
map.insert(None, "unnamed".to_string());
NamePerLanguage(map)
}
}
impl Building {
pub fn sidewalk(&self) -> LaneID {
self.sidewalk_pos.lane()
}
pub fn house_number(&self) -> Option<String> {
let num = self.address.split(' ').next().unwrap();
if num != "???" {
Some(num.to_string())
} else {
None
}
}
/// The polyline goes from the building to the driving position
// TODO Make this handle parking_blackhole
pub fn driving_connection(&self, map: &Map) -> Option<(Position, PolyLine)> {
let lane = map
.get_parent(self.sidewalk())
.find_closest_lane(self.sidewalk(), |l| PathConstraints::Car.can_use(l, map))?;
// TODO Do we need to insist on this buffer, now that we can make cars gradually appear?
let pos = self
.sidewalk_pos
.equiv_pos(lane, map)
.buffer_dist(Distance::meters(7.0), map)?;
Some((pos, self.driveway_geom.clone().optionally_push(pos.pt(map))))
}
/// Returns (biking position, sidewalk position). Could fail if the biking graph is
/// disconnected.
pub fn biking_connection(&self, map: &Map) -> Option<(Position, Position)> {
// Easy case: the building is directly next to a usable lane
if let Some(pair) = sidewalk_to_bike(self.sidewalk_pos, map) {
return Some(pair);
}
// Floodfill the sidewalk graph until we find a sidewalk<->bike connection.
let mut queue: VecDeque<LaneID> = VecDeque::new();
let mut visited: HashSet<LaneID> = HashSet::new();
queue.push_back(self.sidewalk());
loop {
if queue.is_empty() {
return None;
}
let l = queue.pop_front().unwrap();
if visited.contains(&l) {
continue;
}
visited.insert(l);
// TODO Could search by sidewalk endpoint
if let Some(pair) = sidewalk_to_bike(Position::new(l, map.get_l(l).length() / 2.0), map)
{
return Some(pair);
}
for t in map.get_turns_from_lane(l) {
if !visited.contains(&t.id.dst) {
queue.push_back(t.id.dst);
}
}
}
}
pub fn num_parking_spots(&self) -> usize {
match self.parking {
OffstreetParking::PublicGarage(_, n) => n,
OffstreetParking::Private(n, _) => n,
}
}
/// Does this building contain any amenity matching the category?
pub fn has_amenity(&self, category: AmenityType) -> bool {
for amenity in &self.amenities {
if AmenityType::categorize(&amenity.amenity_type) == Some(category) |
}
false
}
}
fn sidewalk_to_bike(sidewalk_pos: Position, map: &Map) -> Option<(Position, Position)> {
let lane = map
.get_parent(sidewalk_pos.lane())
.find_closest_lane(sidewalk_pos.lane(), |l| {
!l.biking_blackhole && PathConstraints::Bike.can_use(l, map)
})?;
// No buffer needed
Some((sidewalk_pos.equiv_pos(lane, map), sidewalk_pos))
}
/// Businesses are categorized into one of these types.
#[derive(Clone, Copy, PartialEq, Eq, PartialOrd, Ord, EnumString, Display, EnumIter)]
pub enum AmenityType {
Bank,
Bar,
Beauty,
Bike,
Cafe,
CarRepair,
CarShare,
Childcare,
ConvenienceStore,
Culture,
Exercise,
FastFood,
Food,
GreenSpace,
Hotel,
Laundry,
Library,
Medical,
Pet,
Playground,
Pool,
PostOffice,
Religious,
School,
Shopping,
Supermarket,
Tourism,
University,
}
impl AmenityType {
fn types(self) -> Vec<&'static str> {
match self {
AmenityType::Bank => vec!["bank"],
AmenityType::Bar => vec!["bar", "pub", "nightclub", "biergarten"],
AmenityType::Beauty => vec!["hairdresser", "beauty", "chemist", "cosmetics"],
AmenityType::Bike => vec!["bicycle"],
AmenityType::Cafe => vec!["cafe", "pastry", "coffee", "tea", "bakery"],
AmenityType::CarRepair => vec!["car_repair"],
AmenityType::CarShare => vec!["car_sharing"],
AmenityType::Childcare => vec!["childcare", "kindergarten"],
AmenityType::ConvenienceStore => vec!["convenience"],
AmenityType::Culture => vec!["arts_centre", "art", "cinema", "theatre"],
AmenityType::Exercise => vec!["fitness_centre", "sports_centre", "track", "pitch"],
AmenityType::FastFood => vec!["fast_food", "food_court"],
AmenityType::Food => vec![
"restaurant",
"farm",
"ice_cream",
"seafood",
"cheese",
"chocolate",
"deli",
"butcher",
"confectionery",
"beverages",
"alcohol",
],
AmenityType::GreenSpace => vec!["park", "garden", "nature_reserve"],
AmenityType::Hotel => vec!["hotel", "hostel", "guest_house", "motel"],
AmenityType::Laundry => vec!["dry_cleaning", "laundry", "tailor"],
AmenityType::Library => vec!["library"],
AmenityType::Medical => vec![
"clinic", | {
return true;
} | conditional_block |
main.go | (key string, addIfMissing, deleteIfPresent bool) *ObjectData {
c.Lock()
defer c.Unlock()
od := c.objects[key]
if od == nil {
od = &ObjectData{}
if addIfMissing {
c.objects[key] = od
}
} else if deleteIfPresent {
delete(c.objects, key)
}
return od
}
func NewController(queue workqueue.RateLimitingInterface, informer cache.Controller, lister corev1listers.ConfigMapLister, compare bool, csvFilename, myAddr string) *Controller {
createHistogram := prometheus.NewHistogramVec(
prometheus.HistogramOpts{
Namespace: HistogramNamespace,
Subsystem: HistogramSubsystem,
Name: CreateHistogramName,
Help: "Configmap creation notification latency, in seconds",
Buckets: []float64{-0.1, 0, 0.03125, 0.0625, 0.125, 0.25, 0.5, 1, 2, 4, 8},
},
[]string{"logger"},
)
if err := prometheus.Register(createHistogram); err != nil {
klog.Error(err)
createHistogram = nil
} else {
createHistogram.With(prometheus.Labels{"logger": myAddr}).Observe(0)
}
updateHistogram := prometheus.NewHistogramVec(
prometheus.HistogramOpts{
Namespace: HistogramNamespace,
Subsystem: HistogramSubsystem,
Name: UpdateHistogramName,
Help: "Configmap creation notification latency, in seconds",
Buckets: []float64{-0.1, 0, 0.03125, 0.0625, 0.125, 0.25, 0.5, 1, 2, 4, 8},
},
[]string{"logger"},
)
if err := prometheus.Register(updateHistogram); err != nil {
klog.Error(err)
updateHistogram = nil
} else {
updateHistogram.With(prometheus.Labels{"logger": myAddr}).Observe(0)
}
updateCounter := prometheus.NewCounterVec(
prometheus.CounterOpts{
Namespace: HistogramNamespace,
Subsystem: HistogramSubsystem,
Name: "updates",
Help: "number of updates dequeued",
},
[]string{"logger"},
)
if err := prometheus.Register(updateCounter); err != nil {
klog.Error(err)
updateCounter = nil
} else {
updateCounter.With(prometheus.Labels{"logger": myAddr}).Add(0)
}
strangeCounter := prometheus.NewCounterVec(
prometheus.CounterOpts{
Namespace: HistogramNamespace,
Subsystem: HistogramSubsystem,
Name: "stranges",
Help: "number of strange situations dequeued",
},
[]string{"logger"},
)
if err := prometheus.Register(strangeCounter); err != nil {
klog.Error(err)
strangeCounter = nil
} else {
strangeCounter.With(prometheus.Labels{"logger": myAddr}).Add(0)
}
duplicateCounter := prometheus.NewCounter(
prometheus.CounterOpts{
Namespace: HistogramNamespace,
Subsystem: HistogramSubsystem,
Name: "duplicates",
Help: "number of duplicates dequeued",
ConstLabels: map[string]string{"logger": myAddr},
})
if err := prometheus.Register(duplicateCounter); err != nil {
klog.Error(err)
duplicateCounter = nil
} else {
duplicateCounter.Add(0)
}
rvGauge := prometheus.NewGauge(
prometheus.GaugeOpts{
Namespace: HistogramNamespace,
Subsystem: HistogramSubsystem,
Name: "resourceVersion",
Help: "latest ResourceVersion observed",
ConstLabels: map[string]string{"logger": myAddr},
})
if err := prometheus.Register(rvGauge); err != nil {
klog.Error(err)
rvGauge = nil
}
return &Controller{
myAddr: myAddr,
compare: compare,
informer: informer,
queue: queue,
lister: lister,
csvFilename: csvFilename,
createLatencyHistogram: createHistogram,
updateLatencyHistogram: updateHistogram,
updateCounter: updateCounter,
strangeCounter: strangeCounter,
duplicateCounter: duplicateCounter,
rvGauge: rvGauge,
objects: make(map[string]*ObjectData),
}
}
func (c *Controller) processNextItem() bool {
// Wait until there is a new item in the working queue
key, quit := c.queue.Get()
if quit {
return false
}
// Tell the queue that we are done with processing this key. This unblocks the key for other workers
// This allows safe parallel processing because two objects with the same key are never processed in
// parallel.
defer c.queue.Done(key)
// Invoke the method containing the business logic
err := c.logDequeue(key.(string))
// Handle the error if something went wrong during the execution of the business logic
c.handleErr(err, key)
return true
}
func (c *Controller) logDequeue(key string) error {
now := time.Now()
namespace, name, err := cache.SplitMetaNamespaceKey(key)
if err != nil {
runtime.HandleError(fmt.Errorf("Failed to split key %q: %v", key, err))
return nil
}
obj, err := c.lister.ConfigMaps(namespace).Get(name)
if err != nil && !apierrors.IsNotFound(err) {
runtime.HandleError(fmt.Errorf("Fetching object with key %s from store failed with %v", key, err))
return nil
}
desireExist := err == nil
od := c.getObjectData(key, desireExist, !desireExist)
op := "delete"
var creationTime time.Time = dummyTime
if obj != nil {
creationTime = obj.ObjectMeta.CreationTimestamp.Time
}
var oqd ObjectQueueData
var lastSeen *corev1.ConfigMap
func() {
od.Lock()
defer od.Unlock()
oqd = od.ObjectQueueData
if c.compare {
lastSeen = od.lastSeen
od.lastSeen = obj.DeepCopy()
}
if desireExist {
if od.actuallyExists {
op = "update"
} else {
op = "create"
}
od.ObjectQueueData = zeroObjectQueueData
od.actuallyExists = true
}
}()
var diff int
if c.compare {
if ConfigMapQuickEqual(lastSeen, obj) {
diff = 2
c.duplicateCounter.Add(1)
} else {
diff = 3
}
}
// Log it
if c.csvFile != nil {
_, err = c.csvFile.Write([]byte(fmt.Sprintf("%s,%s,%q,%s,%d,%d,%d,%s,%s,%d\n",
formatTime(now), op, key, formatTimeNoMillis(creationTime),
oqd.queuedAdds, oqd.queuedUpdates, oqd.queuedDeletes,
formatTime(oqd.firstEnqueue), formatTime(oqd.lastEnqueue),
diff,
)))
if err != nil {
runtime.HandleError(fmt.Errorf("Error writing to CSV file named %q: %+v", c.csvFilename, err))
}
} else {
klog.V(4).Infof("c.csvFile == nil\n")
}
if diff == 2 {
return nil
}
if oqd.queuedAdds+oqd.queuedUpdates+oqd.queuedDeletes != 1 {
if c.strangeCounter != nil {
c.strangeCounter.
With(prometheus.Labels{"logger": c.myAddr}).
Add(1)
}
} else if oqd.queuedUpdates == 1 && c.updateCounter != nil {
c.updateCounter.
With(prometheus.Labels{"logger": c.myAddr}).
Add(1)
}
if op != "delete" && obj != nil && obj.Annotations != nil {
var ctS string
var latencyHistogram *prometheus.HistogramVec
if op == "create" {
ctS = obj.Annotations[CreateTimestampAnnotation]
latencyHistogram = c.createLatencyHistogram
} else {
ctS = obj.Annotations[UpdateTimestampAnnotation]
latencyHistogram = c.updateLatencyHistogram
}
if ctS != "" && latencyHistogram != nil {
clientTime, err := time.Parse(CreateTimestampLayout, ctS)
if err != nil {
return nil
}
latency := now.Sub(clientTime)
klog.V(4).Infof("Latency = %v for op=%s, key=%s, now=%s, clientTime=%s, ts=%s\n", latency, op, key, now, clientTime, ctS)
latencyHistogram.
With(prometheus.Labels{"logger": c.myAddr}).
Observe(latency.Seconds())
}
}
return nil
}
// handleErr | getObjectData | identifier_name |
|
main.go | , informer cache.Controller, lister corev1listers.ConfigMapLister, compare bool, csvFilename, myAddr string) *Controller {
createHistogram := prometheus.NewHistogramVec(
prometheus.HistogramOpts{
Namespace: HistogramNamespace,
Subsystem: HistogramSubsystem,
Name: CreateHistogramName,
Help: "Configmap creation notification latency, in seconds",
Buckets: []float64{-0.1, 0, 0.03125, 0.0625, 0.125, 0.25, 0.5, 1, 2, 4, 8},
},
[]string{"logger"},
)
if err := prometheus.Register(createHistogram); err != nil {
klog.Error(err)
createHistogram = nil
} else {
createHistogram.With(prometheus.Labels{"logger": myAddr}).Observe(0)
}
updateHistogram := prometheus.NewHistogramVec(
prometheus.HistogramOpts{
Namespace: HistogramNamespace,
Subsystem: HistogramSubsystem,
Name: UpdateHistogramName,
Help: "Configmap creation notification latency, in seconds",
Buckets: []float64{-0.1, 0, 0.03125, 0.0625, 0.125, 0.25, 0.5, 1, 2, 4, 8},
},
[]string{"logger"},
)
if err := prometheus.Register(updateHistogram); err != nil {
klog.Error(err)
updateHistogram = nil
} else {
updateHistogram.With(prometheus.Labels{"logger": myAddr}).Observe(0)
}
updateCounter := prometheus.NewCounterVec(
prometheus.CounterOpts{
Namespace: HistogramNamespace,
Subsystem: HistogramSubsystem,
Name: "updates",
Help: "number of updates dequeued",
},
[]string{"logger"},
)
if err := prometheus.Register(updateCounter); err != nil {
klog.Error(err)
updateCounter = nil
} else {
updateCounter.With(prometheus.Labels{"logger": myAddr}).Add(0)
}
strangeCounter := prometheus.NewCounterVec(
prometheus.CounterOpts{
Namespace: HistogramNamespace,
Subsystem: HistogramSubsystem,
Name: "stranges",
Help: "number of strange situations dequeued",
},
[]string{"logger"},
)
if err := prometheus.Register(strangeCounter); err != nil {
klog.Error(err)
strangeCounter = nil
} else {
strangeCounter.With(prometheus.Labels{"logger": myAddr}).Add(0)
}
duplicateCounter := prometheus.NewCounter(
prometheus.CounterOpts{
Namespace: HistogramNamespace,
Subsystem: HistogramSubsystem,
Name: "duplicates",
Help: "number of duplicates dequeued",
ConstLabels: map[string]string{"logger": myAddr},
})
if err := prometheus.Register(duplicateCounter); err != nil {
klog.Error(err)
duplicateCounter = nil
} else {
duplicateCounter.Add(0)
}
rvGauge := prometheus.NewGauge(
prometheus.GaugeOpts{
Namespace: HistogramNamespace,
Subsystem: HistogramSubsystem,
Name: "resourceVersion",
Help: "latest ResourceVersion observed",
ConstLabels: map[string]string{"logger": myAddr},
})
if err := prometheus.Register(rvGauge); err != nil {
klog.Error(err)
rvGauge = nil
}
return &Controller{
myAddr: myAddr,
compare: compare,
informer: informer,
queue: queue,
lister: lister,
csvFilename: csvFilename,
createLatencyHistogram: createHistogram,
updateLatencyHistogram: updateHistogram,
updateCounter: updateCounter,
strangeCounter: strangeCounter,
duplicateCounter: duplicateCounter,
rvGauge: rvGauge,
objects: make(map[string]*ObjectData),
}
}
func (c *Controller) processNextItem() bool |
func (c *Controller) logDequeue(key string) error {
now := time.Now()
namespace, name, err := cache.SplitMetaNamespaceKey(key)
if err != nil {
runtime.HandleError(fmt.Errorf("Failed to split key %q: %v", key, err))
return nil
}
obj, err := c.lister.ConfigMaps(namespace).Get(name)
if err != nil && !apierrors.IsNotFound(err) {
runtime.HandleError(fmt.Errorf("Fetching object with key %s from store failed with %v", key, err))
return nil
}
desireExist := err == nil
od := c.getObjectData(key, desireExist, !desireExist)
op := "delete"
var creationTime time.Time = dummyTime
if obj != nil {
creationTime = obj.ObjectMeta.CreationTimestamp.Time
}
var oqd ObjectQueueData
var lastSeen *corev1.ConfigMap
func() {
od.Lock()
defer od.Unlock()
oqd = od.ObjectQueueData
if c.compare {
lastSeen = od.lastSeen
od.lastSeen = obj.DeepCopy()
}
if desireExist {
if od.actuallyExists {
op = "update"
} else {
op = "create"
}
od.ObjectQueueData = zeroObjectQueueData
od.actuallyExists = true
}
}()
var diff int
if c.compare {
if ConfigMapQuickEqual(lastSeen, obj) {
diff = 2
c.duplicateCounter.Add(1)
} else {
diff = 3
}
}
// Log it
if c.csvFile != nil {
_, err = c.csvFile.Write([]byte(fmt.Sprintf("%s,%s,%q,%s,%d,%d,%d,%s,%s,%d\n",
formatTime(now), op, key, formatTimeNoMillis(creationTime),
oqd.queuedAdds, oqd.queuedUpdates, oqd.queuedDeletes,
formatTime(oqd.firstEnqueue), formatTime(oqd.lastEnqueue),
diff,
)))
if err != nil {
runtime.HandleError(fmt.Errorf("Error writing to CSV file named %q: %+v", c.csvFilename, err))
}
} else {
klog.V(4).Infof("c.csvFile == nil\n")
}
if diff == 2 {
return nil
}
if oqd.queuedAdds+oqd.queuedUpdates+oqd.queuedDeletes != 1 {
if c.strangeCounter != nil {
c.strangeCounter.
With(prometheus.Labels{"logger": c.myAddr}).
Add(1)
}
} else if oqd.queuedUpdates == 1 && c.updateCounter != nil {
c.updateCounter.
With(prometheus.Labels{"logger": c.myAddr}).
Add(1)
}
if op != "delete" && obj != nil && obj.Annotations != nil {
var ctS string
var latencyHistogram *prometheus.HistogramVec
if op == "create" {
ctS = obj.Annotations[CreateTimestampAnnotation]
latencyHistogram = c.createLatencyHistogram
} else {
ctS = obj.Annotations[UpdateTimestampAnnotation]
latencyHistogram = c.updateLatencyHistogram
}
if ctS != "" && latencyHistogram != nil {
clientTime, err := time.Parse(CreateTimestampLayout, ctS)
if err != nil {
return nil
}
latency := now.Sub(clientTime)
klog.V(4).Infof("Latency = %v for op=%s, key=%s, now=%s, clientTime=%s, ts=%s\n", latency, op, key, now, clientTime, ctS)
latencyHistogram.
With(prometheus.Labels{"logger": c.myAddr}).
Observe(latency.Seconds())
}
}
return nil
}
// handleErr checks if an error happened and makes sure we will retry later.
func (c *Controller) handleErr(err error, key interface{}) {
if err == nil {
// Forget about the #AddRateLimited history of the key on every successful synchronization.
// This ensures that future processing of updates for this key is not delayed because of
// an outdated error history.
c.queue.Forget(key)
return
}
k | {
// Wait until there is a new item in the working queue
key, quit := c.queue.Get()
if quit {
return false
}
// Tell the queue that we are done with processing this key. This unblocks the key for other workers
// This allows safe parallel processing because two objects with the same key are never processed in
// parallel.
defer c.queue.Done(key)
// Invoke the method containing the business logic
err := c.logDequeue(key.(string))
// Handle the error if something went wrong during the execution of the business logic
c.handleErr(err, key)
return true
} | identifier_body |
main.go | be processed later again.
c.queue.AddRateLimited(key)
return
}
func (c *Controller) Run(threadiness int, stopCh chan struct{}) {
defer runtime.HandleCrash()
// Let the workers stop when we are done
defer c.queue.ShutDown()
klog.Info("Starting Object Logging controller")
csvFile, err := os.Create(c.csvFilename)
if err != nil {
runtime.HandleError(fmt.Errorf("Failed to create data file named %q: %s", c.csvFilename, err))
} else {
c.csvFile = csvFile
defer csvFile.Close()
}
go c.informer.Run(stopCh)
// Wait for all involved caches to be synced, before processing items from the queue is started
if !cache.WaitForCacheSync(stopCh, c.informer.HasSynced) {
runtime.HandleError(fmt.Errorf("Timed out waiting for caches to sync"))
return
}
for i := 0; i < threadiness; i++ {
go wait.Until(c.runWorker, time.Second, stopCh)
}
<-stopCh
klog.Info("Stopping Object Logging controller")
}
func (c *Controller) runWorker() {
for c.processNextItem() {
}
}
func (c *Controller) ObserveResourceVersion(obj interface{}) {
switch o := obj.(type) {
case cache.DeletedFinalStateUnknown:
klog.V(5).Infof("Recursing for %#+v @ %#p\n", obj, obj)
c.ObserveResourceVersion(o.Obj)
case cache.ExplicitKey:
klog.V(5).Infof("Got ExplicitKey %q\n", o)
return
default:
meta, err := apimeta.Accessor(obj)
if err != nil {
klog.V(5).Infof("apimeta.Accessor(%#+v) threw %#+v\n", obj, err)
return
}
rvS := meta.GetResourceVersion()
rvU, err := strconv.ParseUint(rvS, 10, 64)
if err != nil {
klog.V(5).Infof("Error parsing ResourceVersion %q of %#+v: %#+v\n", rvS, obj, err)
} else {
klog.V(5).Infof("Observing ResourceVersion %d of %#+v @ %#p\n", rvU, obj, obj)
c.rvGauge.Set(float64(rvU))
}
}
}
func main() {
var kubeconfig string
var master string
var useProtobuf bool
var dataFilename string
var numThreads int
var noCompare bool
klog.InitFlags(nil)
flag.StringVar(&kubeconfig, "kubeconfig", "", "absolute path to the kubeconfig file")
flag.StringVar(&master, "master", "", "master url")
flag.BoolVar(&useProtobuf, "useProtobuf", false, "indicates whether to encode objects with protobuf (as opposed to JSON)")
flag.StringVar(&dataFilename, "data-filname", "/tmp/obj-log.csv", "name of CSV file to create")
flag.IntVar(&numThreads, "threads", 1, "number of worker threads")
flag.BoolVar(&noCompare, "no-compare", false, "omit comparing object values")
flag.Set("logtostderr", "true")
flag.Parse()
// creates the connection
config, err := clientcmd.BuildConfigFromFlags(master, kubeconfig)
if err != nil {
klog.Fatal(err)
}
klog.Infof("Config.Host=%q\n", config.Host)
klog.Infof("Config.APIPath=%q\n", config.APIPath)
myAddr := GetHostAddr()
klog.Infof("Using %s as my host address\n", myAddr)
config.UserAgent = fmt.Sprintf("obj-logger@%s", myAddr)
if useProtobuf {
config.ContentType = "application/vnd.kubernetes.protobuf"
}
// creates the clientset
clientset, err := kubernetes.NewForConfig(config)
if err != nil {
klog.Fatal(err)
}
informerFactory := informers.NewSharedInformerFactory(clientset, 0)
cfgMapInformer := informerFactory.Core().V1().ConfigMaps()
informer := cfgMapInformer.Informer()
lister := cfgMapInformer.Lister()
// create the workqueue
queue := workqueue.NewRateLimitingQueue(workqueue.DefaultControllerRateLimiter())
controller := NewController(queue, informer, lister, !noCompare, dataFilename, myAddr)
// Bind the workqueue to a cache with the help of an informer. This way we make sure that
// whenever the cache is updated, the object key is added to the workqueue.
// Note that when we finally process the item from the workqueue, we might see a newer version
// of the object than the version which was responsible for triggering the update.
informer.AddEventHandler(cache.ResourceEventHandlerFuncs{
AddFunc: func(obj interface{}) {
now := time.Now()
klog.V(4).Infof("ADD %+v @ %#p\n", obj, obj)
controller.ObserveResourceVersion(obj)
key, err := cache.MetaNamespaceKeyFunc(obj)
if err == nil {
od := controller.getObjectData(key, true, false)
od.Lock()
defer od.Unlock()
if od.queuedAdds+od.queuedUpdates+od.queuedDeletes == 0 {
od.firstEnqueue = now
}
od.lastEnqueue = now
od.queuedAdds++
queue.Add(key)
} else {
klog.Errorf("Failed to parse key from obj %#v: %v\n", obj, err)
}
},
UpdateFunc: func(oldobj interface{}, newobj interface{}) {
now := time.Now()
klog.V(4).Infof("UPDATE %#v @ %#p\n", newobj, newobj)
controller.ObserveResourceVersion(newobj)
key, err := cache.MetaNamespaceKeyFunc(newobj)
if err == nil {
od := controller.getObjectData(key, true, false)
od.Lock()
defer od.Unlock()
if od.queuedAdds+od.queuedUpdates+od.queuedDeletes == 0 {
od.firstEnqueue = now
}
od.lastEnqueue = now
od.queuedUpdates++
queue.Add(key)
} else {
klog.Errorf("Failed to parse key from obj %#v: %v\n", newobj, err)
}
},
DeleteFunc: func(obj interface{}) {
now := time.Now()
klog.V(4).Infof("DELETE %#v @ %#p\n", obj, obj)
controller.ObserveResourceVersion(obj)
// IndexerInformer uses a delta queue, therefore for deletes we have to use this
// key function.
key, err := cache.DeletionHandlingMetaNamespaceKeyFunc(obj)
if err == nil {
od := controller.getObjectData(key, true, false)
od.Lock()
defer od.Unlock()
if od.queuedAdds+od.queuedUpdates+od.queuedDeletes == 0 {
od.firstEnqueue = now
}
od.lastEnqueue = now
od.queuedDeletes++
queue.Add(key)
} else {
klog.Errorf("Failed to parse key from obj %#v: %v\n", obj, err)
}
},
})
// Now let's start the controller
stop := make(chan struct{})
defer close(stop)
go controller.Run(numThreads, stop)
// Serve Prometheus metrics
http.Handle("/metrics", promhttp.Handler())
go func() {
klog.Error(http.ListenAndServe(MetricsAddr, nil))
}()
// Wait forever
select {}
}
func formatTime(t time.Time) string {
t = t.UTC()
Y, M, D := t.Date()
h, m, s := t.Clock()
ms := t.Nanosecond() / 1000000
return fmt.Sprintf("%d-%02d-%02d %02d:%02d:%02d.%03d", Y, M, D, h, m, s, ms)
}
func formatTimeNoMillis(t time.Time) string {
t = t.UTC()
Y, M, D := t.Date()
h, m, s := t.Clock()
return fmt.Sprintf("%d-%02d-%02d %02d:%02d:%02d", Y, M, D, h, m, s)
}
func ConfigMapQuickEqual(x, y *corev1.ConfigMap) bool {
if x == y {
return true
}
if x == nil || y == nil {
return false
}
return x.Name == y.Name && x.Namespace == y.Namespace &&
x.UID == y.UID && x.ResourceVersion == y.ResourceVersion &&
MapStringStringEqual(x.Data, y.Data) &&
MapStringStringEqual(x.Labels, y.Labels) &&
MapStringStringEqual(x.Annotations, y.Annotations)
}
func MapStringStringEqual(x, y map[string]string) bool {
if x == nil {
return y == nil
} else if y == nil {
return false
}
if len(x) != len(y) {
return false
}
for k, v := range x | {
if y[k] != v {
return false
}
} | conditional_block |
|
main.go | ,%s,%s,%d\n",
formatTime(now), op, key, formatTimeNoMillis(creationTime),
oqd.queuedAdds, oqd.queuedUpdates, oqd.queuedDeletes,
formatTime(oqd.firstEnqueue), formatTime(oqd.lastEnqueue),
diff,
)))
if err != nil {
runtime.HandleError(fmt.Errorf("Error writing to CSV file named %q: %+v", c.csvFilename, err))
}
} else {
klog.V(4).Infof("c.csvFile == nil\n")
}
if diff == 2 {
return nil
}
if oqd.queuedAdds+oqd.queuedUpdates+oqd.queuedDeletes != 1 {
if c.strangeCounter != nil {
c.strangeCounter.
With(prometheus.Labels{"logger": c.myAddr}).
Add(1)
}
} else if oqd.queuedUpdates == 1 && c.updateCounter != nil {
c.updateCounter.
With(prometheus.Labels{"logger": c.myAddr}).
Add(1)
}
if op != "delete" && obj != nil && obj.Annotations != nil {
var ctS string
var latencyHistogram *prometheus.HistogramVec
if op == "create" {
ctS = obj.Annotations[CreateTimestampAnnotation]
latencyHistogram = c.createLatencyHistogram
} else {
ctS = obj.Annotations[UpdateTimestampAnnotation]
latencyHistogram = c.updateLatencyHistogram
}
if ctS != "" && latencyHistogram != nil {
clientTime, err := time.Parse(CreateTimestampLayout, ctS)
if err != nil {
return nil
}
latency := now.Sub(clientTime)
klog.V(4).Infof("Latency = %v for op=%s, key=%s, now=%s, clientTime=%s, ts=%s\n", latency, op, key, now, clientTime, ctS)
latencyHistogram.
With(prometheus.Labels{"logger": c.myAddr}).
Observe(latency.Seconds())
}
}
return nil
}
// handleErr checks if an error happened and makes sure we will retry later.
func (c *Controller) handleErr(err error, key interface{}) {
if err == nil {
// Forget about the #AddRateLimited history of the key on every successful synchronization.
// This ensures that future processing of updates for this key is not delayed because of
// an outdated error history.
c.queue.Forget(key)
return
}
klog.Infof("Error syncing ConfigMap %v: %v", key, err)
// Re-enqueue the key rate limited. Based on the rate limiter on the
// queue and the re-enqueue history, the key will be processed later again.
c.queue.AddRateLimited(key)
return
}
func (c *Controller) Run(threadiness int, stopCh chan struct{}) {
defer runtime.HandleCrash()
// Let the workers stop when we are done
defer c.queue.ShutDown()
klog.Info("Starting Object Logging controller")
csvFile, err := os.Create(c.csvFilename)
if err != nil {
runtime.HandleError(fmt.Errorf("Failed to create data file named %q: %s", c.csvFilename, err))
} else {
c.csvFile = csvFile
defer csvFile.Close()
}
go c.informer.Run(stopCh)
// Wait for all involved caches to be synced, before processing items from the queue is started
if !cache.WaitForCacheSync(stopCh, c.informer.HasSynced) {
runtime.HandleError(fmt.Errorf("Timed out waiting for caches to sync"))
return
}
for i := 0; i < threadiness; i++ {
go wait.Until(c.runWorker, time.Second, stopCh)
}
<-stopCh
klog.Info("Stopping Object Logging controller")
}
func (c *Controller) runWorker() {
for c.processNextItem() {
}
}
func (c *Controller) ObserveResourceVersion(obj interface{}) {
switch o := obj.(type) {
case cache.DeletedFinalStateUnknown:
klog.V(5).Infof("Recursing for %#+v @ %#p\n", obj, obj)
c.ObserveResourceVersion(o.Obj)
case cache.ExplicitKey:
klog.V(5).Infof("Got ExplicitKey %q\n", o)
return
default:
meta, err := apimeta.Accessor(obj)
if err != nil {
klog.V(5).Infof("apimeta.Accessor(%#+v) threw %#+v\n", obj, err)
return
}
rvS := meta.GetResourceVersion()
rvU, err := strconv.ParseUint(rvS, 10, 64)
if err != nil {
klog.V(5).Infof("Error parsing ResourceVersion %q of %#+v: %#+v\n", rvS, obj, err)
} else {
klog.V(5).Infof("Observing ResourceVersion %d of %#+v @ %#p\n", rvU, obj, obj)
c.rvGauge.Set(float64(rvU))
}
}
}
func main() {
var kubeconfig string
var master string
var useProtobuf bool
var dataFilename string
var numThreads int
var noCompare bool
klog.InitFlags(nil)
flag.StringVar(&kubeconfig, "kubeconfig", "", "absolute path to the kubeconfig file")
flag.StringVar(&master, "master", "", "master url")
flag.BoolVar(&useProtobuf, "useProtobuf", false, "indicates whether to encode objects with protobuf (as opposed to JSON)")
flag.StringVar(&dataFilename, "data-filname", "/tmp/obj-log.csv", "name of CSV file to create")
flag.IntVar(&numThreads, "threads", 1, "number of worker threads")
flag.BoolVar(&noCompare, "no-compare", false, "omit comparing object values")
flag.Set("logtostderr", "true")
flag.Parse()
// creates the connection
config, err := clientcmd.BuildConfigFromFlags(master, kubeconfig)
if err != nil {
klog.Fatal(err)
}
klog.Infof("Config.Host=%q\n", config.Host)
klog.Infof("Config.APIPath=%q\n", config.APIPath)
myAddr := GetHostAddr()
klog.Infof("Using %s as my host address\n", myAddr)
config.UserAgent = fmt.Sprintf("obj-logger@%s", myAddr)
if useProtobuf {
config.ContentType = "application/vnd.kubernetes.protobuf"
}
// creates the clientset
clientset, err := kubernetes.NewForConfig(config)
if err != nil {
klog.Fatal(err)
}
informerFactory := informers.NewSharedInformerFactory(clientset, 0)
cfgMapInformer := informerFactory.Core().V1().ConfigMaps()
informer := cfgMapInformer.Informer()
lister := cfgMapInformer.Lister()
// create the workqueue
queue := workqueue.NewRateLimitingQueue(workqueue.DefaultControllerRateLimiter())
controller := NewController(queue, informer, lister, !noCompare, dataFilename, myAddr)
// Bind the workqueue to a cache with the help of an informer. This way we make sure that
// whenever the cache is updated, the object key is added to the workqueue.
// Note that when we finally process the item from the workqueue, we might see a newer version
// of the object than the version which was responsible for triggering the update.
informer.AddEventHandler(cache.ResourceEventHandlerFuncs{
AddFunc: func(obj interface{}) {
now := time.Now()
klog.V(4).Infof("ADD %+v @ %#p\n", obj, obj)
controller.ObserveResourceVersion(obj)
key, err := cache.MetaNamespaceKeyFunc(obj)
if err == nil {
od := controller.getObjectData(key, true, false)
od.Lock()
defer od.Unlock()
if od.queuedAdds+od.queuedUpdates+od.queuedDeletes == 0 {
od.firstEnqueue = now
}
od.lastEnqueue = now
od.queuedAdds++
queue.Add(key)
} else {
klog.Errorf("Failed to parse key from obj %#v: %v\n", obj, err)
}
},
UpdateFunc: func(oldobj interface{}, newobj interface{}) {
now := time.Now()
klog.V(4).Infof("UPDATE %#v @ %#p\n", newobj, newobj)
controller.ObserveResourceVersion(newobj)
key, err := cache.MetaNamespaceKeyFunc(newobj)
if err == nil {
od := controller.getObjectData(key, true, false)
od.Lock()
defer od.Unlock()
if od.queuedAdds+od.queuedUpdates+od.queuedDeletes == 0 {
od.firstEnqueue = now
}
od.lastEnqueue = now
od.queuedUpdates++
queue.Add(key) | } else {
klog.Errorf("Failed to parse key from obj %#v: %v\n", newobj, err)
}
},
DeleteFunc: func(obj interface{}) { | random_line_split |
|
types.rs | apply_sub_type(&subs, &ty);
}
Ok((im::HashMap::new(), ty))
}
type TypeRes<'a> = (im::HashMap<String, Type>, Type);
pub type Scheme = (im::HashSet<String>, Type);
type Subs<'a> = &'a im::HashMap<String, Type>;
fn apply_sub_type(subs: Subs, ty: &Type) -> Type {
match ty {
Type::TyVar(name) => subs.get(name).unwrap_or_else(|| &ty).clone(),
Type::TyArr(t1, t2) => Type::TyArr(
Box::new(apply_sub_type(subs, t1)),
Box::new(apply_sub_type(subs, t2)),
),
_ => ty.clone(),
}
}
fn apply_sub_scheme(subs: Subs, scheme: Scheme) -> Scheme {
let mut subs1 = subs.clone();
for key in scheme.0.iter() {
subs1 = subs1.without(key);
}
let ty = apply_sub_type(&subs1, &scheme.1);
(scheme.0, ty)
}
fn apply_sub_env(
subs: &im::HashMap<String, Type>,
env: &im::HashMap<String, Scheme>,
) -> im::HashMap<String, Scheme> {
let mut h = im::HashMap::new();
for (key, value) in env.into_iter() {
h = h.update(key.to_string(), apply_sub_scheme(subs, value.clone()));
}
h
}
fn | (subs: Subs, subs2: Subs) -> im::HashMap<String, Type> {
let mut h = im::HashMap::new();
for (key, value) in subs.into_iter() {
h = h.update(key.to_string(), apply_sub_type(subs, &value.clone()));
}
h.union(subs2.clone())
}
fn ftv_ty(ty: &Type) -> im::HashSet<String> {
match ty {
Type::TyVar(a) => im::HashSet::unit(a.clone()),
Type::TyArr(ty1, ty2) => {
let x = ftv_ty(ty1);
let y = ftv_ty(ty2);
x.union(y)
}
_ => im::HashSet::new(),
}
}
fn ftv_env(env: &im::HashMap<String, Scheme>) -> im::HashSet<String> {
let ftvs = env.values().map(|x| ftv_ty(&x.1));
im::HashSet::unions(ftvs)
}
fn generalize(env: &im::HashMap<String, Scheme>, ty: &Type) -> Scheme {
let xs = ftv_ty(ty);
let ys = ftv_env(env);
let a = xs.difference(ys);
(a, ty.clone())
}
fn unify(ty1: &Type, ty2: &Type) -> Result<im::HashMap<String, Type>, String> {
match (ty1, ty2) {
(Type::TyArr(l, r), Type::TyArr(l1, r1)) => {
let s1 = unify(l, l1)?;
let s2 = unify(&apply_sub_type(&s1, &r), &apply_sub_type(&s1, &r1))?;
Ok(compose(&s2, &s1))
}
(Type::TyVar(a), t) => bind(&a, &t),
(t, Type::TyVar(a)) => bind(&a, &t),
(t1, t2) => {
if t1 == t2 {
Ok(im::HashMap::new())
} else {
Err("UnificationFail".to_string())
}
}
}
}
fn bind(var: &str, ty: &Type) -> Result<im::HashMap<String, Type>, String> {
if let Type::TyVar(x) = ty {
if var == x {
return Ok(im::HashMap::new());
}
}
if ftv_ty(ty).contains(var) {
return Err("Infinite Type".to_string());
}
Ok(im::HashMap::new().update(var.to_string(), ty.clone()))
}
fn type_pat(
env: &im::HashMap<String, Scheme>,
case_type: &Type,
pattern: &Pattern,
) -> Result<im::HashMap<String, Type>, String> {
// todo vars / wildcards, etc
let (_s, ty) = env.get(pattern.name).unwrap();
unify(case_type, ty)
}
/// Converts inner type of dataset
fn convert_inner(
env: &im::HashMap<String, Scheme>,
key: &str,
items: &[Expr],
) -> Result<(String, Type), String> {
let (_s, ty) = get_item_type(items, env)?;
Ok((key.to_string(), ty))
}
// Type inference using http://dev.stephendiehl.com/fun/006_hindley_milner.html#substitution
impl<'a> Expr<'_> {
pub fn get_type(&self, env: &im::HashMap<String, Scheme>) -> Result<TypeRes, String> {
match self {
Expr::Literal(l) => Ok((im::HashMap::new(), l.get_type())),
Expr::Ref(x) => {
let err = format!("Could not find reference {}", x);
let ty = env.get(*x).cloned().ok_or(err)?;
Ok((im::HashMap::new(), ty.1))
}
Expr::LetIn(x) => {
let (s1, t1) = x.expr1.expr.get_type(env)?;
let env1 = apply_sub_env(&s1, env);
let t2 = generalize(&env1, &t1);
let extended_ty = env.update(x.name.to_string(), t2);
let (s2, t2) = x.expr2.expr.get_type(&extended_ty)?;
Ok((compose(&s1, &s2), t2))
}
Expr::DataSet(items) => {
let d: im::HashMap<String, Type> = items
.iter()
.map(|(k, items)| convert_inner(env, k, items))
.flatten()
.collect();
if d.len() == items.len() {
Ok((im::HashMap::new(), Type::Dataset(d)))
} else {
Err("Not all rows matched in type".to_string())
}
}
Expr::Lambda(name, expr) => {
let type_var = Type::TyVar(get_id().to_string()); //fresh();
let env1 = env.update((*name).to_string(), (im::HashSet::new(), type_var.clone()));
let (sub, t1) = expr.expr.get_type(&env1)?;
let substituted = apply_sub_type(&sub, &type_var);
Ok((sub, Type::TyArr(Box::new(substituted), Box::new(t1))))
}
Expr::App(expr1, expr2) => {
let tv = Type::TyVar(get_id().to_string());
let (s1, t1) = expr1.get_type(env)?;
let (s2, t2) = expr2.get_type(&apply_sub_env(&s1, env))?;
let s3 = unify(
&apply_sub_type(&s2, &t1),
&Type::TyArr(Box::new(t2), Box::new(tv.clone())),
)?;
Ok((compose(&compose(&s3, &s2), &s1), apply_sub_type(&s3, &tv)))
}
Expr::Match(expr, exprs) => {
let (mut subs, case_type) = expr.get_type(env)?;
let mut branch_type = Type::TyVar(get_id().to_string());
for (p, branch) in exprs {
// TODO check, test
let pat_sub = type_pat(env, &case_type, p)?;
subs = compose(&subs, &pat_sub);
let (s, n_branch_type) = branch.get_type(env)?;
subs = compose(&subs, &s);
let cur_branch_type = apply_sub_type(&subs, &n_branch_type);
let s2 = unify(&branch_type, &cur_branch_type)?;
subs = compose(&subs, &s2);
branch_type = apply_sub_type(&subs, &branch_type);
}
Ok((subs, branch_type))
}
Expr::Projection(names, expr) => {
let from_ty = expr.get_type(env)?;
match from_ty {
(_s, Type::Dataset(items)) => {
if names
.iter()
.filter(|x| !items.contains_key(&x.to_string()))
.count()
> 0
{
// TODO; improve error
return Err("Not all fields in dataset".to_string());
}
Ok((
im::HashMap::new(),
Type::Dataset(
items
.iter()
.filter(|(k, _v)| names.contains(&&*k.to_string()))
.map(|(k, v)| (k.to_string(), v.clone()))
.collect(),
),
))
}
_ => Err("Expected dataset".to_string()),
}
}
x => Err(format!("not implemented {:?}", x)),
}
}
}
impl Literal {
fn get_type(&self) -> Type {
match self {
Literal::Int64(_) => Type::Int64,
Literal::Int32(_) => Type::Int32,
| compose | identifier_name |
types.rs | apply_sub_type(&subs, &ty);
}
Ok((im::HashMap::new(), ty))
}
type TypeRes<'a> = (im::HashMap<String, Type>, Type);
pub type Scheme = (im::HashSet<String>, Type);
type Subs<'a> = &'a im::HashMap<String, Type>;
fn apply_sub_type(subs: Subs, ty: &Type) -> Type {
match ty {
Type::TyVar(name) => subs.get(name).unwrap_or_else(|| &ty).clone(),
Type::TyArr(t1, t2) => Type::TyArr(
Box::new(apply_sub_type(subs, t1)),
Box::new(apply_sub_type(subs, t2)),
),
_ => ty.clone(),
}
}
fn apply_sub_scheme(subs: Subs, scheme: Scheme) -> Scheme {
let mut subs1 = subs.clone();
for key in scheme.0.iter() {
subs1 = subs1.without(key);
}
let ty = apply_sub_type(&subs1, &scheme.1);
(scheme.0, ty)
}
fn apply_sub_env(
subs: &im::HashMap<String, Type>,
env: &im::HashMap<String, Scheme>,
) -> im::HashMap<String, Scheme> {
let mut h = im::HashMap::new();
for (key, value) in env.into_iter() {
h = h.update(key.to_string(), apply_sub_scheme(subs, value.clone()));
}
h
}
fn compose(subs: Subs, subs2: Subs) -> im::HashMap<String, Type> {
let mut h = im::HashMap::new();
for (key, value) in subs.into_iter() {
h = h.update(key.to_string(), apply_sub_type(subs, &value.clone()));
}
h.union(subs2.clone())
}
fn ftv_ty(ty: &Type) -> im::HashSet<String> {
match ty {
Type::TyVar(a) => im::HashSet::unit(a.clone()),
Type::TyArr(ty1, ty2) => {
let x = ftv_ty(ty1);
let y = ftv_ty(ty2);
x.union(y)
}
_ => im::HashSet::new(),
}
}
fn ftv_env(env: &im::HashMap<String, Scheme>) -> im::HashSet<String> {
let ftvs = env.values().map(|x| ftv_ty(&x.1));
im::HashSet::unions(ftvs)
}
fn generalize(env: &im::HashMap<String, Scheme>, ty: &Type) -> Scheme |
fn unify(ty1: &Type, ty2: &Type) -> Result<im::HashMap<String, Type>, String> {
match (ty1, ty2) {
(Type::TyArr(l, r), Type::TyArr(l1, r1)) => {
let s1 = unify(l, l1)?;
let s2 = unify(&apply_sub_type(&s1, &r), &apply_sub_type(&s1, &r1))?;
Ok(compose(&s2, &s1))
}
(Type::TyVar(a), t) => bind(&a, &t),
(t, Type::TyVar(a)) => bind(&a, &t),
(t1, t2) => {
if t1 == t2 {
Ok(im::HashMap::new())
} else {
Err("UnificationFail".to_string())
}
}
}
}
fn bind(var: &str, ty: &Type) -> Result<im::HashMap<String, Type>, String> {
if let Type::TyVar(x) = ty {
if var == x {
return Ok(im::HashMap::new());
}
}
if ftv_ty(ty).contains(var) {
return Err("Infinite Type".to_string());
}
Ok(im::HashMap::new().update(var.to_string(), ty.clone()))
}
fn type_pat(
env: &im::HashMap<String, Scheme>,
case_type: &Type,
pattern: &Pattern,
) -> Result<im::HashMap<String, Type>, String> {
// todo vars / wildcards, etc
let (_s, ty) = env.get(pattern.name).unwrap();
unify(case_type, ty)
}
/// Converts inner type of dataset
fn convert_inner(
env: &im::HashMap<String, Scheme>,
key: &str,
items: &[Expr],
) -> Result<(String, Type), String> {
let (_s, ty) = get_item_type(items, env)?;
Ok((key.to_string(), ty))
}
// Type inference using http://dev.stephendiehl.com/fun/006_hindley_milner.html#substitution
impl<'a> Expr<'_> {
pub fn get_type(&self, env: &im::HashMap<String, Scheme>) -> Result<TypeRes, String> {
match self {
Expr::Literal(l) => Ok((im::HashMap::new(), l.get_type())),
Expr::Ref(x) => {
let err = format!("Could not find reference {}", x);
let ty = env.get(*x).cloned().ok_or(err)?;
Ok((im::HashMap::new(), ty.1))
}
Expr::LetIn(x) => {
let (s1, t1) = x.expr1.expr.get_type(env)?;
let env1 = apply_sub_env(&s1, env);
let t2 = generalize(&env1, &t1);
let extended_ty = env.update(x.name.to_string(), t2);
let (s2, t2) = x.expr2.expr.get_type(&extended_ty)?;
Ok((compose(&s1, &s2), t2))
}
Expr::DataSet(items) => {
let d: im::HashMap<String, Type> = items
.iter()
.map(|(k, items)| convert_inner(env, k, items))
.flatten()
.collect();
if d.len() == items.len() {
Ok((im::HashMap::new(), Type::Dataset(d)))
} else {
Err("Not all rows matched in type".to_string())
}
}
Expr::Lambda(name, expr) => {
let type_var = Type::TyVar(get_id().to_string()); //fresh();
let env1 = env.update((*name).to_string(), (im::HashSet::new(), type_var.clone()));
let (sub, t1) = expr.expr.get_type(&env1)?;
let substituted = apply_sub_type(&sub, &type_var);
Ok((sub, Type::TyArr(Box::new(substituted), Box::new(t1))))
}
Expr::App(expr1, expr2) => {
let tv = Type::TyVar(get_id().to_string());
let (s1, t1) = expr1.get_type(env)?;
let (s2, t2) = expr2.get_type(&apply_sub_env(&s1, env))?;
let s3 = unify(
&apply_sub_type(&s2, &t1),
&Type::TyArr(Box::new(t2), Box::new(tv.clone())),
)?;
Ok((compose(&compose(&s3, &s2), &s1), apply_sub_type(&s3, &tv)))
}
Expr::Match(expr, exprs) => {
let (mut subs, case_type) = expr.get_type(env)?;
let mut branch_type = Type::TyVar(get_id().to_string());
for (p, branch) in exprs {
// TODO check, test
let pat_sub = type_pat(env, &case_type, p)?;
subs = compose(&subs, &pat_sub);
let (s, n_branch_type) = branch.get_type(env)?;
subs = compose(&subs, &s);
let cur_branch_type = apply_sub_type(&subs, &n_branch_type);
let s2 = unify(&branch_type, &cur_branch_type)?;
subs = compose(&subs, &s2);
branch_type = apply_sub_type(&subs, &branch_type);
}
Ok((subs, branch_type))
}
Expr::Projection(names, expr) => {
let from_ty = expr.get_type(env)?;
match from_ty {
(_s, Type::Dataset(items)) => {
if names
.iter()
.filter(|x| !items.contains_key(&x.to_string()))
.count()
> 0
{
// TODO; improve error
return Err("Not all fields in dataset".to_string());
}
Ok((
im::HashMap::new(),
Type::Dataset(
items
.iter()
.filter(|(k, _v)| names.contains(&&*k.to_string()))
.map(|(k, v)| (k.to_string(), v.clone()))
.collect(),
),
))
}
_ => Err("Expected dataset".to_string()),
}
}
x => Err(format!("not implemented {:?}", x)),
}
}
}
impl Literal {
fn get_type(&self) -> Type {
match self {
Literal::Int64(_) => Type::Int64,
Literal::Int32(_) => Type::Int32 | {
let xs = ftv_ty(ty);
let ys = ftv_env(env);
let a = xs.difference(ys);
(a, ty.clone())
} | identifier_body |
types.rs | = apply_sub_type(&subs, &ty);
}
Ok((im::HashMap::new(), ty))
}
type TypeRes<'a> = (im::HashMap<String, Type>, Type);
pub type Scheme = (im::HashSet<String>, Type);
type Subs<'a> = &'a im::HashMap<String, Type>;
fn apply_sub_type(subs: Subs, ty: &Type) -> Type {
match ty {
Type::TyVar(name) => subs.get(name).unwrap_or_else(|| &ty).clone(),
Type::TyArr(t1, t2) => Type::TyArr(
Box::new(apply_sub_type(subs, t1)),
Box::new(apply_sub_type(subs, t2)),
),
_ => ty.clone(),
}
}
fn apply_sub_scheme(subs: Subs, scheme: Scheme) -> Scheme {
let mut subs1 = subs.clone();
for key in scheme.0.iter() {
subs1 = subs1.without(key);
}
let ty = apply_sub_type(&subs1, &scheme.1);
(scheme.0, ty)
}
fn apply_sub_env(
subs: &im::HashMap<String, Type>,
env: &im::HashMap<String, Scheme>,
) -> im::HashMap<String, Scheme> {
let mut h = im::HashMap::new();
for (key, value) in env.into_iter() {
h = h.update(key.to_string(), apply_sub_scheme(subs, value.clone()));
}
h
}
fn compose(subs: Subs, subs2: Subs) -> im::HashMap<String, Type> {
let mut h = im::HashMap::new();
for (key, value) in subs.into_iter() {
h = h.update(key.to_string(), apply_sub_type(subs, &value.clone()));
}
h.union(subs2.clone())
}
fn ftv_ty(ty: &Type) -> im::HashSet<String> {
match ty {
Type::TyVar(a) => im::HashSet::unit(a.clone()),
Type::TyArr(ty1, ty2) => {
let x = ftv_ty(ty1);
let y = ftv_ty(ty2);
x.union(y)
}
_ => im::HashSet::new(),
}
}
fn ftv_env(env: &im::HashMap<String, Scheme>) -> im::HashSet<String> {
let ftvs = env.values().map(|x| ftv_ty(&x.1));
im::HashSet::unions(ftvs)
}
fn generalize(env: &im::HashMap<String, Scheme>, ty: &Type) -> Scheme {
let xs = ftv_ty(ty);
let ys = ftv_env(env);
let a = xs.difference(ys);
(a, ty.clone())
}
fn unify(ty1: &Type, ty2: &Type) -> Result<im::HashMap<String, Type>, String> {
match (ty1, ty2) {
(Type::TyArr(l, r), Type::TyArr(l1, r1)) => {
let s1 = unify(l, l1)?;
let s2 = unify(&apply_sub_type(&s1, &r), &apply_sub_type(&s1, &r1))?;
Ok(compose(&s2, &s1))
}
(Type::TyVar(a), t) => bind(&a, &t),
(t, Type::TyVar(a)) => bind(&a, &t),
(t1, t2) => {
if t1 == t2 {
Ok(im::HashMap::new())
} else {
Err("UnificationFail".to_string())
}
}
}
}
fn bind(var: &str, ty: &Type) -> Result<im::HashMap<String, Type>, String> {
if let Type::TyVar(x) = ty {
if var == x {
return Ok(im::HashMap::new());
}
}
if ftv_ty(ty).contains(var) {
return Err("Infinite Type".to_string());
}
Ok(im::HashMap::new().update(var.to_string(), ty.clone()))
}
fn type_pat(
env: &im::HashMap<String, Scheme>,
case_type: &Type,
pattern: &Pattern,
) -> Result<im::HashMap<String, Type>, String> {
// todo vars / wildcards, etc
let (_s, ty) = env.get(pattern.name).unwrap();
unify(case_type, ty)
}
/// Converts inner type of dataset
fn convert_inner(
env: &im::HashMap<String, Scheme>,
key: &str,
items: &[Expr],
) -> Result<(String, Type), String> {
let (_s, ty) = get_item_type(items, env)?;
Ok((key.to_string(), ty))
}
// Type inference using http://dev.stephendiehl.com/fun/006_hindley_milner.html#substitution
impl<'a> Expr<'_> {
pub fn get_type(&self, env: &im::HashMap<String, Scheme>) -> Result<TypeRes, String> {
match self {
Expr::Literal(l) => Ok((im::HashMap::new(), l.get_type())),
Expr::Ref(x) => {
let err = format!("Could not find reference {}", x);
let ty = env.get(*x).cloned().ok_or(err)?;
Ok((im::HashMap::new(), ty.1))
}
Expr::LetIn(x) => {
let (s1, t1) = x.expr1.expr.get_type(env)?;
let env1 = apply_sub_env(&s1, env);
let t2 = generalize(&env1, &t1);
let extended_ty = env.update(x.name.to_string(), t2);
let (s2, t2) = x.expr2.expr.get_type(&extended_ty)?;
Ok((compose(&s1, &s2), t2))
}
Expr::DataSet(items) => { | .iter()
.map(|(k, items)| convert_inner(env, k, items))
.flatten()
.collect();
if d.len() == items.len() {
Ok((im::HashMap::new(), Type::Dataset(d)))
} else {
Err("Not all rows matched in type".to_string())
}
}
Expr::Lambda(name, expr) => {
let type_var = Type::TyVar(get_id().to_string()); //fresh();
let env1 = env.update((*name).to_string(), (im::HashSet::new(), type_var.clone()));
let (sub, t1) = expr.expr.get_type(&env1)?;
let substituted = apply_sub_type(&sub, &type_var);
Ok((sub, Type::TyArr(Box::new(substituted), Box::new(t1))))
}
Expr::App(expr1, expr2) => {
let tv = Type::TyVar(get_id().to_string());
let (s1, t1) = expr1.get_type(env)?;
let (s2, t2) = expr2.get_type(&apply_sub_env(&s1, env))?;
let s3 = unify(
&apply_sub_type(&s2, &t1),
&Type::TyArr(Box::new(t2), Box::new(tv.clone())),
)?;
Ok((compose(&compose(&s3, &s2), &s1), apply_sub_type(&s3, &tv)))
}
Expr::Match(expr, exprs) => {
let (mut subs, case_type) = expr.get_type(env)?;
let mut branch_type = Type::TyVar(get_id().to_string());
for (p, branch) in exprs {
// TODO check, test
let pat_sub = type_pat(env, &case_type, p)?;
subs = compose(&subs, &pat_sub);
let (s, n_branch_type) = branch.get_type(env)?;
subs = compose(&subs, &s);
let cur_branch_type = apply_sub_type(&subs, &n_branch_type);
let s2 = unify(&branch_type, &cur_branch_type)?;
subs = compose(&subs, &s2);
branch_type = apply_sub_type(&subs, &branch_type);
}
Ok((subs, branch_type))
}
Expr::Projection(names, expr) => {
let from_ty = expr.get_type(env)?;
match from_ty {
(_s, Type::Dataset(items)) => {
if names
.iter()
.filter(|x| !items.contains_key(&x.to_string()))
.count()
> 0
{
// TODO; improve error
return Err("Not all fields in dataset".to_string());
}
Ok((
im::HashMap::new(),
Type::Dataset(
items
.iter()
.filter(|(k, _v)| names.contains(&&*k.to_string()))
.map(|(k, v)| (k.to_string(), v.clone()))
.collect(),
),
))
}
_ => Err("Expected dataset".to_string()),
}
}
x => Err(format!("not implemented {:?}", x)),
}
}
}
impl Literal {
fn get_type(&self) -> Type {
match self {
Literal::Int64(_) => Type::Int64,
Literal::Int32(_) => Type::Int32,
| let d: im::HashMap<String, Type> = items | random_line_split |
util.rs | simple contains for small str array
// benchmark shows linear scan takes at most 10ns
// while phf or bsearch takes 30ns
const ALLOWED_GLOBALS: &[&str] = make_list![
Infinity,
undefined,
NaN,
isFinite,
isNaN,
parseFloat,
parseInt,
decodeURI,
decodeURIComponent,
encodeURI,
encodeURIComponent,
Math,
Number,
Date,
Array,
Object,
Boolean,
String,
RegExp,
Map,
Set,
JSON,
Intl,
BigInt
];
pub fn is_global_allow_listed(s: &str) -> bool {
ALLOWED_GLOBALS.contains(&s)
}
// https://github.com/vuejs/rfcs/blob/master/active-rfcs/0008-render-function-api-change.md#special-reserved-props
const RESERVED: &[&str] = make_list![
key,
ref,
onVnodeMounted,
onVnodeUpdated,
onVnodeUnmounted,
onVnodeBeforeMount,
onVnodeBeforeUpdate,
onVnodeBeforeUnmount
];
#[inline]
pub fn is_reserved_prop(tag: &str) -> bool {
RESERVED.contains(&tag)
}
pub fn is_component_tag(tag: &str) -> bool {
tag == "component" || tag == "Component"
}
pub const fn yes(_: &str) -> bool {
true
}
pub const fn no(_: &str) -> bool {
false
}
pub fn get_vnode_call_helper(v: &VNodeIR<BaseConvertInfo>) -> RuntimeHelper {
use RuntimeHelper as RH;
if v.is_block {
return if v.is_component {
RH::CREATE_BLOCK
} else {
RH::CREATE_ELEMENT_BLOCK
};
}
if v.is_component {
RH::CREATE_VNODE
} else {
RH::CREATE_ELEMENT_VNODE
}
}
pub fn is_builtin_symbol(tag: &Js, helper: RuntimeHelper) -> bool {
if let Js::Symbol(r) = tag {
r == &helper
} else {
false
}
}
pub trait PropPattern {
fn matches(&self, name: &str) -> bool;
}
impl PropPattern for &str {
fn matches(&self, name: &str) -> bool {
name == *self
}
}
impl<F> PropPattern for F
where
F: Fn(&str) -> bool,
{
fn matches(&self, name: &str) -> bool {
self(name)
}
}
impl<const N: usize> PropPattern for [&'static str; N] {
fn matches(&self, name: &str) -> bool {
self.contains(&name)
}
}
type NameExp<'a> = Option<(&'a str, Option<VStr<'a>>)>;
pub trait PropMatcher<'a> {
fn get_name_and_exp(prop: &ElemProp<'a>) -> NameExp<'a>;
fn get_ref<'b>(prop: &'b ElemProp<'a>) -> &'b Self;
fn take(prop: ElemProp<'a>) -> Self;
fn is_match<P>(p: &ElemProp<'a>, pat: &P, allow_empty: bool) -> bool
where
P: PropPattern,
{
Self::get_name_and_exp(p).map_or(false, |(name, exp)| {
pat.matches(name) && (allow_empty || !exp.map_or(true, |v| v.is_empty()))
})
}
}
pub fn is_bind_key<'a>(arg: &Option<DirectiveArg<'a>>, name: &str) -> bool {
get_bind_key(arg).map_or(false, |v| v == name)
}
fn get_bind_key<'a>(arg: &Option<DirectiveArg<'a>>) -> Option<&'a str> {
if let DirectiveArg::Static(name) = arg.as_ref()? {
Some(name)
} else {
None
}
}
impl<'a> PropMatcher<'a> for ElemProp<'a> {
fn get_name_and_exp(prop: &ElemProp<'a>) -> NameExp<'a> {
match prop {
ElemProp::Attr(Attribute { name, value, .. }) => {
let exp = value.as_ref().map(|v| v.content);
Some((name, exp))
}
ElemProp::Dir(dir @ Directive { name: "bind", .. }) => {
let name = get_bind_key(&dir.argument)?;
let exp = dir.expression.as_ref().map(|v| v.content);
Some((name, exp))
}
_ => None,
}
}
fn get_ref<'b>(prop: &'b ElemProp<'a>) -> &'b Self {
prop
}
fn take(prop: ElemProp<'a>) -> Self {
prop
}
}
impl<'a> PropMatcher<'a> for Directive<'a> {
fn get_name_and_exp(prop: &ElemProp<'a>) -> NameExp<'a> {
if let ElemProp::Dir(Directive {
name, expression, ..
}) = prop
{
let exp = expression.as_ref().map(|v| v.content);
Some((name, exp))
} else {
None
}
}
fn get_ref<'b>(prop: &'b ElemProp<'a>) -> &'b Self {
if let ElemProp::Dir(dir) = prop {
return dir;
}
unreachable!("invalid call")
}
fn take(prop: ElemProp<'a>) -> Self {
if let ElemProp::Dir(dir) = prop {
return dir;
}
unreachable!("invalid call")
}
}
pub struct PropFound<'a, E, M>
where
E: Borrow<Element<'a>>,
M: PropMatcher<'a>,
{
elem: E,
pos: usize,
m: PhantomData<&'a M>,
}
impl<'a, E, M> PropFound<'a, E, M>
where
E: Borrow<Element<'a>>,
M: PropMatcher<'a>,
{
fn new(elem: E, pos: usize) -> Option<Self> {
Some(Self {
elem,
pos,
m: PhantomData,
})
}
pub fn get_ref(&self) -> &M {
M::get_ref(&self.elem.borrow().properties[self.pos])
}
}
// take is only available when access is mutable
impl<'a, E, M> PropFound<'a, E, M>
where
E: BorrowMut<Element<'a>>,
M: PropMatcher<'a>, | }
type DirFound<'a, E> = PropFound<'a, E, Directive<'a>>;
// sometimes mutable access to the element is not available so
// Borrow is used to refine PropFound so `take` is optional
pub fn dir_finder<'a, E, P>(elem: E, pat: P) -> PropFinder<'a, E, P, Directive<'a>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
{
PropFinder::new(elem, pat)
}
pub fn find_dir<'a, E, P>(elem: E, pat: P) -> Option<DirFound<'a, E>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
{
PropFinder::new(elem, pat).find()
}
pub fn find_dir_empty<'a, E, P>(elem: E, pat: P) -> Option<DirFound<'a, E>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
{
PropFinder::new(elem, pat).allow_empty().find()
}
pub struct PropFinder<'a, E, P, M = ElemProp<'a>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
M: PropMatcher<'a>,
{
elem: E,
pat: P,
allow_empty: bool,
filter: fn(&ElemProp<'a>) -> bool,
m: PhantomData<&'a M>,
}
impl<'a, E, P, M> PropFinder<'a, E, P, M>
where
E: Borrow<Element<'a>>,
P: PropPattern,
M: PropMatcher<'a>,
{
fn new(elem: E, pat: P) -> Self {
Self {
elem,
pat,
allow_empty: false,
filter: |_| true,
m: PhantomData,
}
}
fn is_match(&self, p: &ElemProp<'a>) -> bool {
M::is_match(p, &self.pat, self.allow_empty)
}
pub fn dynamic_only(self) -> Self {
Self {
filter: |p| matches!(p, ElemProp::Dir(..)),
..self
}
}
pub fn attr_only(self) -> Self {
Self {
filter: |p| matches!(p, ElemProp::Attr(..)),
..self
}
}
pub fn find(self) -> Option<PropFound<'a, E, M>> {
let pos = self
.elem
.borrow()
.properties
.iter()
.position(|p| self.is_match(p) && (self.filter)(p))?;
PropFound::new(self.elem, pos)
}
pub fn allow_empty(self) | {
pub fn take(mut self) -> M {
// TODO: avoid O(n) behavior
M::take(self.elem.borrow_mut().properties.remove(self.pos))
} | random_line_split |
util.rs | (tag: &str) -> bool {
tag == "component" || tag == "Component"
}
pub const fn yes(_: &str) -> bool {
true
}
pub const fn no(_: &str) -> bool {
false
}
pub fn get_vnode_call_helper(v: &VNodeIR<BaseConvertInfo>) -> RuntimeHelper {
use RuntimeHelper as RH;
if v.is_block {
return if v.is_component {
RH::CREATE_BLOCK
} else {
RH::CREATE_ELEMENT_BLOCK
};
}
if v.is_component {
RH::CREATE_VNODE
} else {
RH::CREATE_ELEMENT_VNODE
}
}
pub fn is_builtin_symbol(tag: &Js, helper: RuntimeHelper) -> bool {
if let Js::Symbol(r) = tag {
r == &helper
} else {
false
}
}
pub trait PropPattern {
fn matches(&self, name: &str) -> bool;
}
impl PropPattern for &str {
fn matches(&self, name: &str) -> bool {
name == *self
}
}
impl<F> PropPattern for F
where
F: Fn(&str) -> bool,
{
fn matches(&self, name: &str) -> bool {
self(name)
}
}
impl<const N: usize> PropPattern for [&'static str; N] {
fn matches(&self, name: &str) -> bool {
self.contains(&name)
}
}
type NameExp<'a> = Option<(&'a str, Option<VStr<'a>>)>;
pub trait PropMatcher<'a> {
fn get_name_and_exp(prop: &ElemProp<'a>) -> NameExp<'a>;
fn get_ref<'b>(prop: &'b ElemProp<'a>) -> &'b Self;
fn take(prop: ElemProp<'a>) -> Self;
fn is_match<P>(p: &ElemProp<'a>, pat: &P, allow_empty: bool) -> bool
where
P: PropPattern,
{
Self::get_name_and_exp(p).map_or(false, |(name, exp)| {
pat.matches(name) && (allow_empty || !exp.map_or(true, |v| v.is_empty()))
})
}
}
pub fn is_bind_key<'a>(arg: &Option<DirectiveArg<'a>>, name: &str) -> bool {
get_bind_key(arg).map_or(false, |v| v == name)
}
fn get_bind_key<'a>(arg: &Option<DirectiveArg<'a>>) -> Option<&'a str> {
if let DirectiveArg::Static(name) = arg.as_ref()? {
Some(name)
} else {
None
}
}
impl<'a> PropMatcher<'a> for ElemProp<'a> {
fn get_name_and_exp(prop: &ElemProp<'a>) -> NameExp<'a> {
match prop {
ElemProp::Attr(Attribute { name, value, .. }) => {
let exp = value.as_ref().map(|v| v.content);
Some((name, exp))
}
ElemProp::Dir(dir @ Directive { name: "bind", .. }) => {
let name = get_bind_key(&dir.argument)?;
let exp = dir.expression.as_ref().map(|v| v.content);
Some((name, exp))
}
_ => None,
}
}
fn get_ref<'b>(prop: &'b ElemProp<'a>) -> &'b Self {
prop
}
fn take(prop: ElemProp<'a>) -> Self {
prop
}
}
impl<'a> PropMatcher<'a> for Directive<'a> {
fn get_name_and_exp(prop: &ElemProp<'a>) -> NameExp<'a> {
if let ElemProp::Dir(Directive {
name, expression, ..
}) = prop
{
let exp = expression.as_ref().map(|v| v.content);
Some((name, exp))
} else {
None
}
}
fn get_ref<'b>(prop: &'b ElemProp<'a>) -> &'b Self {
if let ElemProp::Dir(dir) = prop {
return dir;
}
unreachable!("invalid call")
}
fn take(prop: ElemProp<'a>) -> Self {
if let ElemProp::Dir(dir) = prop {
return dir;
}
unreachable!("invalid call")
}
}
pub struct PropFound<'a, E, M>
where
E: Borrow<Element<'a>>,
M: PropMatcher<'a>,
{
elem: E,
pos: usize,
m: PhantomData<&'a M>,
}
impl<'a, E, M> PropFound<'a, E, M>
where
E: Borrow<Element<'a>>,
M: PropMatcher<'a>,
{
fn new(elem: E, pos: usize) -> Option<Self> {
Some(Self {
elem,
pos,
m: PhantomData,
})
}
pub fn get_ref(&self) -> &M {
M::get_ref(&self.elem.borrow().properties[self.pos])
}
}
// take is only available when access is mutable
impl<'a, E, M> PropFound<'a, E, M>
where
E: BorrowMut<Element<'a>>,
M: PropMatcher<'a>,
{
pub fn take(mut self) -> M {
// TODO: avoid O(n) behavior
M::take(self.elem.borrow_mut().properties.remove(self.pos))
}
}
type DirFound<'a, E> = PropFound<'a, E, Directive<'a>>;
// sometimes mutable access to the element is not available so
// Borrow is used to refine PropFound so `take` is optional
pub fn dir_finder<'a, E, P>(elem: E, pat: P) -> PropFinder<'a, E, P, Directive<'a>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
{
PropFinder::new(elem, pat)
}
pub fn find_dir<'a, E, P>(elem: E, pat: P) -> Option<DirFound<'a, E>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
{
PropFinder::new(elem, pat).find()
}
pub fn find_dir_empty<'a, E, P>(elem: E, pat: P) -> Option<DirFound<'a, E>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
{
PropFinder::new(elem, pat).allow_empty().find()
}
pub struct PropFinder<'a, E, P, M = ElemProp<'a>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
M: PropMatcher<'a>,
{
elem: E,
pat: P,
allow_empty: bool,
filter: fn(&ElemProp<'a>) -> bool,
m: PhantomData<&'a M>,
}
impl<'a, E, P, M> PropFinder<'a, E, P, M>
where
E: Borrow<Element<'a>>,
P: PropPattern,
M: PropMatcher<'a>,
{
fn new(elem: E, pat: P) -> Self {
Self {
elem,
pat,
allow_empty: false,
filter: |_| true,
m: PhantomData,
}
}
fn is_match(&self, p: &ElemProp<'a>) -> bool {
M::is_match(p, &self.pat, self.allow_empty)
}
pub fn dynamic_only(self) -> Self {
Self {
filter: |p| matches!(p, ElemProp::Dir(..)),
..self
}
}
pub fn attr_only(self) -> Self {
Self {
filter: |p| matches!(p, ElemProp::Attr(..)),
..self
}
}
pub fn find(self) -> Option<PropFound<'a, E, M>> {
let pos = self
.elem
.borrow()
.properties
.iter()
.position(|p| self.is_match(p) && (self.filter)(p))?;
PropFound::new(self.elem, pos)
}
pub fn allow_empty(self) -> Self {
Self {
allow_empty: true,
..self
}
}
}
impl<'a, P> PropFinder<'a, Element<'a>, P, ElemProp<'a>>
where
P: PropPattern + Copy,
{
pub fn find_all(self) -> impl Iterator<Item = Result<ElemProp<'a>, ElemProp<'a>>> {
let PropFinder {
elem,
pat,
allow_empty,
..
} = self;
elem.properties.into_iter().map(move |p| {
if ElemProp::is_match(&p, &pat, allow_empty) {
Ok(p)
} else {
Err(p)
}
})
}
}
pub fn find_prop<'a, E, P>(elem: E, pat: P) -> Option<PropFound<'a, E, ElemProp<'a>>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
{
PropFinder::new(elem, pat).find()
}
pub fn prop_finder<'a, E, P>(elem: E, pat: P) -> PropFinder<'a, E, P>
where
E: Borrow<Element<'a>>,
P: PropPattern,
| {
PropFinder::new(elem, pat)
} | identifier_body |
|
util.rs | simple contains for small str array
// benchmark shows linear scan takes at most 10ns
// while phf or bsearch takes 30ns
const ALLOWED_GLOBALS: &[&str] = make_list![
Infinity,
undefined,
NaN,
isFinite,
isNaN,
parseFloat,
parseInt,
decodeURI,
decodeURIComponent,
encodeURI,
encodeURIComponent,
Math,
Number,
Date,
Array,
Object,
Boolean,
String,
RegExp,
Map,
Set,
JSON,
Intl,
BigInt
];
pub fn is_global_allow_listed(s: &str) -> bool {
ALLOWED_GLOBALS.contains(&s)
}
// https://github.com/vuejs/rfcs/blob/master/active-rfcs/0008-render-function-api-change.md#special-reserved-props
const RESERVED: &[&str] = make_list![
key,
ref,
onVnodeMounted,
onVnodeUpdated,
onVnodeUnmounted,
onVnodeBeforeMount,
onVnodeBeforeUpdate,
onVnodeBeforeUnmount
];
#[inline]
pub fn is_reserved_prop(tag: &str) -> bool {
RESERVED.contains(&tag)
}
pub fn is_component_tag(tag: &str) -> bool {
tag == "component" || tag == "Component"
}
pub const fn yes(_: &str) -> bool {
true
}
pub const fn no(_: &str) -> bool {
false
}
pub fn get_vnode_call_helper(v: &VNodeIR<BaseConvertInfo>) -> RuntimeHelper {
use RuntimeHelper as RH;
if v.is_block {
return if v.is_component {
RH::CREATE_BLOCK
} else {
RH::CREATE_ELEMENT_BLOCK
};
}
if v.is_component {
RH::CREATE_VNODE
} else {
RH::CREATE_ELEMENT_VNODE
}
}
pub fn is_builtin_symbol(tag: &Js, helper: RuntimeHelper) -> bool {
if let Js::Symbol(r) = tag {
r == &helper
} else {
false
}
}
pub trait PropPattern {
fn matches(&self, name: &str) -> bool;
}
impl PropPattern for &str {
fn matches(&self, name: &str) -> bool {
name == *self
}
}
impl<F> PropPattern for F
where
F: Fn(&str) -> bool,
{
fn matches(&self, name: &str) -> bool {
self(name)
}
}
impl<const N: usize> PropPattern for [&'static str; N] {
fn matches(&self, name: &str) -> bool {
self.contains(&name)
}
}
type NameExp<'a> = Option<(&'a str, Option<VStr<'a>>)>;
pub trait PropMatcher<'a> {
fn get_name_and_exp(prop: &ElemProp<'a>) -> NameExp<'a>;
fn get_ref<'b>(prop: &'b ElemProp<'a>) -> &'b Self;
fn take(prop: ElemProp<'a>) -> Self;
fn is_match<P>(p: &ElemProp<'a>, pat: &P, allow_empty: bool) -> bool
where
P: PropPattern,
{
Self::get_name_and_exp(p).map_or(false, |(name, exp)| {
pat.matches(name) && (allow_empty || !exp.map_or(true, |v| v.is_empty()))
})
}
}
pub fn is_bind_key<'a>(arg: &Option<DirectiveArg<'a>>, name: &str) -> bool {
get_bind_key(arg).map_or(false, |v| v == name)
}
fn get_bind_key<'a>(arg: &Option<DirectiveArg<'a>>) -> Option<&'a str> {
if let DirectiveArg::Static(name) = arg.as_ref()? {
Some(name)
} else |
}
impl<'a> PropMatcher<'a> for ElemProp<'a> {
fn get_name_and_exp(prop: &ElemProp<'a>) -> NameExp<'a> {
match prop {
ElemProp::Attr(Attribute { name, value, .. }) => {
let exp = value.as_ref().map(|v| v.content);
Some((name, exp))
}
ElemProp::Dir(dir @ Directive { name: "bind", .. }) => {
let name = get_bind_key(&dir.argument)?;
let exp = dir.expression.as_ref().map(|v| v.content);
Some((name, exp))
}
_ => None,
}
}
fn get_ref<'b>(prop: &'b ElemProp<'a>) -> &'b Self {
prop
}
fn take(prop: ElemProp<'a>) -> Self {
prop
}
}
impl<'a> PropMatcher<'a> for Directive<'a> {
fn get_name_and_exp(prop: &ElemProp<'a>) -> NameExp<'a> {
if let ElemProp::Dir(Directive {
name, expression, ..
}) = prop
{
let exp = expression.as_ref().map(|v| v.content);
Some((name, exp))
} else {
None
}
}
fn get_ref<'b>(prop: &'b ElemProp<'a>) -> &'b Self {
if let ElemProp::Dir(dir) = prop {
return dir;
}
unreachable!("invalid call")
}
fn take(prop: ElemProp<'a>) -> Self {
if let ElemProp::Dir(dir) = prop {
return dir;
}
unreachable!("invalid call")
}
}
pub struct PropFound<'a, E, M>
where
E: Borrow<Element<'a>>,
M: PropMatcher<'a>,
{
elem: E,
pos: usize,
m: PhantomData<&'a M>,
}
impl<'a, E, M> PropFound<'a, E, M>
where
E: Borrow<Element<'a>>,
M: PropMatcher<'a>,
{
fn new(elem: E, pos: usize) -> Option<Self> {
Some(Self {
elem,
pos,
m: PhantomData,
})
}
pub fn get_ref(&self) -> &M {
M::get_ref(&self.elem.borrow().properties[self.pos])
}
}
// take is only available when access is mutable
impl<'a, E, M> PropFound<'a, E, M>
where
E: BorrowMut<Element<'a>>,
M: PropMatcher<'a>,
{
pub fn take(mut self) -> M {
// TODO: avoid O(n) behavior
M::take(self.elem.borrow_mut().properties.remove(self.pos))
}
}
type DirFound<'a, E> = PropFound<'a, E, Directive<'a>>;
// sometimes mutable access to the element is not available so
// Borrow is used to refine PropFound so `take` is optional
pub fn dir_finder<'a, E, P>(elem: E, pat: P) -> PropFinder<'a, E, P, Directive<'a>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
{
PropFinder::new(elem, pat)
}
pub fn find_dir<'a, E, P>(elem: E, pat: P) -> Option<DirFound<'a, E>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
{
PropFinder::new(elem, pat).find()
}
pub fn find_dir_empty<'a, E, P>(elem: E, pat: P) -> Option<DirFound<'a, E>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
{
PropFinder::new(elem, pat).allow_empty().find()
}
pub struct PropFinder<'a, E, P, M = ElemProp<'a>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
M: PropMatcher<'a>,
{
elem: E,
pat: P,
allow_empty: bool,
filter: fn(&ElemProp<'a>) -> bool,
m: PhantomData<&'a M>,
}
impl<'a, E, P, M> PropFinder<'a, E, P, M>
where
E: Borrow<Element<'a>>,
P: PropPattern,
M: PropMatcher<'a>,
{
fn new(elem: E, pat: P) -> Self {
Self {
elem,
pat,
allow_empty: false,
filter: |_| true,
m: PhantomData,
}
}
fn is_match(&self, p: &ElemProp<'a>) -> bool {
M::is_match(p, &self.pat, self.allow_empty)
}
pub fn dynamic_only(self) -> Self {
Self {
filter: |p| matches!(p, ElemProp::Dir(..)),
..self
}
}
pub fn attr_only(self) -> Self {
Self {
filter: |p| matches!(p, ElemProp::Attr(..)),
..self
}
}
pub fn find(self) -> Option<PropFound<'a, E, M>> {
let pos = self
.elem
.borrow()
.properties
.iter()
.position(|p| self.is_match(p) && (self.filter)(p))?;
PropFound::new(self.elem, pos)
}
pub fn allow_empty | {
None
} | conditional_block |
util.rs | simple contains for small str array
// benchmark shows linear scan takes at most 10ns
// while phf or bsearch takes 30ns
const ALLOWED_GLOBALS: &[&str] = make_list![
Infinity,
undefined,
NaN,
isFinite,
isNaN,
parseFloat,
parseInt,
decodeURI,
decodeURIComponent,
encodeURI,
encodeURIComponent,
Math,
Number,
Date,
Array,
Object,
Boolean,
String,
RegExp,
Map,
Set,
JSON,
Intl,
BigInt
];
pub fn is_global_allow_listed(s: &str) -> bool {
ALLOWED_GLOBALS.contains(&s)
}
// https://github.com/vuejs/rfcs/blob/master/active-rfcs/0008-render-function-api-change.md#special-reserved-props
const RESERVED: &[&str] = make_list![
key,
ref,
onVnodeMounted,
onVnodeUpdated,
onVnodeUnmounted,
onVnodeBeforeMount,
onVnodeBeforeUpdate,
onVnodeBeforeUnmount
];
#[inline]
pub fn is_reserved_prop(tag: &str) -> bool {
RESERVED.contains(&tag)
}
pub fn is_component_tag(tag: &str) -> bool {
tag == "component" || tag == "Component"
}
pub const fn yes(_: &str) -> bool {
true
}
pub const fn no(_: &str) -> bool {
false
}
pub fn get_vnode_call_helper(v: &VNodeIR<BaseConvertInfo>) -> RuntimeHelper {
use RuntimeHelper as RH;
if v.is_block {
return if v.is_component {
RH::CREATE_BLOCK
} else {
RH::CREATE_ELEMENT_BLOCK
};
}
if v.is_component {
RH::CREATE_VNODE
} else {
RH::CREATE_ELEMENT_VNODE
}
}
pub fn is_builtin_symbol(tag: &Js, helper: RuntimeHelper) -> bool {
if let Js::Symbol(r) = tag {
r == &helper
} else {
false
}
}
pub trait PropPattern {
fn matches(&self, name: &str) -> bool;
}
impl PropPattern for &str {
fn matches(&self, name: &str) -> bool {
name == *self
}
}
impl<F> PropPattern for F
where
F: Fn(&str) -> bool,
{
fn matches(&self, name: &str) -> bool {
self(name)
}
}
impl<const N: usize> PropPattern for [&'static str; N] {
fn matches(&self, name: &str) -> bool {
self.contains(&name)
}
}
type NameExp<'a> = Option<(&'a str, Option<VStr<'a>>)>;
pub trait PropMatcher<'a> {
fn get_name_and_exp(prop: &ElemProp<'a>) -> NameExp<'a>;
fn get_ref<'b>(prop: &'b ElemProp<'a>) -> &'b Self;
fn take(prop: ElemProp<'a>) -> Self;
fn is_match<P>(p: &ElemProp<'a>, pat: &P, allow_empty: bool) -> bool
where
P: PropPattern,
{
Self::get_name_and_exp(p).map_or(false, |(name, exp)| {
pat.matches(name) && (allow_empty || !exp.map_or(true, |v| v.is_empty()))
})
}
}
pub fn is_bind_key<'a>(arg: &Option<DirectiveArg<'a>>, name: &str) -> bool {
get_bind_key(arg).map_or(false, |v| v == name)
}
fn get_bind_key<'a>(arg: &Option<DirectiveArg<'a>>) -> Option<&'a str> {
if let DirectiveArg::Static(name) = arg.as_ref()? {
Some(name)
} else {
None
}
}
impl<'a> PropMatcher<'a> for ElemProp<'a> {
fn get_name_and_exp(prop: &ElemProp<'a>) -> NameExp<'a> {
match prop {
ElemProp::Attr(Attribute { name, value, .. }) => {
let exp = value.as_ref().map(|v| v.content);
Some((name, exp))
}
ElemProp::Dir(dir @ Directive { name: "bind", .. }) => {
let name = get_bind_key(&dir.argument)?;
let exp = dir.expression.as_ref().map(|v| v.content);
Some((name, exp))
}
_ => None,
}
}
fn get_ref<'b>(prop: &'b ElemProp<'a>) -> &'b Self {
prop
}
fn take(prop: ElemProp<'a>) -> Self {
prop
}
}
impl<'a> PropMatcher<'a> for Directive<'a> {
fn get_name_and_exp(prop: &ElemProp<'a>) -> NameExp<'a> {
if let ElemProp::Dir(Directive {
name, expression, ..
}) = prop
{
let exp = expression.as_ref().map(|v| v.content);
Some((name, exp))
} else {
None
}
}
fn get_ref<'b>(prop: &'b ElemProp<'a>) -> &'b Self {
if let ElemProp::Dir(dir) = prop {
return dir;
}
unreachable!("invalid call")
}
fn take(prop: ElemProp<'a>) -> Self {
if let ElemProp::Dir(dir) = prop {
return dir;
}
unreachable!("invalid call")
}
}
pub struct | <'a, E, M>
where
E: Borrow<Element<'a>>,
M: PropMatcher<'a>,
{
elem: E,
pos: usize,
m: PhantomData<&'a M>,
}
impl<'a, E, M> PropFound<'a, E, M>
where
E: Borrow<Element<'a>>,
M: PropMatcher<'a>,
{
fn new(elem: E, pos: usize) -> Option<Self> {
Some(Self {
elem,
pos,
m: PhantomData,
})
}
pub fn get_ref(&self) -> &M {
M::get_ref(&self.elem.borrow().properties[self.pos])
}
}
// take is only available when access is mutable
impl<'a, E, M> PropFound<'a, E, M>
where
E: BorrowMut<Element<'a>>,
M: PropMatcher<'a>,
{
pub fn take(mut self) -> M {
// TODO: avoid O(n) behavior
M::take(self.elem.borrow_mut().properties.remove(self.pos))
}
}
type DirFound<'a, E> = PropFound<'a, E, Directive<'a>>;
// sometimes mutable access to the element is not available so
// Borrow is used to refine PropFound so `take` is optional
pub fn dir_finder<'a, E, P>(elem: E, pat: P) -> PropFinder<'a, E, P, Directive<'a>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
{
PropFinder::new(elem, pat)
}
pub fn find_dir<'a, E, P>(elem: E, pat: P) -> Option<DirFound<'a, E>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
{
PropFinder::new(elem, pat).find()
}
pub fn find_dir_empty<'a, E, P>(elem: E, pat: P) -> Option<DirFound<'a, E>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
{
PropFinder::new(elem, pat).allow_empty().find()
}
pub struct PropFinder<'a, E, P, M = ElemProp<'a>>
where
E: Borrow<Element<'a>>,
P: PropPattern,
M: PropMatcher<'a>,
{
elem: E,
pat: P,
allow_empty: bool,
filter: fn(&ElemProp<'a>) -> bool,
m: PhantomData<&'a M>,
}
impl<'a, E, P, M> PropFinder<'a, E, P, M>
where
E: Borrow<Element<'a>>,
P: PropPattern,
M: PropMatcher<'a>,
{
fn new(elem: E, pat: P) -> Self {
Self {
elem,
pat,
allow_empty: false,
filter: |_| true,
m: PhantomData,
}
}
fn is_match(&self, p: &ElemProp<'a>) -> bool {
M::is_match(p, &self.pat, self.allow_empty)
}
pub fn dynamic_only(self) -> Self {
Self {
filter: |p| matches!(p, ElemProp::Dir(..)),
..self
}
}
pub fn attr_only(self) -> Self {
Self {
filter: |p| matches!(p, ElemProp::Attr(..)),
..self
}
}
pub fn find(self) -> Option<PropFound<'a, E, M>> {
let pos = self
.elem
.borrow()
.properties
.iter()
.position(|p| self.is_match(p) && (self.filter)(p))?;
PropFound::new(self.elem, pos)
}
pub fn allow_empty(self | PropFound | identifier_name |
lib.rs | windres.exe and ar.exe.
/// This could be something like: "C:\Program Files\mingw-w64\x86_64-5.3.0-win32-seh-rt_v4-rev0\mingw64\bin"
///
/// For MSVC the Windows SDK has to be installed. It comes with the resource compiler rc.exe.
/// This should be set to the root directory of the Windows SDK, e.g., "C:\Program Files (x86)\Windows Kits\10" or,
/// if multiple 10 versions are installed, set it directly to the correct bin directory "C:\Program Files (x86)\Windows Kits\10\bin\10.0.14393.0\x64"
///
/// If it is left unset, it will look up a path in the registry, i.e. HKLM\SOFTWARE\Microsoft\Windows Kits\Installed Roots
sdk_dir: Option<PathBuf>,
/// A string containing an [application manifest] to be included with the application on Windows.
///
/// Defaults to:
/// ```ignore
#[doc = include_str!("window-app-manifest.xml")]
/// ```
///
/// [application manifest]: https://learn.microsoft.com/en-us/windows/win32/sbscs/application-manifests
app_manifest: Option<String>,
}
impl WindowsAttributes {
/// Creates the default attribute set.
pub fn new() -> Self {
Self::default()
}
/// Sets the icon to use on the window. Currently only used on Windows.
/// It must be in `ico` format. Defaults to `icons/icon.ico`.
#[must_use]
pub fn window_icon_path<P: AsRef<Path>>(mut self, window_icon_path: P) -> Self {
self
.window_icon_path
.replace(window_icon_path.as_ref().into());
self
}
/// Sets the sdk dir for windows. Currently only used on Windows. This must be a valid UTF-8
/// path. Defaults to whatever the `winres` crate determines is best.
#[must_use]
pub fn sdk_dir<P: AsRef<Path>>(mut self, sdk_dir: P) -> Self {
self.sdk_dir = Some(sdk_dir.as_ref().into());
self
}
/// Sets the Windows app [manifest].
///
/// # Example
///
/// The following manifest will brand the exe as requesting administrator privileges.
/// Thus, everytime it is executed, a Windows UAC dialog will appear.
///
/// Note that you can move the manifest contents to a separate file and use `include_str!("manifest.xml")`
/// instead of the inline string.
///
/// ```rust,no_run
/// let mut windows = tauri_build::WindowsAttributes::new();
/// windows = windows.app_manifest(r#"
/// <assembly xmlns="urn:schemas-microsoft-com:asm.v1" manifestVersion="1.0">
/// <trustInfo xmlns="urn:schemas-microsoft-com:asm.v3">
/// <security>
/// <requestedPrivileges>
/// <requestedExecutionLevel level="requireAdministrator" uiAccess="false" />
/// </requestedPrivileges>
/// </security>
/// </trustInfo>
/// </assembly>
/// "#);
/// tauri_build::try_build(
/// tauri_build::Attributes::new().windows_attributes(windows)
/// ).expect("failed to run build script");
/// ```
///
/// Defaults to:
/// ```ignore
#[doc = include_str!("window-app-manifest.xml")]
/// [manifest]: https://learn.microsoft.com/en-us/windows/win32/sbscs/application-manifests
/// ```
#[must_use]
pub fn app_manifest<S: AsRef<str>>(mut self, manifest: S) -> Self {
self.app_manifest = Some(manifest.as_ref().to_string());
self
}
}
/// The attributes used on the build.
#[derive(Debug, Default)]
pub struct Attributes {
#[allow(dead_code)]
windows_attributes: WindowsAttributes,
}
impl Attributes {
/// Creates the default attribute set.
pub fn new() -> Self |
/// Sets the icon to use on the window. Currently only used on Windows.
#[must_use]
pub fn windows_attributes(mut self, windows_attributes: WindowsAttributes) -> Self {
self.windows_attributes = windows_attributes;
self
}
}
/// Run all build time helpers for your Tauri Application.
///
/// The current helpers include the following:
/// * Generates a Windows Resource file when targeting Windows.
///
/// # Platforms
///
/// [`build()`] should be called inside of `build.rs` regardless of the platform:
/// * New helpers may target more platforms in the future.
/// * Platform specific code is handled by the helpers automatically.
/// * A build script is required in order to activate some cargo environmental variables that are
/// used when generating code and embedding assets - so [`build()`] may as well be called.
///
/// In short, this is saying don't put the call to [`build()`] behind a `#[cfg(windows)]`.
///
/// # Panics
///
/// If any of the build time helpers fail, they will [`std::panic!`] with the related error message.
/// This is typically desirable when running inside a build script; see [`try_build`] for no panics.
pub fn build() {
if let Err(error) = try_build(Attributes::default()) {
let error = format!("{error:#}");
println!("{error}");
if error.starts_with("unknown field") {
print!("found an unknown configuration field. This usually means that you are using a CLI version that is newer than `tauri-build` and is incompatible. ");
println!(
"Please try updating the Rust crates by running `cargo update` in the Tauri app folder."
);
}
std::process::exit(1);
}
}
/// Non-panicking [`build()`].
#[allow(unused_variables)]
pub fn try_build(attributes: Attributes) -> Result<()> {
use anyhow::anyhow;
println!("cargo:rerun-if-env-changed=TAURI_CONFIG");
println!("cargo:rerun-if-changed=tauri.conf.json");
#[cfg(feature = "config-json5")]
println!("cargo:rerun-if-changed=tauri.conf.json5");
#[cfg(feature = "config-toml")]
println!("cargo:rerun-if-changed=Tauri.toml");
let target_os = std::env::var("CARGO_CFG_TARGET_OS").unwrap();
let mobile = target_os == "ios" || target_os == "android";
cfg_alias("desktop", !mobile);
cfg_alias("mobile", mobile);
let mut config = serde_json::from_value(tauri_utils::config::parse::read_from(
std::env::current_dir().unwrap(),
)?)?;
if let Ok(env) = std::env::var("TAURI_CONFIG") {
let merge_config: serde_json::Value = serde_json::from_str(&env)?;
json_patch::merge(&mut config, &merge_config);
}
let config: Config = serde_json::from_value(config)?;
cfg_alias("dev", !has_feature("custom-protocol"));
let ws_path = get_workspace_dir()?;
let mut manifest =
Manifest::<cargo_toml::Value>::from_slice_with_metadata(&std::fs::read("Cargo.toml")?)?;
if let Ok(ws_manifest) = Manifest::from_path(ws_path.join("Cargo.toml")) {
Manifest::complete_from_path_and_workspace(
&mut manifest,
Path::new("Cargo.toml"),
Some((&ws_manifest, ws_path.as_path())),
)?;
} else {
Manifest::complete_from_path(&mut manifest, Path::new("Cargo.toml"))?;
}
if let Some(tauri_build) = manifest.build_dependencies.remove("tauri-build") {
let error_message = check_features(&config, tauri_build, true);
if !error_message.is_empty() {
return Err(anyhow!("
The `tauri-build` dependency features on the `Cargo.toml` file does not match the allowlist defined under `tauri.conf.json`.
Please run `tauri dev` or `tauri build` or {}.
", error_message));
}
}
if let Some(tauri) = manifest.dependencies.remove("tauri") {
let error_message = check_features(&config, tauri, false);
if !error_message.is_empty() {
return Err(anyhow!("
The `tauri` dependency features on the `Cargo.toml` file does not match the allowlist defined under `tauri.conf.json`.
Please run `tauri dev` or `tauri build` or {}.
", error_message));
}
}
let target_triple = std::env::var("TARGET").unwrap();
let out_dir = PathBuf::from(std::env::var("OUT_DIR").unwrap());
// TODO: far from ideal, but there's no other way to get the target dir, see <https://github.com/rust-lang/cargo/issues/5457>
let target_dir = out_dir
.parent()
.unwrap()
.parent()
.unwrap()
.parent()
.unwrap();
if let Some(paths) = &config.tauri.bundle.external_bin {
copy_binaries(
| {
Self::default()
} | identifier_body |
lib.rs | put windres.exe and ar.exe.
/// This could be something like: "C:\Program Files\mingw-w64\x86_64-5.3.0-win32-seh-rt_v4-rev0\mingw64\bin"
///
/// For MSVC the Windows SDK has to be installed. It comes with the resource compiler rc.exe.
/// This should be set to the root directory of the Windows SDK, e.g., "C:\Program Files (x86)\Windows Kits\10" or,
/// if multiple 10 versions are installed, set it directly to the correct bin directory "C:\Program Files (x86)\Windows Kits\10\bin\10.0.14393.0\x64"
///
/// If it is left unset, it will look up a path in the registry, i.e. HKLM\SOFTWARE\Microsoft\Windows Kits\Installed Roots
sdk_dir: Option<PathBuf>,
/// A string containing an [application manifest] to be included with the application on Windows.
///
/// Defaults to:
/// ```ignore
#[doc = include_str!("window-app-manifest.xml")]
/// ```
///
/// [application manifest]: https://learn.microsoft.com/en-us/windows/win32/sbscs/application-manifests
app_manifest: Option<String>,
}
impl WindowsAttributes {
/// Creates the default attribute set.
pub fn new() -> Self {
Self::default()
}
/// Sets the icon to use on the window. Currently only used on Windows.
/// It must be in `ico` format. Defaults to `icons/icon.ico`.
#[must_use]
pub fn window_icon_path<P: AsRef<Path>>(mut self, window_icon_path: P) -> Self {
self
.window_icon_path
.replace(window_icon_path.as_ref().into());
self
}
/// Sets the sdk dir for windows. Currently only used on Windows. This must be a valid UTF-8
/// path. Defaults to whatever the `winres` crate determines is best.
#[must_use]
pub fn sdk_dir<P: AsRef<Path>>(mut self, sdk_dir: P) -> Self {
self.sdk_dir = Some(sdk_dir.as_ref().into());
self
}
/// Sets the Windows app [manifest].
///
/// # Example
///
/// The following manifest will brand the exe as requesting administrator privileges.
/// Thus, everytime it is executed, a Windows UAC dialog will appear.
///
/// Note that you can move the manifest contents to a separate file and use `include_str!("manifest.xml")`
/// instead of the inline string.
///
/// ```rust,no_run
/// let mut windows = tauri_build::WindowsAttributes::new();
/// windows = windows.app_manifest(r#"
/// <assembly xmlns="urn:schemas-microsoft-com:asm.v1" manifestVersion="1.0">
/// <trustInfo xmlns="urn:schemas-microsoft-com:asm.v3">
/// <security>
/// <requestedPrivileges>
/// <requestedExecutionLevel level="requireAdministrator" uiAccess="false" />
/// </requestedPrivileges>
/// </security>
/// </trustInfo>
/// </assembly>
/// "#);
/// tauri_build::try_build(
/// tauri_build::Attributes::new().windows_attributes(windows)
/// ).expect("failed to run build script");
/// ```
///
/// Defaults to:
/// ```ignore
#[doc = include_str!("window-app-manifest.xml")]
/// [manifest]: https://learn.microsoft.com/en-us/windows/win32/sbscs/application-manifests
/// ```
#[must_use]
pub fn app_manifest<S: AsRef<str>>(mut self, manifest: S) -> Self {
self.app_manifest = Some(manifest.as_ref().to_string());
self
}
}
/// The attributes used on the build.
#[derive(Debug, Default)]
pub struct Attributes {
#[allow(dead_code)]
windows_attributes: WindowsAttributes,
}
impl Attributes {
/// Creates the default attribute set.
pub fn new() -> Self {
Self::default()
}
/// Sets the icon to use on the window. Currently only used on Windows.
#[must_use]
pub fn | (mut self, windows_attributes: WindowsAttributes) -> Self {
self.windows_attributes = windows_attributes;
self
}
}
/// Run all build time helpers for your Tauri Application.
///
/// The current helpers include the following:
/// * Generates a Windows Resource file when targeting Windows.
///
/// # Platforms
///
/// [`build()`] should be called inside of `build.rs` regardless of the platform:
/// * New helpers may target more platforms in the future.
/// * Platform specific code is handled by the helpers automatically.
/// * A build script is required in order to activate some cargo environmental variables that are
/// used when generating code and embedding assets - so [`build()`] may as well be called.
///
/// In short, this is saying don't put the call to [`build()`] behind a `#[cfg(windows)]`.
///
/// # Panics
///
/// If any of the build time helpers fail, they will [`std::panic!`] with the related error message.
/// This is typically desirable when running inside a build script; see [`try_build`] for no panics.
pub fn build() {
if let Err(error) = try_build(Attributes::default()) {
let error = format!("{error:#}");
println!("{error}");
if error.starts_with("unknown field") {
print!("found an unknown configuration field. This usually means that you are using a CLI version that is newer than `tauri-build` and is incompatible. ");
println!(
"Please try updating the Rust crates by running `cargo update` in the Tauri app folder."
);
}
std::process::exit(1);
}
}
/// Non-panicking [`build()`].
#[allow(unused_variables)]
pub fn try_build(attributes: Attributes) -> Result<()> {
use anyhow::anyhow;
println!("cargo:rerun-if-env-changed=TAURI_CONFIG");
println!("cargo:rerun-if-changed=tauri.conf.json");
#[cfg(feature = "config-json5")]
println!("cargo:rerun-if-changed=tauri.conf.json5");
#[cfg(feature = "config-toml")]
println!("cargo:rerun-if-changed=Tauri.toml");
let target_os = std::env::var("CARGO_CFG_TARGET_OS").unwrap();
let mobile = target_os == "ios" || target_os == "android";
cfg_alias("desktop", !mobile);
cfg_alias("mobile", mobile);
let mut config = serde_json::from_value(tauri_utils::config::parse::read_from(
std::env::current_dir().unwrap(),
)?)?;
if let Ok(env) = std::env::var("TAURI_CONFIG") {
let merge_config: serde_json::Value = serde_json::from_str(&env)?;
json_patch::merge(&mut config, &merge_config);
}
let config: Config = serde_json::from_value(config)?;
cfg_alias("dev", !has_feature("custom-protocol"));
let ws_path = get_workspace_dir()?;
let mut manifest =
Manifest::<cargo_toml::Value>::from_slice_with_metadata(&std::fs::read("Cargo.toml")?)?;
if let Ok(ws_manifest) = Manifest::from_path(ws_path.join("Cargo.toml")) {
Manifest::complete_from_path_and_workspace(
&mut manifest,
Path::new("Cargo.toml"),
Some((&ws_manifest, ws_path.as_path())),
)?;
} else {
Manifest::complete_from_path(&mut manifest, Path::new("Cargo.toml"))?;
}
if let Some(tauri_build) = manifest.build_dependencies.remove("tauri-build") {
let error_message = check_features(&config, tauri_build, true);
if !error_message.is_empty() {
return Err(anyhow!("
The `tauri-build` dependency features on the `Cargo.toml` file does not match the allowlist defined under `tauri.conf.json`.
Please run `tauri dev` or `tauri build` or {}.
", error_message));
}
}
if let Some(tauri) = manifest.dependencies.remove("tauri") {
let error_message = check_features(&config, tauri, false);
if !error_message.is_empty() {
return Err(anyhow!("
The `tauri` dependency features on the `Cargo.toml` file does not match the allowlist defined under `tauri.conf.json`.
Please run `tauri dev` or `tauri build` or {}.
", error_message));
}
}
let target_triple = std::env::var("TARGET").unwrap();
let out_dir = PathBuf::from(std::env::var("OUT_DIR").unwrap());
// TODO: far from ideal, but there's no other way to get the target dir, see <https://github.com/rust-lang/cargo/issues/5457>
let target_dir = out_dir
.parent()
.unwrap()
.parent()
.unwrap()
.parent()
.unwrap();
if let Some(paths) = &config.tauri.bundle.external_bin {
copy_binaries(
| windows_attributes | identifier_name |
lib.rs | -json5")]
println!("cargo:rerun-if-changed=tauri.conf.json5");
#[cfg(feature = "config-toml")]
println!("cargo:rerun-if-changed=Tauri.toml");
let target_os = std::env::var("CARGO_CFG_TARGET_OS").unwrap();
let mobile = target_os == "ios" || target_os == "android";
cfg_alias("desktop", !mobile);
cfg_alias("mobile", mobile);
let mut config = serde_json::from_value(tauri_utils::config::parse::read_from(
std::env::current_dir().unwrap(),
)?)?;
if let Ok(env) = std::env::var("TAURI_CONFIG") {
let merge_config: serde_json::Value = serde_json::from_str(&env)?;
json_patch::merge(&mut config, &merge_config);
}
let config: Config = serde_json::from_value(config)?;
cfg_alias("dev", !has_feature("custom-protocol"));
let ws_path = get_workspace_dir()?;
let mut manifest =
Manifest::<cargo_toml::Value>::from_slice_with_metadata(&std::fs::read("Cargo.toml")?)?;
if let Ok(ws_manifest) = Manifest::from_path(ws_path.join("Cargo.toml")) {
Manifest::complete_from_path_and_workspace(
&mut manifest,
Path::new("Cargo.toml"),
Some((&ws_manifest, ws_path.as_path())),
)?;
} else {
Manifest::complete_from_path(&mut manifest, Path::new("Cargo.toml"))?;
}
if let Some(tauri_build) = manifest.build_dependencies.remove("tauri-build") {
let error_message = check_features(&config, tauri_build, true);
if !error_message.is_empty() {
return Err(anyhow!("
The `tauri-build` dependency features on the `Cargo.toml` file does not match the allowlist defined under `tauri.conf.json`.
Please run `tauri dev` or `tauri build` or {}.
", error_message));
}
}
if let Some(tauri) = manifest.dependencies.remove("tauri") {
let error_message = check_features(&config, tauri, false);
if !error_message.is_empty() {
return Err(anyhow!("
The `tauri` dependency features on the `Cargo.toml` file does not match the allowlist defined under `tauri.conf.json`.
Please run `tauri dev` or `tauri build` or {}.
", error_message));
}
}
let target_triple = std::env::var("TARGET").unwrap();
let out_dir = PathBuf::from(std::env::var("OUT_DIR").unwrap());
// TODO: far from ideal, but there's no other way to get the target dir, see <https://github.com/rust-lang/cargo/issues/5457>
let target_dir = out_dir
.parent()
.unwrap()
.parent()
.unwrap()
.parent()
.unwrap();
if let Some(paths) = &config.tauri.bundle.external_bin {
copy_binaries(
ResourcePaths::new(external_binaries(paths, &target_triple).as_slice(), true),
&target_triple,
target_dir,
manifest.package.as_ref().map(|p| &p.name),
)?;
}
#[allow(unused_mut, clippy::redundant_clone)]
let mut resources = config.tauri.bundle.resources.clone().unwrap_or_default();
if target_triple.contains("windows") {
if let Some(fixed_webview2_runtime_path) =
&config.tauri.bundle.windows.webview_fixed_runtime_path
{
resources.push(fixed_webview2_runtime_path.display().to_string());
}
}
copy_resources(ResourcePaths::new(resources.as_slice(), true), target_dir)?;
if target_triple.contains("darwin") {
if let Some(version) = &config.tauri.bundle.macos.minimum_system_version {
println!("cargo:rustc-env=MACOSX_DEPLOYMENT_TARGET={}", version);
}
}
if target_triple.contains("windows") {
use anyhow::Context;
use semver::Version;
use tauri_winres::{VersionInfo, WindowsResource};
fn find_icon<F: Fn(&&String) -> bool>(config: &Config, predicate: F, default: &str) -> PathBuf {
let icon_path = config
.tauri
.bundle
.icon
.iter()
.find(|i| predicate(i))
.cloned()
.unwrap_or_else(|| default.to_string());
icon_path.into()
}
let window_icon_path = attributes
.windows_attributes
.window_icon_path
.unwrap_or_else(|| find_icon(&config, |i| i.ends_with(".ico"), "icons/icon.ico"));
if window_icon_path.exists() {
let mut res = WindowsResource::new();
if let Some(manifest) = attributes.windows_attributes.app_manifest {
res.set_manifest(&manifest);
} else {
res.set_manifest(include_str!("window-app-manifest.xml"));
}
if let Some(sdk_dir) = &attributes.windows_attributes.sdk_dir {
if let Some(sdk_dir_str) = sdk_dir.to_str() {
res.set_toolkit_path(sdk_dir_str);
} else {
return Err(anyhow!(
"sdk_dir path is not valid; only UTF-8 characters are allowed"
));
}
}
if let Some(version) = &config.package.version {
if let Ok(v) = Version::parse(version) {
let version = v.major << 48 | v.minor << 32 | v.patch << 16;
res.set_version_info(VersionInfo::FILEVERSION, version);
res.set_version_info(VersionInfo::PRODUCTVERSION, version);
}
res.set("FileVersion", version);
res.set("ProductVersion", version);
}
if let Some(product_name) = &config.package.product_name {
res.set("ProductName", product_name);
res.set("FileDescription", product_name);
}
res.set_icon_with_id(&window_icon_path.display().to_string(), "32512");
res.compile().with_context(|| {
format!(
"failed to compile `{}` into a Windows Resource file during tauri-build",
window_icon_path.display()
)
})?;
} else {
return Err(anyhow!(format!(
"`{}` not found; required for generating a Windows Resource file during tauri-build",
window_icon_path.display()
)));
}
let target_env = std::env::var("CARGO_CFG_TARGET_ENV").unwrap();
match target_env.as_str() {
"gnu" => {
let target_arch = match std::env::var("CARGO_CFG_TARGET_ARCH").unwrap().as_str() {
"x86_64" => Some("x64"),
"x86" => Some("x86"),
"aarch64" => Some("arm64"),
arch => None,
};
if let Some(target_arch) = target_arch {
for entry in std::fs::read_dir(target_dir.join("build"))? {
let path = entry?.path();
let webview2_loader_path = path
.join("out")
.join(target_arch)
.join("WebView2Loader.dll");
if path.to_string_lossy().contains("webview2-com-sys") && webview2_loader_path.exists()
{
std::fs::copy(webview2_loader_path, target_dir.join("WebView2Loader.dll"))?;
break;
}
}
}
}
"msvc" => {
if std::env::var("STATIC_VCRUNTIME").map_or(false, |v| v == "true") {
static_vcruntime::build();
}
}
_ => (),
}
}
Ok(())
}
#[derive(Debug, Default, PartialEq, Eq)]
struct Diff {
remove: Vec<String>,
add: Vec<String>,
}
fn features_diff(current: &[String], expected: &[String]) -> Diff {
let mut remove = Vec::new();
let mut add = Vec::new();
for feature in current {
if !expected.contains(feature) {
remove.push(feature.clone());
}
}
for feature in expected {
if !current.contains(feature) {
add.push(feature.clone());
}
}
Diff { remove, add }
}
fn check_features(config: &Config, dependency: Dependency, is_tauri_build: bool) -> String {
use tauri_utils::config::{PatternKind, TauriConfig};
let features = match dependency {
Dependency::Simple(_) => Vec::new(),
Dependency::Detailed(dep) => dep.features,
Dependency::Inherited(dep) => dep.features,
};
let all_cli_managed_features = if is_tauri_build {
vec!["isolation"]
} else {
TauriConfig::all_features()
};
let expected = if is_tauri_build {
match config.tauri.pattern {
PatternKind::Isolation { .. } => vec!["isolation".to_string()],
_ => vec![],
}
} else | {
config
.tauri
.features()
.into_iter()
.map(|f| f.to_string())
.collect::<Vec<String>>()
} | conditional_block |
|
lib.rs | _manifest: Option<String>,
}
impl WindowsAttributes {
/// Creates the default attribute set.
pub fn new() -> Self {
Self::default()
}
/// Sets the icon to use on the window. Currently only used on Windows.
/// It must be in `ico` format. Defaults to `icons/icon.ico`.
#[must_use]
pub fn window_icon_path<P: AsRef<Path>>(mut self, window_icon_path: P) -> Self {
self
.window_icon_path
.replace(window_icon_path.as_ref().into());
self
}
/// Sets the sdk dir for windows. Currently only used on Windows. This must be a valid UTF-8
/// path. Defaults to whatever the `winres` crate determines is best.
#[must_use]
pub fn sdk_dir<P: AsRef<Path>>(mut self, sdk_dir: P) -> Self {
self.sdk_dir = Some(sdk_dir.as_ref().into());
self
}
/// Sets the Windows app [manifest].
///
/// # Example
///
/// The following manifest will brand the exe as requesting administrator privileges.
/// Thus, everytime it is executed, a Windows UAC dialog will appear.
///
/// Note that you can move the manifest contents to a separate file and use `include_str!("manifest.xml")`
/// instead of the inline string.
///
/// ```rust,no_run
/// let mut windows = tauri_build::WindowsAttributes::new();
/// windows = windows.app_manifest(r#"
/// <assembly xmlns="urn:schemas-microsoft-com:asm.v1" manifestVersion="1.0">
/// <trustInfo xmlns="urn:schemas-microsoft-com:asm.v3">
/// <security>
/// <requestedPrivileges>
/// <requestedExecutionLevel level="requireAdministrator" uiAccess="false" />
/// </requestedPrivileges>
/// </security>
/// </trustInfo>
/// </assembly>
/// "#);
/// tauri_build::try_build(
/// tauri_build::Attributes::new().windows_attributes(windows)
/// ).expect("failed to run build script");
/// ```
///
/// Defaults to:
/// ```ignore
#[doc = include_str!("window-app-manifest.xml")]
/// [manifest]: https://learn.microsoft.com/en-us/windows/win32/sbscs/application-manifests
/// ```
#[must_use]
pub fn app_manifest<S: AsRef<str>>(mut self, manifest: S) -> Self {
self.app_manifest = Some(manifest.as_ref().to_string());
self
}
}
/// The attributes used on the build.
#[derive(Debug, Default)]
pub struct Attributes {
#[allow(dead_code)]
windows_attributes: WindowsAttributes,
}
impl Attributes {
/// Creates the default attribute set.
pub fn new() -> Self {
Self::default()
}
/// Sets the icon to use on the window. Currently only used on Windows.
#[must_use]
pub fn windows_attributes(mut self, windows_attributes: WindowsAttributes) -> Self {
self.windows_attributes = windows_attributes;
self
}
}
/// Run all build time helpers for your Tauri Application.
///
/// The current helpers include the following:
/// * Generates a Windows Resource file when targeting Windows.
///
/// # Platforms
///
/// [`build()`] should be called inside of `build.rs` regardless of the platform:
/// * New helpers may target more platforms in the future.
/// * Platform specific code is handled by the helpers automatically.
/// * A build script is required in order to activate some cargo environmental variables that are
/// used when generating code and embedding assets - so [`build()`] may as well be called.
///
/// In short, this is saying don't put the call to [`build()`] behind a `#[cfg(windows)]`.
///
/// # Panics
///
/// If any of the build time helpers fail, they will [`std::panic!`] with the related error message.
/// This is typically desirable when running inside a build script; see [`try_build`] for no panics.
pub fn build() {
if let Err(error) = try_build(Attributes::default()) {
let error = format!("{error:#}");
println!("{error}");
if error.starts_with("unknown field") {
print!("found an unknown configuration field. This usually means that you are using a CLI version that is newer than `tauri-build` and is incompatible. ");
println!(
"Please try updating the Rust crates by running `cargo update` in the Tauri app folder."
);
}
std::process::exit(1);
}
}
/// Non-panicking [`build()`].
#[allow(unused_variables)]
pub fn try_build(attributes: Attributes) -> Result<()> {
use anyhow::anyhow;
println!("cargo:rerun-if-env-changed=TAURI_CONFIG");
println!("cargo:rerun-if-changed=tauri.conf.json");
#[cfg(feature = "config-json5")]
println!("cargo:rerun-if-changed=tauri.conf.json5");
#[cfg(feature = "config-toml")]
println!("cargo:rerun-if-changed=Tauri.toml");
let target_os = std::env::var("CARGO_CFG_TARGET_OS").unwrap();
let mobile = target_os == "ios" || target_os == "android";
cfg_alias("desktop", !mobile);
cfg_alias("mobile", mobile);
let mut config = serde_json::from_value(tauri_utils::config::parse::read_from(
std::env::current_dir().unwrap(),
)?)?;
if let Ok(env) = std::env::var("TAURI_CONFIG") {
let merge_config: serde_json::Value = serde_json::from_str(&env)?;
json_patch::merge(&mut config, &merge_config);
}
let config: Config = serde_json::from_value(config)?;
cfg_alias("dev", !has_feature("custom-protocol"));
let ws_path = get_workspace_dir()?;
let mut manifest =
Manifest::<cargo_toml::Value>::from_slice_with_metadata(&std::fs::read("Cargo.toml")?)?;
if let Ok(ws_manifest) = Manifest::from_path(ws_path.join("Cargo.toml")) {
Manifest::complete_from_path_and_workspace(
&mut manifest,
Path::new("Cargo.toml"),
Some((&ws_manifest, ws_path.as_path())),
)?;
} else {
Manifest::complete_from_path(&mut manifest, Path::new("Cargo.toml"))?;
}
if let Some(tauri_build) = manifest.build_dependencies.remove("tauri-build") {
let error_message = check_features(&config, tauri_build, true);
if !error_message.is_empty() {
return Err(anyhow!("
The `tauri-build` dependency features on the `Cargo.toml` file does not match the allowlist defined under `tauri.conf.json`.
Please run `tauri dev` or `tauri build` or {}.
", error_message));
}
}
if let Some(tauri) = manifest.dependencies.remove("tauri") {
let error_message = check_features(&config, tauri, false);
if !error_message.is_empty() {
return Err(anyhow!("
The `tauri` dependency features on the `Cargo.toml` file does not match the allowlist defined under `tauri.conf.json`.
Please run `tauri dev` or `tauri build` or {}.
", error_message));
}
}
let target_triple = std::env::var("TARGET").unwrap();
let out_dir = PathBuf::from(std::env::var("OUT_DIR").unwrap());
// TODO: far from ideal, but there's no other way to get the target dir, see <https://github.com/rust-lang/cargo/issues/5457>
let target_dir = out_dir
.parent()
.unwrap()
.parent()
.unwrap()
.parent()
.unwrap();
if let Some(paths) = &config.tauri.bundle.external_bin {
copy_binaries(
ResourcePaths::new(external_binaries(paths, &target_triple).as_slice(), true),
&target_triple,
target_dir,
manifest.package.as_ref().map(|p| &p.name),
)?;
}
#[allow(unused_mut, clippy::redundant_clone)]
let mut resources = config.tauri.bundle.resources.clone().unwrap_or_default();
if target_triple.contains("windows") {
if let Some(fixed_webview2_runtime_path) =
&config.tauri.bundle.windows.webview_fixed_runtime_path
{
resources.push(fixed_webview2_runtime_path.display().to_string());
}
}
copy_resources(ResourcePaths::new(resources.as_slice(), true), target_dir)?;
if target_triple.contains("darwin") {
if let Some(version) = &config.tauri.bundle.macos.minimum_system_version {
println!("cargo:rustc-env=MACOSX_DEPLOYMENT_TARGET={}", version);
}
}
if target_triple.contains("windows") {
use anyhow::Context;
use semver::Version;
use tauri_winres::{VersionInfo, WindowsResource};
fn find_icon<F: Fn(&&String) -> bool>(config: &Config, predicate: F, default: &str) -> PathBuf {
let icon_path = config | .tauri
.bundle | random_line_split |
|
main.rs | 210317")
.about("Enables services that service consumers want to use on Google Cloud Platform, lists the available or enabled services, or disables services that service consumers no longer use.")
.after_help("All documentation details can be found at <TODO figure out URL>")
.arg(Arg::with_name("scope")
.long("scope")
.help("Specify the authentication method should be executed in. Each scope requires the user to grant this application permission to use it. If unset, it defaults to the shortest scope url for a particular method.")
.multiple(true)
.takes_value(true))
.arg(Arg::with_name("folder")
.long("config-dir")
.help("A directory into which we will store our persistent data. Defaults to a user-writable directory that we will create during the first invocation." )
.multiple(false)
.takes_value(true))
.arg(Arg::with_name("debug")
.long("debug")
.help("Provide more output to aid with debugging")
.multiple(false)
.takes_value(false));
let mut operations0 = SubCommand::with_name("operations")
.setting(AppSettings::ColoredHelp)
.about("methods: get and list");
{
let mcmd = SubCommand::with_name("get").about("Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.");
operations0 = operations0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("list").about("Lists operations that match the specified filter in the request. If the server doesn\'t support this method, it returns `UNIMPLEMENTED`. NOTE: the `name` binding allows API services to override the binding to use different resource name schemes, such as `users/*/operations`. To override the binding, API services can add a binding such as `\"/v1/{name=users/*}/operations\"` to their service configuration. For backwards compatibility, the default name includes the operations collection id, however overriding users must ensure the name binding is the parent resource, without the operations collection id.");
operations0 = operations0.subcommand(mcmd);
}
let mut services0 = SubCommand::with_name("services")
.setting(AppSettings::ColoredHelp)
.about(
"methods: batch_enable, disable, enable, generate_service_identity, get and list",
);
{
let mcmd = SubCommand::with_name("batch_enable").about("Enable multiple services on a project. The operation is atomic: if enabling any service fails, then the entire batch fails, and no state changes occur. Operation");
services0 = services0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("disable").about("Disable a service so that it can no longer be used with a project. This prevents unintended usage that may cause unexpected billing charges or security leaks. It is not valid to call the disable method on a service that is not currently enabled. Callers will receive a `FAILED_PRECONDITION` status if the target service is not currently enabled. Operation");
services0 = services0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("enable")
.about("Enable a service so that it can be used with a project. Operation");
services0 = services0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("generate_service_identity")
.about("Generate service identity for service.");
services0 = services0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("get")
.about("Returns the service configuration and enabled state for a given service.");
services0 = services0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("list").about("List all services available to the specified project, and the current state of those services with respect to the project. The list includes all public services, all services for which the calling user has the `servicemanagement.services.bind` permission, and all services that have already been enabled on the project. The list can be filtered to only include services in a specific state, for example to only include services enabled on the project.");
services0 = services0.subcommand(mcmd);
}
let mut consumer_quota_metrics1 = SubCommand::with_name("consumer_quota_metrics")
.setting(AppSettings::ColoredHelp)
.about("methods: get, import_admin_overrides, import_consumer_overrides and list");
{
let mcmd = SubCommand::with_name("get")
.about("Retrieves a summary of quota information for a specific quota metric");
consumer_quota_metrics1 = consumer_quota_metrics1.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("import_admin_overrides").about("Create or update multiple admin overrides atomically, all on the same consumer, but on many different metrics or limits. The name field in the quota override message should not be set.");
consumer_quota_metrics1 = consumer_quota_metrics1.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("import_consumer_overrides").about("Create or update multiple consumer overrides atomically, all on the same consumer, but on many different metrics or limits. The name field in the quota override message should not be set.");
consumer_quota_metrics1 = consumer_quota_metrics1.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("list").about("Retrieves a summary of all quota information visible to the service consumer, organized by service metric. Each metric includes information about all of its defined limits. Each limit includes the limit configuration (quota unit, preciseness, default value), the current effective limit value, and all of the overrides applied to the limit.");
consumer_quota_metrics1 = consumer_quota_metrics1.subcommand(mcmd);
}
let mut limits2 = SubCommand::with_name("limits")
.setting(AppSettings::ColoredHelp)
.about("methods: get");
{
let mcmd = SubCommand::with_name("get")
.about("Retrieves a summary of quota information for a specific quota limit.");
limits2 = limits2.subcommand(mcmd);
}
let mut admin_overrides3 = SubCommand::with_name("admin_overrides")
.setting(AppSettings::ColoredHelp)
.about("methods: create, delete, list and patch");
{
let mcmd = SubCommand::with_name("create").about("Creates an admin override. An admin override is applied by an administrator of a parent folder or parent organization of the consumer receiving the override. An admin override is intended to limit the amount of quota the consumer can use out of the total quota pool allocated to all children of the folder or organization.");
admin_overrides3 = admin_overrides3.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("delete").about("Deletes an admin override.");
admin_overrides3 = admin_overrides3.subcommand(mcmd);
}
{
let mcmd =
SubCommand::with_name("list").about("Lists all admin overrides on this limit.");
admin_overrides3 = admin_overrides3.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("patch").about("Updates an admin override.");
admin_overrides3 = admin_overrides3.subcommand(mcmd);
}
let mut consumer_overrides3 = SubCommand::with_name("consumer_overrides")
.setting(AppSettings::ColoredHelp)
.about("methods: create, delete, list and patch");
{
let mcmd = SubCommand::with_name("create").about("Creates a consumer override. A consumer override is applied to the consumer on its own authority to limit its own quota usage. Consumer overrides cannot be used to grant more quota than would be allowed by admin overrides, producer overrides, or the default limit of the service.");
consumer_overrides3 = consumer_overrides3.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("delete").about("Deletes a consumer override.");
consumer_overrides3 = consumer_overrides3.subcommand(mcmd);
}
{
let mcmd =
SubCommand::with_name("list").about("Lists all consumer overrides on this limit.");
consumer_overrides3 = consumer_overrides3.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("patch").about("Updates a consumer override.");
consumer_overrides3 = consumer_overrides3.subcommand(mcmd);
}
limits2 = limits2.subcommand(consumer_overrides3);
limits2 = limits2.subcommand(admin_overrides3);
consumer_quota_metrics1 = consumer_quota_metrics1.subcommand(limits2);
services0 = services0.subcommand(consumer_quota_metrics1);
app = app.subcommand(services0);
app = app.subcommand(operations0);
Self { app }
}
}
use google_serviceusage1_beta1 as api;
fn main() | {
// TODO: set homedir afterwards, once the address is unmovable, or use Pin for the very first time
// to allow a self-referential structure :D!
let _home_dir = dirs::config_dir()
.expect("configuration directory can be obtained")
.join("google-service-cli");
let outer = Outer::default_boxed();
let app = outer.inner.app;
let _matches = app.get_matches();
} | identifier_body |
|
main.rs | , 'b>,
}
struct HeapApp<'a, 'b> {
app: App<'a, 'b>,
}
impl<'a, 'b> Default for HeapApp<'a, 'b> {
fn default() -> Self {
let mut app = App::new("serviceusage1_beta1")
.setting(clap::AppSettings::ColoredHelp)
.author("Sebastian Thiel <[email protected]>")
.version("0.1.0-20210317")
.about("Enables services that service consumers want to use on Google Cloud Platform, lists the available or enabled services, or disables services that service consumers no longer use.")
.after_help("All documentation details can be found at <TODO figure out URL>")
.arg(Arg::with_name("scope")
.long("scope")
.help("Specify the authentication method should be executed in. Each scope requires the user to grant this application permission to use it. If unset, it defaults to the shortest scope url for a particular method.")
.multiple(true)
.takes_value(true))
.arg(Arg::with_name("folder")
.long("config-dir")
.help("A directory into which we will store our persistent data. Defaults to a user-writable directory that we will create during the first invocation." )
.multiple(false)
.takes_value(true))
.arg(Arg::with_name("debug")
.long("debug")
.help("Provide more output to aid with debugging")
.multiple(false)
.takes_value(false));
let mut operations0 = SubCommand::with_name("operations")
.setting(AppSettings::ColoredHelp)
.about("methods: get and list");
{
let mcmd = SubCommand::with_name("get").about("Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.");
operations0 = operations0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("list").about("Lists operations that match the specified filter in the request. If the server doesn\'t support this method, it returns `UNIMPLEMENTED`. NOTE: the `name` binding allows API services to override the binding to use different resource name schemes, such as `users/*/operations`. To override the binding, API services can add a binding such as `\"/v1/{name=users/*}/operations\"` to their service configuration. For backwards compatibility, the default name includes the operations collection id, however overriding users must ensure the name binding is the parent resource, without the operations collection id.");
operations0 = operations0.subcommand(mcmd);
}
let mut services0 = SubCommand::with_name("services")
.setting(AppSettings::ColoredHelp)
.about(
"methods: batch_enable, disable, enable, generate_service_identity, get and list",
);
{
let mcmd = SubCommand::with_name("batch_enable").about("Enable multiple services on a project. The operation is atomic: if enabling any service fails, then the entire batch fails, and no state changes occur. Operation");
services0 = services0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("disable").about("Disable a service so that it can no longer be used with a project. This prevents unintended usage that may cause unexpected billing charges or security leaks. It is not valid to call the disable method on a service that is not currently enabled. Callers will receive a `FAILED_PRECONDITION` status if the target service is not currently enabled. Operation");
services0 = services0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("enable")
.about("Enable a service so that it can be used with a project. Operation");
services0 = services0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("generate_service_identity")
.about("Generate service identity for service.");
services0 = services0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("get")
.about("Returns the service configuration and enabled state for a given service.");
services0 = services0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("list").about("List all services available to the specified project, and the current state of those services with respect to the project. The list includes all public services, all services for which the calling user has the `servicemanagement.services.bind` permission, and all services that have already been enabled on the project. The list can be filtered to only include services in a specific state, for example to only include services enabled on the project.");
services0 = services0.subcommand(mcmd);
}
let mut consumer_quota_metrics1 = SubCommand::with_name("consumer_quota_metrics")
.setting(AppSettings::ColoredHelp)
.about("methods: get, import_admin_overrides, import_consumer_overrides and list");
{
let mcmd = SubCommand::with_name("get")
.about("Retrieves a summary of quota information for a specific quota metric");
consumer_quota_metrics1 = consumer_quota_metrics1.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("import_admin_overrides").about("Create or update multiple admin overrides atomically, all on the same consumer, but on many different metrics or limits. The name field in the quota override message should not be set.");
consumer_quota_metrics1 = consumer_quota_metrics1.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("import_consumer_overrides").about("Create or update multiple consumer overrides atomically, all on the same consumer, but on many different metrics or limits. The name field in the quota override message should not be set.");
consumer_quota_metrics1 = consumer_quota_metrics1.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("list").about("Retrieves a summary of all quota information visible to the service consumer, organized by service metric. Each metric includes information about all of its defined limits. Each limit includes the limit configuration (quota unit, preciseness, default value), the current effective limit value, and all of the overrides applied to the limit.");
consumer_quota_metrics1 = consumer_quota_metrics1.subcommand(mcmd);
}
let mut limits2 = SubCommand::with_name("limits")
.setting(AppSettings::ColoredHelp)
.about("methods: get");
{
let mcmd = SubCommand::with_name("get")
.about("Retrieves a summary of quota information for a specific quota limit.");
limits2 = limits2.subcommand(mcmd);
}
let mut admin_overrides3 = SubCommand::with_name("admin_overrides")
.setting(AppSettings::ColoredHelp)
.about("methods: create, delete, list and patch");
{
let mcmd = SubCommand::with_name("create").about("Creates an admin override. An admin override is applied by an administrator of a parent folder or parent organization of the consumer receiving the override. An admin override is intended to limit the amount of quota the consumer can use out of the total quota pool allocated to all children of the folder or organization.");
admin_overrides3 = admin_overrides3.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("delete").about("Deletes an admin override.");
admin_overrides3 = admin_overrides3.subcommand(mcmd);
}
{
let mcmd =
SubCommand::with_name("list").about("Lists all admin overrides on this limit.");
admin_overrides3 = admin_overrides3.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("patch").about("Updates an admin override.");
admin_overrides3 = admin_overrides3.subcommand(mcmd);
} | {
let mcmd = SubCommand::with_name("create").about("Creates a consumer override. A consumer override is applied to the consumer on its own authority to limit its own quota usage. Consumer overrides cannot be used to grant more quota than would be allowed by admin overrides, producer overrides, or the default limit of the service.");
consumer_overrides3 = consumer_overrides3.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("delete").about("Deletes a consumer override.");
consumer_overrides3 = consumer_overrides3.subcommand(mcmd);
}
{
let mcmd =
SubCommand::with_name("list").about("Lists all consumer overrides on this limit.");
consumer_overrides3 = consumer_overrides3.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("patch").about("Updates a consumer override.");
consumer_overrides3 = consumer_overrides3.subcommand(mcmd);
}
limits2 = limits2.subcommand(consumer_overrides3);
limits2 = limits2.subcommand(admin_overrides3);
consumer_quota_metrics1 = consumer_quota_metrics1.subcommand(limits2);
services0 = services0.subcommand(consumer_quota_metrics1);
app = app.subcommand(services0);
app = app.subcommand(operations0);
Self { app }
}
}
use google_serviceusage1_beta | let mut consumer_overrides3 = SubCommand::with_name("consumer_overrides")
.setting(AppSettings::ColoredHelp)
.about("methods: create, delete, list and patch"); | random_line_split |
main.rs | 'b>,
}
struct | <'a, 'b> {
app: App<'a, 'b>,
}
impl<'a, 'b> Default for HeapApp<'a, 'b> {
fn default() -> Self {
let mut app = App::new("serviceusage1_beta1")
.setting(clap::AppSettings::ColoredHelp)
.author("Sebastian Thiel <[email protected]>")
.version("0.1.0-20210317")
.about("Enables services that service consumers want to use on Google Cloud Platform, lists the available or enabled services, or disables services that service consumers no longer use.")
.after_help("All documentation details can be found at <TODO figure out URL>")
.arg(Arg::with_name("scope")
.long("scope")
.help("Specify the authentication method should be executed in. Each scope requires the user to grant this application permission to use it. If unset, it defaults to the shortest scope url for a particular method.")
.multiple(true)
.takes_value(true))
.arg(Arg::with_name("folder")
.long("config-dir")
.help("A directory into which we will store our persistent data. Defaults to a user-writable directory that we will create during the first invocation." )
.multiple(false)
.takes_value(true))
.arg(Arg::with_name("debug")
.long("debug")
.help("Provide more output to aid with debugging")
.multiple(false)
.takes_value(false));
let mut operations0 = SubCommand::with_name("operations")
.setting(AppSettings::ColoredHelp)
.about("methods: get and list");
{
let mcmd = SubCommand::with_name("get").about("Gets the latest state of a long-running operation. Clients can use this method to poll the operation result at intervals as recommended by the API service.");
operations0 = operations0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("list").about("Lists operations that match the specified filter in the request. If the server doesn\'t support this method, it returns `UNIMPLEMENTED`. NOTE: the `name` binding allows API services to override the binding to use different resource name schemes, such as `users/*/operations`. To override the binding, API services can add a binding such as `\"/v1/{name=users/*}/operations\"` to their service configuration. For backwards compatibility, the default name includes the operations collection id, however overriding users must ensure the name binding is the parent resource, without the operations collection id.");
operations0 = operations0.subcommand(mcmd);
}
let mut services0 = SubCommand::with_name("services")
.setting(AppSettings::ColoredHelp)
.about(
"methods: batch_enable, disable, enable, generate_service_identity, get and list",
);
{
let mcmd = SubCommand::with_name("batch_enable").about("Enable multiple services on a project. The operation is atomic: if enabling any service fails, then the entire batch fails, and no state changes occur. Operation");
services0 = services0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("disable").about("Disable a service so that it can no longer be used with a project. This prevents unintended usage that may cause unexpected billing charges or security leaks. It is not valid to call the disable method on a service that is not currently enabled. Callers will receive a `FAILED_PRECONDITION` status if the target service is not currently enabled. Operation");
services0 = services0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("enable")
.about("Enable a service so that it can be used with a project. Operation");
services0 = services0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("generate_service_identity")
.about("Generate service identity for service.");
services0 = services0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("get")
.about("Returns the service configuration and enabled state for a given service.");
services0 = services0.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("list").about("List all services available to the specified project, and the current state of those services with respect to the project. The list includes all public services, all services for which the calling user has the `servicemanagement.services.bind` permission, and all services that have already been enabled on the project. The list can be filtered to only include services in a specific state, for example to only include services enabled on the project.");
services0 = services0.subcommand(mcmd);
}
let mut consumer_quota_metrics1 = SubCommand::with_name("consumer_quota_metrics")
.setting(AppSettings::ColoredHelp)
.about("methods: get, import_admin_overrides, import_consumer_overrides and list");
{
let mcmd = SubCommand::with_name("get")
.about("Retrieves a summary of quota information for a specific quota metric");
consumer_quota_metrics1 = consumer_quota_metrics1.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("import_admin_overrides").about("Create or update multiple admin overrides atomically, all on the same consumer, but on many different metrics or limits. The name field in the quota override message should not be set.");
consumer_quota_metrics1 = consumer_quota_metrics1.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("import_consumer_overrides").about("Create or update multiple consumer overrides atomically, all on the same consumer, but on many different metrics or limits. The name field in the quota override message should not be set.");
consumer_quota_metrics1 = consumer_quota_metrics1.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("list").about("Retrieves a summary of all quota information visible to the service consumer, organized by service metric. Each metric includes information about all of its defined limits. Each limit includes the limit configuration (quota unit, preciseness, default value), the current effective limit value, and all of the overrides applied to the limit.");
consumer_quota_metrics1 = consumer_quota_metrics1.subcommand(mcmd);
}
let mut limits2 = SubCommand::with_name("limits")
.setting(AppSettings::ColoredHelp)
.about("methods: get");
{
let mcmd = SubCommand::with_name("get")
.about("Retrieves a summary of quota information for a specific quota limit.");
limits2 = limits2.subcommand(mcmd);
}
let mut admin_overrides3 = SubCommand::with_name("admin_overrides")
.setting(AppSettings::ColoredHelp)
.about("methods: create, delete, list and patch");
{
let mcmd = SubCommand::with_name("create").about("Creates an admin override. An admin override is applied by an administrator of a parent folder or parent organization of the consumer receiving the override. An admin override is intended to limit the amount of quota the consumer can use out of the total quota pool allocated to all children of the folder or organization.");
admin_overrides3 = admin_overrides3.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("delete").about("Deletes an admin override.");
admin_overrides3 = admin_overrides3.subcommand(mcmd);
}
{
let mcmd =
SubCommand::with_name("list").about("Lists all admin overrides on this limit.");
admin_overrides3 = admin_overrides3.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("patch").about("Updates an admin override.");
admin_overrides3 = admin_overrides3.subcommand(mcmd);
}
let mut consumer_overrides3 = SubCommand::with_name("consumer_overrides")
.setting(AppSettings::ColoredHelp)
.about("methods: create, delete, list and patch");
{
let mcmd = SubCommand::with_name("create").about("Creates a consumer override. A consumer override is applied to the consumer on its own authority to limit its own quota usage. Consumer overrides cannot be used to grant more quota than would be allowed by admin overrides, producer overrides, or the default limit of the service.");
consumer_overrides3 = consumer_overrides3.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("delete").about("Deletes a consumer override.");
consumer_overrides3 = consumer_overrides3.subcommand(mcmd);
}
{
let mcmd =
SubCommand::with_name("list").about("Lists all consumer overrides on this limit.");
consumer_overrides3 = consumer_overrides3.subcommand(mcmd);
}
{
let mcmd = SubCommand::with_name("patch").about("Updates a consumer override.");
consumer_overrides3 = consumer_overrides3.subcommand(mcmd);
}
limits2 = limits2.subcommand(consumer_overrides3);
limits2 = limits2.subcommand(admin_overrides3);
consumer_quota_metrics1 = consumer_quota_metrics1.subcommand(limits2);
services0 = services0.subcommand(consumer_quota_metrics1);
app = app.subcommand(services0);
app = app.subcommand(operations0);
Self { app }
}
}
use google_serviceusage1_beta | HeapApp | identifier_name |
v3.go | (core.Volume{
Name: secret.Source,
VolumeSource: vSrc,
})
serviceContainer.VolumeMounts = append(serviceContainer.VolumeMounts, core.VolumeMount{
Name: secret.Source,
MountPath: target,
})
}
for _, c := range composeServiceConfig.Configs {
target := c.Target
if target == "" {
target = "/" + c.Source
}
vSrc := core.ConfigMapVolumeSource{}
vSrc.Name = common.MakeFileNameCompliant(c.Source)
if o, ok := composeObject.Configs[c.Source]; ok {
if o.External.External {
logrus.Errorf("Config metadata %s has an external source", c.Source)
} else {
srcBaseName := filepath.Base(o.File)
vSrc.Items = []core.KeyToPath{{Key: srcBaseName, Path: filepath.Base(target)}}
if c.Mode != nil {
signedMode := int32(*c.Mode)
vSrc.DefaultMode = &signedMode
}
}
} else {
logrus.Errorf("Unable to find configmap object for %s", vSrc.Name)
}
serviceConfig.AddVolume(core.Volume{
Name: vSrc.Name,
VolumeSource: core.VolumeSource{ConfigMap: &vSrc},
})
serviceContainer.VolumeMounts = append(serviceContainer.VolumeMounts,
core.VolumeMount{
Name: vSrc.Name,
MountPath: target,
SubPath: filepath.Base(target),
})
}
for _, vol := range composeServiceConfig.Volumes {
if isPath(vol.Source) {
hPath := vol.Source
if !filepath.IsAbs(vol.Source) {
hPath, err := filepath.Abs(vol.Source)
if err != nil {
logrus.Debugf("Could not create an absolute path for [%s]", hPath)
}
}
// Generate a hash Id for the given source file path to be mounted.
hashID := getHash([]byte(hPath))
volumeName := fmt.Sprintf("%s%d", common.VolumePrefix, hashID)
serviceContainer.VolumeMounts = append(serviceContainer.VolumeMounts, core.VolumeMount{
Name: volumeName,
MountPath: vol.Target,
})
serviceConfig.AddVolume(core.Volume{
Name: volumeName,
VolumeSource: core.VolumeSource{
HostPath: &core.HostPathVolumeSource{Path: vol.Source},
},
})
} else {
serviceContainer.VolumeMounts = append(serviceContainer.VolumeMounts, core.VolumeMount{
Name: vol.Source,
MountPath: vol.Target,
})
serviceConfig.AddVolume(core.Volume{
Name: vol.Source,
VolumeSource: core.VolumeSource{
PersistentVolumeClaim: &core.PersistentVolumeClaimVolumeSource{
ClaimName: vol.Source,
},
},
})
storageObj := irtypes.Storage{StorageType: irtypes.PVCKind, Name: vol.Source, Content: nil}
ir.AddStorage(storageObj)
}
}
serviceConfig.Containers = []core.Container{serviceContainer}
ir.Services[name] = serviceConfig
}
return ir, nil
}
func (c *V3Loader) getSecretStorages(secrets map[string]types.SecretConfig) []irtypes.Storage {
storages := make([]irtypes.Storage, len(secrets))
for secretName, secretObj := range secrets {
storage := irtypes.Storage{
Name: secretName,
StorageType: irtypes.SecretKind,
}
if !secretObj.External.External {
content, err := ioutil.ReadFile(secretObj.File)
if err != nil {
logrus.Warnf("Could not read the secret file [%s]", secretObj.File)
} else {
storage.Content = map[string][]byte{secretName: content}
}
}
storages = append(storages, storage)
}
return storages
}
func (c *V3Loader) getConfigStorages(configs map[string]types.ConfigObjConfig) []irtypes.Storage {
Storages := make([]irtypes.Storage, len(configs))
for cfgName, cfgObj := range configs {
storage := irtypes.Storage{
Name: cfgName,
StorageType: irtypes.ConfigMapKind,
}
if !cfgObj.External.External {
fileInfo, err := os.Stat(cfgObj.File)
if err != nil {
logrus.Warnf("Could not identify the type of secret artifact [%s]. Encountered [%s]", cfgObj.File, err)
} else {
if !fileInfo.IsDir() {
content, err := ioutil.ReadFile(cfgObj.File)
if err != nil {
logrus.Warnf("Could not read the secret file [%s]. Encountered [%s]", cfgObj.File, err)
} else {
storage.Content = map[string][]byte{cfgName: content}
}
} else {
dataMap, err := c.getAllDirContentAsMap(cfgObj.File)
if err != nil {
logrus.Warnf("Could not read the secret directory [%s]. Encountered [%s]", cfgObj.File, err)
} else {
storage.Content = dataMap
}
}
}
}
Storages = append(Storages, storage)
}
return Storages
}
func (*V3Loader) getPorts(ports []types.ServicePortConfig, expose []string) []core.ContainerPort {
containerPorts := []core.ContainerPort{}
exist := map[string]bool{}
for _, port := range ports {
proto := core.ProtocolTCP
if strings.EqualFold(string(core.ProtocolUDP), port.Protocol) {
proto = core.ProtocolUDP
}
// Add the port to the k8s pod.
containerPorts = append(containerPorts, core.ContainerPort{
ContainerPort: int32(port.Target),
Protocol: proto,
})
exist[cast.ToString(port.Target)] = true
}
for _, port := range expose {
portValue := port
protocol := core.ProtocolTCP
if strings.Contains(portValue, "/") {
splits := strings.Split(port, "/")
portValue = splits[0]
protocol = core.Protocol(strings.ToUpper(splits[1]))
}
if exist[portValue] {
continue
}
// Add the port to the k8s pod.
containerPorts = append(containerPorts, core.ContainerPort{
ContainerPort: cast.ToInt32(portValue),
Protocol: protocol,
})
}
return containerPorts
}
func (*V3Loader) addPorts(ports []types.ServicePortConfig, expose []string, service *irtypes.Service) {
exist := map[string]bool{}
for _, port := range ports {
// Forward the port on the k8s service to the k8s pod.
podPort := networking.ServiceBackendPort{
Number: int32(port.Target),
}
servicePort := networking.ServiceBackendPort{
Number: int32(port.Published),
}
service.AddPortForwarding(servicePort, podPort, "")
exist[cast.ToString(port.Target)] = true
}
for _, port := range expose {
portValue := port
if strings.Contains(portValue, "/") {
splits := strings.Split(port, "/")
portValue = splits[0]
}
if exist[portValue] {
continue
}
// Forward the port on the k8s service to the k8s pod.
portNumber := cast.ToInt32(portValue)
podPort := networking.ServiceBackendPort{
Number: portNumber,
}
servicePort := networking.ServiceBackendPort{
Number: portNumber,
}
service.AddPortForwarding(servicePort, podPort, "")
}
}
func (c *V3Loader) getNetworks(composeServiceConfig types.ServiceConfig, composeObject types.Config) (networks []string) {
networks = []string{}
for key := range composeServiceConfig.Networks {
netName := composeObject.Networks[key].Name
if netName == "" {
netName = key
}
networks = append(networks, netName)
}
return networks
}
func (c *V3Loader) getHealthCheck(composeHealthCheck types.HealthCheckConfig) (core.Probe, error) | {
probe := core.Probe{}
if len(composeHealthCheck.Test) > 1 {
probe.Handler = core.Handler{
Exec: &core.ExecAction{
// docker/cli adds "CMD-SHELL" to the struct, hence we remove the first element of composeHealthCheck.Test
Command: composeHealthCheck.Test[1:],
},
}
} else {
logrus.Warnf("Could not find command to execute in probe : %s", composeHealthCheck.Test)
}
if composeHealthCheck.Timeout != nil {
parse, err := time.ParseDuration(composeHealthCheck.Timeout.String())
if err != nil {
return probe, errors.Wrap(err, "unable to parse health check timeout variable")
}
probe.TimeoutSeconds = int32(parse.Seconds())
} | identifier_body |
|
v3.go |
}
return config, nil
}
// ConvertToIR loads an v3 compose file into IR
func (c *V3Loader) ConvertToIR(composefilepath string, serviceName string) (irtypes.IR, error) {
logrus.Debugf("About to load configuration from docker compose file at path %s", composefilepath)
config, err := ParseV3(composefilepath)
if err != nil {
logrus.Warnf("Error while loading docker compose config : %s", err)
return irtypes.IR{}, err
}
logrus.Debugf("About to start loading docker compose to intermediate rep")
return c.convertToIR(filepath.Dir(composefilepath), *config, serviceName)
}
func (c *V3Loader) convertToIR(filedir string, composeObject types.Config, serviceName string) (irtypes.IR, error) {
ir := irtypes.IR{
Services: map[string]irtypes.Service{},
}
//Secret volumes transformed to IR
ir.Storages = c.getSecretStorages(composeObject.Secrets)
//ConfigMap volumes transformed to IR
ir.Storages = append(ir.Storages, c.getConfigStorages(composeObject.Configs)...)
for _, composeServiceConfig := range composeObject.Services {
if composeServiceConfig.Name != serviceName {
continue
}
name := common.NormalizeForServiceName(composeServiceConfig.Name)
serviceConfig := irtypes.NewServiceWithName(name)
serviceContainer := core.Container{}
serviceContainer.Image = composeServiceConfig.Image
if serviceContainer.Image == "" {
serviceContainer.Image = name + ":latest"
}
serviceContainer.WorkingDir = composeServiceConfig.WorkingDir
serviceContainer.Command = composeServiceConfig.Entrypoint
serviceContainer.Args = composeServiceConfig.Command
serviceContainer.Stdin = composeServiceConfig.StdinOpen
serviceContainer.Name = strings.ToLower(composeServiceConfig.ContainerName)
if serviceContainer.Name == "" {
serviceContainer.Name = strings.ToLower(serviceConfig.Name)
}
serviceContainer.TTY = composeServiceConfig.Tty
serviceContainer.Ports = c.getPorts(composeServiceConfig.Ports, composeServiceConfig.Expose)
c.addPorts(composeServiceConfig.Ports, composeServiceConfig.Expose, &serviceConfig)
serviceConfig.Annotations = map[string]string(composeServiceConfig.Labels)
serviceConfig.Labels = common.MergeStringMaps(composeServiceConfig.Labels, composeServiceConfig.Deploy.Labels)
if composeServiceConfig.Hostname != "" {
serviceConfig.Hostname = composeServiceConfig.Hostname
}
if composeServiceConfig.DomainName != "" {
serviceConfig.Subdomain = composeServiceConfig.DomainName
}
if composeServiceConfig.Pid != "" {
if composeServiceConfig.Pid == "host" {
serviceConfig.SecurityContext.HostPID = true
} else {
logrus.Warnf("Ignoring PID key for service \"%v\". Invalid value \"%v\".", name, composeServiceConfig.Pid)
}
}
securityContext := &core.SecurityContext{}
if composeServiceConfig.Privileged {
securityContext.Privileged = &composeServiceConfig.Privileged
}
if composeServiceConfig.User != "" {
uid, err := cast.ToInt64E(composeServiceConfig.User)
if err != nil {
logrus.Warn("Ignoring user directive. User to be specified as a UID (numeric).")
} else {
securityContext.RunAsUser = &uid
}
}
capsAdd := []core.Capability{}
capsDrop := []core.Capability{}
for _, capAdd := range composeServiceConfig.CapAdd {
capsAdd = append(capsAdd, core.Capability(capAdd))
}
for _, capDrop := range composeServiceConfig.CapDrop {
capsDrop = append(capsDrop, core.Capability(capDrop))
}
//set capabilities if it is not empty
if len(capsAdd) > 0 || len(capsDrop) > 0 {
securityContext.Capabilities = &core.Capabilities{
Add: capsAdd,
Drop: capsDrop,
}
}
// update template only if securityContext is not empty
if *securityContext != (core.SecurityContext{}) {
serviceContainer.SecurityContext = securityContext
}
podSecurityContext := &core.PodSecurityContext{}
if !cmp.Equal(*podSecurityContext, core.PodSecurityContext{}) {
serviceConfig.SecurityContext = podSecurityContext
}
if composeServiceConfig.Deploy.Mode == "global" {
serviceConfig.Daemon = true
}
serviceConfig.Networks = c.getNetworks(composeServiceConfig, composeObject)
if (composeServiceConfig.Deploy.Resources != types.Resources{}) {
if composeServiceConfig.Deploy.Resources.Limits != nil {
resourceLimit := core.ResourceList{}
memLimit := libcomposeyaml.MemStringorInt(composeServiceConfig.Deploy.Resources.Limits.MemoryBytes)
if memLimit != 0 {
resourceLimit[core.ResourceMemory] = *resource.NewQuantity(int64(memLimit), "RandomStringForFormat")
}
if composeServiceConfig.Deploy.Resources.Limits.NanoCPUs != "" {
cpuLimit, err := cast.ToFloat64E(composeServiceConfig.Deploy.Resources.Limits.NanoCPUs)
if err != nil {
logrus.Warnf("Unable to convert cpu limits resources value : %s", err)
}
CPULimit := int64(cpuLimit * 1000)
if CPULimit != 0 {
resourceLimit[core.ResourceCPU] = *resource.NewMilliQuantity(CPULimit, resource.DecimalSI)
}
}
serviceContainer.Resources.Limits = resourceLimit
}
if composeServiceConfig.Deploy.Resources.Reservations != nil {
resourceRequests := core.ResourceList{}
MemReservation := libcomposeyaml.MemStringorInt(composeServiceConfig.Deploy.Resources.Reservations.MemoryBytes)
if MemReservation != 0 {
resourceRequests[core.ResourceMemory] = *resource.NewQuantity(int64(MemReservation), "RandomStringForFormat")
}
if composeServiceConfig.Deploy.Resources.Reservations.NanoCPUs != "" {
cpuReservation, err := cast.ToFloat64E(composeServiceConfig.Deploy.Resources.Reservations.NanoCPUs)
if err != nil {
logrus.Warnf("Unable to convert cpu limits reservation value : %s", err)
}
CPUReservation := int64(cpuReservation * 1000)
if CPUReservation != 0 {
resourceRequests[core.ResourceCPU] = *resource.NewMilliQuantity(CPUReservation, resource.DecimalSI)
}
}
serviceContainer.Resources.Requests = resourceRequests
}
}
// HealthCheck
if composeServiceConfig.HealthCheck != nil && !composeServiceConfig.HealthCheck.Disable {
probe, err := c.getHealthCheck(*composeServiceConfig.HealthCheck)
if err != nil {
logrus.Warnf("Unable to parse health check : %s", err)
} else {
serviceContainer.LivenessProbe = &probe
}
}
restart := composeServiceConfig.Restart
if composeServiceConfig.Deploy.RestartPolicy != nil {
restart = composeServiceConfig.Deploy.RestartPolicy.Condition
}
if restart == "unless-stopped" {
logrus.Warnf("Restart policy 'unless-stopped' in service %s is not supported, convert it to 'always'", name)
serviceConfig.RestartPolicy = core.RestartPolicyAlways
}
// replicas:
if composeServiceConfig.Deploy.Replicas != nil {
serviceConfig.Replicas = int(*composeServiceConfig.Deploy.Replicas)
}
serviceContainer.Env = c.getEnvs(composeServiceConfig)
vml, vl := makeVolumesFromTmpFS(name, composeServiceConfig.Tmpfs)
for _, v := range vl {
serviceConfig.AddVolume(v)
}
serviceContainer.VolumeMounts = append(serviceContainer.VolumeMounts, vml...)
for _, secret := range composeServiceConfig.Secrets {
target := filepath.Join(defaultSecretBasePath, secret.Source)
src := secret.Source
if secret.Target != "" {
tokens := strings.Split(secret.Source, "/")
var prefix string
if !strings.HasPrefix(secret.Target, "/") {
prefix = defaultSecretBasePath + "/"
}
if tokens[len(tokens)-1] == secret.Target {
target = prefix + secret.Source
} else {
target = prefix + strings.TrimSuffix(secret.Target, "/"+tokens[len(tokens)-1])
}
src = tokens[len(tokens)-1]
}
vSrc := core.VolumeSource{
Secret: &core.SecretVolumeSource{ | },
}
if secret.Mode != nil {
mode := cast.ToInt32(*secret.Mode)
vSrc.Secret.DefaultMode = &mode
}
serviceConfig.AddVolume(core.Volume{
Name: secret.Source,
VolumeSource: vSrc,
})
serviceContainer.VolumeMounts = append(serviceContainer.VolumeMounts, core.VolumeMount{
Name: | SecretName: secret.Source,
Items: []core.KeyToPath{{
Key: secret.Source,
Path: src,
}}, | random_line_split |
v3.go | if restart == "unless-stopped" {
logrus.Warnf("Restart policy 'unless-stopped' in service %s is not supported, convert it to 'always'", name)
serviceConfig.RestartPolicy = core.RestartPolicyAlways
}
// replicas:
if composeServiceConfig.Deploy.Replicas != nil {
serviceConfig.Replicas = int(*composeServiceConfig.Deploy.Replicas)
}
serviceContainer.Env = c.getEnvs(composeServiceConfig)
vml, vl := makeVolumesFromTmpFS(name, composeServiceConfig.Tmpfs)
for _, v := range vl {
serviceConfig.AddVolume(v)
}
serviceContainer.VolumeMounts = append(serviceContainer.VolumeMounts, vml...)
for _, secret := range composeServiceConfig.Secrets {
target := filepath.Join(defaultSecretBasePath, secret.Source)
src := secret.Source
if secret.Target != "" {
tokens := strings.Split(secret.Source, "/")
var prefix string
if !strings.HasPrefix(secret.Target, "/") {
prefix = defaultSecretBasePath + "/"
}
if tokens[len(tokens)-1] == secret.Target {
target = prefix + secret.Source
} else {
target = prefix + strings.TrimSuffix(secret.Target, "/"+tokens[len(tokens)-1])
}
src = tokens[len(tokens)-1]
}
vSrc := core.VolumeSource{
Secret: &core.SecretVolumeSource{
SecretName: secret.Source,
Items: []core.KeyToPath{{
Key: secret.Source,
Path: src,
}},
},
}
if secret.Mode != nil {
mode := cast.ToInt32(*secret.Mode)
vSrc.Secret.DefaultMode = &mode
}
serviceConfig.AddVolume(core.Volume{
Name: secret.Source,
VolumeSource: vSrc,
})
serviceContainer.VolumeMounts = append(serviceContainer.VolumeMounts, core.VolumeMount{
Name: secret.Source,
MountPath: target,
})
}
for _, c := range composeServiceConfig.Configs {
target := c.Target
if target == "" {
target = "/" + c.Source
}
vSrc := core.ConfigMapVolumeSource{}
vSrc.Name = common.MakeFileNameCompliant(c.Source)
if o, ok := composeObject.Configs[c.Source]; ok {
if o.External.External {
logrus.Errorf("Config metadata %s has an external source", c.Source)
} else {
srcBaseName := filepath.Base(o.File)
vSrc.Items = []core.KeyToPath{{Key: srcBaseName, Path: filepath.Base(target)}}
if c.Mode != nil {
signedMode := int32(*c.Mode)
vSrc.DefaultMode = &signedMode
}
}
} else {
logrus.Errorf("Unable to find configmap object for %s", vSrc.Name)
}
serviceConfig.AddVolume(core.Volume{
Name: vSrc.Name,
VolumeSource: core.VolumeSource{ConfigMap: &vSrc},
})
serviceContainer.VolumeMounts = append(serviceContainer.VolumeMounts,
core.VolumeMount{
Name: vSrc.Name,
MountPath: target,
SubPath: filepath.Base(target),
})
}
for _, vol := range composeServiceConfig.Volumes {
if isPath(vol.Source) {
hPath := vol.Source
if !filepath.IsAbs(vol.Source) {
hPath, err := filepath.Abs(vol.Source)
if err != nil {
logrus.Debugf("Could not create an absolute path for [%s]", hPath)
}
}
// Generate a hash Id for the given source file path to be mounted.
hashID := getHash([]byte(hPath))
volumeName := fmt.Sprintf("%s%d", common.VolumePrefix, hashID)
serviceContainer.VolumeMounts = append(serviceContainer.VolumeMounts, core.VolumeMount{
Name: volumeName,
MountPath: vol.Target,
})
serviceConfig.AddVolume(core.Volume{
Name: volumeName,
VolumeSource: core.VolumeSource{
HostPath: &core.HostPathVolumeSource{Path: vol.Source},
},
})
} else {
serviceContainer.VolumeMounts = append(serviceContainer.VolumeMounts, core.VolumeMount{
Name: vol.Source,
MountPath: vol.Target,
})
serviceConfig.AddVolume(core.Volume{
Name: vol.Source,
VolumeSource: core.VolumeSource{
PersistentVolumeClaim: &core.PersistentVolumeClaimVolumeSource{
ClaimName: vol.Source,
},
},
})
storageObj := irtypes.Storage{StorageType: irtypes.PVCKind, Name: vol.Source, Content: nil}
ir.AddStorage(storageObj)
}
}
serviceConfig.Containers = []core.Container{serviceContainer}
ir.Services[name] = serviceConfig
}
return ir, nil
}
func (c *V3Loader) getSecretStorages(secrets map[string]types.SecretConfig) []irtypes.Storage {
storages := make([]irtypes.Storage, len(secrets))
for secretName, secretObj := range secrets {
storage := irtypes.Storage{
Name: secretName,
StorageType: irtypes.SecretKind,
}
if !secretObj.External.External {
content, err := ioutil.ReadFile(secretObj.File)
if err != nil {
logrus.Warnf("Could not read the secret file [%s]", secretObj.File)
} else {
storage.Content = map[string][]byte{secretName: content}
}
}
storages = append(storages, storage)
}
return storages
}
func (c *V3Loader) getConfigStorages(configs map[string]types.ConfigObjConfig) []irtypes.Storage {
Storages := make([]irtypes.Storage, len(configs))
for cfgName, cfgObj := range configs {
storage := irtypes.Storage{
Name: cfgName,
StorageType: irtypes.ConfigMapKind,
}
if !cfgObj.External.External {
fileInfo, err := os.Stat(cfgObj.File)
if err != nil {
logrus.Warnf("Could not identify the type of secret artifact [%s]. Encountered [%s]", cfgObj.File, err)
} else {
if !fileInfo.IsDir() {
content, err := ioutil.ReadFile(cfgObj.File)
if err != nil {
logrus.Warnf("Could not read the secret file [%s]. Encountered [%s]", cfgObj.File, err)
} else {
storage.Content = map[string][]byte{cfgName: content}
}
} else {
dataMap, err := c.getAllDirContentAsMap(cfgObj.File)
if err != nil {
logrus.Warnf("Could not read the secret directory [%s]. Encountered [%s]", cfgObj.File, err)
} else {
storage.Content = dataMap
}
}
}
}
Storages = append(Storages, storage)
}
return Storages
}
func (*V3Loader) getPorts(ports []types.ServicePortConfig, expose []string) []core.ContainerPort {
containerPorts := []core.ContainerPort{}
exist := map[string]bool{}
for _, port := range ports {
proto := core.ProtocolTCP
if strings.EqualFold(string(core.ProtocolUDP), port.Protocol) {
proto = core.ProtocolUDP
}
// Add the port to the k8s pod.
containerPorts = append(containerPorts, core.ContainerPort{
ContainerPort: int32(port.Target),
Protocol: proto,
})
exist[cast.ToString(port.Target)] = true
}
for _, port := range expose {
portValue := port
protocol := core.ProtocolTCP
if strings.Contains(portValue, "/") {
splits := strings.Split(port, "/")
portValue = splits[0]
protocol = core.Protocol(strings.ToUpper(splits[1]))
}
if exist[portValue] {
continue
}
// Add the port to the k8s pod.
containerPorts = append(containerPorts, core.ContainerPort{
ContainerPort: cast.ToInt32(portValue),
Protocol: protocol,
})
}
return containerPorts
}
func (*V3Loader) addPorts(ports []types.ServicePortConfig, expose []string, service *irtypes.Service) {
exist := map[string]bool{}
for _, port := range ports {
// Forward the port on the k8s service to the k8s pod.
podPort := networking.ServiceBackendPort{
Number: int32(port.Target),
}
servicePort := networking.ServiceBackendPort{
Number: int32(port.Published),
}
service.AddPortForwarding(servicePort, podPort, "")
exist[cast.ToString(port.Target)] = true
}
for _, port := range expose {
portValue := port
if strings.Contains(portValue, "/") | {
splits := strings.Split(port, "/")
portValue = splits[0]
} | conditional_block |
|
v3.go |
}
return config, nil
}
// ConvertToIR loads an v3 compose file into IR
func (c *V3Loader) ConvertToIR(composefilepath string, serviceName string) (irtypes.IR, error) {
logrus.Debugf("About to load configuration from docker compose file at path %s", composefilepath)
config, err := ParseV3(composefilepath)
if err != nil {
logrus.Warnf("Error while loading docker compose config : %s", err)
return irtypes.IR{}, err
}
logrus.Debugf("About to start loading docker compose to intermediate rep")
return c.convertToIR(filepath.Dir(composefilepath), *config, serviceName)
}
func (c *V3Loader) | (filedir string, composeObject types.Config, serviceName string) (irtypes.IR, error) {
ir := irtypes.IR{
Services: map[string]irtypes.Service{},
}
//Secret volumes transformed to IR
ir.Storages = c.getSecretStorages(composeObject.Secrets)
//ConfigMap volumes transformed to IR
ir.Storages = append(ir.Storages, c.getConfigStorages(composeObject.Configs)...)
for _, composeServiceConfig := range composeObject.Services {
if composeServiceConfig.Name != serviceName {
continue
}
name := common.NormalizeForServiceName(composeServiceConfig.Name)
serviceConfig := irtypes.NewServiceWithName(name)
serviceContainer := core.Container{}
serviceContainer.Image = composeServiceConfig.Image
if serviceContainer.Image == "" {
serviceContainer.Image = name + ":latest"
}
serviceContainer.WorkingDir = composeServiceConfig.WorkingDir
serviceContainer.Command = composeServiceConfig.Entrypoint
serviceContainer.Args = composeServiceConfig.Command
serviceContainer.Stdin = composeServiceConfig.StdinOpen
serviceContainer.Name = strings.ToLower(composeServiceConfig.ContainerName)
if serviceContainer.Name == "" {
serviceContainer.Name = strings.ToLower(serviceConfig.Name)
}
serviceContainer.TTY = composeServiceConfig.Tty
serviceContainer.Ports = c.getPorts(composeServiceConfig.Ports, composeServiceConfig.Expose)
c.addPorts(composeServiceConfig.Ports, composeServiceConfig.Expose, &serviceConfig)
serviceConfig.Annotations = map[string]string(composeServiceConfig.Labels)
serviceConfig.Labels = common.MergeStringMaps(composeServiceConfig.Labels, composeServiceConfig.Deploy.Labels)
if composeServiceConfig.Hostname != "" {
serviceConfig.Hostname = composeServiceConfig.Hostname
}
if composeServiceConfig.DomainName != "" {
serviceConfig.Subdomain = composeServiceConfig.DomainName
}
if composeServiceConfig.Pid != "" {
if composeServiceConfig.Pid == "host" {
serviceConfig.SecurityContext.HostPID = true
} else {
logrus.Warnf("Ignoring PID key for service \"%v\". Invalid value \"%v\".", name, composeServiceConfig.Pid)
}
}
securityContext := &core.SecurityContext{}
if composeServiceConfig.Privileged {
securityContext.Privileged = &composeServiceConfig.Privileged
}
if composeServiceConfig.User != "" {
uid, err := cast.ToInt64E(composeServiceConfig.User)
if err != nil {
logrus.Warn("Ignoring user directive. User to be specified as a UID (numeric).")
} else {
securityContext.RunAsUser = &uid
}
}
capsAdd := []core.Capability{}
capsDrop := []core.Capability{}
for _, capAdd := range composeServiceConfig.CapAdd {
capsAdd = append(capsAdd, core.Capability(capAdd))
}
for _, capDrop := range composeServiceConfig.CapDrop {
capsDrop = append(capsDrop, core.Capability(capDrop))
}
//set capabilities if it is not empty
if len(capsAdd) > 0 || len(capsDrop) > 0 {
securityContext.Capabilities = &core.Capabilities{
Add: capsAdd,
Drop: capsDrop,
}
}
// update template only if securityContext is not empty
if *securityContext != (core.SecurityContext{}) {
serviceContainer.SecurityContext = securityContext
}
podSecurityContext := &core.PodSecurityContext{}
if !cmp.Equal(*podSecurityContext, core.PodSecurityContext{}) {
serviceConfig.SecurityContext = podSecurityContext
}
if composeServiceConfig.Deploy.Mode == "global" {
serviceConfig.Daemon = true
}
serviceConfig.Networks = c.getNetworks(composeServiceConfig, composeObject)
if (composeServiceConfig.Deploy.Resources != types.Resources{}) {
if composeServiceConfig.Deploy.Resources.Limits != nil {
resourceLimit := core.ResourceList{}
memLimit := libcomposeyaml.MemStringorInt(composeServiceConfig.Deploy.Resources.Limits.MemoryBytes)
if memLimit != 0 {
resourceLimit[core.ResourceMemory] = *resource.NewQuantity(int64(memLimit), "RandomStringForFormat")
}
if composeServiceConfig.Deploy.Resources.Limits.NanoCPUs != "" {
cpuLimit, err := cast.ToFloat64E(composeServiceConfig.Deploy.Resources.Limits.NanoCPUs)
if err != nil {
logrus.Warnf("Unable to convert cpu limits resources value : %s", err)
}
CPULimit := int64(cpuLimit * 1000)
if CPULimit != 0 {
resourceLimit[core.ResourceCPU] = *resource.NewMilliQuantity(CPULimit, resource.DecimalSI)
}
}
serviceContainer.Resources.Limits = resourceLimit
}
if composeServiceConfig.Deploy.Resources.Reservations != nil {
resourceRequests := core.ResourceList{}
MemReservation := libcomposeyaml.MemStringorInt(composeServiceConfig.Deploy.Resources.Reservations.MemoryBytes)
if MemReservation != 0 {
resourceRequests[core.ResourceMemory] = *resource.NewQuantity(int64(MemReservation), "RandomStringForFormat")
}
if composeServiceConfig.Deploy.Resources.Reservations.NanoCPUs != "" {
cpuReservation, err := cast.ToFloat64E(composeServiceConfig.Deploy.Resources.Reservations.NanoCPUs)
if err != nil {
logrus.Warnf("Unable to convert cpu limits reservation value : %s", err)
}
CPUReservation := int64(cpuReservation * 1000)
if CPUReservation != 0 {
resourceRequests[core.ResourceCPU] = *resource.NewMilliQuantity(CPUReservation, resource.DecimalSI)
}
}
serviceContainer.Resources.Requests = resourceRequests
}
}
// HealthCheck
if composeServiceConfig.HealthCheck != nil && !composeServiceConfig.HealthCheck.Disable {
probe, err := c.getHealthCheck(*composeServiceConfig.HealthCheck)
if err != nil {
logrus.Warnf("Unable to parse health check : %s", err)
} else {
serviceContainer.LivenessProbe = &probe
}
}
restart := composeServiceConfig.Restart
if composeServiceConfig.Deploy.RestartPolicy != nil {
restart = composeServiceConfig.Deploy.RestartPolicy.Condition
}
if restart == "unless-stopped" {
logrus.Warnf("Restart policy 'unless-stopped' in service %s is not supported, convert it to 'always'", name)
serviceConfig.RestartPolicy = core.RestartPolicyAlways
}
// replicas:
if composeServiceConfig.Deploy.Replicas != nil {
serviceConfig.Replicas = int(*composeServiceConfig.Deploy.Replicas)
}
serviceContainer.Env = c.getEnvs(composeServiceConfig)
vml, vl := makeVolumesFromTmpFS(name, composeServiceConfig.Tmpfs)
for _, v := range vl {
serviceConfig.AddVolume(v)
}
serviceContainer.VolumeMounts = append(serviceContainer.VolumeMounts, vml...)
for _, secret := range composeServiceConfig.Secrets {
target := filepath.Join(defaultSecretBasePath, secret.Source)
src := secret.Source
if secret.Target != "" {
tokens := strings.Split(secret.Source, "/")
var prefix string
if !strings.HasPrefix(secret.Target, "/") {
prefix = defaultSecretBasePath + "/"
}
if tokens[len(tokens)-1] == secret.Target {
target = prefix + secret.Source
} else {
target = prefix + strings.TrimSuffix(secret.Target, "/"+tokens[len(tokens)-1])
}
src = tokens[len(tokens)-1]
}
vSrc := core.VolumeSource{
Secret: &core.SecretVolumeSource{
SecretName: secret.Source,
Items: []core.KeyToPath{{
Key: secret.Source,
Path: src,
}},
},
}
if secret.Mode != nil {
mode := cast.ToInt32(*secret.Mode)
vSrc.Secret.DefaultMode = &mode
}
serviceConfig.AddVolume(core.Volume{
Name: secret.Source,
VolumeSource: vSrc,
})
serviceContainer.VolumeMounts = append(serviceContainer.VolumeMounts, core.VolumeMount{
Name: | convertToIR | identifier_name |
fmt.rs | return Err(err);
}
let result = func(writer);
writer.reset()?;
result
}
/// Adapter struct implementing [Write] over types implementing [WriteColor]
pub struct Termcolor<W>(pub W);
impl<W> Write for Termcolor<W>
where
W: WriteColor,
{
fn write_str(&mut self, elements: &MarkupElements, content: &str) -> io::Result<()> {
with_format(&mut self.0, elements, |writer| {
let mut adapter = SanitizeAdapter {
writer,
error: Ok(()),
};
match adapter.write_str(content) {
Ok(()) => Ok(()),
Err(..) => {
if adapter.error.is_err() {
adapter.error
} else {
// SanitizeAdapter can only fail if the underlying
// writer returns an error
unreachable!()
}
}
}
})
}
fn write_fmt(&mut self, elements: &MarkupElements, content: fmt::Arguments) -> io::Result<()> {
with_format(&mut self.0, elements, |writer| {
let mut adapter = SanitizeAdapter {
writer,
error: Ok(()),
};
match adapter.write_fmt(content) {
Ok(()) => Ok(()),
Err(..) => {
if adapter.error.is_err() {
adapter.error
} else {
Err(io::Error::new(
io::ErrorKind::Other,
"a Display formatter returned an error",
))
}
}
}
})
}
}
/// Adapter [fmt::Write] calls to [io::Write] with sanitization,
/// implemented as an internal struct to avoid exposing [fmt::Write] on
/// [Termcolor]
struct SanitizeAdapter<W> {
writer: W,
error: io::Result<()>,
}
impl<W: io::Write> fmt::Write for SanitizeAdapter<W> {
fn write_str(&mut self, content: &str) -> fmt::Result {
let mut buffer = [0; 4];
for item in content.chars() {
// Replace non-whitespace, zero-width characters with the Unicode replacement character
let is_whitespace = item.is_whitespace();
let is_zero_width = UnicodeWidthChar::width(item).map_or(true, |width| width == 0);
let item = if !is_whitespace && is_zero_width {
char::REPLACEMENT_CHARACTER
} else {
item
};
item.encode_utf8(&mut buffer);
if let Err(err) = self.writer.write_all(&buffer[..item.len_utf8()]) {
self.error = Err(err);
return Err(fmt::Error);
}
}
Ok(())
}
}
/// The [Formatter] is the `rome_console` equivalent to [std::fmt::Formatter]:
/// it's never constructed directly by consumers, and can only be used through
/// the mutable reference passed to implementations of the [Display] trait).
/// It manages the state of the markup to print, and implementations of
/// [Display] can call into its methods to append content into the current
/// printing session
pub struct Formatter<'fmt> {
/// Stack of markup elements currently applied to the text being printed
state: MarkupElements<'fmt>,
/// Inner IO writer this [Formatter] will print text into
writer: &'fmt mut dyn Write,
}
impl<'fmt> Formatter<'fmt> {
/// Create a new instance of the [Formatter] using the provided `writer` for printing
pub fn new(writer: &'fmt mut dyn Write) -> Self {
Self {
state: MarkupElements::Root,
writer,
}
}
/// Return a new instance of the [Formatter] with `elements` appended to its element stack
fn with_elements<'b>(&'b mut self, elements: &'b [MarkupElement]) -> Formatter<'b> {
Formatter {
state: MarkupElements::Node(&self.state, elements),
writer: self.writer,
}
}
/// Write a piece of markup into this formatter
pub fn write_markup(&mut self, markup: Markup) -> io::Result<()> {
for node in markup.0 {
let mut fmt = self.with_elements(node.elements);
node.content.fmt(&mut fmt)?;
}
Ok(())
}
/// Write a slice of text into this formatter
pub fn write_str(&mut self, content: &str) -> io::Result<()> {
self.writer.write_str(&self.state, content)
}
/// Write formatted text into this formatter
pub fn write_fmt(&mut self, content: fmt::Arguments) -> io::Result<()> {
self.writer.write_fmt(&self.state, content)
}
}
/// Formatting trait for types to be displayed as markup, the `rome_console`
/// equivalent to [std::fmt::Display]
///
/// # Example
/// Implementing `Display` on a custom struct
/// ```
/// use std::io;
/// use rome_console::{fmt::{Display, Formatter}, markup};
///
/// struct Warning(String);
///
/// impl Display for Warning {
/// fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
/// fmt.write_markup(markup! {
/// <Warn>{self.0}</Warn>
/// })
/// }
/// }
///
/// let warning = Warning(String::from("content"));
/// markup! {
/// <Emphasis>{warning}</Emphasis>
/// };
/// ```
pub trait Display {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()>;
}
// Blanket implementations of Display for reference types
impl<'a, T> Display for &'a T
where
T: Display + ?Sized,
{
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
T::fmt(self, fmt)
}
}
impl<'a, T> Display for Cow<'a, T>
where
T: Display + ToOwned + ?Sized,
{
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
T::fmt(self, fmt)
}
}
// Simple implementations of Display calling through to write_str for types
// that implement Deref<str>
impl Display for str {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
fmt.write_str(self)
}
}
impl Display for String {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
fmt.write_str(self)
}
}
// Implement Display for Markup and Rust format Arguments
impl<'a> Display for Markup<'a> {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
fmt.write_markup(*self)
}
}
impl<'a> Display for std::fmt::Arguments<'a> {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
fmt.write_fmt(*self)
}
}
/// Implement [Display] for types that implement [std::fmt::Display] by calling
/// through to [Formatter::write_fmt]
macro_rules! impl_std_display {
($ty:ty) => {
impl Display for $ty {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
write!(fmt, "{self}")
}
}
};
}
impl_std_display!(char);
impl_std_display!(i8);
impl_std_display!(i16);
impl_std_display!(i32);
impl_std_display!(i64);
impl_std_display!(i128);
impl_std_display!(isize);
impl_std_display!(u8);
impl_std_display!(u16);
impl_std_display!(u32);
impl_std_display!(u64);
impl_std_display!(u128);
impl_std_display!(usize);
impl Display for Duration {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
use crate as rome_console;
let secs = self.as_secs();
if secs > 1 {
return fmt.write_markup(markup! {
{secs}<Dim>"s"</Dim>
});
}
let millis = self.as_millis();
if millis > 1 {
return fmt.write_markup(markup! {
{millis}<Dim>"ms"</Dim>
});
}
let micros = self.as_micros();
if micros > 1 {
return fmt.write_markup(markup! {
{micros}<Dim>"µs"</Dim>
});
}
let nanos = self.as_nanos();
fmt.write_markup(markup! {
{nanos}<Dim>"ns"</Dim>
})
}
}
#[cfg(test)]
mod tests {
use std::{fmt::Write, str::from_utf8};
use super::SanitizeAdapter;
#[test]
fn test_sanitize() {
// Sanitization should leave whitespace control characters (space,
// tabs, newline, ...) and non-ASCII unicode characters as-is but
// redact zero-width characters (RTL override, null character, bell,
// zero-width space, ...) | const INPUT: &str = "t\tes t\r\n\u{202D}t\0es\x07t\u{202E}\nt\u{200B}es🐛t";
const OUTPUT: &str = "t\tes t\r\n\u{FFFD}t\u{FFFD}es\u{FFFD}t\u{FFFD}\nt\u{FFFD}es🐛t";
let mut buffer = Vec::new();
| random_line_split |
|
fmt.rs | impl FnMut(&'a [MarkupElement])) {
if let Self::Node(parent, elem) = self {
parent.for_each(func);
func(elem);
}
}
}
pub trait Write {
fn write_str(&mut self, elements: &MarkupElements, content: &str) -> io::Result<()>;
fn write_fmt(&mut self, elements: &MarkupElements, content: fmt::Arguments) -> io::Result<()>;
}
/// Applies the current format in `state` to `writer`, calls `func` to
/// print a piece of text, then reset the printing format
fn with_format<W>(
writer: &mut W,
state: &MarkupElements,
func: impl FnOnce(&mut W) -> io::Result<()>,
) -> io::Result<()>
where
W: WriteColor,
{
let mut color = ColorSpec::new();
state.for_each(&mut |elements| {
for element in elements {
element.update_color(&mut color);
}
});
if let Err(err) = writer.set_color(&color) {
writer.reset()?;
return Err(err);
}
let result = func(writer);
writer.reset()?;
result
}
/// Adapter struct implementing [Write] over types implementing [WriteColor]
pub struct Termcolor<W>(pub W);
impl<W> Write for Termcolor<W>
where
W: WriteColor,
{
fn write_str(&mut self, elements: &MarkupElements, content: &str) -> io::Result<()> {
with_format(&mut self.0, elements, |writer| {
let mut adapter = SanitizeAdapter {
writer,
error: Ok(()),
};
match adapter.write_str(content) {
Ok(()) => Ok(()),
Err(..) => {
if adapter.error.is_err() {
adapter.error
} else {
// SanitizeAdapter can only fail if the underlying
// writer returns an error
unreachable!()
}
}
}
})
}
fn write_fmt(&mut self, elements: &MarkupElements, content: fmt::Arguments) -> io::Result<()> {
with_format(&mut self.0, elements, |writer| {
let mut adapter = SanitizeAdapter {
writer,
error: Ok(()),
};
match adapter.write_fmt(content) {
Ok(()) => Ok(()),
Err(..) => {
if adapter.error.is_err() {
adapter.error
} else {
Err(io::Error::new(
io::ErrorKind::Other,
"a Display formatter returned an error",
))
}
}
}
})
}
}
/// Adapter [fmt::Write] calls to [io::Write] with sanitization,
/// implemented as an internal struct to avoid exposing [fmt::Write] on
/// [Termcolor]
struct SanitizeAdapter<W> {
writer: W,
error: io::Result<()>,
}
impl<W: io::Write> fmt::Write for SanitizeAdapter<W> {
fn write_str(&mut self, content: &str) -> fmt::Result {
let mut buffer = [0; 4];
for item in content.chars() {
// Replace non-whitespace, zero-width characters with the Unicode replacement character
let is_whitespace = item.is_whitespace();
let is_zero_width = UnicodeWidthChar::width(item).map_or(true, |width| width == 0);
let item = if !is_whitespace && is_zero_width {
char::REPLACEMENT_CHARACTER
} else {
item
};
item.encode_utf8(&mut buffer);
if let Err(err) = self.writer.write_all(&buffer[..item.len_utf8()]) {
self.error = Err(err);
return Err(fmt::Error);
}
}
Ok(())
}
}
/// The [Formatter] is the `rome_console` equivalent to [std::fmt::Formatter]:
/// it's never constructed directly by consumers, and can only be used through
/// the mutable reference passed to implementations of the [Display] trait).
/// It manages the state of the markup to print, and implementations of
/// [Display] can call into its methods to append content into the current
/// printing session
pub struct Formatter<'fmt> {
/// Stack of markup elements currently applied to the text being printed
state: MarkupElements<'fmt>,
/// Inner IO writer this [Formatter] will print text into
writer: &'fmt mut dyn Write,
}
impl<'fmt> Formatter<'fmt> {
/// Create a new instance of the [Formatter] using the provided `writer` for printing
pub fn new(writer: &'fmt mut dyn Write) -> Self {
Self {
state: MarkupElements::Root,
writer,
}
}
/// Return a new instance of the [Formatter] with `elements` appended to its element stack
fn with_elements<'b>(&'b mut self, elements: &'b [MarkupElement]) -> Formatter<'b> |
/// Write a piece of markup into this formatter
pub fn write_markup(&mut self, markup: Markup) -> io::Result<()> {
for node in markup.0 {
let mut fmt = self.with_elements(node.elements);
node.content.fmt(&mut fmt)?;
}
Ok(())
}
/// Write a slice of text into this formatter
pub fn write_str(&mut self, content: &str) -> io::Result<()> {
self.writer.write_str(&self.state, content)
}
/// Write formatted text into this formatter
pub fn write_fmt(&mut self, content: fmt::Arguments) -> io::Result<()> {
self.writer.write_fmt(&self.state, content)
}
}
/// Formatting trait for types to be displayed as markup, the `rome_console`
/// equivalent to [std::fmt::Display]
///
/// # Example
/// Implementing `Display` on a custom struct
/// ```
/// use std::io;
/// use rome_console::{fmt::{Display, Formatter}, markup};
///
/// struct Warning(String);
///
/// impl Display for Warning {
/// fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
/// fmt.write_markup(markup! {
/// <Warn>{self.0}</Warn>
/// })
/// }
/// }
///
/// let warning = Warning(String::from("content"));
/// markup! {
/// <Emphasis>{warning}</Emphasis>
/// };
/// ```
pub trait Display {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()>;
}
// Blanket implementations of Display for reference types
impl<'a, T> Display for &'a T
where
T: Display + ?Sized,
{
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
T::fmt(self, fmt)
}
}
impl<'a, T> Display for Cow<'a, T>
where
T: Display + ToOwned + ?Sized,
{
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
T::fmt(self, fmt)
}
}
// Simple implementations of Display calling through to write_str for types
// that implement Deref<str>
impl Display for str {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
fmt.write_str(self)
}
}
impl Display for String {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
fmt.write_str(self)
}
}
// Implement Display for Markup and Rust format Arguments
impl<'a> Display for Markup<'a> {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
fmt.write_markup(*self)
}
}
impl<'a> Display for std::fmt::Arguments<'a> {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
fmt.write_fmt(*self)
}
}
/// Implement [Display] for types that implement [std::fmt::Display] by calling
/// through to [Formatter::write_fmt]
macro_rules! impl_std_display {
($ty:ty) => {
impl Display for $ty {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
write!(fmt, "{self}")
}
}
};
}
impl_std_display!(char);
impl_std_display!(i8);
impl_std_display!(i16);
impl_std_display!(i32);
impl_std_display!(i64);
impl_std_display!(i128);
impl_std_display!(isize);
impl_std_display!(u8);
impl_std_display!(u16);
impl_std_display!(u32);
impl_std_display!(u64);
impl_std_display!(u128);
impl_std_display!(usize);
impl Display for Duration {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
use crate as rome_console;
let secs = self.as_secs();
if secs > 1 {
return fmt.write_markup(markup! {
{secs}<Dim>"s"</Dim>
});
}
let millis = self.as_millis();
if millis > 1 {
return fmt.write_markup(markup! {
{millis}<Dim>"ms"</Dim>
});
}
let micros = self.as_micros();
if micros > 1 {
return fmt.write_markup(markup! {
{micros}<Dim>"µ | {
Formatter {
state: MarkupElements::Node(&self.state, elements),
writer: self.writer,
}
} | identifier_body |
fmt.rs | mut impl FnMut(&'a [MarkupElement])) {
if let Self::Node(parent, elem) = self {
parent.for_each(func);
func(elem);
}
}
}
pub trait Write {
fn write_str(&mut self, elements: &MarkupElements, content: &str) -> io::Result<()>;
fn write_fmt(&mut self, elements: &MarkupElements, content: fmt::Arguments) -> io::Result<()>;
}
/// Applies the current format in `state` to `writer`, calls `func` to
/// print a piece of text, then reset the printing format
fn with_format<W>(
writer: &mut W,
state: &MarkupElements,
func: impl FnOnce(&mut W) -> io::Result<()>,
) -> io::Result<()>
where
W: WriteColor,
{
let mut color = ColorSpec::new();
state.for_each(&mut |elements| {
for element in elements {
element.update_color(&mut color);
}
});
if let Err(err) = writer.set_color(&color) {
writer.reset()?;
return Err(err);
}
let result = func(writer);
writer.reset()?;
result
}
/// Adapter struct implementing [Write] over types implementing [WriteColor]
pub struct Termcolor<W>(pub W);
impl<W> Write for Termcolor<W>
where
W: WriteColor,
{
fn write_str(&mut self, elements: &MarkupElements, content: &str) -> io::Result<()> {
with_format(&mut self.0, elements, |writer| {
let mut adapter = SanitizeAdapter {
writer,
error: Ok(()),
};
match adapter.write_str(content) {
Ok(()) => Ok(()),
Err(..) => {
if adapter.error.is_err() | else {
// SanitizeAdapter can only fail if the underlying
// writer returns an error
unreachable!()
}
}
}
})
}
fn write_fmt(&mut self, elements: &MarkupElements, content: fmt::Arguments) -> io::Result<()> {
with_format(&mut self.0, elements, |writer| {
let mut adapter = SanitizeAdapter {
writer,
error: Ok(()),
};
match adapter.write_fmt(content) {
Ok(()) => Ok(()),
Err(..) => {
if adapter.error.is_err() {
adapter.error
} else {
Err(io::Error::new(
io::ErrorKind::Other,
"a Display formatter returned an error",
))
}
}
}
})
}
}
/// Adapter [fmt::Write] calls to [io::Write] with sanitization,
/// implemented as an internal struct to avoid exposing [fmt::Write] on
/// [Termcolor]
struct SanitizeAdapter<W> {
writer: W,
error: io::Result<()>,
}
impl<W: io::Write> fmt::Write for SanitizeAdapter<W> {
fn write_str(&mut self, content: &str) -> fmt::Result {
let mut buffer = [0; 4];
for item in content.chars() {
// Replace non-whitespace, zero-width characters with the Unicode replacement character
let is_whitespace = item.is_whitespace();
let is_zero_width = UnicodeWidthChar::width(item).map_or(true, |width| width == 0);
let item = if !is_whitespace && is_zero_width {
char::REPLACEMENT_CHARACTER
} else {
item
};
item.encode_utf8(&mut buffer);
if let Err(err) = self.writer.write_all(&buffer[..item.len_utf8()]) {
self.error = Err(err);
return Err(fmt::Error);
}
}
Ok(())
}
}
/// The [Formatter] is the `rome_console` equivalent to [std::fmt::Formatter]:
/// it's never constructed directly by consumers, and can only be used through
/// the mutable reference passed to implementations of the [Display] trait).
/// It manages the state of the markup to print, and implementations of
/// [Display] can call into its methods to append content into the current
/// printing session
pub struct Formatter<'fmt> {
/// Stack of markup elements currently applied to the text being printed
state: MarkupElements<'fmt>,
/// Inner IO writer this [Formatter] will print text into
writer: &'fmt mut dyn Write,
}
impl<'fmt> Formatter<'fmt> {
/// Create a new instance of the [Formatter] using the provided `writer` for printing
pub fn new(writer: &'fmt mut dyn Write) -> Self {
Self {
state: MarkupElements::Root,
writer,
}
}
/// Return a new instance of the [Formatter] with `elements` appended to its element stack
fn with_elements<'b>(&'b mut self, elements: &'b [MarkupElement]) -> Formatter<'b> {
Formatter {
state: MarkupElements::Node(&self.state, elements),
writer: self.writer,
}
}
/// Write a piece of markup into this formatter
pub fn write_markup(&mut self, markup: Markup) -> io::Result<()> {
for node in markup.0 {
let mut fmt = self.with_elements(node.elements);
node.content.fmt(&mut fmt)?;
}
Ok(())
}
/// Write a slice of text into this formatter
pub fn write_str(&mut self, content: &str) -> io::Result<()> {
self.writer.write_str(&self.state, content)
}
/// Write formatted text into this formatter
pub fn write_fmt(&mut self, content: fmt::Arguments) -> io::Result<()> {
self.writer.write_fmt(&self.state, content)
}
}
/// Formatting trait for types to be displayed as markup, the `rome_console`
/// equivalent to [std::fmt::Display]
///
/// # Example
/// Implementing `Display` on a custom struct
/// ```
/// use std::io;
/// use rome_console::{fmt::{Display, Formatter}, markup};
///
/// struct Warning(String);
///
/// impl Display for Warning {
/// fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
/// fmt.write_markup(markup! {
/// <Warn>{self.0}</Warn>
/// })
/// }
/// }
///
/// let warning = Warning(String::from("content"));
/// markup! {
/// <Emphasis>{warning}</Emphasis>
/// };
/// ```
pub trait Display {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()>;
}
// Blanket implementations of Display for reference types
impl<'a, T> Display for &'a T
where
T: Display + ?Sized,
{
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
T::fmt(self, fmt)
}
}
impl<'a, T> Display for Cow<'a, T>
where
T: Display + ToOwned + ?Sized,
{
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
T::fmt(self, fmt)
}
}
// Simple implementations of Display calling through to write_str for types
// that implement Deref<str>
impl Display for str {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
fmt.write_str(self)
}
}
impl Display for String {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
fmt.write_str(self)
}
}
// Implement Display for Markup and Rust format Arguments
impl<'a> Display for Markup<'a> {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
fmt.write_markup(*self)
}
}
impl<'a> Display for std::fmt::Arguments<'a> {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
fmt.write_fmt(*self)
}
}
/// Implement [Display] for types that implement [std::fmt::Display] by calling
/// through to [Formatter::write_fmt]
macro_rules! impl_std_display {
($ty:ty) => {
impl Display for $ty {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
write!(fmt, "{self}")
}
}
};
}
impl_std_display!(char);
impl_std_display!(i8);
impl_std_display!(i16);
impl_std_display!(i32);
impl_std_display!(i64);
impl_std_display!(i128);
impl_std_display!(isize);
impl_std_display!(u8);
impl_std_display!(u16);
impl_std_display!(u32);
impl_std_display!(u64);
impl_std_display!(u128);
impl_std_display!(usize);
impl Display for Duration {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
use crate as rome_console;
let secs = self.as_secs();
if secs > 1 {
return fmt.write_markup(markup! {
{secs}<Dim>"s"</Dim>
});
}
let millis = self.as_millis();
if millis > 1 {
return fmt.write_markup(markup! {
{millis}<Dim>"ms"</Dim>
});
}
let micros = self.as_micros();
if micros > 1 {
return fmt.write_markup(markup! {
{micros}<Dim>"µ | {
adapter.error
} | conditional_block |
fmt.rs | mut impl FnMut(&'a [MarkupElement])) {
if let Self::Node(parent, elem) = self {
parent.for_each(func);
func(elem);
}
}
}
pub trait Write {
fn write_str(&mut self, elements: &MarkupElements, content: &str) -> io::Result<()>;
fn write_fmt(&mut self, elements: &MarkupElements, content: fmt::Arguments) -> io::Result<()>;
}
/// Applies the current format in `state` to `writer`, calls `func` to
/// print a piece of text, then reset the printing format
fn with_format<W>(
writer: &mut W,
state: &MarkupElements,
func: impl FnOnce(&mut W) -> io::Result<()>,
) -> io::Result<()>
where
W: WriteColor,
{
let mut color = ColorSpec::new();
state.for_each(&mut |elements| {
for element in elements {
element.update_color(&mut color);
}
});
if let Err(err) = writer.set_color(&color) {
writer.reset()?;
return Err(err);
}
let result = func(writer);
writer.reset()?;
result
}
/// Adapter struct implementing [Write] over types implementing [WriteColor]
pub struct Termcolor<W>(pub W);
impl<W> Write for Termcolor<W>
where
W: WriteColor,
{
fn write_str(&mut self, elements: &MarkupElements, content: &str) -> io::Result<()> {
with_format(&mut self.0, elements, |writer| {
let mut adapter = SanitizeAdapter {
writer,
error: Ok(()),
};
match adapter.write_str(content) {
Ok(()) => Ok(()),
Err(..) => {
if adapter.error.is_err() {
adapter.error
} else {
// SanitizeAdapter can only fail if the underlying
// writer returns an error
unreachable!()
}
}
}
})
}
fn write_fmt(&mut self, elements: &MarkupElements, content: fmt::Arguments) -> io::Result<()> {
with_format(&mut self.0, elements, |writer| {
let mut adapter = SanitizeAdapter {
writer,
error: Ok(()),
};
match adapter.write_fmt(content) {
Ok(()) => Ok(()),
Err(..) => {
if adapter.error.is_err() {
adapter.error
} else {
Err(io::Error::new(
io::ErrorKind::Other,
"a Display formatter returned an error",
))
}
}
}
})
}
}
/// Adapter [fmt::Write] calls to [io::Write] with sanitization,
/// implemented as an internal struct to avoid exposing [fmt::Write] on
/// [Termcolor]
struct SanitizeAdapter<W> {
writer: W,
error: io::Result<()>,
}
impl<W: io::Write> fmt::Write for SanitizeAdapter<W> {
fn | (&mut self, content: &str) -> fmt::Result {
let mut buffer = [0; 4];
for item in content.chars() {
// Replace non-whitespace, zero-width characters with the Unicode replacement character
let is_whitespace = item.is_whitespace();
let is_zero_width = UnicodeWidthChar::width(item).map_or(true, |width| width == 0);
let item = if !is_whitespace && is_zero_width {
char::REPLACEMENT_CHARACTER
} else {
item
};
item.encode_utf8(&mut buffer);
if let Err(err) = self.writer.write_all(&buffer[..item.len_utf8()]) {
self.error = Err(err);
return Err(fmt::Error);
}
}
Ok(())
}
}
/// The [Formatter] is the `rome_console` equivalent to [std::fmt::Formatter]:
/// it's never constructed directly by consumers, and can only be used through
/// the mutable reference passed to implementations of the [Display] trait).
/// It manages the state of the markup to print, and implementations of
/// [Display] can call into its methods to append content into the current
/// printing session
pub struct Formatter<'fmt> {
/// Stack of markup elements currently applied to the text being printed
state: MarkupElements<'fmt>,
/// Inner IO writer this [Formatter] will print text into
writer: &'fmt mut dyn Write,
}
impl<'fmt> Formatter<'fmt> {
/// Create a new instance of the [Formatter] using the provided `writer` for printing
pub fn new(writer: &'fmt mut dyn Write) -> Self {
Self {
state: MarkupElements::Root,
writer,
}
}
/// Return a new instance of the [Formatter] with `elements` appended to its element stack
fn with_elements<'b>(&'b mut self, elements: &'b [MarkupElement]) -> Formatter<'b> {
Formatter {
state: MarkupElements::Node(&self.state, elements),
writer: self.writer,
}
}
/// Write a piece of markup into this formatter
pub fn write_markup(&mut self, markup: Markup) -> io::Result<()> {
for node in markup.0 {
let mut fmt = self.with_elements(node.elements);
node.content.fmt(&mut fmt)?;
}
Ok(())
}
/// Write a slice of text into this formatter
pub fn write_str(&mut self, content: &str) -> io::Result<()> {
self.writer.write_str(&self.state, content)
}
/// Write formatted text into this formatter
pub fn write_fmt(&mut self, content: fmt::Arguments) -> io::Result<()> {
self.writer.write_fmt(&self.state, content)
}
}
/// Formatting trait for types to be displayed as markup, the `rome_console`
/// equivalent to [std::fmt::Display]
///
/// # Example
/// Implementing `Display` on a custom struct
/// ```
/// use std::io;
/// use rome_console::{fmt::{Display, Formatter}, markup};
///
/// struct Warning(String);
///
/// impl Display for Warning {
/// fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
/// fmt.write_markup(markup! {
/// <Warn>{self.0}</Warn>
/// })
/// }
/// }
///
/// let warning = Warning(String::from("content"));
/// markup! {
/// <Emphasis>{warning}</Emphasis>
/// };
/// ```
pub trait Display {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()>;
}
// Blanket implementations of Display for reference types
impl<'a, T> Display for &'a T
where
T: Display + ?Sized,
{
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
T::fmt(self, fmt)
}
}
impl<'a, T> Display for Cow<'a, T>
where
T: Display + ToOwned + ?Sized,
{
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
T::fmt(self, fmt)
}
}
// Simple implementations of Display calling through to write_str for types
// that implement Deref<str>
impl Display for str {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
fmt.write_str(self)
}
}
impl Display for String {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
fmt.write_str(self)
}
}
// Implement Display for Markup and Rust format Arguments
impl<'a> Display for Markup<'a> {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
fmt.write_markup(*self)
}
}
impl<'a> Display for std::fmt::Arguments<'a> {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
fmt.write_fmt(*self)
}
}
/// Implement [Display] for types that implement [std::fmt::Display] by calling
/// through to [Formatter::write_fmt]
macro_rules! impl_std_display {
($ty:ty) => {
impl Display for $ty {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
write!(fmt, "{self}")
}
}
};
}
impl_std_display!(char);
impl_std_display!(i8);
impl_std_display!(i16);
impl_std_display!(i32);
impl_std_display!(i64);
impl_std_display!(i128);
impl_std_display!(isize);
impl_std_display!(u8);
impl_std_display!(u16);
impl_std_display!(u32);
impl_std_display!(u64);
impl_std_display!(u128);
impl_std_display!(usize);
impl Display for Duration {
fn fmt(&self, fmt: &mut Formatter) -> io::Result<()> {
use crate as rome_console;
let secs = self.as_secs();
if secs > 1 {
return fmt.write_markup(markup! {
{secs}<Dim>"s"</Dim>
});
}
let millis = self.as_millis();
if millis > 1 {
return fmt.write_markup(markup! {
{millis}<Dim>"ms"</Dim>
});
}
let micros = self.as_micros();
if micros > 1 {
return fmt.write_markup(markup! {
{micros}<Dim>"µ | write_str | identifier_name |
serial.rs | {
fn eq(&self, other: &Self) -> bool
{
self.id == other.id
}
}
impl fmt::Debug for SerialQueue
{
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result
{
write!(f, "SerialQueue ({})", self.id)
}
}
unsafe impl Send for SerialQueue
{}
unsafe impl Sync for SerialQueue
{}
impl SerialQueue
{
/// Create a new SerialQueue and assign it to the global thread pool
pub fn new() -> SerialQueue
{
let (tx, rx) = mailbox();
let (deadlock_tx, deadlock_rx) = mailbox();
let internal_tx = tx.clone();
new_coroutine(move || {
SerialQueue::do_loop(internal_tx, rx, deadlock_rx);
Ok(())
});
let queue = SerialQueue {
id: ID.fetch_add(1, Ordering::SeqCst),
tx: tx,
deadlock_tx: deadlock_tx,
};
info!("Queue created ({:?})", queue);
queue
}
/// Create a new SerialQueue and assign it solely to a newly created OS Thread
///
/// A SerialQueue created through this method will spawn a new native OS Thread
/// and the queue will be the only utilizing it. The thread will be destructed,
/// when the queue is dropped.
///
/// The purpose of this constructor is to provide a way to use blocking IO with TaskQueue.
/// The use of this method however is discouraged, as the new thread may influence
/// the scheduler negatively and evented IO, where possible performs a lot better in combination
/// with the TaskQueue library
pub fn new_native() -> SerialQueue
{
let (tx, rx) = mailbox();
let (deadlock_tx, deadlock_rx) = mailbox();
let internal_tx = tx.clone();
thread::spawn(move || {
Mioco::new_configured({
let mut config = Config::new();
config.set_thread_num(1);
config
}).start(move || {
SerialQueue::do_loop(internal_tx, rx, deadlock_rx);
Ok(())
});
});
let queue = SerialQueue {
id: ID.fetch_add(1, Ordering::SeqCst),
tx: tx,
deadlock_tx: deadlock_tx,
};
info!("Native Queue created ({:?})", queue);
queue
}
fn do_loop(queue_tx: MailboxOuterEnd<Command>,
rx: MailboxInnerEnd<Command>,
deadlock_rx: MailboxInnerEnd<()>)
{
debug!("loop: spawing serial loop");
loop {
trace!("loop: next iteration");
match rx.read() {
Command::End => break,
Command::Wait(routine) => {
trace!("loop: handling previous deadlocked coroutine");
let tx_clone = queue_tx.clone();
loop {
select!(
routine:r => {
if routine.try_read().is_some() {
trace!("loop: task ended");
break;
} else {
continue;
}
},
deadlock_rx:r => {
if deadlock_rx.try_read().is_some() {
trace!("loop: deadlock detected");
tx_clone.send(Command::Wait(routine));
break;
} else {
continue;
}
},
);
}
}
Command::Run(task) => {
let tx_clone = queue_tx.clone();
mioco::set_children_userdata(Some(Userdata::SameThread));
let routine = mioco::spawn_ext(move || {
trace!("loop: spawned new coroutine for task");
task.call_box(());
Ok(())
})
.exit_notificator();
trace!("loop: wait for deadlock notification or coroutine finish");
loop {
select!(
routine:r => {
if routine.try_read().is_some() {
trace!("loop: task ended");
break;
} else {
continue;
}
},
deadlock_rx:r => {
if deadlock_rx.try_read().is_some() {
trace!("loop: deadlock detected");
tx_clone.send(Command::Wait(routine));
break;
} else {
continue;
}
},
);
}
}
}
}
debug!("loop: queue ended");
}
/// Bind this queue to a variable
///
/// This function allows to create a `BoundSerialQueue`.
/// Its purpose is to bind variables to a queue, so they can be used by the tasks submitted.
///
/// # Example
/// ```
/// # use taskqueue::*;
/// # let queue = SerialQueue::new();
/// let x = 5;
/// let bound = queue.with(move || x);
/// bound.scoped_with(|x| println!("{}", x));
/// // x gets dropped inside the queues thread before the queue gets dropped
/// ```
///
/// You can create multiple bindings at once.
/// Through tuples you may bind multiple variables at once.
///
/// It is even possible to move the creation of the bound variable into the queue by creating
/// it inside the passed constructor, which is then executed on the queue.
/// And because SerialQueues never change their underlying OS Thread,
/// this allows to use variables that are not Send and Sync in a thread-safe but shared way.
///
/// # Example
/// ```
/// # use taskqueue::*;
/// # let queue = SerialQueue::new();
/// # fn my_ffi_function() -> *mut () { &mut () };
/// let bound = queue.with(|| {
/// let raw_ptr = my_ffi_function();
/// raw_ptr
/// });
/// bound.scoped_with(|raw_ptr| println!("{}", raw_ptr.is_null()));
/// // raw_ptr gets dropped inside the queues thread.
/// // This way raw_ptr is never moved between threads.
/// ```
pub fn with<'queue, R: 'static, F>(&'queue self,
constructor: F)
-> BoundSerialQueue<'queue, R>
where F: FnOnce() -> R + Send
{
let binding = self.sync(move || NotThreadSafe::new(constructor()));
BoundSerialQueue {
queue: self,
binding: binding,
}
}
}
impl Queue for SerialQueue
{
fn async<R, F>(&self, operation: F) -> Future<R>
where R: Send + 'static,
F: FnOnce() -> R + Send + 'static
{
let (tx, rx) = mailbox();
let operation: Box<FnBox() + Send + 'static> = Stack::assemble(self, move || {
tx.send(operation());
});
debug!("Queue ({:?}) queued task", self);
self.tx.send(Command::Run(operation));
Future::new_from_serial(self, Some(self.deadlock_tx.clone()), rx)
}
}
/// A bound SerialQueue holding a queue-bound variable
///
/// Create a BoundSerialQueue using `SerialQueue::with`.
/// BoundSerialQueue's hold variables that may be used through
/// tasks executed on this queue though `scoped_with`, `sync_with`
/// `foreach_with` or `loop_while_with`.
///
/// `async_with` cannot be provided, as the bound variable is
/// dropped, when the BoundSerialQueue gets dropped.
///
/// Internally BoundSerialQueue refer to the same queue, they were created from.
/// Multiple BoundSerialQueue's may exist for one queue at once.
pub struct BoundSerialQueue<'queue, T: 'static>
{
queue: &'queue SerialQueue,
binding: NotThreadSafe<T>,
}
impl<'queue, T: 'static> fmt::Debug for BoundSerialQueue<'queue, T>
{
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result
{
write!(f, "BoundSerialQueue ({:p})", self)
}
}
impl<'queue, T: 'static> BoundSerialQueue<'queue, T>
{
/// Like `Queue::scoped` but provides a mutable reference to the bound variable
///
/// # Safety
///
/// The same rules as for `Queue::scoped` to apply.
///
/// # Example
///
/// ```
/// # use taskqueue::*;
/// # let queue = SerialQueue::new();
/// let bound = queue.with(|| { "Hello".to_string() });
/// let name = "Stack".to_string();
/// bound.scoped_with(|message| { println!("{} {}!", message, name) });
/// ```
pub fn scoped_with<R, F>(&'queue self, operation: F) -> FutureGuard<R>
where R: Send + 'static,
F: FnOnce(&'queue mut T) -> R + Send + 'queue
{
self.queue.scoped(move || operation(unsafe { self.binding.get_mut() }))
}
/// Like `Queue::sync` but provides a mutable reference to the bound variable
///
/// # Example
///
/// ```
/// # use taskqueue::*;
/// # let queue = SerialQueue::new();
/// let bound = queue.with(|| { "a bound queue".to_string() });
/// bound.sync_with(|name| { println!("Hello {}", name) });
/// ```
pub fn sync_with<R, F>(&'queue self, operation: F) -> R
where R: Send + 'static,
F: FnOnce(&'queue mut T) -> R + Send + 'queue
| {
self.queue.sync(move || operation(unsafe { self.binding.get_mut() }))
} | identifier_body |
|
serial.rs | this task is running before...");
/// });
/// main.sync(|| {
/// println!("...this task and...");
/// assert_eq!(future_one.get() + future_two.get(), 138);
/// });
/// println!("...this is running last");
/// });
/// ```
pub struct SerialQueue
{
id: usize,
tx: MailboxOuterEnd<Command>,
deadlock_tx: MailboxOuterEnd<()>,
}
impl PartialEq for SerialQueue
{
fn eq(&self, other: &Self) -> bool
{
self.id == other.id
}
}
impl fmt::Debug for SerialQueue
{
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result
{
write!(f, "SerialQueue ({})", self.id)
}
}
unsafe impl Send for SerialQueue
{}
unsafe impl Sync for SerialQueue
{}
impl SerialQueue
{
/// Create a new SerialQueue and assign it to the global thread pool
pub fn new() -> SerialQueue
{
let (tx, rx) = mailbox();
let (deadlock_tx, deadlock_rx) = mailbox();
let internal_tx = tx.clone();
new_coroutine(move || {
SerialQueue::do_loop(internal_tx, rx, deadlock_rx);
Ok(())
});
let queue = SerialQueue {
id: ID.fetch_add(1, Ordering::SeqCst),
tx: tx,
deadlock_tx: deadlock_tx,
};
info!("Queue created ({:?})", queue);
queue
}
/// Create a new SerialQueue and assign it solely to a newly created OS Thread
///
/// A SerialQueue created through this method will spawn a new native OS Thread
/// and the queue will be the only utilizing it. The thread will be destructed,
/// when the queue is dropped.
///
/// The purpose of this constructor is to provide a way to use blocking IO with TaskQueue.
/// The use of this method however is discouraged, as the new thread may influence
/// the scheduler negatively and evented IO, where possible performs a lot better in combination
/// with the TaskQueue library
pub fn new_native() -> SerialQueue
{
let (tx, rx) = mailbox();
let (deadlock_tx, deadlock_rx) = mailbox();
let internal_tx = tx.clone();
thread::spawn(move || {
Mioco::new_configured({
let mut config = Config::new();
config.set_thread_num(1);
config
}).start(move || {
SerialQueue::do_loop(internal_tx, rx, deadlock_rx);
Ok(())
});
});
let queue = SerialQueue {
id: ID.fetch_add(1, Ordering::SeqCst),
tx: tx,
deadlock_tx: deadlock_tx,
};
info!("Native Queue created ({:?})", queue);
queue
}
fn do_loop(queue_tx: MailboxOuterEnd<Command>,
rx: MailboxInnerEnd<Command>,
deadlock_rx: MailboxInnerEnd<()>)
{
debug!("loop: spawing serial loop");
loop {
trace!("loop: next iteration");
match rx.read() {
Command::End => break,
Command::Wait(routine) => {
trace!("loop: handling previous deadlocked coroutine");
let tx_clone = queue_tx.clone();
loop {
select!(
routine:r => {
if routine.try_read().is_some() {
trace!("loop: task ended");
break;
} else {
continue;
}
},
deadlock_rx:r => {
if deadlock_rx.try_read().is_some() {
trace!("loop: deadlock detected");
tx_clone.send(Command::Wait(routine));
break;
} else {
continue;
}
},
);
}
}
Command::Run(task) => {
let tx_clone = queue_tx.clone();
mioco::set_children_userdata(Some(Userdata::SameThread));
let routine = mioco::spawn_ext(move || {
trace!("loop: spawned new coroutine for task");
task.call_box(());
Ok(())
})
.exit_notificator();
trace!("loop: wait for deadlock notification or coroutine finish");
loop {
select!(
routine:r => {
if routine.try_read().is_some() {
trace!("loop: task ended");
break;
} else {
continue;
}
},
deadlock_rx:r => {
if deadlock_rx.try_read().is_some() {
trace!("loop: deadlock detected");
tx_clone.send(Command::Wait(routine));
break;
} else {
continue;
}
},
);
}
}
}
}
debug!("loop: queue ended");
}
/// Bind this queue to a variable
///
/// This function allows to create a `BoundSerialQueue`.
/// Its purpose is to bind variables to a queue, so they can be used by the tasks submitted.
///
/// # Example
/// ```
/// # use taskqueue::*;
/// # let queue = SerialQueue::new();
/// let x = 5;
/// let bound = queue.with(move || x);
/// bound.scoped_with(|x| println!("{}", x));
/// // x gets dropped inside the queues thread before the queue gets dropped
/// ```
///
/// You can create multiple bindings at once.
/// Through tuples you may bind multiple variables at once.
///
/// It is even possible to move the creation of the bound variable into the queue by creating
/// it inside the passed constructor, which is then executed on the queue.
/// And because SerialQueues never change their underlying OS Thread,
/// this allows to use variables that are not Send and Sync in a thread-safe but shared way.
///
/// # Example
/// ```
/// # use taskqueue::*;
/// # let queue = SerialQueue::new();
/// # fn my_ffi_function() -> *mut () { &mut () };
/// let bound = queue.with(|| {
/// let raw_ptr = my_ffi_function();
/// raw_ptr
/// });
/// bound.scoped_with(|raw_ptr| println!("{}", raw_ptr.is_null()));
/// // raw_ptr gets dropped inside the queues thread.
/// // This way raw_ptr is never moved between threads.
/// ```
pub fn with<'queue, R: 'static, F>(&'queue self,
constructor: F)
-> BoundSerialQueue<'queue, R>
where F: FnOnce() -> R + Send
{
let binding = self.sync(move || NotThreadSafe::new(constructor()));
BoundSerialQueue {
queue: self,
binding: binding,
}
}
}
impl Queue for SerialQueue
{
fn | <R, F>(&self, operation: F) -> Future<R>
where R: Send + 'static,
F: FnOnce() -> R + Send + 'static
{
let (tx, rx) = mailbox();
let operation: Box<FnBox() + Send + 'static> = Stack::assemble(self, move || {
tx.send(operation());
});
debug!("Queue ({:?}) queued task", self);
self.tx.send(Command::Run(operation));
Future::new_from_serial(self, Some(self.deadlock_tx.clone()), rx)
}
}
/// A bound SerialQueue holding a queue-bound variable
///
/// Create a BoundSerialQueue using `SerialQueue::with`.
/// BoundSerialQueue's hold variables that may be used through
/// tasks executed on this queue though `scoped_with`, `sync_with`
/// `foreach_with` or `loop_while_with`.
///
/// `async_with` cannot be provided, as the bound variable is
/// dropped, when the BoundSerialQueue gets dropped.
///
/// Internally BoundSerialQueue refer to the same queue, they were created from.
/// Multiple BoundSerialQueue's may exist for one queue at once.
pub struct BoundSerialQueue<'queue, T: 'static>
{
queue: &'queue SerialQueue,
binding: NotThreadSafe<T>,
}
impl<'queue, T: 'static> fmt::Debug for BoundSerialQueue<'queue, T>
{
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result
{
write!(f, "BoundSerialQueue ({:p})", self)
}
}
impl<'queue, T: 'static> BoundSerialQueue<'queue, T>
{
/// Like `Queue::scoped` but provides a mutable reference to the bound variable
///
/// # Safety
///
/// The same rules as for `Queue::scoped` to apply.
///
/// # Example
///
/// ```
/// # use taskqueue::*;
/// # let queue = SerialQueue::new();
/// let bound = queue.with(|| { "Hello".to_string() });
/// let name = "Stack".to_string();
/// bound.scoped_with(|message| { println!("{} {}!", message, name) });
/// ```
pub fn scoped_with<R, F>(&'queue self, operation: F) -> FutureGuard<R>
where R: Send + 'static,
F: FnOnce(&'queue mut T) -> R + Send + 'queue
{
self.queue.scoped(move || operation(unsafe { self.binding.get_mut() }))
}
/// Like `Queue::sync` but provides a mutable reference to the bound variable
///
/// # Example
///
/// ```
/// # use taskqueue::*;
/// # let queue = SerialQueue::new();
/// let bound = | async | identifier_name |
serial.rs | this task is running before...");
/// });
/// main.sync(|| {
/// println!("...this task and...");
/// assert_eq!(future_one.get() + future_two.get(), 138);
/// });
/// println!("...this is running last");
/// });
/// ```
pub struct SerialQueue
{
id: usize,
tx: MailboxOuterEnd<Command>,
deadlock_tx: MailboxOuterEnd<()>,
}
impl PartialEq for SerialQueue
{
fn eq(&self, other: &Self) -> bool
{
self.id == other.id
}
} | {
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result
{
write!(f, "SerialQueue ({})", self.id)
}
}
unsafe impl Send for SerialQueue
{}
unsafe impl Sync for SerialQueue
{}
impl SerialQueue
{
/// Create a new SerialQueue and assign it to the global thread pool
pub fn new() -> SerialQueue
{
let (tx, rx) = mailbox();
let (deadlock_tx, deadlock_rx) = mailbox();
let internal_tx = tx.clone();
new_coroutine(move || {
SerialQueue::do_loop(internal_tx, rx, deadlock_rx);
Ok(())
});
let queue = SerialQueue {
id: ID.fetch_add(1, Ordering::SeqCst),
tx: tx,
deadlock_tx: deadlock_tx,
};
info!("Queue created ({:?})", queue);
queue
}
/// Create a new SerialQueue and assign it solely to a newly created OS Thread
///
/// A SerialQueue created through this method will spawn a new native OS Thread
/// and the queue will be the only utilizing it. The thread will be destructed,
/// when the queue is dropped.
///
/// The purpose of this constructor is to provide a way to use blocking IO with TaskQueue.
/// The use of this method however is discouraged, as the new thread may influence
/// the scheduler negatively and evented IO, where possible performs a lot better in combination
/// with the TaskQueue library
pub fn new_native() -> SerialQueue
{
let (tx, rx) = mailbox();
let (deadlock_tx, deadlock_rx) = mailbox();
let internal_tx = tx.clone();
thread::spawn(move || {
Mioco::new_configured({
let mut config = Config::new();
config.set_thread_num(1);
config
}).start(move || {
SerialQueue::do_loop(internal_tx, rx, deadlock_rx);
Ok(())
});
});
let queue = SerialQueue {
id: ID.fetch_add(1, Ordering::SeqCst),
tx: tx,
deadlock_tx: deadlock_tx,
};
info!("Native Queue created ({:?})", queue);
queue
}
fn do_loop(queue_tx: MailboxOuterEnd<Command>,
rx: MailboxInnerEnd<Command>,
deadlock_rx: MailboxInnerEnd<()>)
{
debug!("loop: spawing serial loop");
loop {
trace!("loop: next iteration");
match rx.read() {
Command::End => break,
Command::Wait(routine) => {
trace!("loop: handling previous deadlocked coroutine");
let tx_clone = queue_tx.clone();
loop {
select!(
routine:r => {
if routine.try_read().is_some() {
trace!("loop: task ended");
break;
} else {
continue;
}
},
deadlock_rx:r => {
if deadlock_rx.try_read().is_some() {
trace!("loop: deadlock detected");
tx_clone.send(Command::Wait(routine));
break;
} else {
continue;
}
},
);
}
}
Command::Run(task) => {
let tx_clone = queue_tx.clone();
mioco::set_children_userdata(Some(Userdata::SameThread));
let routine = mioco::spawn_ext(move || {
trace!("loop: spawned new coroutine for task");
task.call_box(());
Ok(())
})
.exit_notificator();
trace!("loop: wait for deadlock notification or coroutine finish");
loop {
select!(
routine:r => {
if routine.try_read().is_some() {
trace!("loop: task ended");
break;
} else {
continue;
}
},
deadlock_rx:r => {
if deadlock_rx.try_read().is_some() {
trace!("loop: deadlock detected");
tx_clone.send(Command::Wait(routine));
break;
} else {
continue;
}
},
);
}
}
}
}
debug!("loop: queue ended");
}
/// Bind this queue to a variable
///
/// This function allows to create a `BoundSerialQueue`.
/// Its purpose is to bind variables to a queue, so they can be used by the tasks submitted.
///
/// # Example
/// ```
/// # use taskqueue::*;
/// # let queue = SerialQueue::new();
/// let x = 5;
/// let bound = queue.with(move || x);
/// bound.scoped_with(|x| println!("{}", x));
/// // x gets dropped inside the queues thread before the queue gets dropped
/// ```
///
/// You can create multiple bindings at once.
/// Through tuples you may bind multiple variables at once.
///
/// It is even possible to move the creation of the bound variable into the queue by creating
/// it inside the passed constructor, which is then executed on the queue.
/// And because SerialQueues never change their underlying OS Thread,
/// this allows to use variables that are not Send and Sync in a thread-safe but shared way.
///
/// # Example
/// ```
/// # use taskqueue::*;
/// # let queue = SerialQueue::new();
/// # fn my_ffi_function() -> *mut () { &mut () };
/// let bound = queue.with(|| {
/// let raw_ptr = my_ffi_function();
/// raw_ptr
/// });
/// bound.scoped_with(|raw_ptr| println!("{}", raw_ptr.is_null()));
/// // raw_ptr gets dropped inside the queues thread.
/// // This way raw_ptr is never moved between threads.
/// ```
pub fn with<'queue, R: 'static, F>(&'queue self,
constructor: F)
-> BoundSerialQueue<'queue, R>
where F: FnOnce() -> R + Send
{
let binding = self.sync(move || NotThreadSafe::new(constructor()));
BoundSerialQueue {
queue: self,
binding: binding,
}
}
}
impl Queue for SerialQueue
{
fn async<R, F>(&self, operation: F) -> Future<R>
where R: Send + 'static,
F: FnOnce() -> R + Send + 'static
{
let (tx, rx) = mailbox();
let operation: Box<FnBox() + Send + 'static> = Stack::assemble(self, move || {
tx.send(operation());
});
debug!("Queue ({:?}) queued task", self);
self.tx.send(Command::Run(operation));
Future::new_from_serial(self, Some(self.deadlock_tx.clone()), rx)
}
}
/// A bound SerialQueue holding a queue-bound variable
///
/// Create a BoundSerialQueue using `SerialQueue::with`.
/// BoundSerialQueue's hold variables that may be used through
/// tasks executed on this queue though `scoped_with`, `sync_with`
/// `foreach_with` or `loop_while_with`.
///
/// `async_with` cannot be provided, as the bound variable is
/// dropped, when the BoundSerialQueue gets dropped.
///
/// Internally BoundSerialQueue refer to the same queue, they were created from.
/// Multiple BoundSerialQueue's may exist for one queue at once.
pub struct BoundSerialQueue<'queue, T: 'static>
{
queue: &'queue SerialQueue,
binding: NotThreadSafe<T>,
}
impl<'queue, T: 'static> fmt::Debug for BoundSerialQueue<'queue, T>
{
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result
{
write!(f, "BoundSerialQueue ({:p})", self)
}
}
impl<'queue, T: 'static> BoundSerialQueue<'queue, T>
{
/// Like `Queue::scoped` but provides a mutable reference to the bound variable
///
/// # Safety
///
/// The same rules as for `Queue::scoped` to apply.
///
/// # Example
///
/// ```
/// # use taskqueue::*;
/// # let queue = SerialQueue::new();
/// let bound = queue.with(|| { "Hello".to_string() });
/// let name = "Stack".to_string();
/// bound.scoped_with(|message| { println!("{} {}!", message, name) });
/// ```
pub fn scoped_with<R, F>(&'queue self, operation: F) -> FutureGuard<R>
where R: Send + 'static,
F: FnOnce(&'queue mut T) -> R + Send + 'queue
{
self.queue.scoped(move || operation(unsafe { self.binding.get_mut() }))
}
/// Like `Queue::sync` but provides a mutable reference to the bound variable
///
/// # Example
///
/// ```
/// # use taskqueue::*;
/// # let queue = SerialQueue::new();
/// let bound = queue |
impl fmt::Debug for SerialQueue | random_line_split |
paths.py | ,
"PathSpec", # TODO(user): recursive definition.
]
@classmethod
def OS(cls, **kwargs):
|
@classmethod
def TSK(cls, **kwargs):
return cls(pathtype=PathSpec.PathType.TSK, **kwargs)
@classmethod
def NTFS(cls, **kwargs):
return cls(pathtype=PathSpec.PathType.NTFS, **kwargs)
@classmethod
def Registry(cls, **kwargs):
return cls(pathtype=PathSpec.PathType.REGISTRY, **kwargs)
@classmethod
def Temp(cls, **kwargs):
return cls(pathtype=PathSpec.PathType.TMPFILE, **kwargs)
def CopyConstructor(self, other):
# pylint: disable=protected-access
self.SetRawData(other._CopyRawData())
# pylint: enable=protected-access
def __len__(self):
"""Return the total number of path components."""
i = -1
# TODO(user):pytype: type checker doesn't treat self as iterable.
for i, _ in enumerate(self): # pytype: disable=wrong-arg-types
pass
return i + 1
def __getitem__(self, item):
# TODO(user):pytype: type checker doesn't treat self as iterable.
for i, element in enumerate(self): # pytype: disable=wrong-arg-types
if i == item:
return element
raise IndexError("Pathspec index (%s) out of range" % item)
def __iter__(self):
"""Only iterate over all components from the current pointer."""
element = self
while element.HasField("pathtype"):
yield element
if element.HasField("nested_path"):
element = element.nested_path
else:
break
def Insert(self, index, rdfpathspec=None, **kwarg):
"""Insert a single component at index."""
if rdfpathspec is None:
rdfpathspec = self.__class__(**kwarg)
if index == 0:
# Copy ourselves to a temp copy.
nested_proto = self.__class__()
nested_proto.SetRawData(self.GetRawData())
# Replace ourselves with the new object.
self.SetRawData(rdfpathspec.GetRawData())
# Append the temp copy to the end.
self.last.nested_path = nested_proto
else:
previous = self[index - 1]
rdfpathspec.last.nested_path = previous.nested_path
previous.nested_path = rdfpathspec
def Append(self, component=None, **kwarg):
"""Append a new pathspec component to this pathspec."""
if component is None:
component = self.__class__(**kwarg)
if self.HasField("pathtype"):
self.last.nested_path = component
else:
for k, v in kwarg.items():
setattr(self, k, v)
self.SetRawData(component.GetRawData())
return self
def CollapsePath(self):
return utils.JoinPath(*[x.path for x in self])
def Pop(self, index=0):
"""Removes and returns the pathspec at the specified index."""
if index < 0:
index += len(self)
if index == 0:
result = self.__class__()
result.SetRawData(self.GetRawData())
self.SetRawData(self.nested_path.GetRawData())
else:
# Get the raw protobufs for the previous member.
previous = self[index - 1]
result = previous.nested_path
# Manipulate the previous members protobuf to patch the next component in.
previous.nested_path = result.nested_path
result.nested_path = None
return result
@property
def first(self):
return self
@property
def last(self):
if self.HasField("pathtype") and self.pathtype != self.PathType.UNSET:
# TODO(user):pytype: type checker doesn't treat self as iterable.
return list(self)[-1] # pytype: disable=wrong-arg-types
return self
def Dirname(self):
"""Get a new copied object with only the directory path."""
result = self.Copy()
while 1:
last_directory = posixpath.dirname(result.last.path)
if last_directory != "/" or len(result) <= 1:
result.last.path = last_directory
# Make sure to clear the inode information.
result.last.inode = None
break
result.Pop(-1)
return result
def Basename(self):
# TODO(user):pytype: type checker doesn't treat self as reversible.
for component in reversed(self): # pytype: disable=wrong-arg-types
basename = posixpath.basename(component.path)
if basename:
return basename
return ""
def Validate(self):
if not self.HasField("pathtype") or self.pathtype == self.PathType.UNSET:
raise ValueError("No path type set in PathSpec.")
AFF4_PREFIXES = {
0: "/fs/os", # PathSpec.PathType.OS
1: "/fs/tsk", # PathSpec.PathType.TSK
2: "/registry", # PathSpec.PathType.REGISTRY
4: "/temp", # PathSpec.PathType.TMPFILE
5: "/fs/ntfs", # PathSpec.PathType.NTFS
}
def AFF4Path(self, client_urn):
"""Returns the AFF4 URN this pathspec will be stored under.
Args:
client_urn: A ClientURN.
Returns:
A urn that corresponds to this pathspec.
Raises:
ValueError: If pathspec is not of the correct type.
"""
# If the first level is OS and the second level is TSK its probably a mount
# point resolution. We map it into the tsk branch. For example if we get:
# path: \\\\.\\Volume{1234}\\
# pathtype: OS
# mount_point: /c:/
# nested_path {
# path: /windows/
# pathtype: TSK
# }
# We map this to aff4://client_id/fs/tsk/\\\\.\\Volume{1234}\\/windows/
# (The same applies for NTFS)
if not self.HasField("pathtype"):
raise ValueError(
"Can't determine AFF4 path without a valid pathtype for {}.".format(
self))
first_component = self[0]
dev = first_component.path
if first_component.HasField("offset"):
# We divide here just to get prettier numbers in the GUI
dev += ":{}".format(first_component.offset // 512)
if (len(self) > 1 and first_component.pathtype == PathSpec.PathType.OS and
self[1].pathtype in (PathSpec.PathType.TSK, PathSpec.PathType.NTFS)):
result = [self.AFF4_PREFIXES[self[1].pathtype], dev]
# Skip the top level pathspec.
start = 1
else:
# For now just map the top level prefix based on the first pathtype
result = [self.AFF4_PREFIXES[first_component.pathtype]]
start = 0
for p in self[start]:
component = p.path
# The following encode different pathspec properties into the AFF4 path in
# such a way that unique files on the client are mapped to unique URNs in
# the AFF4 space. Note that this transformation does not need to be
# reversible since we always use the PathSpec when accessing files on the
# client.
if p.HasField("offset"):
component += ":{}".format(p.offset // 512)
# Support ADS names.
if p.HasField("stream_name"):
component += ":" + p.stream_name
result.append(component)
return client_urn.Add("/".join(result))
def _unique(iterable):
"""Returns a list of unique values in preserved order."""
return list(dict.fromkeys(iterable))
class GlobComponentExplanation(rdf_structs.RDFProtoStruct):
"""A sub-part of a GlobExpression with examples."""
protobuf = flows_pb2.GlobComponentExplanation
# Grouping pattern: e.g. {test.exe,foo.doc,bar.txt}
GROUPING_PATTERN = re.compile("{([^}]+,[^}]+)}")
_VAR_PATTERN = re.compile("(" + "|".join([r"%%\w+%%", r"%%\w+\.\w+%%"]) + ")")
_REGEX_SPLIT_PATTERN = re.compile(
"(" + "|".join(["{[^}]+,[^}]+}", "\\?", "\\*\\*\\/?", "\\*"]) + ")")
_COMPONENT_SPLIT_PATTERN = re.compile("(" + "|".join([
r"{[^}]+,[^}]+}", r"\?", r"\*\*\d*/?", r"\*", r"%%\w+%%", r"%%\w+\.\w+%%"
]) + ")")
class GlobExpression(rdfvalue.RDFString):
"""A glob expression for a client path.
A glob expression represents a set of regular expressions which match files on
the client. The Glob expression supports the following expansions:
| return cls(pathtype=PathSpec.PathType.OS, **kwargs) | identifier_body |
paths.py | ,
"PathSpec", # TODO(user): recursive definition.
]
@classmethod
def OS(cls, **kwargs):
return cls(pathtype=PathSpec.PathType.OS, **kwargs)
@classmethod
def TSK(cls, **kwargs):
return cls(pathtype=PathSpec.PathType.TSK, **kwargs)
@classmethod
def NTFS(cls, **kwargs):
return cls(pathtype=PathSpec.PathType.NTFS, **kwargs)
@classmethod
def Registry(cls, **kwargs):
return cls(pathtype=PathSpec.PathType.REGISTRY, **kwargs)
@classmethod
def Temp(cls, **kwargs):
return cls(pathtype=PathSpec.PathType.TMPFILE, **kwargs)
def CopyConstructor(self, other):
# pylint: disable=protected-access
self.SetRawData(other._CopyRawData())
# pylint: enable=protected-access
def __len__(self):
"""Return the total number of path components."""
i = -1
# TODO(user):pytype: type checker doesn't treat self as iterable.
for i, _ in enumerate(self): # pytype: disable=wrong-arg-types
pass
return i + 1
def __getitem__(self, item):
# TODO(user):pytype: type checker doesn't treat self as iterable.
for i, element in enumerate(self): # pytype: disable=wrong-arg-types
if i == item:
return element
raise IndexError("Pathspec index (%s) out of range" % item)
def __iter__(self):
"""Only iterate over all components from the current pointer."""
element = self
while element.HasField("pathtype"):
yield element
if element.HasField("nested_path"):
element = element.nested_path
else:
break
def Insert(self, index, rdfpathspec=None, **kwarg):
"""Insert a single component at index."""
if rdfpathspec is None:
rdfpathspec = self.__class__(**kwarg)
if index == 0:
# Copy ourselves to a temp copy.
nested_proto = self.__class__()
nested_proto.SetRawData(self.GetRawData())
# Replace ourselves with the new object.
self.SetRawData(rdfpathspec.GetRawData())
# Append the temp copy to the end.
self.last.nested_path = nested_proto
else:
previous = self[index - 1]
rdfpathspec.last.nested_path = previous.nested_path
previous.nested_path = rdfpathspec
def Append(self, component=None, **kwarg):
"""Append a new pathspec component to this pathspec."""
if component is None:
component = self.__class__(**kwarg)
if self.HasField("pathtype"):
self.last.nested_path = component
else:
for k, v in kwarg.items():
setattr(self, k, v)
self.SetRawData(component.GetRawData())
return self
def CollapsePath(self):
return utils.JoinPath(*[x.path for x in self])
def Pop(self, index=0):
"""Removes and returns the pathspec at the specified index."""
if index < 0:
index += len(self)
if index == 0:
result = self.__class__()
result.SetRawData(self.GetRawData())
self.SetRawData(self.nested_path.GetRawData())
else:
# Get the raw protobufs for the previous member.
previous = self[index - 1]
result = previous.nested_path
# Manipulate the previous members protobuf to patch the next component in.
previous.nested_path = result.nested_path
result.nested_path = None
return result
@property
def first(self):
return self
@property
def last(self):
if self.HasField("pathtype") and self.pathtype != self.PathType.UNSET:
# TODO(user):pytype: type checker doesn't treat self as iterable.
return list(self)[-1] # pytype: disable=wrong-arg-types
return self
def Dirname(self):
"""Get a new copied object with only the directory path."""
result = self.Copy()
while 1:
last_directory = posixpath.dirname(result.last.path)
if last_directory != "/" or len(result) <= 1:
result.last.path = last_directory
# Make sure to clear the inode information.
result.last.inode = None
break
result.Pop(-1)
return result
def Basename(self):
# TODO(user):pytype: type checker doesn't treat self as reversible.
for component in reversed(self): # pytype: disable=wrong-arg-types
basename = posixpath.basename(component.path)
if basename:
return basename
return ""
def Validate(self):
if not self.HasField("pathtype") or self.pathtype == self.PathType.UNSET:
raise ValueError("No path type set in PathSpec.")
AFF4_PREFIXES = {
0: "/fs/os", # PathSpec.PathType.OS
1: "/fs/tsk", # PathSpec.PathType.TSK
2: "/registry", # PathSpec.PathType.REGISTRY
4: "/temp", # PathSpec.PathType.TMPFILE
5: "/fs/ntfs", # PathSpec.PathType.NTFS
}
def AFF4Path(self, client_urn):
"""Returns the AFF4 URN this pathspec will be stored under.
Args:
client_urn: A ClientURN.
Returns:
A urn that corresponds to this pathspec.
Raises:
ValueError: If pathspec is not of the correct type.
"""
# If the first level is OS and the second level is TSK its probably a mount
# point resolution. We map it into the tsk branch. For example if we get:
# path: \\\\.\\Volume{1234}\\
# pathtype: OS
# mount_point: /c:/
# nested_path {
# path: /windows/
# pathtype: TSK
# }
# We map this to aff4://client_id/fs/tsk/\\\\.\\Volume{1234}\\/windows/
# (The same applies for NTFS)
if not self.HasField("pathtype"):
raise ValueError(
"Can't determine AFF4 path without a valid pathtype for {}.".format(
self))
first_component = self[0]
dev = first_component.path
if first_component.HasField("offset"):
# We divide here just to get prettier numbers in the GUI
dev += ":{}".format(first_component.offset // 512)
if (len(self) > 1 and first_component.pathtype == PathSpec.PathType.OS and
self[1].pathtype in (PathSpec.PathType.TSK, PathSpec.PathType.NTFS)):
result = [self.AFF4_PREFIXES[self[1].pathtype], dev]
# Skip the top level pathspec.
start = 1
else:
# For now just map the top level prefix based on the first pathtype
result = [self.AFF4_PREFIXES[first_component.pathtype]]
start = 0
for p in self[start]:
component = p.path
# The following encode different pathspec properties into the AFF4 path in
# such a way that unique files on the client are mapped to unique URNs in
# the AFF4 space. Note that this transformation does not need to be
# reversible since we always use the PathSpec when accessing files on the
# client.
if p.HasField("offset"):
component += ":{}".format(p.offset // 512)
# Support ADS names.
if p.HasField("stream_name"):
component += ":" + p.stream_name
result.append(component)
return client_urn.Add("/".join(result))
def _unique(iterable):
"""Returns a list of unique values in preserved order."""
return list(dict.fromkeys(iterable))
class | (rdf_structs.RDFProtoStruct):
"""A sub-part of a GlobExpression with examples."""
protobuf = flows_pb2.GlobComponentExplanation
# Grouping pattern: e.g. {test.exe,foo.doc,bar.txt}
GROUPING_PATTERN = re.compile("{([^}]+,[^}]+)}")
_VAR_PATTERN = re.compile("(" + "|".join([r"%%\w+%%", r"%%\w+\.\w+%%"]) + ")")
_REGEX_SPLIT_PATTERN = re.compile(
"(" + "|".join(["{[^}]+,[^}]+}", "\\?", "\\*\\*\\/?", "\\*"]) + ")")
_COMPONENT_SPLIT_PATTERN = re.compile("(" + "|".join([
r"{[^}]+,[^}]+}", r"\?", r"\*\*\d*/?", r"\*", r"%%\w+%%", r"%%\w+\.\w+%%"
]) + ")")
class GlobExpression(rdfvalue.RDFString):
"""A glob expression for a client path.
A glob expression represents a set of regular expressions which match files on
the client. The Glob expression supports the following expansions:
| GlobComponentExplanation | identifier_name |
paths.py |
@classmethod
def OS(cls, **kwargs):
return cls(pathtype=PathSpec.PathType.OS, **kwargs)
@classmethod
def TSK(cls, **kwargs):
return cls(pathtype=PathSpec.PathType.TSK, **kwargs)
@classmethod
def NTFS(cls, **kwargs):
return cls(pathtype=PathSpec.PathType.NTFS, **kwargs)
@classmethod
def Registry(cls, **kwargs):
return cls(pathtype=PathSpec.PathType.REGISTRY, **kwargs)
@classmethod
def Temp(cls, **kwargs):
return cls(pathtype=PathSpec.PathType.TMPFILE, **kwargs)
def CopyConstructor(self, other):
# pylint: disable=protected-access
self.SetRawData(other._CopyRawData())
# pylint: enable=protected-access
def __len__(self):
"""Return the total number of path components."""
i = -1
# TODO(user):pytype: type checker doesn't treat self as iterable.
for i, _ in enumerate(self): # pytype: disable=wrong-arg-types
pass
return i + 1
def __getitem__(self, item):
# TODO(user):pytype: type checker doesn't treat self as iterable.
for i, element in enumerate(self): # pytype: disable=wrong-arg-types
if i == item:
return element
raise IndexError("Pathspec index (%s) out of range" % item)
def __iter__(self):
"""Only iterate over all components from the current pointer."""
element = self
while element.HasField("pathtype"):
yield element
if element.HasField("nested_path"):
element = element.nested_path
else:
break
def Insert(self, index, rdfpathspec=None, **kwarg):
"""Insert a single component at index."""
if rdfpathspec is None:
rdfpathspec = self.__class__(**kwarg)
if index == 0:
# Copy ourselves to a temp copy.
nested_proto = self.__class__()
nested_proto.SetRawData(self.GetRawData())
# Replace ourselves with the new object.
self.SetRawData(rdfpathspec.GetRawData())
# Append the temp copy to the end.
self.last.nested_path = nested_proto
else:
previous = self[index - 1]
rdfpathspec.last.nested_path = previous.nested_path
previous.nested_path = rdfpathspec
def Append(self, component=None, **kwarg):
"""Append a new pathspec component to this pathspec."""
if component is None:
component = self.__class__(**kwarg)
if self.HasField("pathtype"):
self.last.nested_path = component
else:
for k, v in kwarg.items():
setattr(self, k, v)
self.SetRawData(component.GetRawData())
return self
def CollapsePath(self):
return utils.JoinPath(*[x.path for x in self])
def Pop(self, index=0):
"""Removes and returns the pathspec at the specified index."""
if index < 0:
index += len(self)
if index == 0:
result = self.__class__()
result.SetRawData(self.GetRawData())
self.SetRawData(self.nested_path.GetRawData())
else:
# Get the raw protobufs for the previous member.
previous = self[index - 1]
result = previous.nested_path
# Manipulate the previous members protobuf to patch the next component in.
previous.nested_path = result.nested_path
result.nested_path = None
return result
@property
def first(self):
return self
@property
def last(self):
if self.HasField("pathtype") and self.pathtype != self.PathType.UNSET:
# TODO(user):pytype: type checker doesn't treat self as iterable.
return list(self)[-1] # pytype: disable=wrong-arg-types
return self
def Dirname(self):
"""Get a new copied object with only the directory path."""
result = self.Copy()
while 1:
last_directory = posixpath.dirname(result.last.path)
if last_directory != "/" or len(result) <= 1:
result.last.path = last_directory
# Make sure to clear the inode information.
result.last.inode = None
break
result.Pop(-1)
return result
def Basename(self):
# TODO(user):pytype: type checker doesn't treat self as reversible.
for component in reversed(self): # pytype: disable=wrong-arg-types
basename = posixpath.basename(component.path)
if basename:
return basename
return ""
def Validate(self):
if not self.HasField("pathtype") or self.pathtype == self.PathType.UNSET:
raise ValueError("No path type set in PathSpec.")
AFF4_PREFIXES = {
0: "/fs/os", # PathSpec.PathType.OS
1: "/fs/tsk", # PathSpec.PathType.TSK
2: "/registry", # PathSpec.PathType.REGISTRY
4: "/temp", # PathSpec.PathType.TMPFILE
5: "/fs/ntfs", # PathSpec.PathType.NTFS
}
def AFF4Path(self, client_urn):
"""Returns the AFF4 URN this pathspec will be stored under.
Args:
client_urn: A ClientURN.
Returns:
A urn that corresponds to this pathspec.
Raises:
ValueError: If pathspec is not of the correct type.
"""
# If the first level is OS and the second level is TSK its probably a mount
# point resolution. We map it into the tsk branch. For example if we get:
# path: \\\\.\\Volume{1234}\\
# pathtype: OS
# mount_point: /c:/
# nested_path {
# path: /windows/
# pathtype: TSK
# }
# We map this to aff4://client_id/fs/tsk/\\\\.\\Volume{1234}\\/windows/
# (The same applies for NTFS)
if not self.HasField("pathtype"):
raise ValueError(
"Can't determine AFF4 path without a valid pathtype for {}.".format(
self))
first_component = self[0]
dev = first_component.path
if first_component.HasField("offset"):
# We divide here just to get prettier numbers in the GUI
dev += ":{}".format(first_component.offset // 512)
if (len(self) > 1 and first_component.pathtype == PathSpec.PathType.OS and
self[1].pathtype in (PathSpec.PathType.TSK, PathSpec.PathType.NTFS)):
result = [self.AFF4_PREFIXES[self[1].pathtype], dev]
# Skip the top level pathspec.
start = 1
else:
# For now just map the top level prefix based on the first pathtype
result = [self.AFF4_PREFIXES[first_component.pathtype]]
start = 0
for p in self[start]:
component = p.path
# The following encode different pathspec properties into the AFF4 path in
# such a way that unique files on the client are mapped to unique URNs in
# the AFF4 space. Note that this transformation does not need to be
# reversible since we always use the PathSpec when accessing files on the
# client.
if p.HasField("offset"):
component += ":{}".format(p.offset // 512)
# Support ADS names.
if p.HasField("stream_name"):
component += ":" + p.stream_name
result.append(component)
return client_urn.Add("/".join(result))
def _unique(iterable):
"""Returns a list of unique values in preserved order."""
return list(dict.fromkeys(iterable))
class GlobComponentExplanation(rdf_structs.RDFProtoStruct):
"""A sub-part of a GlobExpression with examples."""
protobuf = flows_pb2.GlobComponentExplanation
# Grouping pattern: e.g. {test.exe,foo.doc,bar.txt}
GROUPING_PATTERN = re.compile("{([^}]+,[^}]+)}")
_VAR_PATTERN = re.compile("(" + "|".join([r"%%\w+%%", r"%%\w+\.\w+%%"]) + ")")
_REGEX_SPLIT_PATTERN = re.compile(
"(" + "|".join(["{[^}]+,[^}]+}", "\\?", "\\*\\*\\/?", "\\*"]) + ")")
_COMPONENT_SPLIT_PATTERN = re.compile("(" + "|".join([
r"{[^}]+,[^}]+}", r"\?", r"\*\*\d*/?", r"\*", r"%%\w+%%", r"%%\w+\.\w+%%"
]) + ")")
class GlobExpression(rdfvalue.RDFString):
"""A glob expression for a client path.
A glob expression represents a set of regular expressions which match files on
the client | rdf_deps = [
rdfvalue.ByteSize,
"PathSpec", # TODO(user): recursive definition.
] | random_line_split |
|
paths.py | .__class__()
nested_proto.SetRawData(self.GetRawData())
# Replace ourselves with the new object.
self.SetRawData(rdfpathspec.GetRawData())
# Append the temp copy to the end.
self.last.nested_path = nested_proto
else:
previous = self[index - 1]
rdfpathspec.last.nested_path = previous.nested_path
previous.nested_path = rdfpathspec
def Append(self, component=None, **kwarg):
"""Append a new pathspec component to this pathspec."""
if component is None:
component = self.__class__(**kwarg)
if self.HasField("pathtype"):
self.last.nested_path = component
else:
for k, v in kwarg.items():
setattr(self, k, v)
self.SetRawData(component.GetRawData())
return self
def CollapsePath(self):
return utils.JoinPath(*[x.path for x in self])
def Pop(self, index=0):
"""Removes and returns the pathspec at the specified index."""
if index < 0:
index += len(self)
if index == 0:
result = self.__class__()
result.SetRawData(self.GetRawData())
self.SetRawData(self.nested_path.GetRawData())
else:
# Get the raw protobufs for the previous member.
previous = self[index - 1]
result = previous.nested_path
# Manipulate the previous members protobuf to patch the next component in.
previous.nested_path = result.nested_path
result.nested_path = None
return result
@property
def first(self):
return self
@property
def last(self):
if self.HasField("pathtype") and self.pathtype != self.PathType.UNSET:
# TODO(user):pytype: type checker doesn't treat self as iterable.
return list(self)[-1] # pytype: disable=wrong-arg-types
return self
def Dirname(self):
"""Get a new copied object with only the directory path."""
result = self.Copy()
while 1:
last_directory = posixpath.dirname(result.last.path)
if last_directory != "/" or len(result) <= 1:
result.last.path = last_directory
# Make sure to clear the inode information.
result.last.inode = None
break
result.Pop(-1)
return result
def Basename(self):
# TODO(user):pytype: type checker doesn't treat self as reversible.
for component in reversed(self): # pytype: disable=wrong-arg-types
basename = posixpath.basename(component.path)
if basename:
return basename
return ""
def Validate(self):
if not self.HasField("pathtype") or self.pathtype == self.PathType.UNSET:
raise ValueError("No path type set in PathSpec.")
AFF4_PREFIXES = {
0: "/fs/os", # PathSpec.PathType.OS
1: "/fs/tsk", # PathSpec.PathType.TSK
2: "/registry", # PathSpec.PathType.REGISTRY
4: "/temp", # PathSpec.PathType.TMPFILE
5: "/fs/ntfs", # PathSpec.PathType.NTFS
}
def AFF4Path(self, client_urn):
"""Returns the AFF4 URN this pathspec will be stored under.
Args:
client_urn: A ClientURN.
Returns:
A urn that corresponds to this pathspec.
Raises:
ValueError: If pathspec is not of the correct type.
"""
# If the first level is OS and the second level is TSK its probably a mount
# point resolution. We map it into the tsk branch. For example if we get:
# path: \\\\.\\Volume{1234}\\
# pathtype: OS
# mount_point: /c:/
# nested_path {
# path: /windows/
# pathtype: TSK
# }
# We map this to aff4://client_id/fs/tsk/\\\\.\\Volume{1234}\\/windows/
# (The same applies for NTFS)
if not self.HasField("pathtype"):
raise ValueError(
"Can't determine AFF4 path without a valid pathtype for {}.".format(
self))
first_component = self[0]
dev = first_component.path
if first_component.HasField("offset"):
# We divide here just to get prettier numbers in the GUI
dev += ":{}".format(first_component.offset // 512)
if (len(self) > 1 and first_component.pathtype == PathSpec.PathType.OS and
self[1].pathtype in (PathSpec.PathType.TSK, PathSpec.PathType.NTFS)):
result = [self.AFF4_PREFIXES[self[1].pathtype], dev]
# Skip the top level pathspec.
start = 1
else:
# For now just map the top level prefix based on the first pathtype
result = [self.AFF4_PREFIXES[first_component.pathtype]]
start = 0
for p in self[start]:
component = p.path
# The following encode different pathspec properties into the AFF4 path in
# such a way that unique files on the client are mapped to unique URNs in
# the AFF4 space. Note that this transformation does not need to be
# reversible since we always use the PathSpec when accessing files on the
# client.
if p.HasField("offset"):
component += ":{}".format(p.offset // 512)
# Support ADS names.
if p.HasField("stream_name"):
component += ":" + p.stream_name
result.append(component)
return client_urn.Add("/".join(result))
def _unique(iterable):
"""Returns a list of unique values in preserved order."""
return list(dict.fromkeys(iterable))
class GlobComponentExplanation(rdf_structs.RDFProtoStruct):
"""A sub-part of a GlobExpression with examples."""
protobuf = flows_pb2.GlobComponentExplanation
# Grouping pattern: e.g. {test.exe,foo.doc,bar.txt}
GROUPING_PATTERN = re.compile("{([^}]+,[^}]+)}")
_VAR_PATTERN = re.compile("(" + "|".join([r"%%\w+%%", r"%%\w+\.\w+%%"]) + ")")
_REGEX_SPLIT_PATTERN = re.compile(
"(" + "|".join(["{[^}]+,[^}]+}", "\\?", "\\*\\*\\/?", "\\*"]) + ")")
_COMPONENT_SPLIT_PATTERN = re.compile("(" + "|".join([
r"{[^}]+,[^}]+}", r"\?", r"\*\*\d*/?", r"\*", r"%%\w+%%", r"%%\w+\.\w+%%"
]) + ")")
class GlobExpression(rdfvalue.RDFString):
"""A glob expression for a client path.
A glob expression represents a set of regular expressions which match files on
the client. The Glob expression supports the following expansions:
1) Client attribute expansions are surrounded with %% characters. They will be
expanded from the client AFF4 object.
2) Groupings are collections of alternates. e.g. {foo.exe,bar.sys}
3) Wild cards like * and ?
"""
context_help_url = "investigating-with-grr/flows/specifying-file-paths.html"
RECURSION_REGEX = re.compile(r"\*\*(\d*)")
def Validate(self):
"""GlobExpression is valid."""
if len(self.RECURSION_REGEX.findall(self._value)) > 1:
raise ValueError("Only one ** is permitted per path: %s." % self._value)
def Interpolate(self, knowledge_base=None):
kb = knowledge_base
patterns = artifact_utils.InterpolateKbAttributes(self._value, kb)
for pattern in patterns:
# Normalize the component path (this allows us to resolve ../
# sequences).
pattern = utils.NormalizePath(pattern.replace("\\", "/"))
for p in self.InterpolateGrouping(pattern):
yield p
def InterpolateGrouping(self, pattern):
"""Interpolate inline globbing groups."""
components = []
offset = 0
for match in GROUPING_PATTERN.finditer(pattern):
components.append([pattern[offset:match.start()]])
# Expand the attribute into the set of possibilities:
alternatives = match.group(1).split(",")
components.append(_unique(alternatives))
offset = match.end()
components.append([pattern[offset:]])
# Now calculate the cartesian products of all these sets to form all
# strings.
for vector in itertools.product(*components):
yield u"".join(vector)
def _ReplaceRegExGrouping(self, grouping):
alternatives = grouping.group(1).split(",")
return "(" + "|".join(re.escape(s) for s in alternatives) + ")"
def _ReplaceRegExPart(self, part):
if part == "**/":
return "(?:.*\\/)?"
elif part == "*":
return "[^\\/]*"
elif part == "?":
return "[^\\/]"
elif GROUPING_PATTERN.match(part):
| return GROUPING_PATTERN.sub(self._ReplaceRegExGrouping, part) | conditional_block |
|
images.go | I: %v", err)
}
aciFile, err := os.Open(acis[0])
if err != nil {
return nil, nil, nil, fmt.Errorf("error opening squashed ACI file: %v", err)
}
return nil, aciFile, nil, nil
}
// attempt to automatically fetch the public key in case it is available on a TLS connection.
if globalFlags.TrustKeysFromHttps && !globalFlags.InsecureSkipVerify && appName != "" {
pkls, err := getPubKeyLocations(appName, false, globalFlags.Debug)
if err != nil {
stderr("Error determining key location: %v", err)
} else {
// no http, don't ask user for accepting the key, no overriding
if err := addKeys(pkls, appName, false, true, false); err != nil {
stderr("Error adding keys: %v", err)
}
}
}
var retrySignature bool
if f.ks != nil && ascFile == nil {
u, err := url.Parse(ascURL)
if err != nil {
return nil, nil, nil, fmt.Errorf("error parsing ASC url: %v", err)
}
if u.Scheme == "file" {
ascFile, err = os.Open(u.Path)
if err != nil {
return nil, nil, nil, fmt.Errorf("error opening signature file: %v", err)
}
} else {
stderr("Downloading signature from %v\n", ascURL)
ascFile, err = f.s.TmpFile()
if err != nil {
return nil, nil, nil, fmt.Errorf("error setting up temporary file: %v", err)
}
defer os.Remove(ascFile.Name())
err = f.downloadSignatureFile(ascURL, ascFile)
switch err {
case errStatusAccepted:
retrySignature = true
stderr("rkt: server requested deferring the signature download")
case nil:
break
default:
return nil, nil, nil, fmt.Errorf("error downloading the signature file: %v", err)
}
}
defer ascFile.Close()
}
// check if the identity used by the signature is in the store before a
// possibly expensive download. This is only an optimization and it's
// ok to skip the test if the signature will be downloaded later.
if !retrySignature && f.ks != nil && appName != "" {
if _, err := ascFile.Seek(0, 0); err != nil {
return nil, nil, nil, fmt.Errorf("error seeking signature file: %v", err)
}
if entity, err = f.ks.CheckSignature(appName, bytes.NewReader([]byte{}), ascFile); err != nil {
if _, ok := err.(pgperrors.SignatureError); !ok {
return nil, nil, nil, err
}
}
}
var aciFile *os.File
if u.Scheme == "file" {
aciFile, err = os.Open(u.Path)
if err != nil {
return nil, nil, nil, fmt.Errorf("error opening ACI file: %v", err)
}
} else {
aciFile, err = f.s.TmpFile()
if err != nil {
return nil, aciFile, nil, fmt.Errorf("error setting up temporary file: %v", err)
}
defer os.Remove(aciFile.Name())
if cd, err = f.downloadACI(aciURL, aciFile, etag); err != nil {
return nil, nil, nil, fmt.Errorf("error downloading ACI: %v", err)
}
if cd.useCached {
return nil, nil, cd, nil
}
}
if retrySignature {
if err = f.downloadSignatureFile(ascURL, ascFile); err != nil {
return nil, aciFile, nil, fmt.Errorf("error downloading the signature file: %v", err)
}
}
manifest, err := aci.ManifestFromImage(aciFile)
if err != nil {
return nil, aciFile, nil, err
}
// Check if the downloaded ACI has the correct app name.
// The check is only performed when the aci is downloaded through the
// discovery protocol, but not with local files or full URL.
if appName != "" && manifest.Name.String() != appName {
return nil, aciFile, nil,
fmt.Errorf("error when reading the app name: %q expected but %q found",
appName, manifest.Name.String())
}
if f.ks != nil {
if _, err := aciFile.Seek(0, 0); err != nil {
return nil, aciFile, nil, fmt.Errorf("error seeking ACI file: %v", err)
}
if _, err := ascFile.Seek(0, 0); err != nil {
return nil, aciFile, nil, fmt.Errorf("error seeking signature file: %v", err)
}
if entity, err = f.ks.CheckSignature(manifest.Name.String(), aciFile, ascFile); err != nil {
return nil, aciFile, nil, err
}
}
if _, err := aciFile.Seek(0, 0); err != nil {
return nil, aciFile, nil, fmt.Errorf("error seeking ACI file: %v", err)
}
return entity, aciFile, cd, nil
}
type writeSyncer interface {
io.Writer
Sync() error
}
// downloadACI gets the aci specified at aciurl
func (f *fetcher) downloadACI(aciurl string, out writeSyncer, etag string) (*cacheData, error) {
return f.downloadHTTP(aciurl, "ACI", out, etag)
}
// downloadSignatureFile gets the signature specified at sigurl
func (f *fetcher) downloadSignatureFile(sigurl string, out writeSyncer) error {
_, err := f.downloadHTTP(sigurl, "signature", out, "")
return err
}
// downloadHTTP retrieves url, creating a temp file using getTempFile
// http:// and https:// urls supported
func (f *fetcher) downloadHTTP(url, label string, out writeSyncer, etag string) (*cacheData, error) {
req, err := http.NewRequest("GET", url, nil)
if err != nil {
return nil, err
}
options := make(http.Header)
// Send credentials only over secure channel
if req.URL.Scheme == "https" {
if hostOpts, ok := f.headers[req.URL.Host]; ok {
options = hostOpts.Header()
}
}
for k, v := range options {
for _, e := range v {
req.Header.Add(k, e)
}
}
transport := http.DefaultTransport
if f.insecureSkipVerify {
transport = &http.Transport{
TLSClientConfig: &tls.Config{InsecureSkipVerify: true},
}
}
if etag != "" {
req.Header.Add("If-None-Match", etag)
}
req.Header.Add("User-Agent", fmt.Sprintf("rkt/%s", version.Version))
client := &http.Client{Transport: transport}
res, err := client.Do(req)
if err != nil {
return nil, err
}
defer res.Body.Close()
cd := &cacheData{}
// TODO(jonboulle): handle http more robustly (redirects?)
switch res.StatusCode {
case http.StatusAccepted:
// If the server returns Status Accepted (HTTP 202), we should retry
// downloading the signature later.
return nil, errStatusAccepted
case http.StatusOK:
fallthrough
case http.StatusNotModified:
cd.etag = res.Header.Get("ETag")
cd.maxAge = getMaxAge(res.Header.Get("Cache-Control"))
cd.useCached = (res.StatusCode == http.StatusNotModified)
if cd.useCached {
return cd, nil
}
default:
return nil, fmt.Errorf("bad HTTP status code: %d", res.StatusCode)
}
prefix := "Downloading " + label
fmtBytesSize := 18
barSize := int64(80 - len(prefix) - fmtBytesSize)
bar := ioprogress.DrawTextFormatBar(barSize)
fmtfunc := func(progress, total int64) string {
// Content-Length is set to -1 when unknown.
if total == -1 {
return fmt.Sprintf(
"%s: %v of an unknown total size",
prefix,
ioprogress.ByteUnitStr(progress),
)
}
return fmt.Sprintf(
"%s: %s %s",
prefix,
bar(progress, total),
ioprogress.DrawTextFormatBytes(progress, total),
)
}
reader := &ioprogress.Reader{
Reader: res.Body,
Size: res.ContentLength,
DrawFunc: ioprogress.DrawTerminalf(os.Stdout, fmtfunc),
DrawInterval: time.Second,
}
if _, err := io.Copy(out, reader); err != nil {
return nil, fmt.Errorf("error copying %s: %v", label, err)
} |
if err := out.Sync(); err != nil { | random_line_split |
|
images.go | ); err != nil {
if _, ok := err.(pgperrors.SignatureError); !ok {
return nil, nil, nil, err
}
}
}
var aciFile *os.File
if u.Scheme == "file" {
aciFile, err = os.Open(u.Path)
if err != nil {
return nil, nil, nil, fmt.Errorf("error opening ACI file: %v", err)
}
} else {
aciFile, err = f.s.TmpFile()
if err != nil {
return nil, aciFile, nil, fmt.Errorf("error setting up temporary file: %v", err)
}
defer os.Remove(aciFile.Name())
if cd, err = f.downloadACI(aciURL, aciFile, etag); err != nil {
return nil, nil, nil, fmt.Errorf("error downloading ACI: %v", err)
}
if cd.useCached {
return nil, nil, cd, nil
}
}
if retrySignature {
if err = f.downloadSignatureFile(ascURL, ascFile); err != nil {
return nil, aciFile, nil, fmt.Errorf("error downloading the signature file: %v", err)
}
}
manifest, err := aci.ManifestFromImage(aciFile)
if err != nil {
return nil, aciFile, nil, err
}
// Check if the downloaded ACI has the correct app name.
// The check is only performed when the aci is downloaded through the
// discovery protocol, but not with local files or full URL.
if appName != "" && manifest.Name.String() != appName {
return nil, aciFile, nil,
fmt.Errorf("error when reading the app name: %q expected but %q found",
appName, manifest.Name.String())
}
if f.ks != nil {
if _, err := aciFile.Seek(0, 0); err != nil {
return nil, aciFile, nil, fmt.Errorf("error seeking ACI file: %v", err)
}
if _, err := ascFile.Seek(0, 0); err != nil {
return nil, aciFile, nil, fmt.Errorf("error seeking signature file: %v", err)
}
if entity, err = f.ks.CheckSignature(manifest.Name.String(), aciFile, ascFile); err != nil {
return nil, aciFile, nil, err
}
}
if _, err := aciFile.Seek(0, 0); err != nil {
return nil, aciFile, nil, fmt.Errorf("error seeking ACI file: %v", err)
}
return entity, aciFile, cd, nil
}
type writeSyncer interface {
io.Writer
Sync() error
}
// downloadACI gets the aci specified at aciurl
func (f *fetcher) downloadACI(aciurl string, out writeSyncer, etag string) (*cacheData, error) {
return f.downloadHTTP(aciurl, "ACI", out, etag)
}
// downloadSignatureFile gets the signature specified at sigurl
func (f *fetcher) downloadSignatureFile(sigurl string, out writeSyncer) error {
_, err := f.downloadHTTP(sigurl, "signature", out, "")
return err
}
// downloadHTTP retrieves url, creating a temp file using getTempFile
// http:// and https:// urls supported
func (f *fetcher) downloadHTTP(url, label string, out writeSyncer, etag string) (*cacheData, error) {
req, err := http.NewRequest("GET", url, nil)
if err != nil {
return nil, err
}
options := make(http.Header)
// Send credentials only over secure channel
if req.URL.Scheme == "https" {
if hostOpts, ok := f.headers[req.URL.Host]; ok {
options = hostOpts.Header()
}
}
for k, v := range options {
for _, e := range v {
req.Header.Add(k, e)
}
}
transport := http.DefaultTransport
if f.insecureSkipVerify {
transport = &http.Transport{
TLSClientConfig: &tls.Config{InsecureSkipVerify: true},
}
}
if etag != "" {
req.Header.Add("If-None-Match", etag)
}
req.Header.Add("User-Agent", fmt.Sprintf("rkt/%s", version.Version))
client := &http.Client{Transport: transport}
res, err := client.Do(req)
if err != nil {
return nil, err
}
defer res.Body.Close()
cd := &cacheData{}
// TODO(jonboulle): handle http more robustly (redirects?)
switch res.StatusCode {
case http.StatusAccepted:
// If the server returns Status Accepted (HTTP 202), we should retry
// downloading the signature later.
return nil, errStatusAccepted
case http.StatusOK:
fallthrough
case http.StatusNotModified:
cd.etag = res.Header.Get("ETag")
cd.maxAge = getMaxAge(res.Header.Get("Cache-Control"))
cd.useCached = (res.StatusCode == http.StatusNotModified)
if cd.useCached {
return cd, nil
}
default:
return nil, fmt.Errorf("bad HTTP status code: %d", res.StatusCode)
}
prefix := "Downloading " + label
fmtBytesSize := 18
barSize := int64(80 - len(prefix) - fmtBytesSize)
bar := ioprogress.DrawTextFormatBar(barSize)
fmtfunc := func(progress, total int64) string {
// Content-Length is set to -1 when unknown.
if total == -1 {
return fmt.Sprintf(
"%s: %v of an unknown total size",
prefix,
ioprogress.ByteUnitStr(progress),
)
}
return fmt.Sprintf(
"%s: %s %s",
prefix,
bar(progress, total),
ioprogress.DrawTextFormatBytes(progress, total),
)
}
reader := &ioprogress.Reader{
Reader: res.Body,
Size: res.ContentLength,
DrawFunc: ioprogress.DrawTerminalf(os.Stdout, fmtfunc),
DrawInterval: time.Second,
}
if _, err := io.Copy(out, reader); err != nil {
return nil, fmt.Errorf("error copying %s: %v", label, err)
}
if err := out.Sync(); err != nil {
return nil, fmt.Errorf("error writing %s: %v", label, err)
}
return cd, nil
}
func ascURLFromImgURL(imgurl string) string {
s := strings.TrimSuffix(imgurl, ".aci")
return s + ".aci.asc"
}
// newDiscoveryApp creates a discovery app if the given img is an app name and
// has a URL-like structure, for example example.com/reduce-worker.
// Or it returns nil.
func newDiscoveryApp(img string) *discovery.App {
app, err := discovery.NewAppFromString(img)
if err != nil {
return nil
}
u, err := url.Parse(app.Name.String())
if err != nil || u.Scheme != "" {
return nil
}
if _, ok := app.Labels["arch"]; !ok {
app.Labels["arch"] = defaultArch
}
if _, ok := app.Labels["os"]; !ok {
app.Labels["os"] = defaultOS
}
return app
}
func discoverApp(app *discovery.App, insecure bool) (*discovery.Endpoints, error) {
ep, attempts, err := discovery.DiscoverEndpoints(*app, insecure)
if globalFlags.Debug {
for _, a := range attempts {
stderr("meta tag 'ac-discovery' not found on %s: %v", a.Prefix, a.Error)
}
}
if err != nil {
return nil, err
}
if len(ep.ACIEndpoints) == 0 {
return nil, fmt.Errorf("no endpoints discovered")
}
return ep, nil
}
func getStoreKeyFromApp(s *store.Store, img string) (string, error) {
app, err := discovery.NewAppFromString(img)
if err != nil {
return "", fmt.Errorf("cannot parse the image name: %v", err)
}
labels, err := types.LabelsFromMap(app.Labels)
if err != nil {
return "", fmt.Errorf("invalid labels in the name: %v", err)
}
key, err := s.GetACI(app.Name, labels)
if err != nil {
return "", fmt.Errorf("cannot find image: %v", err)
}
return key, nil
}
func getStoreKeyFromAppOrHash(s *store.Store, input string) (string, error) {
var key string
if _, err := types.NewHash(input); err == nil {
key, err = s.ResolveKey(input)
if err != nil {
return "", fmt.Errorf("cannot resolve key: %v", err)
}
} else {
key, err = getStoreKeyFromApp(s, input)
if err != nil {
return "", fmt.Errorf("cannot find image: %v", err)
}
}
return key, nil
}
type cacheData struct {
useCached bool
etag string
maxAge int
}
func | getMaxAge | identifier_name |
|
images.go | File *os.File, latest bool) (string, error) {
return f.fetchImageFrom(appName, ep.ACIEndpoints[0].ACI, ep.ACIEndpoints[0].ASC, "", ascFile, latest)
}
func (f *fetcher) fetchImageFromURL(imgurl string, scheme string, ascFile *os.File, latest bool) (string, error) {
return f.fetchImageFrom("", imgurl, ascURLFromImgURL(imgurl), scheme, ascFile, latest)
}
// fetchImageFrom fetches an image from the aciURL.
// If the aciURL is a file path (scheme == 'file'), then we bypass the on-disk store.
// If the `--local` flag is provided, then we will only fetch from the on-disk store (unless aciURL is a file path).
// If the label is 'latest', then we will bypass the on-disk store (unless '--local' is specified).
// Otherwise if '--local' is false, aciURL is not a file path, and the label is not 'latest' or empty, we will first
// try to fetch from the on-disk store, if not found, then fetch from the internet.
func (f *fetcher) fetchImageFrom(appName string, aciURL, ascURL, scheme string, ascFile *os.File, latest bool) (string, error) {
var rem *store.Remote
if f.insecureSkipVerify {
if f.ks != nil {
stderr("rkt: warning: TLS verification and signature verification has been disabled")
}
} else if scheme == "docker" {
return "", fmt.Errorf("signature verification for docker images is not supported (try --insecure-skip-verify)")
}
if (f.local && scheme != "file") || (scheme != "file" && !latest) {
var err error
ok := false
rem, ok, err = f.s.GetRemote(aciURL)
if err != nil {
return "", err
}
if ok {
if f.local {
stderr("rkt: using image in local store for app %s", appName)
return rem.BlobKey, nil
}
if useCached(rem.DownloadTime, rem.CacheMaxAge) {
stderr("rkt: found image in local store, skipping fetching from %s", aciURL)
return rem.BlobKey, nil
}
}
if f.local {
return "", fmt.Errorf("url %s not available in local store", aciURL)
}
}
if scheme != "file" && f.debug {
stderr("rkt: fetching image from %s", aciURL)
}
var etag string
if rem != nil {
etag = rem.ETag
}
entity, aciFile, cd, err := f.fetch(appName, aciURL, ascURL, ascFile, etag)
if err != nil {
return "", err
}
if cd != nil && cd.useCached {
if rem != nil {
return rem.BlobKey, nil
} else {
// should never happen
panic("asked to use cached image but remote is nil")
}
}
if scheme != "file" {
defer os.Remove(aciFile.Name())
}
if entity != nil && !f.insecureSkipVerify {
stderr("rkt: signature verified:")
for _, v := range entity.Identities {
stderr(" %s", v.Name)
}
}
key, err := f.s.WriteACI(aciFile, latest)
if err != nil {
return "", err
}
if scheme != "file" {
rem := store.NewRemote(aciURL, ascURL)
rem.BlobKey = key
rem.DownloadTime = time.Now()
if cd != nil {
rem.ETag = cd.etag
rem.CacheMaxAge = cd.maxAge
}
err = f.s.WriteRemote(rem)
if err != nil {
return "", err
}
}
return key, nil
}
// fetch opens/downloads and verifies the remote ACI.
// If appName is not "", it will be used to check that the manifest contain the correct appName
// If ascFile is not nil, it will be used as the signature file and ascURL will be ignored.
// If Keystore is nil signature verification will be skipped, regardless of ascFile.
// fetch returns the signer, an *os.File representing the ACI, and an error if any.
// err will be nil if the ACI fetches successfully and the ACI is verified.
func (f *fetcher) fetch(appName string, aciURL, ascURL string, ascFile *os.File, etag string) (*openpgp.Entity, *os.File, *cacheData, error) {
var (
entity *openpgp.Entity
cd *cacheData
)
u, err := url.Parse(aciURL)
if err != nil {
return nil, nil, nil, fmt.Errorf("error parsing ACI url: %v", err)
}
if u.Scheme == "docker" {
registryURL := strings.TrimPrefix(aciURL, "docker://")
storeTmpDir, err := f.s.TmpDir()
if err != nil {
return nil, nil, nil, fmt.Errorf("error creating temporary dir for docker to ACI conversion: %v", err)
}
tmpDir, err := ioutil.TempDir(storeTmpDir, "docker2aci-")
if err != nil {
return nil, nil, nil, err
}
defer os.RemoveAll(tmpDir)
indexName := docker2aci.GetIndexName(registryURL)
user := ""
password := ""
if creds, ok := f.dockerAuth[indexName]; ok {
user = creds.User
password = creds.Password
}
acis, err := docker2aci.Convert(registryURL, true, tmpDir, tmpDir, user, password)
if err != nil {
return nil, nil, nil, fmt.Errorf("error converting docker image to ACI: %v", err)
}
aciFile, err := os.Open(acis[0])
if err != nil {
return nil, nil, nil, fmt.Errorf("error opening squashed ACI file: %v", err)
}
return nil, aciFile, nil, nil
}
// attempt to automatically fetch the public key in case it is available on a TLS connection.
if globalFlags.TrustKeysFromHttps && !globalFlags.InsecureSkipVerify && appName != "" {
pkls, err := getPubKeyLocations(appName, false, globalFlags.Debug)
if err != nil {
stderr("Error determining key location: %v", err)
} else {
// no http, don't ask user for accepting the key, no overriding
if err := addKeys(pkls, appName, false, true, false); err != nil {
stderr("Error adding keys: %v", err)
}
}
}
var retrySignature bool
if f.ks != nil && ascFile == nil {
u, err := url.Parse(ascURL)
if err != nil {
return nil, nil, nil, fmt.Errorf("error parsing ASC url: %v", err)
}
if u.Scheme == "file" {
ascFile, err = os.Open(u.Path)
if err != nil {
return nil, nil, nil, fmt.Errorf("error opening signature file: %v", err)
}
} else {
stderr("Downloading signature from %v\n", ascURL)
ascFile, err = f.s.TmpFile()
if err != nil {
return nil, nil, nil, fmt.Errorf("error setting up temporary file: %v", err)
}
defer os.Remove(ascFile.Name())
err = f.downloadSignatureFile(ascURL, ascFile)
switch err {
case errStatusAccepted:
retrySignature = true
stderr("rkt: server requested deferring the signature download")
case nil:
break
default:
return nil, nil, nil, fmt.Errorf("error downloading the signature file: %v", err)
}
}
defer ascFile.Close()
}
// check if the identity used by the signature is in the store before a
// possibly expensive download. This is only an optimization and it's
// ok to skip the test if the signature will be downloaded later.
if !retrySignature && f.ks != nil && appName != "" {
if _, err := ascFile.Seek(0, 0); err != nil {
return nil, nil, nil, fmt.Errorf("error seeking signature file: %v", err)
}
if entity, err = f.ks.CheckSignature(appName, bytes.NewReader([]byte{}), ascFile); err != nil {
if _, ok := err.(pgperrors.SignatureError); !ok {
return nil, nil, nil, err
}
}
}
var aciFile *os.File
if u.Scheme == "file" | {
aciFile, err = os.Open(u.Path)
if err != nil {
return nil, nil, nil, fmt.Errorf("error opening ACI file: %v", err)
}
} | conditional_block |
|
images.go | rem.BlobKey = key
rem.DownloadTime = time.Now()
if cd != nil {
rem.ETag = cd.etag
rem.CacheMaxAge = cd.maxAge
}
err = f.s.WriteRemote(rem)
if err != nil {
return "", err
}
}
return key, nil
}
// fetch opens/downloads and verifies the remote ACI.
// If appName is not "", it will be used to check that the manifest contain the correct appName
// If ascFile is not nil, it will be used as the signature file and ascURL will be ignored.
// If Keystore is nil signature verification will be skipped, regardless of ascFile.
// fetch returns the signer, an *os.File representing the ACI, and an error if any.
// err will be nil if the ACI fetches successfully and the ACI is verified.
func (f *fetcher) fetch(appName string, aciURL, ascURL string, ascFile *os.File, etag string) (*openpgp.Entity, *os.File, *cacheData, error) {
var (
entity *openpgp.Entity
cd *cacheData
)
u, err := url.Parse(aciURL)
if err != nil {
return nil, nil, nil, fmt.Errorf("error parsing ACI url: %v", err)
}
if u.Scheme == "docker" {
registryURL := strings.TrimPrefix(aciURL, "docker://")
storeTmpDir, err := f.s.TmpDir()
if err != nil {
return nil, nil, nil, fmt.Errorf("error creating temporary dir for docker to ACI conversion: %v", err)
}
tmpDir, err := ioutil.TempDir(storeTmpDir, "docker2aci-")
if err != nil {
return nil, nil, nil, err
}
defer os.RemoveAll(tmpDir)
indexName := docker2aci.GetIndexName(registryURL)
user := ""
password := ""
if creds, ok := f.dockerAuth[indexName]; ok {
user = creds.User
password = creds.Password
}
acis, err := docker2aci.Convert(registryURL, true, tmpDir, tmpDir, user, password)
if err != nil {
return nil, nil, nil, fmt.Errorf("error converting docker image to ACI: %v", err)
}
aciFile, err := os.Open(acis[0])
if err != nil {
return nil, nil, nil, fmt.Errorf("error opening squashed ACI file: %v", err)
}
return nil, aciFile, nil, nil
}
// attempt to automatically fetch the public key in case it is available on a TLS connection.
if globalFlags.TrustKeysFromHttps && !globalFlags.InsecureSkipVerify && appName != "" {
pkls, err := getPubKeyLocations(appName, false, globalFlags.Debug)
if err != nil {
stderr("Error determining key location: %v", err)
} else {
// no http, don't ask user for accepting the key, no overriding
if err := addKeys(pkls, appName, false, true, false); err != nil {
stderr("Error adding keys: %v", err)
}
}
}
var retrySignature bool
if f.ks != nil && ascFile == nil {
u, err := url.Parse(ascURL)
if err != nil {
return nil, nil, nil, fmt.Errorf("error parsing ASC url: %v", err)
}
if u.Scheme == "file" {
ascFile, err = os.Open(u.Path)
if err != nil {
return nil, nil, nil, fmt.Errorf("error opening signature file: %v", err)
}
} else {
stderr("Downloading signature from %v\n", ascURL)
ascFile, err = f.s.TmpFile()
if err != nil {
return nil, nil, nil, fmt.Errorf("error setting up temporary file: %v", err)
}
defer os.Remove(ascFile.Name())
err = f.downloadSignatureFile(ascURL, ascFile)
switch err {
case errStatusAccepted:
retrySignature = true
stderr("rkt: server requested deferring the signature download")
case nil:
break
default:
return nil, nil, nil, fmt.Errorf("error downloading the signature file: %v", err)
}
}
defer ascFile.Close()
}
// check if the identity used by the signature is in the store before a
// possibly expensive download. This is only an optimization and it's
// ok to skip the test if the signature will be downloaded later.
if !retrySignature && f.ks != nil && appName != "" {
if _, err := ascFile.Seek(0, 0); err != nil {
return nil, nil, nil, fmt.Errorf("error seeking signature file: %v", err)
}
if entity, err = f.ks.CheckSignature(appName, bytes.NewReader([]byte{}), ascFile); err != nil {
if _, ok := err.(pgperrors.SignatureError); !ok {
return nil, nil, nil, err
}
}
}
var aciFile *os.File
if u.Scheme == "file" {
aciFile, err = os.Open(u.Path)
if err != nil {
return nil, nil, nil, fmt.Errorf("error opening ACI file: %v", err)
}
} else {
aciFile, err = f.s.TmpFile()
if err != nil {
return nil, aciFile, nil, fmt.Errorf("error setting up temporary file: %v", err)
}
defer os.Remove(aciFile.Name())
if cd, err = f.downloadACI(aciURL, aciFile, etag); err != nil {
return nil, nil, nil, fmt.Errorf("error downloading ACI: %v", err)
}
if cd.useCached {
return nil, nil, cd, nil
}
}
if retrySignature {
if err = f.downloadSignatureFile(ascURL, ascFile); err != nil {
return nil, aciFile, nil, fmt.Errorf("error downloading the signature file: %v", err)
}
}
manifest, err := aci.ManifestFromImage(aciFile)
if err != nil {
return nil, aciFile, nil, err
}
// Check if the downloaded ACI has the correct app name.
// The check is only performed when the aci is downloaded through the
// discovery protocol, but not with local files or full URL.
if appName != "" && manifest.Name.String() != appName {
return nil, aciFile, nil,
fmt.Errorf("error when reading the app name: %q expected but %q found",
appName, manifest.Name.String())
}
if f.ks != nil {
if _, err := aciFile.Seek(0, 0); err != nil {
return nil, aciFile, nil, fmt.Errorf("error seeking ACI file: %v", err)
}
if _, err := ascFile.Seek(0, 0); err != nil {
return nil, aciFile, nil, fmt.Errorf("error seeking signature file: %v", err)
}
if entity, err = f.ks.CheckSignature(manifest.Name.String(), aciFile, ascFile); err != nil {
return nil, aciFile, nil, err
}
}
if _, err := aciFile.Seek(0, 0); err != nil {
return nil, aciFile, nil, fmt.Errorf("error seeking ACI file: %v", err)
}
return entity, aciFile, cd, nil
}
type writeSyncer interface {
io.Writer
Sync() error
}
// downloadACI gets the aci specified at aciurl
func (f *fetcher) downloadACI(aciurl string, out writeSyncer, etag string) (*cacheData, error) {
return f.downloadHTTP(aciurl, "ACI", out, etag)
}
// downloadSignatureFile gets the signature specified at sigurl
func (f *fetcher) downloadSignatureFile(sigurl string, out writeSyncer) error {
_, err := f.downloadHTTP(sigurl, "signature", out, "")
return err
}
// downloadHTTP retrieves url, creating a temp file using getTempFile
// http:// and https:// urls supported
func (f *fetcher) downloadHTTP(url, label string, out writeSyncer, etag string) (*cacheData, error) | {
req, err := http.NewRequest("GET", url, nil)
if err != nil {
return nil, err
}
options := make(http.Header)
// Send credentials only over secure channel
if req.URL.Scheme == "https" {
if hostOpts, ok := f.headers[req.URL.Host]; ok {
options = hostOpts.Header()
}
}
for k, v := range options {
for _, e := range v {
req.Header.Add(k, e)
}
}
transport := http.DefaultTransport
if f.insecureSkipVerify {
transport = &http.Transport{ | identifier_body |
|
query.go | &queryCommandeer{
rootCommandeer: rootCommandeer,
}
cmd := &cobra.Command{
Aliases: []string{"get"},
Use: "query [<metrics>] [flags]",
Short: "Query a TSDB instance",
Long: `Query a TSDB instance (table).`,
Example: `The examples assume that the endpoint of the web-gateway service, the login credentials, and
the name of the data container are configured in the default configuration file (` + config.DefaultConfigurationFileName + `)
instead of using the -s|--server, -u|--username, -p|--password, and -c|--container flags.
- tsdbctl query temperature -t mytsdb
- tsdbctl query -t performance -f "starts(__name__, 'cpu') AND os=='win'"
- tsdbctl query metric2,metric3,metric4 -t pmetrics -b 0 -e now-1h -a "sum,avg" -i 20m
- tsdbctl query -t mytsdb -f "LabelA==8.1" -l 1d -o json
- tsdbctl query metric1 -t my_tsdb -l 1d -a "count,sum,avg" --groupBy LabelA,LabelB
- tsdbctl query metric1 -t my_tsdb -l 1d -a "count_all,sum_all"
Notes:
- You must set the metric-name argument (<metrics>) and/or the query-filter flag (-f|--filter).
- Queries that set the metric-name argument (<metrics>) use range scan and are therefore faster.
- To query the full TSDB content, set the -f|--filter to a query filter that always evaluates
to true (such as "1==1"), don't set the <metrics> argument, and set the -b|--begin flag to 0.
- You can use either over-time aggregates or cross series (*_all) aggregates, but not both in the same query.
Arguments:
<metrics> (string) Comma-separated list of metric names to query. If you don't set this argument, you must
provide a query filter using the -f|--filter flag.`,
RunE: func(cmd *cobra.Command, args []string) error {
// Save the metric name if provided as a positional argument ($1)
if len(args) > 0 {
commandeer.name = args[0]
}
return commandeer.query()
},
}
cmd.Flags().StringVarP(&commandeer.to, "end", "e", "",
"End (maximum) time for the query, as a string containing an\nRFC 3339 time string, a Unix timestamp in milliseconds, or\na relative time of the format \"now\" or \"now-[0-9]+[mhd]\"\n(where 'm' = minutes, 'h' = hours, and 'd' = days).\nExamples: \"2018-09-26T14:10:20Z\"; \"1537971006000\";\n\"now-3h\"; \"now-7d\". (default \"now\")")
cmd.Flags().StringVarP(&commandeer.from, "begin", "b", "",
"Start (minimum) time for the query, as a string containing\nan RFC 3339 time, a Unix timestamp in milliseconds, a\nrelative time of the format \"now\" or \"now-[0-9]+[mhd]\"\n(where 'm' = minutes, 'h' = hours, and 'd' = days), or 0\nfor the earliest time. Examples: \"2016-01-02T15:34:26Z\";\n\"1451748866\"; \"now-90m\"; \"0\". (default = <end time> - 1h)")
cmd.Flags().StringVarP(&commandeer.output, "output", "o", formatter.DefaultOutputFormat,
"Output format in which to display the query results -\n\"text\" | \"csv\" | \"json\".")
cmd.Flags().StringVarP(&commandeer.filter, "filter", "f", "",
"Query filter, as an Iguazio Data Science Platform\nfilter expression. To reference a metric name from within\nthe query filter, use the \"__name__\" attribute.\nExamples: \"method=='get'\"; \"__name__='cpu' AND os=='win'\".")
cmd.Flags().StringVarP(&commandeer.last, "last", "l", "",
"Return data for the specified time period before the\ncurrent time, of the format \"[0-9]+[mhd]\" (where\n'm' = minutes, 'h' = hours, and 'd' = days>). When setting\nthis flag, don't set the -b|--begin or -e|--end flags.\nExamples: \"1h\"; \"15m\"; \"30d\" to return data for the last\n1 hour, 15 minutes, or 30 days.")
cmd.Flags().StringVarP(&commandeer.aggregationWindow, "aggregation-window", "w", "",
"Sliding time window for aggregation. Must be used in conjunction with `-a <aggr>`. Examples: \"1h\"; \"150m\".")
cmd.Flags().StringVarP(&commandeer.functions, "aggregates", "a", "",
"Aggregation information to return, as a comma-separated\nlist of supported aggregation functions - count | avg |\nsum | min | max | stddev | stdvar | last | rate.\nFor cross series aggregations add an \"_all\" suffix for the wanted aggregate.\nNote: you can query either over time aggregates or cross series aggregate but not both in the same query.\nExample: \"sum,min,max,count\", \"sum_all,avg_all\".")
cmd.Flags().StringVarP(&commandeer.step, "aggregation-interval", "i", "",
"Aggregation interval for applying the aggregation functions\n(if set - see the -a|--aggregates flag), of the format\n\"[0-9]+[mhd]\" (where 'm' = minutes, 'h' = hours, and\n'd' = days). Examples: \"1h\"; \"150m\". (default =\n<end time> - <start time>)")
cmd.Flags().StringVar(&commandeer.groupBy, "groupBy", "",
"Comma separated list of labels to group the result by")
cmd.Flags().BoolVar(&commandeer.usePreciseAggregations, "use-precise-aggregations", false,
"Disable server aggregation optimizations for more accurate results.")
cmd.Flags().BoolVarP(&commandeer.oldQuerier, "oldQuerier", "q", false, "use old querier")
cmd.Flags().Lookup("oldQuerier").Hidden = true
commandeer.cmd = cmd
return commandeer
}
func (qc *queryCommandeer) query() error {
if qc.name == "" && qc.filter == "" {
return errors.New("the query command must receive either a metric-name parameter (<metrics>) or a query filter (set via the -f|--filter flag)")
}
if qc.last != "" && (qc.from != "" || qc.to != "") {
return errors.New("the -l|--last flag cannot be set together with the -b|--begin and/or -e|--end flags")
}
// Initialize parameters and adapter
if err := qc.rootCommandeer.initialize(); err != nil {
return err
}
if err := qc.rootCommandeer.startAdapter(); err != nil {
return err
}
step, err := utils.Str2duration(qc.step)
if err != nil {
return err
}
// Set start & end times
to := time.Now().Unix() * 1000
if qc.to != "" {
to, err = utils.Str2unixTime(qc.to)
if err != nil {
return err
}
}
from := to - 1000*3600 // Default start time = one hour before the end time
if qc.from != "" {
from, err = utils.Str2unixTime(qc.from)
if err != nil {
return err
}
}
if qc.last != "" {
last, err := utils.Str2duration(qc.last)
if err != nil {
return err
}
from = to - last
}
qc.rootCommandeer.logger.DebugWith("Query", "from", from, "to", to, "name", qc.name,
"filter", qc.filter, "functions", qc.functions, "step", qc.step, "groupBy", qc.groupBy)
if !qc.oldQuerier {
return qc.newQuery(from, to, step)
}
return qc.oldQuery(from, to, step)
}
func (qc *queryCommandeer) newQuery(from, to, step int64) error {
qry, err := qc.rootCommandeer.adapter.QuerierV2()
if err != nil {
return errors.Wrap(err, "Failed to initialize the Querier object.")
}
aggregationWindow, err := utils.Str2duration(qc.aggregationWindow)
if err != nil | {
return errors.Wrap(err, "Failed to parse aggregation window")
} | conditional_block |
|
query.go | series (*_all) aggregates, but not both in the same query.
Arguments:
<metrics> (string) Comma-separated list of metric names to query. If you don't set this argument, you must
provide a query filter using the -f|--filter flag.`,
RunE: func(cmd *cobra.Command, args []string) error {
// Save the metric name if provided as a positional argument ($1)
if len(args) > 0 {
commandeer.name = args[0]
}
return commandeer.query()
},
}
cmd.Flags().StringVarP(&commandeer.to, "end", "e", "",
"End (maximum) time for the query, as a string containing an\nRFC 3339 time string, a Unix timestamp in milliseconds, or\na relative time of the format \"now\" or \"now-[0-9]+[mhd]\"\n(where 'm' = minutes, 'h' = hours, and 'd' = days).\nExamples: \"2018-09-26T14:10:20Z\"; \"1537971006000\";\n\"now-3h\"; \"now-7d\". (default \"now\")")
cmd.Flags().StringVarP(&commandeer.from, "begin", "b", "",
"Start (minimum) time for the query, as a string containing\nan RFC 3339 time, a Unix timestamp in milliseconds, a\nrelative time of the format \"now\" or \"now-[0-9]+[mhd]\"\n(where 'm' = minutes, 'h' = hours, and 'd' = days), or 0\nfor the earliest time. Examples: \"2016-01-02T15:34:26Z\";\n\"1451748866\"; \"now-90m\"; \"0\". (default = <end time> - 1h)")
cmd.Flags().StringVarP(&commandeer.output, "output", "o", formatter.DefaultOutputFormat,
"Output format in which to display the query results -\n\"text\" | \"csv\" | \"json\".")
cmd.Flags().StringVarP(&commandeer.filter, "filter", "f", "",
"Query filter, as an Iguazio Data Science Platform\nfilter expression. To reference a metric name from within\nthe query filter, use the \"__name__\" attribute.\nExamples: \"method=='get'\"; \"__name__='cpu' AND os=='win'\".")
cmd.Flags().StringVarP(&commandeer.last, "last", "l", "",
"Return data for the specified time period before the\ncurrent time, of the format \"[0-9]+[mhd]\" (where\n'm' = minutes, 'h' = hours, and 'd' = days>). When setting\nthis flag, don't set the -b|--begin or -e|--end flags.\nExamples: \"1h\"; \"15m\"; \"30d\" to return data for the last\n1 hour, 15 minutes, or 30 days.")
cmd.Flags().StringVarP(&commandeer.aggregationWindow, "aggregation-window", "w", "",
"Sliding time window for aggregation. Must be used in conjunction with `-a <aggr>`. Examples: \"1h\"; \"150m\".")
cmd.Flags().StringVarP(&commandeer.functions, "aggregates", "a", "",
"Aggregation information to return, as a comma-separated\nlist of supported aggregation functions - count | avg |\nsum | min | max | stddev | stdvar | last | rate.\nFor cross series aggregations add an \"_all\" suffix for the wanted aggregate.\nNote: you can query either over time aggregates or cross series aggregate but not both in the same query.\nExample: \"sum,min,max,count\", \"sum_all,avg_all\".")
cmd.Flags().StringVarP(&commandeer.step, "aggregation-interval", "i", "",
"Aggregation interval for applying the aggregation functions\n(if set - see the -a|--aggregates flag), of the format\n\"[0-9]+[mhd]\" (where 'm' = minutes, 'h' = hours, and\n'd' = days). Examples: \"1h\"; \"150m\". (default =\n<end time> - <start time>)")
cmd.Flags().StringVar(&commandeer.groupBy, "groupBy", "",
"Comma separated list of labels to group the result by")
cmd.Flags().BoolVar(&commandeer.usePreciseAggregations, "use-precise-aggregations", false,
"Disable server aggregation optimizations for more accurate results.")
cmd.Flags().BoolVarP(&commandeer.oldQuerier, "oldQuerier", "q", false, "use old querier")
cmd.Flags().Lookup("oldQuerier").Hidden = true
commandeer.cmd = cmd
return commandeer
}
func (qc *queryCommandeer) query() error {
if qc.name == "" && qc.filter == "" {
return errors.New("the query command must receive either a metric-name parameter (<metrics>) or a query filter (set via the -f|--filter flag)")
}
if qc.last != "" && (qc.from != "" || qc.to != "") {
return errors.New("the -l|--last flag cannot be set together with the -b|--begin and/or -e|--end flags")
}
// Initialize parameters and adapter
if err := qc.rootCommandeer.initialize(); err != nil {
return err
}
if err := qc.rootCommandeer.startAdapter(); err != nil {
return err
}
step, err := utils.Str2duration(qc.step)
if err != nil {
return err
}
// Set start & end times
to := time.Now().Unix() * 1000
if qc.to != "" {
to, err = utils.Str2unixTime(qc.to)
if err != nil {
return err
}
}
from := to - 1000*3600 // Default start time = one hour before the end time
if qc.from != "" {
from, err = utils.Str2unixTime(qc.from)
if err != nil {
return err
}
}
if qc.last != "" {
last, err := utils.Str2duration(qc.last)
if err != nil {
return err
}
from = to - last
}
qc.rootCommandeer.logger.DebugWith("Query", "from", from, "to", to, "name", qc.name,
"filter", qc.filter, "functions", qc.functions, "step", qc.step, "groupBy", qc.groupBy)
if !qc.oldQuerier {
return qc.newQuery(from, to, step)
}
return qc.oldQuery(from, to, step)
}
func (qc *queryCommandeer) newQuery(from, to, step int64) error {
qry, err := qc.rootCommandeer.adapter.QuerierV2()
if err != nil {
return errors.Wrap(err, "Failed to initialize the Querier object.")
}
aggregationWindow, err := utils.Str2duration(qc.aggregationWindow)
if err != nil {
return errors.Wrap(err, "Failed to parse aggregation window")
}
var selectParams *pquerier.SelectParams
if strings.HasPrefix(qc.name, "select") {
selectParams, _, err = pquerier.ParseQuery(qc.name)
if err != nil {
return errors.Wrap(err, "failed to parse sql")
}
selectParams.Step = step
selectParams.From = from
selectParams.To = to
selectParams.UseOnlyClientAggr = qc.usePreciseAggregations
selectParams.AggregationWindow = aggregationWindow
} else {
selectParams = &pquerier.SelectParams{Name: qc.name, Functions: qc.functions,
Step: step, Filter: qc.filter, From: from, To: to, GroupBy: qc.groupBy,
UseOnlyClientAggr: qc.usePreciseAggregations,
AggregationWindow: aggregationWindow}
}
set, err := qry.Select(selectParams)
if err != nil {
return errors.Wrap(err, "The query selection failed.")
}
f, err := formatter.NewFormatter(qc.output, nil)
if err != nil {
return errors.Wrapf(err, "Failed to start formatter '%s'.", qc.output)
}
err = f.Write(qc.cmd.OutOrStdout(), set)
return err
}
func (qc *queryCommandeer) oldQuery(from, to, step int64) error {
qry, err := qc.rootCommandeer.adapter.Querier(context.TODO(), from, to)
if err != nil {
return errors.Wrap(err, "Failed to initialize the Querier object.")
}
var set utils.SeriesSet
set, err = qry.Select(qc.name, qc.functions, step, qc.filter)
if err != nil {
return errors.Wrap(err, "The query selection failed.")
}
f, err := formatter.NewFormatter(qc.output, nil)
if err != nil { | return errors.Wrapf(err, "Failed to start formatter '%s'.", qc.output)
}
| random_line_split |
|
query.go | *RootCommandeer
name string
filter string
to string
from string
last string
functions string
step string
output string
oldQuerier bool
groupBy string
usePreciseAggregations bool
aggregationWindow string
}
func newQueryCommandeer(rootCommandeer *RootCommandeer) *queryCommandeer {
commandeer := &queryCommandeer{
rootCommandeer: rootCommandeer,
}
cmd := &cobra.Command{
Aliases: []string{"get"},
Use: "query [<metrics>] [flags]",
Short: "Query a TSDB instance",
Long: `Query a TSDB instance (table).`,
Example: `The examples assume that the endpoint of the web-gateway service, the login credentials, and
the name of the data container are configured in the default configuration file (` + config.DefaultConfigurationFileName + `)
instead of using the -s|--server, -u|--username, -p|--password, and -c|--container flags.
- tsdbctl query temperature -t mytsdb
- tsdbctl query -t performance -f "starts(__name__, 'cpu') AND os=='win'"
- tsdbctl query metric2,metric3,metric4 -t pmetrics -b 0 -e now-1h -a "sum,avg" -i 20m
- tsdbctl query -t mytsdb -f "LabelA==8.1" -l 1d -o json
- tsdbctl query metric1 -t my_tsdb -l 1d -a "count,sum,avg" --groupBy LabelA,LabelB
- tsdbctl query metric1 -t my_tsdb -l 1d -a "count_all,sum_all"
Notes:
- You must set the metric-name argument (<metrics>) and/or the query-filter flag (-f|--filter).
- Queries that set the metric-name argument (<metrics>) use range scan and are therefore faster.
- To query the full TSDB content, set the -f|--filter to a query filter that always evaluates
to true (such as "1==1"), don't set the <metrics> argument, and set the -b|--begin flag to 0.
- You can use either over-time aggregates or cross series (*_all) aggregates, but not both in the same query.
Arguments:
<metrics> (string) Comma-separated list of metric names to query. If you don't set this argument, you must
provide a query filter using the -f|--filter flag.`,
RunE: func(cmd *cobra.Command, args []string) error {
// Save the metric name if provided as a positional argument ($1)
if len(args) > 0 {
commandeer.name = args[0]
}
return commandeer.query()
},
}
cmd.Flags().StringVarP(&commandeer.to, "end", "e", "",
"End (maximum) time for the query, as a string containing an\nRFC 3339 time string, a Unix timestamp in milliseconds, or\na relative time of the format \"now\" or \"now-[0-9]+[mhd]\"\n(where 'm' = minutes, 'h' = hours, and 'd' = days).\nExamples: \"2018-09-26T14:10:20Z\"; \"1537971006000\";\n\"now-3h\"; \"now-7d\". (default \"now\")")
cmd.Flags().StringVarP(&commandeer.from, "begin", "b", "",
"Start (minimum) time for the query, as a string containing\nan RFC 3339 time, a Unix timestamp in milliseconds, a\nrelative time of the format \"now\" or \"now-[0-9]+[mhd]\"\n(where 'm' = minutes, 'h' = hours, and 'd' = days), or 0\nfor the earliest time. Examples: \"2016-01-02T15:34:26Z\";\n\"1451748866\"; \"now-90m\"; \"0\". (default = <end time> - 1h)")
cmd.Flags().StringVarP(&commandeer.output, "output", "o", formatter.DefaultOutputFormat,
"Output format in which to display the query results -\n\"text\" | \"csv\" | \"json\".")
cmd.Flags().StringVarP(&commandeer.filter, "filter", "f", "",
"Query filter, as an Iguazio Data Science Platform\nfilter expression. To reference a metric name from within\nthe query filter, use the \"__name__\" attribute.\nExamples: \"method=='get'\"; \"__name__='cpu' AND os=='win'\".")
cmd.Flags().StringVarP(&commandeer.last, "last", "l", "",
"Return data for the specified time period before the\ncurrent time, of the format \"[0-9]+[mhd]\" (where\n'm' = minutes, 'h' = hours, and 'd' = days>). When setting\nthis flag, don't set the -b|--begin or -e|--end flags.\nExamples: \"1h\"; \"15m\"; \"30d\" to return data for the last\n1 hour, 15 minutes, or 30 days.")
cmd.Flags().StringVarP(&commandeer.aggregationWindow, "aggregation-window", "w", "",
"Sliding time window for aggregation. Must be used in conjunction with `-a <aggr>`. Examples: \"1h\"; \"150m\".")
cmd.Flags().StringVarP(&commandeer.functions, "aggregates", "a", "",
"Aggregation information to return, as a comma-separated\nlist of supported aggregation functions - count | avg |\nsum | min | max | stddev | stdvar | last | rate.\nFor cross series aggregations add an \"_all\" suffix for the wanted aggregate.\nNote: you can query either over time aggregates or cross series aggregate but not both in the same query.\nExample: \"sum,min,max,count\", \"sum_all,avg_all\".")
cmd.Flags().StringVarP(&commandeer.step, "aggregation-interval", "i", "",
"Aggregation interval for applying the aggregation functions\n(if set - see the -a|--aggregates flag), of the format\n\"[0-9]+[mhd]\" (where 'm' = minutes, 'h' = hours, and\n'd' = days). Examples: \"1h\"; \"150m\". (default =\n<end time> - <start time>)")
cmd.Flags().StringVar(&commandeer.groupBy, "groupBy", "",
"Comma separated list of labels to group the result by")
cmd.Flags().BoolVar(&commandeer.usePreciseAggregations, "use-precise-aggregations", false,
"Disable server aggregation optimizations for more accurate results.")
cmd.Flags().BoolVarP(&commandeer.oldQuerier, "oldQuerier", "q", false, "use old querier")
cmd.Flags().Lookup("oldQuerier").Hidden = true
commandeer.cmd = cmd
return commandeer
}
func (qc *queryCommandeer) query() error {
if qc.name == "" && qc.filter == "" {
return errors.New("the query command must receive either a metric-name parameter (<metrics>) or a query filter (set via the -f|--filter flag)")
}
if qc.last != "" && (qc.from != "" || qc.to != "") {
return errors.New("the -l|--last flag cannot be set together with the -b|--begin and/or -e|--end flags")
}
// Initialize parameters and adapter
if err := qc.rootCommandeer.initialize(); err != nil {
return err
}
if err := qc.rootCommandeer.startAdapter(); err != nil {
return err
}
step, err := utils.Str2duration(qc.step)
if err != nil {
return err
}
// Set start & end times
to := time.Now().Unix() * 1000
if qc.to != "" {
to, err = utils.Str2unixTime(qc.to)
if err != nil {
return err
}
}
from := to - 1000*3600 // Default start time = one hour before the end time
if qc.from != "" {
from, err = utils.Str2unixTime(qc.from)
if err != nil {
return err
}
}
if qc.last != "" {
last, err := utils.Str2duration(qc.last)
if err != nil {
return err
}
from = to - last
}
qc.rootCommandeer.logger.DebugWith("Query", "from", from, "to", to, "name", qc.name,
"filter", qc.filter, "functions", qc.functions, "step", qc.step, "groupBy", qc.groupBy)
if !qc.oldQuerier {
return qc.newQuery(from, to, step)
}
return qc.oldQuery(from, to, step)
}
func (qc *queryCommandeer) | newQuery | identifier_name |
|
query.go | "github.com/v3io/v3io-tsdb/pkg/pquerier"
"github.com/v3io/v3io-tsdb/pkg/utils"
)
type queryCommandeer struct {
cmd *cobra.Command
rootCommandeer *RootCommandeer
name string
filter string
to string
from string
last string
functions string
step string
output string
oldQuerier bool
groupBy string
usePreciseAggregations bool
aggregationWindow string
}
func newQueryCommandeer(rootCommandeer *RootCommandeer) *queryCommandeer {
commandeer := &queryCommandeer{
rootCommandeer: rootCommandeer,
}
cmd := &cobra.Command{
Aliases: []string{"get"},
Use: "query [<metrics>] [flags]",
Short: "Query a TSDB instance",
Long: `Query a TSDB instance (table).`,
Example: `The examples assume that the endpoint of the web-gateway service, the login credentials, and
the name of the data container are configured in the default configuration file (` + config.DefaultConfigurationFileName + `)
instead of using the -s|--server, -u|--username, -p|--password, and -c|--container flags.
- tsdbctl query temperature -t mytsdb
- tsdbctl query -t performance -f "starts(__name__, 'cpu') AND os=='win'"
- tsdbctl query metric2,metric3,metric4 -t pmetrics -b 0 -e now-1h -a "sum,avg" -i 20m
- tsdbctl query -t mytsdb -f "LabelA==8.1" -l 1d -o json
- tsdbctl query metric1 -t my_tsdb -l 1d -a "count,sum,avg" --groupBy LabelA,LabelB
- tsdbctl query metric1 -t my_tsdb -l 1d -a "count_all,sum_all"
Notes:
- You must set the metric-name argument (<metrics>) and/or the query-filter flag (-f|--filter).
- Queries that set the metric-name argument (<metrics>) use range scan and are therefore faster.
- To query the full TSDB content, set the -f|--filter to a query filter that always evaluates
to true (such as "1==1"), don't set the <metrics> argument, and set the -b|--begin flag to 0.
- You can use either over-time aggregates or cross series (*_all) aggregates, but not both in the same query.
Arguments:
<metrics> (string) Comma-separated list of metric names to query. If you don't set this argument, you must
provide a query filter using the -f|--filter flag.`,
RunE: func(cmd *cobra.Command, args []string) error {
// Save the metric name if provided as a positional argument ($1)
if len(args) > 0 {
commandeer.name = args[0]
}
return commandeer.query()
},
}
cmd.Flags().StringVarP(&commandeer.to, "end", "e", "",
"End (maximum) time for the query, as a string containing an\nRFC 3339 time string, a Unix timestamp in milliseconds, or\na relative time of the format \"now\" or \"now-[0-9]+[mhd]\"\n(where 'm' = minutes, 'h' = hours, and 'd' = days).\nExamples: \"2018-09-26T14:10:20Z\"; \"1537971006000\";\n\"now-3h\"; \"now-7d\". (default \"now\")")
cmd.Flags().StringVarP(&commandeer.from, "begin", "b", "",
"Start (minimum) time for the query, as a string containing\nan RFC 3339 time, a Unix timestamp in milliseconds, a\nrelative time of the format \"now\" or \"now-[0-9]+[mhd]\"\n(where 'm' = minutes, 'h' = hours, and 'd' = days), or 0\nfor the earliest time. Examples: \"2016-01-02T15:34:26Z\";\n\"1451748866\"; \"now-90m\"; \"0\". (default = <end time> - 1h)")
cmd.Flags().StringVarP(&commandeer.output, "output", "o", formatter.DefaultOutputFormat,
"Output format in which to display the query results -\n\"text\" | \"csv\" | \"json\".")
cmd.Flags().StringVarP(&commandeer.filter, "filter", "f", "",
"Query filter, as an Iguazio Data Science Platform\nfilter expression. To reference a metric name from within\nthe query filter, use the \"__name__\" attribute.\nExamples: \"method=='get'\"; \"__name__='cpu' AND os=='win'\".")
cmd.Flags().StringVarP(&commandeer.last, "last", "l", "",
"Return data for the specified time period before the\ncurrent time, of the format \"[0-9]+[mhd]\" (where\n'm' = minutes, 'h' = hours, and 'd' = days>). When setting\nthis flag, don't set the -b|--begin or -e|--end flags.\nExamples: \"1h\"; \"15m\"; \"30d\" to return data for the last\n1 hour, 15 minutes, or 30 days.")
cmd.Flags().StringVarP(&commandeer.aggregationWindow, "aggregation-window", "w", "",
"Sliding time window for aggregation. Must be used in conjunction with `-a <aggr>`. Examples: \"1h\"; \"150m\".")
cmd.Flags().StringVarP(&commandeer.functions, "aggregates", "a", "",
"Aggregation information to return, as a comma-separated\nlist of supported aggregation functions - count | avg |\nsum | min | max | stddev | stdvar | last | rate.\nFor cross series aggregations add an \"_all\" suffix for the wanted aggregate.\nNote: you can query either over time aggregates or cross series aggregate but not both in the same query.\nExample: \"sum,min,max,count\", \"sum_all,avg_all\".")
cmd.Flags().StringVarP(&commandeer.step, "aggregation-interval", "i", "",
"Aggregation interval for applying the aggregation functions\n(if set - see the -a|--aggregates flag), of the format\n\"[0-9]+[mhd]\" (where 'm' = minutes, 'h' = hours, and\n'd' = days). Examples: \"1h\"; \"150m\". (default =\n<end time> - <start time>)")
cmd.Flags().StringVar(&commandeer.groupBy, "groupBy", "",
"Comma separated list of labels to group the result by")
cmd.Flags().BoolVar(&commandeer.usePreciseAggregations, "use-precise-aggregations", false,
"Disable server aggregation optimizations for more accurate results.")
cmd.Flags().BoolVarP(&commandeer.oldQuerier, "oldQuerier", "q", false, "use old querier")
cmd.Flags().Lookup("oldQuerier").Hidden = true
commandeer.cmd = cmd
return commandeer
}
func (qc *queryCommandeer) query() error | if err != nil {
return err
}
// Set start & end times
to := time.Now().Unix() * 1000
if qc.to != "" {
to, err = utils.Str2unixTime(qc.to)
if err != nil {
return err
}
}
from := to - 1000*3600 // Default start time = one hour before the end time
if qc.from != "" {
from, err = utils.Str2unixTime(qc.from)
if err != nil {
return err
}
}
if qc.last != "" {
last, err := utils.Str2duration(qc.last)
if err != nil {
return err
}
from = to - last
}
qc.rootCommandeer.logger.DebugWith("Query", "from", from, "to", to, "name", qc.name,
"filter", qc.filter, "functions", qc.functions, "step", qc.step, "groupBy | {
if qc.name == "" && qc.filter == "" {
return errors.New("the query command must receive either a metric-name parameter (<metrics>) or a query filter (set via the -f|--filter flag)")
}
if qc.last != "" && (qc.from != "" || qc.to != "") {
return errors.New("the -l|--last flag cannot be set together with the -b|--begin and/or -e|--end flags")
}
// Initialize parameters and adapter
if err := qc.rootCommandeer.initialize(); err != nil {
return err
}
if err := qc.rootCommandeer.startAdapter(); err != nil {
return err
}
step, err := utils.Str2duration(qc.step) | identifier_body |
main0.rs | mut arr[..]);
// let data = vec![0x83, b'c', b'a', b't'];
// let aa: String = rlp::decode(&data).unwrap();
// println!("aa = {:?}", aa);
// let pk = hex::decode("ee5495585eff78f2fcf95bab21ef1a598c54d1e3c672e23b3bb97a4fc7490660").unwrap();
let private_key = SecretKey::parse_slice(&arr[0..arr.len()]).unwrap();
// let private_key = SecretKey::parse_slice(&pk).unwrap();
let pubkey = PublicKey::from_secret_key(&private_key);
let id = &pubkey.serialize().to_vec()[1..];
println!("id is {:?}", hex::encode(&id));
const CLIENT: Token = Token(0);
const SENDER: Token = Token(0);
let udp_server_ip = "35.180.217.147";
let udp_server_port = "30304";
let upd_server_addr = "35.180.217.147:30304";
let local_udp_addr = "192.168.31.125:30309";
let mut udp_socket = UdpSocket::bind(local_udp_addr.parse()?)?;
// let local_addr = udp_socket.local_addr()?;
//
// println!("local_addr = {:?}", local_addr);
println!("private_key is {:?}", private_key);
let mut poll = Poll::new()?;
let mut events = Events::with_capacity(1024);
let addr = "192.168.31.248:30303".parse()?;
let peer = PeerInfo::from_sock_addr(&upd_server_addr.parse()?);
let local_peer = PeerInfo::from_sock_addr(&local_udp_addr.parse()?);
let mut sent_ping = false;
// message::encode_ping(&peer, &peer, &private_key);
// println!("peer ip {:?}", peer.encode());
// let addr = "127.0.0.1:9000".parse()?;
let mut send_queue: VecDeque<Vec<u8>> = VecDeque::new();
let mut client = TcpStream::connect(addr)?;
let mut status_sent = false;
poll.registry().register(&mut client, CLIENT, Interest::READABLE | Interest::WRITABLE)?;
poll.registry().register(&mut udp_socket, SENDER, Interest::READABLE | Interest::WRITABLE)?;
let mut received_data = Vec::with_capacity(4096);
send_queue.push_back(message::encode_ping(&local_peer, &peer, &private_key));
loop {
poll.poll(&mut events, None)?;
for event in events.iter() {
match event.token() {
SENDER => {
println!("udp socket is active");
if event.is_writable() {
'inner: loop {
if let Some(buf) = send_queue.pop_front() {
match udp_socket.send_to(&buf, upd_server_addr.parse()?) {
Ok(size) => {
println!("sent {:?} bytes(total {:?})", size, buf.len());
// we have some buf remain for next time
if size < buf.len() {
if size == 0 {
send_queue.push_front(buf);
}
break 'inner;
}
},
Err(e) => {
println!("send error {:?}", e);
break 'inner;
}
}
} else {
println!("no data to send, reregister for next writable event");
break 'inner;
}
}
}
if event.is_readable() {
'read: loop {
let mut buf = [0; 1024];
match udp_socket.recv_from(&mut buf) {
Ok((size, addr)) => {
println!("read {:?} bytes from {:?}", size, addr);
if (size > 0) {
let read_buf = &buf[..size];
let hash_signed = keccak(&read_buf[32..]);
println!("hash_signed = {:?}", hash_signed);
println!("check_sum = {:?}", hex::encode(&read_buf[0..32]));
// if hash_signed.as_bytes() != &read_buf[0..32] {
// // return Box::new(Err("bad protocol"));
// break;
// }
let signed = &read_buf[(32 + 65)..];
let message_type = signed[0];
print_message_type(message_type);
println!("message_type is {:?}", message_type);
let recover_id = RecoveryId::parse(read_buf[32 + 64]).expect("can not get recover id");
println!("recover_id = {:?}", recover_id);
let signature = Signature::parse_slice(&read_buf[32..(32 + 64)]).expect("can not get signature");
let hash = keccak(signed);
let pubkey = recover(&Message::parse_slice(&hash).unwrap(), &signature, &recover_id).expect("can not recover pubkey");
println!("pubkey is {:?}", hex::encode(&pubkey.serialize_compressed().to_vec()));
let rlp = Rlp::new(&signed[1..]);
if message_type == 0x01 {
// got a ping message
let version: u8 = rlp.val_at(0)?;
let from_peer = PeerInfo::decode_rlp(&rlp.at(1)?)?;
let to_peer = PeerInfo::decode_rlp(&rlp.at(2)?)?;
println!("from_peer = {:?}, to_peer = {:?}", from_peer, to_peer);
let timestamp: u64 = rlp.val_at(3)?;
println!("version = {:?}, timestamp = {:?}", version, timestamp);
// send pong message
let from = PeerInfo::from_sock_addr(&addr);
let bytes = message::encode_pong(&from, &read_buf[0..32].to_vec(), ×tamp, &private_key); |
} else if message_type == 0x02 {
// got a pong message
let from_peer = PeerInfo::decode_rlp(&rlp.at(0)?)?;
let hash_bytes = rlp.at(1)?.data()?;
let timestamp: u64 = rlp.val_at(2)?;
println!("got a pong message {:?} {:?}", from_peer, timestamp);
// start send findneighbours packet
let bytes = message::encode_find_node(&private_key);
println!("find node bytes is {:?}", bytes.len());
send_queue.push_back(bytes);
} else if message_type == 0x03 {
println!("got a find node message");
} else if message_type == 0x04 {
println!("got a node message");
}
poll.registry().reregister(&mut udp_socket, event.token(), Interest::WRITABLE)?;
// we have read all data
if (size < buf.len()) {
println!("no more data read");
break 'read;
}
} else {
println!("no data read");
break 'read;
}
},
Err(e) => {
println!("read error {:?}", e);
break 'read;
}
}
}
}
},
CLIENT => {
if event.is_readable() {
println!("client socket is readable");
// read buf
let mut buf = [0; 1024];
match client.read(&mut buf) {
Ok(n) => {
if (n > 0) {
received_data.extend_from_slice(&buf[..n]);
println!("read data: {:?}", String::from_utf8_lossy(&received_data));
}
println!("read {:?} bytes", n);
},
Err(err) => {
println!("read data error {:?}", err);
}
}
}
if event.is_writable() {
// send auth info
let auth = "0196045fa704aa5f5a85f36c6b399b08d823083228d63c4346f382f78a18b684f3a4e64a671de498abf20cba88dd8f3f0a11443bed18248895b981e0c842e9e4fafe387cf9ad619ba89fe7dbfa6f504725bb673a804f3526df31c68a69caf9bc7a9eed62fe73dffdeae5e21f55e2a1ec28e17ad5f98bd0a61759fe25f8f96665278197413d86ab84ea2f3adbf70634b49d13b4b550 | println!("pong bytes is {:?}", bytes.len());
send_queue.push_back(bytes);
// send_queue | random_line_split |
main0.rs | mut send_queue: VecDeque<Vec<u8>> = VecDeque::new();
let mut client = TcpStream::connect(addr)?;
let mut status_sent = false;
poll.registry().register(&mut client, CLIENT, Interest::READABLE | Interest::WRITABLE)?;
poll.registry().register(&mut udp_socket, SENDER, Interest::READABLE | Interest::WRITABLE)?;
let mut received_data = Vec::with_capacity(4096);
send_queue.push_back(message::encode_ping(&local_peer, &peer, &private_key));
loop {
poll.poll(&mut events, None)?;
for event in events.iter() {
match event.token() {
SENDER => {
println!("udp socket is active");
if event.is_writable() {
'inner: loop {
if let Some(buf) = send_queue.pop_front() {
match udp_socket.send_to(&buf, upd_server_addr.parse()?) {
Ok(size) => {
println!("sent {:?} bytes(total {:?})", size, buf.len());
// we have some buf remain for next time
if size < buf.len() {
if size == 0 {
send_queue.push_front(buf);
}
break 'inner;
}
},
Err(e) => {
println!("send error {:?}", e);
break 'inner;
}
}
} else {
println!("no data to send, reregister for next writable event");
break 'inner;
}
}
}
if event.is_readable() {
'read: loop {
let mut buf = [0; 1024];
match udp_socket.recv_from(&mut buf) {
Ok((size, addr)) => {
println!("read {:?} bytes from {:?}", size, addr);
if (size > 0) {
let read_buf = &buf[..size];
let hash_signed = keccak(&read_buf[32..]);
println!("hash_signed = {:?}", hash_signed);
println!("check_sum = {:?}", hex::encode(&read_buf[0..32]));
// if hash_signed.as_bytes() != &read_buf[0..32] {
// // return Box::new(Err("bad protocol"));
// break;
// }
let signed = &read_buf[(32 + 65)..];
let message_type = signed[0];
print_message_type(message_type);
println!("message_type is {:?}", message_type);
let recover_id = RecoveryId::parse(read_buf[32 + 64]).expect("can not get recover id");
println!("recover_id = {:?}", recover_id);
let signature = Signature::parse_slice(&read_buf[32..(32 + 64)]).expect("can not get signature");
let hash = keccak(signed);
let pubkey = recover(&Message::parse_slice(&hash).unwrap(), &signature, &recover_id).expect("can not recover pubkey");
println!("pubkey is {:?}", hex::encode(&pubkey.serialize_compressed().to_vec()));
let rlp = Rlp::new(&signed[1..]);
if message_type == 0x01 {
// got a ping message
let version: u8 = rlp.val_at(0)?;
let from_peer = PeerInfo::decode_rlp(&rlp.at(1)?)?;
let to_peer = PeerInfo::decode_rlp(&rlp.at(2)?)?;
println!("from_peer = {:?}, to_peer = {:?}", from_peer, to_peer);
let timestamp: u64 = rlp.val_at(3)?;
println!("version = {:?}, timestamp = {:?}", version, timestamp);
// send pong message
let from = PeerInfo::from_sock_addr(&addr);
let bytes = message::encode_pong(&from, &read_buf[0..32].to_vec(), ×tamp, &private_key);
println!("pong bytes is {:?}", bytes.len());
send_queue.push_back(bytes);
// send_queue
} else if message_type == 0x02 {
// got a pong message
let from_peer = PeerInfo::decode_rlp(&rlp.at(0)?)?;
let hash_bytes = rlp.at(1)?.data()?;
let timestamp: u64 = rlp.val_at(2)?;
println!("got a pong message {:?} {:?}", from_peer, timestamp);
// start send findneighbours packet
let bytes = message::encode_find_node(&private_key);
println!("find node bytes is {:?}", bytes.len());
send_queue.push_back(bytes);
} else if message_type == 0x03 {
println!("got a find node message");
} else if message_type == 0x04 {
println!("got a node message");
}
poll.registry().reregister(&mut udp_socket, event.token(), Interest::WRITABLE)?;
// we have read all data
if (size < buf.len()) {
println!("no more data read");
break 'read;
}
} else {
println!("no data read");
break 'read;
}
},
Err(e) => {
println!("read error {:?}", e);
break 'read;
}
}
}
}
},
CLIENT => {
if event.is_readable() {
println!("client socket is readable");
// read buf
let mut buf = [0; 1024];
match client.read(&mut buf) {
Ok(n) => {
if (n > 0) {
received_data.extend_from_slice(&buf[..n]);
println!("read data: {:?}", String::from_utf8_lossy(&received_data));
}
println!("read {:?} bytes", n);
},
Err(err) => {
println!("read data error {:?}", err);
}
}
}
if event.is_writable() {
// send auth info
let auth = "0196045fa704aa5f5a85f36c6b399b08d823083228d63c4346f382f78a18b684f3a4e64a671de498abf20cba88dd8f3f0a11443bed18248895b981e0c842e9e4fafe387cf9ad619ba89fe7dbfa6f504725bb673a804f3526df31c68a69caf9bc7a9eed62fe73dffdeae5e21f55e2a1ec28e17ad5f98bd0a61759fe25f8f96665278197413d86ab84ea2f3adbf70634b49d13b4b55037e23f393ddc2ae46e63d4c3d1b67945bcf22d03183a1b1ff3b9b74cf3d83a8093489b508759c5042ca0d7de29aa6eb024800868594f848f646f1488c7bbf2a598d411a7333db52168f53e04e28b260e218233e9641232304625ba67cbaa7b6a3703161235ab41758d466701beac1a08e5edc612e42cb7235d43cbdd51ff7bb3cbe4720dfa165f084dafce2c84795eb619016647c9aef4d6d9b31e1a4b1e3b18e856a025ab99275b8b860816259ddf86cdc20c22e0f6f70445258113fade6d38814cb88d8c0693a64880088563cb02ff15236bca24720aaaa9da219c0f2fa71f8a4b1e34793a330b31ccfbdcbaf0026c881d5761b198be428feb93b170afe95174722f";
let buf = hex::decode(auth).unwrap();
if !status_sent {
status_sent = true;
match client.write_all(&buf) {
Ok(_) => | {
println!("write ok");
} | conditional_block |
|
main0.rs | },
}
}
fn main() -> Result<(), Box<dyn Error>> {
let mut arr = [0u8; 32];
thread_rng().fill(&mut arr[..]);
// let data = vec![0x83, b'c', b'a', b't'];
// let aa: String = rlp::decode(&data).unwrap();
// println!("aa = {:?}", aa);
// let pk = hex::decode("ee5495585eff78f2fcf95bab21ef1a598c54d1e3c672e23b3bb97a4fc7490660").unwrap();
let private_key = SecretKey::parse_slice(&arr[0..arr.len()]).unwrap();
// let private_key = SecretKey::parse_slice(&pk).unwrap();
let pubkey = PublicKey::from_secret_key(&private_key);
let id = &pubkey.serialize().to_vec()[1..];
println!("id is {:?}", hex::encode(&id));
const CLIENT: Token = Token(0);
const SENDER: Token = Token(0);
let udp_server_ip = "35.180.217.147";
let udp_server_port = "30304";
let upd_server_addr = "35.180.217.147:30304";
let local_udp_addr = "192.168.31.125:30309";
let mut udp_socket = UdpSocket::bind(local_udp_addr.parse()?)?;
// let local_addr = udp_socket.local_addr()?;
//
// println!("local_addr = {:?}", local_addr);
println!("private_key is {:?}", private_key);
let mut poll = Poll::new()?;
let mut events = Events::with_capacity(1024);
let addr = "192.168.31.248:30303".parse()?;
let peer = PeerInfo::from_sock_addr(&upd_server_addr.parse()?);
let local_peer = PeerInfo::from_sock_addr(&local_udp_addr.parse()?);
let mut sent_ping = false;
// message::encode_ping(&peer, &peer, &private_key);
// println!("peer ip {:?}", peer.encode());
// let addr = "127.0.0.1:9000".parse()?;
let mut send_queue: VecDeque<Vec<u8>> = VecDeque::new();
let mut client = TcpStream::connect(addr)?;
let mut status_sent = false;
poll.registry().register(&mut client, CLIENT, Interest::READABLE | Interest::WRITABLE)?;
poll.registry().register(&mut udp_socket, SENDER, Interest::READABLE | Interest::WRITABLE)?;
let mut received_data = Vec::with_capacity(4096);
send_queue.push_back(message::encode_ping(&local_peer, &peer, &private_key));
loop {
poll.poll(&mut events, None)?;
for event in events.iter() {
match event.token() {
SENDER => {
println!("udp socket is active");
if event.is_writable() {
'inner: loop {
if let Some(buf) = send_queue.pop_front() {
match udp_socket.send_to(&buf, upd_server_addr.parse()?) {
Ok(size) => {
println!("sent {:?} bytes(total {:?})", size, buf.len());
// we have some buf remain for next time
if size < buf.len() {
if size == 0 {
send_queue.push_front(buf);
}
break 'inner;
}
},
Err(e) => {
println!("send error {:?}", e);
break 'inner;
}
}
} else {
println!("no data to send, reregister for next writable event");
break 'inner;
}
}
}
if event.is_readable() {
'read: loop {
let mut buf = [0; 1024];
match udp_socket.recv_from(&mut buf) {
Ok((size, addr)) => {
println!("read {:?} bytes from {:?}", size, addr);
if (size > 0) {
let read_buf = &buf[..size];
let hash_signed = keccak(&read_buf[32..]);
println!("hash_signed = {:?}", hash_signed);
println!("check_sum = {:?}", hex::encode(&read_buf[0..32]));
// if hash_signed.as_bytes() != &read_buf[0..32] {
// // return Box::new(Err("bad protocol"));
// break;
// }
let signed = &read_buf[(32 + 65)..];
let message_type = signed[0];
print_message_type(message_type);
println!("message_type is {:?}", message_type);
let recover_id = RecoveryId::parse(read_buf[32 + 64]).expect("can not get recover id");
println!("recover_id = {:?}", recover_id);
let signature = Signature::parse_slice(&read_buf[32..(32 + 64)]).expect("can not get signature");
let hash = keccak(signed);
let pubkey = recover(&Message::parse_slice(&hash).unwrap(), &signature, &recover_id).expect("can not recover pubkey");
println!("pubkey is {:?}", hex::encode(&pubkey.serialize_compressed().to_vec()));
let rlp = Rlp::new(&signed[1..]);
if message_type == 0x01 {
// got a ping message
let version: u8 = rlp.val_at(0)?;
let from_peer = PeerInfo::decode_rlp(&rlp.at(1)?)?;
let to_peer = PeerInfo::decode_rlp(&rlp.at(2)?)?;
println!("from_peer = {:?}, to_peer = {:?}", from_peer, to_peer);
let timestamp: u64 = rlp.val_at(3)?;
println!("version = {:?}, timestamp = {:?}", version, timestamp);
// send pong message
let from = PeerInfo::from_sock_addr(&addr);
let bytes = message::encode_pong(&from, &read_buf[0..32].to_vec(), ×tamp, &private_key);
println!("pong bytes is {:?}", bytes.len());
send_queue.push_back(bytes);
// send_queue
} else if message_type == 0x02 {
// got a pong message
let from_peer = PeerInfo::decode_rlp(&rlp.at(0)?)?;
let hash_bytes = rlp.at(1)?.data()?;
let timestamp: u64 = rlp.val_at(2)?;
println!("got a pong message {:?} {:?}", from_peer, timestamp);
// start send findneighbours packet
let bytes = message::encode_find_node(&private_key);
println!("find node bytes is {:?}", bytes.len());
send_queue.push_back(bytes);
} else if message_type == 0x03 {
println!("got a find node message");
} else if message_type == 0x04 {
println!("got a node message");
}
poll.registry().reregister(&mut udp_socket, event.token(), Interest::WRITABLE)?;
// we have read all data
if (size < buf.len()) {
println!("no more data read");
break 'read;
}
} else {
println!("no data read");
break 'read;
}
},
Err(e) => {
println!("read error {:?}", e);
break 'read;
}
}
}
}
},
CLIENT => {
if event.is_readable() {
println!("client socket is readable");
// read buf
let mut buf = [0; 1024];
match client.read(&mut buf) {
Ok(n) => {
if (n > 0) {
received_data.extend_from_slice(&buf[..n]);
println!("read data: {:?}", String::from_utf8_lossy(&received_data));
}
println!("read {:?} bytes", n);
},
Err(err) => {
println!("read data error {:?}", err);
}
}
}
if event.is_writable() {
// send auth info
let auth = "0196045fa704aa5f5a85f36c6b399b08d823083228d63c4346f382f78a18b684f3a4e64a671de498abf20cba88dd8f3f0a11443bed18248895b981e0c842e9e4fafe387cf9ad619ba89fe7dbfa6f504 | {
match message_type {
0x01 => {
println!("ping message");
},
0x02 => {
println!("pong message");
},
0x03 => {
println!("find neighbours message");
},
0x04 => {
println!("neighbours message");
},
_ => {
println!("unknow message"); | identifier_body |
|
main0.rs | () -> Result<(), Box<dyn Error>> {
let mut arr = [0u8; 32];
thread_rng().fill(&mut arr[..]);
// let data = vec![0x83, b'c', b'a', b't'];
// let aa: String = rlp::decode(&data).unwrap();
// println!("aa = {:?}", aa);
// let pk = hex::decode("ee5495585eff78f2fcf95bab21ef1a598c54d1e3c672e23b3bb97a4fc7490660").unwrap();
let private_key = SecretKey::parse_slice(&arr[0..arr.len()]).unwrap();
// let private_key = SecretKey::parse_slice(&pk).unwrap();
let pubkey = PublicKey::from_secret_key(&private_key);
let id = &pubkey.serialize().to_vec()[1..];
println!("id is {:?}", hex::encode(&id));
const CLIENT: Token = Token(0);
const SENDER: Token = Token(0);
let udp_server_ip = "35.180.217.147";
let udp_server_port = "30304";
let upd_server_addr = "35.180.217.147:30304";
let local_udp_addr = "192.168.31.125:30309";
let mut udp_socket = UdpSocket::bind(local_udp_addr.parse()?)?;
// let local_addr = udp_socket.local_addr()?;
//
// println!("local_addr = {:?}", local_addr);
println!("private_key is {:?}", private_key);
let mut poll = Poll::new()?;
let mut events = Events::with_capacity(1024);
let addr = "192.168.31.248:30303".parse()?;
let peer = PeerInfo::from_sock_addr(&upd_server_addr.parse()?);
let local_peer = PeerInfo::from_sock_addr(&local_udp_addr.parse()?);
let mut sent_ping = false;
// message::encode_ping(&peer, &peer, &private_key);
// println!("peer ip {:?}", peer.encode());
// let addr = "127.0.0.1:9000".parse()?;
let mut send_queue: VecDeque<Vec<u8>> = VecDeque::new();
let mut client = TcpStream::connect(addr)?;
let mut status_sent = false;
poll.registry().register(&mut client, CLIENT, Interest::READABLE | Interest::WRITABLE)?;
poll.registry().register(&mut udp_socket, SENDER, Interest::READABLE | Interest::WRITABLE)?;
let mut received_data = Vec::with_capacity(4096);
send_queue.push_back(message::encode_ping(&local_peer, &peer, &private_key));
loop {
poll.poll(&mut events, None)?;
for event in events.iter() {
match event.token() {
SENDER => {
println!("udp socket is active");
if event.is_writable() {
'inner: loop {
if let Some(buf) = send_queue.pop_front() {
match udp_socket.send_to(&buf, upd_server_addr.parse()?) {
Ok(size) => {
println!("sent {:?} bytes(total {:?})", size, buf.len());
// we have some buf remain for next time
if size < buf.len() {
if size == 0 {
send_queue.push_front(buf);
}
break 'inner;
}
},
Err(e) => {
println!("send error {:?}", e);
break 'inner;
}
}
} else {
println!("no data to send, reregister for next writable event");
break 'inner;
}
}
}
if event.is_readable() {
'read: loop {
let mut buf = [0; 1024];
match udp_socket.recv_from(&mut buf) {
Ok((size, addr)) => {
println!("read {:?} bytes from {:?}", size, addr);
if (size > 0) {
let read_buf = &buf[..size];
let hash_signed = keccak(&read_buf[32..]);
println!("hash_signed = {:?}", hash_signed);
println!("check_sum = {:?}", hex::encode(&read_buf[0..32]));
// if hash_signed.as_bytes() != &read_buf[0..32] {
// // return Box::new(Err("bad protocol"));
// break;
// }
let signed = &read_buf[(32 + 65)..];
let message_type = signed[0];
print_message_type(message_type);
println!("message_type is {:?}", message_type);
let recover_id = RecoveryId::parse(read_buf[32 + 64]).expect("can not get recover id");
println!("recover_id = {:?}", recover_id);
let signature = Signature::parse_slice(&read_buf[32..(32 + 64)]).expect("can not get signature");
let hash = keccak(signed);
let pubkey = recover(&Message::parse_slice(&hash).unwrap(), &signature, &recover_id).expect("can not recover pubkey");
println!("pubkey is {:?}", hex::encode(&pubkey.serialize_compressed().to_vec()));
let rlp = Rlp::new(&signed[1..]);
if message_type == 0x01 {
// got a ping message
let version: u8 = rlp.val_at(0)?;
let from_peer = PeerInfo::decode_rlp(&rlp.at(1)?)?;
let to_peer = PeerInfo::decode_rlp(&rlp.at(2)?)?;
println!("from_peer = {:?}, to_peer = {:?}", from_peer, to_peer);
let timestamp: u64 = rlp.val_at(3)?;
println!("version = {:?}, timestamp = {:?}", version, timestamp);
// send pong message
let from = PeerInfo::from_sock_addr(&addr);
let bytes = message::encode_pong(&from, &read_buf[0..32].to_vec(), ×tamp, &private_key);
println!("pong bytes is {:?}", bytes.len());
send_queue.push_back(bytes);
// send_queue
} else if message_type == 0x02 {
// got a pong message
let from_peer = PeerInfo::decode_rlp(&rlp.at(0)?)?;
let hash_bytes = rlp.at(1)?.data()?;
let timestamp: u64 = rlp.val_at(2)?;
println!("got a pong message {:?} {:?}", from_peer, timestamp);
// start send findneighbours packet
let bytes = message::encode_find_node(&private_key);
println!("find node bytes is {:?}", bytes.len());
send_queue.push_back(bytes);
} else if message_type == 0x03 {
println!("got a find node message");
} else if message_type == 0x04 {
println!("got a node message");
}
poll.registry().reregister(&mut udp_socket, event.token(), Interest::WRITABLE)?;
// we have read all data
if (size < buf.len()) {
println!("no more data read");
break 'read;
}
} else {
println!("no data read");
break 'read;
}
},
Err(e) => {
println!("read error {:?}", e);
break 'read;
}
}
}
}
},
CLIENT => {
if event.is_readable() {
println!("client socket is readable");
// read buf
let mut buf = [0; 1024];
match client.read(&mut buf) {
Ok(n) => {
if (n > 0) {
received_data.extend_from_slice(&buf[..n]);
println!("read data: {:?}", String::from_utf8_lossy(&received_data));
}
println!("read {:?} bytes", n);
},
Err(err) => {
println!("read data error {:?}", err);
}
}
}
if event.is_writable() {
// send auth info
let auth = "0196045fa704aa5f5a85f36c6b399b08d823083228d63c4346f382f78a18b684f3a4e64a671de498abf20cba88dd8f3f0a11443bed18248895b981e0c842e9e4fafe387cf9ad619ba89fe7dbfa6f504725bb673a804f3526df31c68a69caf9bc7a9eed62fe73dffdeae5e21f55e2a1ec28e17ad5f98bd0a61759fe25f8f9666527819741 | main | identifier_name |
|
csvload.go | String
}
s = strings.TrimSpace(s)
if len(s) == 0 {
return Unknown
}
var hasNonDigit bool
var dotCount int
var length int
_ = strings.Map(func(r rune) rune {
length++
if r == '.' {
dotCount++
} else if !hasNonDigit {
hasNonDigit = !('0' <= r && r <= '9')
}
return -1
},
s)
if !hasNonDigit && s[0] != '0' {
if dotCount == 1 {
return Float
}
if dotCount == 0 {
return Int
}
}
if 10 <= len(s) && len(s) <= len(dateFormat) {
if _, err := time.Parse(dateFormat[:len(s)], s); err == nil {
return Date
}
}
return String
}
func CreateTable(ctx context.Context, db *sql.DB, tbl string, rows <-chan dbcsv.Row, truncate bool, tablespace string) ([]Column, error) {
tbl = strings.ToUpper(tbl)
qry := "SELECT COUNT(0) FROM user_tables WHERE UPPER(table_name) = :1"
var n int64
var cols []Column
if err := db.QueryRowContext(ctx, qry, tbl).Scan(&n); err != nil {
return cols, errors.Wrap(err, qry)
}
if n > 0 && truncate {
qry = `TRUNCATE TABLE ` + tbl
if _, err := db.ExecContext(ctx, qry); err != nil {
return cols, errors.Wrap(err, qry)
}
}
if n == 0 {
row := <-rows
log.Printf("row: %q", row.Values)
cols = make([]Column, len(row.Values))
for i, v := range row.Values {
v = strings.Map(func(r rune) rune {
r = unicode.ToLower(r)
switch r {
case 'á':
return 'a'
case 'é':
return 'e'
case 'í':
return 'i'
case 'ö', 'ő', 'ó':
return 'o'
case 'ü', 'ű', 'ú':
return 'u'
case '_':
return '_'
default:
if 'a' <= r && r <= 'z' || '0' <= r && r <= '9' {
return r
}
return '_'
}
},
v)
if len(v) > 30 {
v = fmt.Sprintf("%s_%02d", v[:27], i)
}
cols[i].Name = v
}
if ForceString {
for i := range cols {
cols[i].Type = String
}
}
for row := range rows {
for i, v := range row.Values {
if len(v) > cols[i].Length {
cols[i].Length = len(v)
}
if cols[i].Type == String {
continue
}
typ := typeOf(v)
if cols[i].Type == Unknown {
cols[i].Type = typ
} else if typ != cols[i].Type {
cols[i].Type = String
}
}
}
var buf bytes.Buffer
buf.WriteString(`CREATE TABLE "` + tbl + `" (`)
for i, c := range cols {
if i != 0 {
buf.WriteString(",\n")
}
if c.Type == Date {
fmt.Fprintf(&buf, " %s DATE", c.Name)
continue
}
length := c.Length
if length == 0 {
length = 1
}
fmt.Fprintf(&buf, " %s %s(%d)", c.Name, c.Type.String(), length)
}
buf.WriteString("\n)")
if tablespace != "" {
buf.WriteString(" TABLESPACE ")
buf.WriteString(tablespace)
}
qry = buf.String()
log.Println(qry)
if _, err := db.Exec(qry); err != nil {
return cols, errors.Wrap(err, qry)
}
cols = cols[:0]
}
qry = `SELECT column_name, data_type, NVL(data_length, 0), NVL(data_precision, 0), NVL(data_scale, 0), nullable
FROM user_tab_cols WHERE table_name = :1
ORDER BY column_id`
tRows, err := db.QueryContext(ctx, qry, tbl)
if err != nil {
return cols, errors.Wrap(err, qry)
}
defer tRows.Close()
for tRows.Next() {
var c Column
var nullable string
if err = tRows.Scan(&c.Name, &c.DataType, &c.Length, &c.Precision, &c.Scale, &nullable); err != nil {
return cols, err
}
c.Nullable = nullable != "N"
cols = append(cols, c)
}
return cols, nil
}
type Column struct {
Length int
Name string
Type Type
DataType string
Precision, Scale int
Nullable bool
}
type Type uint8
const (
Unknown = Type(0)
String = Type(1)
Int = Type(2)
Float = Type(3)
Date = Type(4)
)
func (t Type) String() string {
switch t {
case Int, Float:
return "NUMBER"
case Date:
return "DATE"
default:
return "VARCHAR2"
}
}
func (c Column) FromString(ss []string) (interface{}, error) {
if c.DataType == "DATE" || c.Type == Date {
res := make([]time.Time, len(ss))
for i, s := range ss {
if s == "" {
continue
}
var err error
if res[i], err = time.Parse(dateFormat[:len(s)], s); err != nil {
return res, errors.Wrapf(err, "%d. %q", i, s)
}
}
return res, nil
}
if strings.HasPrefix(c.DataType, "VARCHAR2") {
for i, s := range ss {
if len(s) > c.Length {
ss[i] = s[:c.Length]
return ss, errors.Errorf("%d. %q is longer (%d) then allowed (%d) for column %v", i, s, len(s), c.Length, c)
}
}
return ss, nil
}
if c.Type == Int {
for i, s := range ss {
e := strings.Map(func(r rune) rune {
if !('0' <= r && r <= '9' || r == '-') {
return r
}
return -1
}, s)
if e != "" {
ss[i] = ""
return ss, errors.Errorf("%d. %q is not integer (%q)", i, s, e)
}
}
return ss, nil
}
if c.Type == Float {
for i, s := range ss {
e := strings.Map(func(r rune) rune {
if !('0' <= r && r <= '9' || r == '-' || r == '.') {
return r
}
return -1
}, s)
if e != "" {
ss[i] = ""
return ss, errors.Errorf("%d. %q is not float (%q)", i, s, e)
}
}
return ss, nil
}
return ss, nil
}
func getColumns(ctx context.Context, db *sql.DB, tbl string) ([]Column, error) {
// TODO(tgulacsi): this is Oracle-specific!
const qry = "SELECT column_name, data_type, data_length, data_precision, data_scale, nullable FROM user_tab_cols WHERE table_name = UPPER(:1) ORDER BY column_id"
rows, err := db.QueryContext(ctx, qry, tbl)
if err != nil {
return nil, errors.Wrap(err, qry)
}
defer rows.Close()
var cols []Column
for rows.Next() {
var c Column
var prec, scale sql.NullInt64
var nullable string
if err = rows.Scan(&c.Name, &c.DataType, &c.Length, &prec, &scale, &nullable); err != nil {
return nil, err
}
c.Nullable = nullable == "Y"
switch c.DataType {
case "DATE":
c.Type = Date
c.Length = 8
case "NUMBER":
c.Precision, c.Scale = int(prec.Int64), int(scale.Int64)
if c.Scale > 0 {
c.Type = Float
c.Length = c.Precision + 1
} else {
c.Type = Int
c.Length = c.Precision
}
default:
c.Type = String
}
cols = append(cols, c)
}
return cols, rows.Close()
}
var qRepl = strings.NewReplacer(
"'", "''",
"&", "'||CHR(38)||'",
)
func quote(w io.Writer, s string) error {
if _, err := w.Write([]byte{'\''}); err != nil {
return | err
}
if _, err | conditional_block |
|
csvload.go | %s.", n, inserted, src, tbl, dur)
return err
}
func typeOf(s string) Type {
if ForceString {
return String
}
s = strings.TrimSpace(s)
if len(s) == 0 {
return Unknown
}
var hasNonDigit bool
var dotCount int
var length int
_ = strings.Map(func(r rune) rune {
length++
if r == '.' {
dotCount++
} else if !hasNonDigit {
hasNonDigit = !('0' <= r && r <= '9')
}
return -1
},
s)
if !hasNonDigit && s[0] != '0' {
if dotCount == 1 {
return Float
}
if dotCount == 0 {
return Int
}
}
if 10 <= len(s) && len(s) <= len(dateFormat) {
if _, err := time.Parse(dateFormat[:len(s)], s); err == nil {
return Date
}
}
return String
}
func CreateTable(ctx context.Context, db *sql.DB, tbl string, rows <-chan dbcsv.Row, truncate bool, tablespace string) ([]Column, error) {
tbl = strings.ToUpper(tbl)
qry := "SELECT COUNT(0) FROM user_tables WHERE UPPER(table_name) = :1"
var n int64
var cols []Column
if err := db.QueryRowContext(ctx, qry, tbl).Scan(&n); err != nil {
return cols, errors.Wrap(err, qry)
}
if n > 0 && truncate {
qry = `TRUNCATE TABLE ` + tbl
if _, err := db.ExecContext(ctx, qry); err != nil {
return cols, errors.Wrap(err, qry)
}
}
if n == 0 {
row := <-rows
log.Printf("row: %q", row.Values)
cols = make([]Column, len(row.Values))
for i, v := range row.Values {
v = strings.Map(func(r rune) rune {
r = unicode.ToLower(r)
switch r {
case 'á':
return 'a'
case 'é':
return 'e'
case 'í':
return 'i'
case 'ö', 'ő', 'ó':
return 'o'
case 'ü', 'ű', 'ú':
return 'u'
case '_':
return '_'
default:
if 'a' <= r && r <= 'z' || '0' <= r && r <= '9' {
return r
}
return '_'
}
},
v)
if len(v) > 30 {
v = fmt.Sprintf("%s_%02d", v[:27], i)
}
cols[i].Name = v
}
if ForceString {
for i := range cols {
cols[i].Type = String
}
}
for row := range rows {
for i, v := range row.Values {
if len(v) > cols[i].Length {
cols[i].Length = len(v)
}
if cols[i].Type == String {
continue
}
typ := typeOf(v)
if cols[i].Type == Unknown {
cols[i].Type = typ
} else if typ != cols[i].Type {
cols[i].Type = String
}
}
}
var buf bytes.Buffer
buf.WriteString(`CREATE TABLE "` + tbl + `" (`)
for i, c := range cols {
if i != 0 {
buf.WriteString(",\n")
}
if c.Type == Date {
fmt.Fprintf(&buf, " %s DATE", c.Name)
continue
}
length := c.Length
if length == 0 {
length = 1
}
fmt.Fprintf(&buf, " %s %s(%d)", c.Name, c.Type.String(), length)
}
buf.WriteString("\n)")
if tablespace != "" {
buf.WriteString(" TABLESPACE ")
buf.WriteString(tablespace)
}
qry = buf.String()
log.Println(qry)
if _, err := db.Exec(qry); err != nil {
return cols, errors.Wrap(err, qry)
}
cols = cols[:0]
}
qry = `SELECT column_name, data_type, NVL(data_length, 0), NVL(data_precision, 0), NVL(data_scale, 0), nullable
FROM user_tab_cols WHERE table_name = :1
ORDER BY column_id`
tRows, err := db.QueryContext(ctx, qry, tbl)
if err != nil {
return cols, errors.Wrap(err, qry)
}
defer tRows.Close()
for tRows.Next() {
var c Column
var nullable string
if err = tRows.Scan(&c.Name, &c.DataType, &c.Length, &c.Precision, &c.Scale, &nullable); err != nil {
return cols, err
}
c.Nullable = nullable != "N"
cols = append(cols, c)
}
return cols, nil
}
type Column struct {
Length int
Name string
Type Type
DataType string
Precision, Scale int
Nullable bool
}
type Type uint8
const (
Unknown = Type(0)
String = Type(1)
Int = Type(2)
Float = Type(3)
Date = Type(4)
)
func (t Type) String() string {
switch t {
case Int, Float:
return "NUMBER"
case Date:
return "DATE"
default:
return "VARCHAR2"
}
}
func (c Column) FromString(ss []string) (interface{}, error) {
if c.DataType == "DATE" || c.Type == Date {
res := make([]time.Time, len(ss))
for i, s := range ss {
if s == "" {
continue
}
var err error
if res[i], err = time.Parse(dateFormat[:len(s)], s); err != nil {
return res, errors.Wrapf(err, "%d. %q", i, s)
}
}
return res, nil
}
if strings.HasPrefix(c.DataType, "VARCHAR2") {
for i, s := range ss {
if len(s) > c.Length {
ss[i] = s[:c.Length]
return ss, errors.Errorf("%d. %q is longer (%d) then allowed (%d) for column %v", i, s, len(s), c.Length, c)
}
}
return ss, nil
}
if c.Type == Int {
for i, s := range ss {
e := strings.Map(func(r rune) rune {
if !('0' <= r && r <= '9' || r == '-') {
return r
}
return -1
}, s)
if e != "" {
ss[i] = ""
return ss, errors.Errorf("%d. %q is not integer (%q)", i, s, e)
}
}
return ss, nil
}
if c.Type == Float {
for i, s := range ss {
e := strings.Map(func(r rune) rune {
if !('0' <= r && r <= '9' || r == '-' || r == '.') {
return r
}
return -1
}, s)
if e != "" {
ss[i] = ""
return ss, errors.Errorf("%d. %q is not float (%q)", i, s, e)
}
}
return ss, nil
}
return ss, nil
}
func getColumns(ctx context.Context, db *sql.DB, tbl string) ([]Column, error) {
// TODO(tgulacsi): this is Oracle-specific!
const qry = "SELECT column_name, data_type, data_length, data_precision, data_scale, nullable FROM user_tab_cols WHERE table_name = UPPER(:1) ORDER BY column_id"
rows, err := db.QueryContext(ctx, qry, tbl)
if err != nil {
return nil, errors.Wrap(err, qry)
}
defer rows.Close()
var cols []Column
for rows.Next() {
var c Column
var prec, scale sql.NullInt64
var nullable string
if err = rows.Scan(&c.Name, &c.DataType, &c.Length, &prec, &scale, &nullable); err != nil {
return nil, err
}
c.Nullable = nullable == "Y"
switch c.DataType {
case "DATE":
c.Type = Date
c.Length = 8
case "NUMBER":
c.Precision, c.Scale = int(prec.Int64), int(scale.Int64)
if c.Scale > 0 {
c.Type = Float
c.Length = c.Precision + 1
} else {
c.Type = Int
c.Length = c.Precision
}
default:
c.Type = String
}
cols = append(cols, c)
}
return cols, rows.Close()
}
var qRepl = strings.NewReplacer(
"'", "''",
"&", "'||CHR(38)||'",
)
func quote(w io. | Write | identifier_name |
|
csvload.go | ns(*flagConcurrency)
tbl := strings.ToUpper(flag.Arg(0))
src := flag.Arg(1)
if ForceString {
err = cfg.OpenVolatile(flag.Arg(1))
} else {
err = cfg.Open(flag.Arg(1))
}
if err != nil {
return err
}
defer cfg.Close()
rows := make(chan dbcsv.Row)
ctx, cancel := context.WithCancel(context.Background())
go func() {
defer close(rows)
cfg.ReadRows(ctx,
func(_ string, row dbcsv.Row) error {
select {
case <-ctx.Done():
return ctx.Err()
case rows <- row:
}
return nil
},
)
}()
if *flagJustPrint {
cols, err := getColumns(ctx, db, tbl)
if err != nil {
return err
}
var buf strings.Builder
for i, col := range cols {
if i != 0 {
buf.Write([]byte{',', ' '})
}
buf.WriteString(col.Name)
}
fmt.Println("INSERT ALL")
prefix := " INTO " + tbl + " (" + buf.String() + ")"
colMap := make(map[string]Column, len(cols))
for _, col := range cols {
colMap[col.Name] = col
}
cols = cols[:0]
for _, nm := range (<-rows).Values {
cols = append(cols, colMap[strings.ToUpper(nm)])
}
dRepl := strings.NewReplacer(".", "", "-", "")
for row := range rows {
buf.Reset()
for j, s := range row.Values {
if j != 0 {
buf.Write([]byte{',', ' '})
}
col := cols[j]
if col.Type != Date {
if err = quote(&buf, s); err != nil {
return err
}
} else {
buf.WriteString("TO_DATE('")
d := dRepl.Replace(s)
if len(d) == 6 {
d = "20" + d
}
buf.WriteString(d)
buf.WriteString("','YYYYMMDD')")
}
}
fmt.Printf("%s VALUES (%s)\n", prefix, buf.String())
}
fmt.Println("SELECT 1 FROM DUAL;")
return nil
}
columns, err := CreateTable(ctx, db, tbl, rows, *flagTruncate, *flagTablespace)
cancel()
if err != nil {
return err
}
var buf strings.Builder
fmt.Fprintf(&buf, `INSERT INTO "%s" (`, tbl)
for i, c := range columns {
if i != 0 {
buf.WriteString(", ")
}
buf.WriteString(c.Name)
}
buf.WriteString(") VALUES (")
for i := range columns {
if i != 0 {
buf.WriteString(", ")
}
fmt.Fprintf(&buf, ":%d", i+1)
}
buf.WriteString(")")
qry := buf.String()
log.Println(qry)
start := time.Now()
ctx, cancel = context.WithCancel(context.Background())
defer cancel()
grp, ctx := errgroup.WithContext(ctx)
type rowsType struct {
Rows [][]string
Start int64
}
rowsCh := make(chan rowsType, *flagConcurrency)
chunkPool := sync.Pool{New: func() interface{} { z := make([][]string, 0, chunkSize); return &z }}
var inserted int64
for i := 0; i < *flagConcurrency; i++ {
grp.Go(func() error {
tx, txErr := db.BeginTx(ctx, nil)
if txErr != nil {
return txErr
}
defer tx.Rollback()
stmt, prepErr := tx.Prepare(qry)
if prepErr != nil {
return errors.Wrap(prepErr, qry)
}
nCols := len(columns)
cols := make([][]string, nCols)
rowsI := make([]interface{}, nCols)
for rs := range rowsCh {
chunk := rs.Rows
if err = ctx.Err(); err != nil {
return err
}
if len(chunk) == 0 {
continue
}
nRows := len(chunk)
for j := range cols {
if cap(cols[j]) < nRows {
cols[j] = make([]string, nRows)
} else {
cols[j] = cols[j][:nRows]
for i := range cols[j] {
cols[j][i] = ""
}
}
}
for k, row := range chunk {
if len(row) > len(cols) {
log.Printf("%d. more elements in the row (%d) then columns (%d)!", rs.Start+int64(k), len(row), len(cols))
row = row[:len(cols)]
}
for j, v := range row {
cols[j][k] = v
}
}
for i, col := range cols {
if rowsI[i], err = columns[i].FromString(col); err != nil {
log.Printf("%d. col: %+v", i, err)
for k, row := range chunk {
if _, err = columns[i].FromString(col[k : k+1]); err != nil {
log.Printf("%d.%q %q: %q", rs.Start+int64(k), columns[i].Name, col[k:k+1], row)
break
}
}
return errors.Wrapf(err, columns[i].Name)
}
}
_, err = stmt.Exec(rowsI...)
{
z := chunk[:0]
chunkPool.Put(&z)
}
if err == nil {
atomic.AddInt64(&inserted, int64(len(chunk)))
continue
}
err = errors.Wrapf(err, "%s", qry)
log.Println(err)
rowsR := make([]reflect.Value, len(rowsI))
rowsI2 := make([]interface{}, len(rowsI))
for j, I := range rowsI {
rowsR[j] = reflect.ValueOf(I)
rowsI2[j] = ""
}
R2 := reflect.ValueOf(rowsI2)
for j := range cols[0] { // rows
for i, r := range rowsR { // cols
if r.Len() <= j {
log.Printf("%d[%q]=%d", j, columns[i].Name, r.Len())
rowsI2[i] = ""
continue
}
R2.Index(i).Set(r.Index(j))
}
if _, err = stmt.Exec(rowsI2...); err != nil {
err = errors.Wrapf(err, "%s, %q", qry, rowsI2)
log.Println(err)
return err | return tx.Commit()
})
}
var n int64
var headerSeen bool
chunk := (*(chunkPool.Get().(*[][]string)))[:0]
if err = cfg.ReadRows(ctx,
func(_ string, row dbcsv.Row) error {
if err = ctx.Err(); err != nil {
chunk = chunk[:0]
return err
}
if !headerSeen {
headerSeen = true
return nil
} else if n%10000 == 0 {
writeHeapProf()
}
for i, s := range row.Values {
row.Values[i] = strings.TrimSpace(s)
}
chunk = append(chunk, row.Values)
if len(chunk) < chunkSize {
return nil
}
select {
case rowsCh <- rowsType{Rows: chunk, Start: n}:
n += int64(len(chunk))
case <-ctx.Done():
return ctx.Err()
}
chunk = (*chunkPool.Get().(*[][]string))[:0]
return nil
},
); err != nil {
return err
}
if len(chunk) != 0 {
rowsCh <- rowsType{Rows: chunk, Start: n}
n += int64(len(chunk))
}
close(rowsCh)
err = grp.Wait()
dur := time.Since(start)
log.Printf("Read %d, inserted %d rows from %q to %q in %s.", n, inserted, src, tbl, dur)
return err
}
func typeOf(s string) Type {
if ForceString {
return String
}
s = strings.TrimSpace(s)
if len(s) == 0 {
return Unknown
}
var hasNonDigit bool
var dotCount int
var length int
_ = strings.Map(func(r rune) rune {
length++
if r == '.' {
dotCount++
} else if !hasNonDigit {
hasNonDigit = !('0' <= r && r <= '9')
}
return -1
},
s)
if !hasNonDigit && s[0] != '0' {
if dotCount == 1 {
return Float
}
if dotCount == 0 {
return Int
}
}
if 10 <= len(s) && len(s) <= len(dateFormat) {
if _, err := time.Parse(dateFormat[:len(s)], s); err == nil {
| }
}
return err
} | random_line_split |
csvload.go | [:0]
chunkPool.Put(&z)
}
if err == nil {
atomic.AddInt64(&inserted, int64(len(chunk)))
continue
}
err = errors.Wrapf(err, "%s", qry)
log.Println(err)
rowsR := make([]reflect.Value, len(rowsI))
rowsI2 := make([]interface{}, len(rowsI))
for j, I := range rowsI {
rowsR[j] = reflect.ValueOf(I)
rowsI2[j] = ""
}
R2 := reflect.ValueOf(rowsI2)
for j := range cols[0] { // rows
for i, r := range rowsR { // cols
if r.Len() <= j {
log.Printf("%d[%q]=%d", j, columns[i].Name, r.Len())
rowsI2[i] = ""
continue
}
R2.Index(i).Set(r.Index(j))
}
if _, err = stmt.Exec(rowsI2...); err != nil {
err = errors.Wrapf(err, "%s, %q", qry, rowsI2)
log.Println(err)
return err
}
}
return err
}
return tx.Commit()
})
}
var n int64
var headerSeen bool
chunk := (*(chunkPool.Get().(*[][]string)))[:0]
if err = cfg.ReadRows(ctx,
func(_ string, row dbcsv.Row) error {
if err = ctx.Err(); err != nil {
chunk = chunk[:0]
return err
}
if !headerSeen {
headerSeen = true
return nil
} else if n%10000 == 0 {
writeHeapProf()
}
for i, s := range row.Values {
row.Values[i] = strings.TrimSpace(s)
}
chunk = append(chunk, row.Values)
if len(chunk) < chunkSize {
return nil
}
select {
case rowsCh <- rowsType{Rows: chunk, Start: n}:
n += int64(len(chunk))
case <-ctx.Done():
return ctx.Err()
}
chunk = (*chunkPool.Get().(*[][]string))[:0]
return nil
},
); err != nil {
return err
}
if len(chunk) != 0 {
rowsCh <- rowsType{Rows: chunk, Start: n}
n += int64(len(chunk))
}
close(rowsCh)
err = grp.Wait()
dur := time.Since(start)
log.Printf("Read %d, inserted %d rows from %q to %q in %s.", n, inserted, src, tbl, dur)
return err
}
func typeOf(s string) Type {
if ForceString {
return String
}
s = strings.TrimSpace(s)
if len(s) == 0 {
return Unknown
}
var hasNonDigit bool
var dotCount int
var length int
_ = strings.Map(func(r rune) rune {
length++
if r == '.' {
dotCount++
} else if !hasNonDigit {
hasNonDigit = !('0' <= r && r <= '9')
}
return -1
},
s)
if !hasNonDigit && s[0] != '0' {
if dotCount == 1 {
return Float
}
if dotCount == 0 {
return Int
}
}
if 10 <= len(s) && len(s) <= len(dateFormat) {
if _, err := time.Parse(dateFormat[:len(s)], s); err == nil {
return Date
}
}
return String
}
func CreateTable(ctx context.Context, db *sql.DB, tbl string, rows <-chan dbcsv.Row, truncate bool, tablespace string) ([]Column, error) {
tbl = strings.ToUpper(tbl)
qry := "SELECT COUNT(0) FROM user_tables WHERE UPPER(table_name) = :1"
var n int64
var cols []Column
if err := db.QueryRowContext(ctx, qry, tbl).Scan(&n); err != nil {
return cols, errors.Wrap(err, qry)
}
if n > 0 && truncate {
qry = `TRUNCATE TABLE ` + tbl
if _, err := db.ExecContext(ctx, qry); err != nil {
return cols, errors.Wrap(err, qry)
}
}
if n == 0 {
row := <-rows
log.Printf("row: %q", row.Values)
cols = make([]Column, len(row.Values))
for i, v := range row.Values {
v = strings.Map(func(r rune) rune {
r = unicode.ToLower(r)
switch r {
case 'á':
return 'a'
case 'é':
return 'e'
case 'í':
return 'i'
case 'ö', 'ő', 'ó':
return 'o'
case 'ü', 'ű', 'ú':
return 'u'
case '_':
return '_'
default:
if 'a' <= r && r <= 'z' || '0' <= r && r <= '9' {
return r
}
return '_'
}
},
v)
if len(v) > 30 {
v = fmt.Sprintf("%s_%02d", v[:27], i)
}
cols[i].Name = v
}
if ForceString {
for i := range cols {
cols[i].Type = String
}
}
for row := range rows {
for i, v := range row.Values {
if len(v) > cols[i].Length {
cols[i].Length = len(v)
}
if cols[i].Type == String {
continue
}
typ := typeOf(v)
if cols[i].Type == Unknown {
cols[i].Type = typ
} else if typ != cols[i].Type {
cols[i].Type = String
}
}
}
var buf bytes.Buffer
buf.WriteString(`CREATE TABLE "` + tbl + `" (`)
for i, c := range cols {
if i != 0 {
buf.WriteString(",\n")
}
if c.Type == Date {
fmt.Fprintf(&buf, " %s DATE", c.Name)
continue
}
length := c.Length
if length == 0 {
length = 1
}
fmt.Fprintf(&buf, " %s %s(%d)", c.Name, c.Type.String(), length)
}
buf.WriteString("\n)")
if tablespace != "" {
buf.WriteString(" TABLESPACE ")
buf.WriteString(tablespace)
}
qry = buf.String()
log.Println(qry)
if _, err := db.Exec(qry); err != nil {
return cols, errors.Wrap(err, qry)
}
cols = cols[:0]
}
qry = `SELECT column_name, data_type, NVL(data_length, 0), NVL(data_precision, 0), NVL(data_scale, 0), nullable
FROM user_tab_cols WHERE table_name = :1
ORDER BY column_id`
tRows, err := db.QueryContext(ctx, qry, tbl)
if err != nil {
return cols, errors.Wrap(err, qry)
}
defer tRows.Close()
for tRows.Next() {
var c Column
var nullable string
if err = tRows.Scan(&c.Name, &c.DataType, &c.Length, &c.Precision, &c.Scale, &nullable); err != nil {
return cols, err
}
c.Nullable = nullable != "N"
cols = append(cols, c)
}
return cols, nil
}
type Column struct {
Length int
Name string
Type Type
DataType string
Precision, Scale int
Nullable bool
}
type Type uint8
const (
Unknown = Type(0)
String = Type(1)
Int = Type(2)
Float = Type(3)
Date = Type(4)
)
func (t Type) String() string {
switch t {
case Int, Float:
return "NUMBER"
case Date:
return "DATE"
default:
return "VARCHAR2"
}
}
func (c Column) FromString(ss []string) (interface{}, error) {
if c.Dat | aType == "DATE" || c.Type == Date {
res := make([]time.Time, len(ss))
for i, s := range ss {
if s == "" {
continue
}
var err error
if res[i], err = time.Parse(dateFormat[:len(s)], s); err != nil {
return res, errors.Wrapf(err, "%d. %q", i, s)
}
}
return res, nil
}
if strings.HasPrefix(c.DataType, "VARCHAR2") {
for i, s := range ss {
if len(s) > c.Length {
ss[i] = s[:c.Length]
return ss, errors.Errorf("%d. %q is longer (%d) then allowed (%d) for column %v", i, s, len(s), c.Length, c)
} | identifier_body |
|
prediction.py | dir_4[(ww_dir > 45) & (ww_dir < 135)] = 1
dir_4[(ww_dir > 225) & (ww_dir < 315)] = 3
if ship_dir in ("W"):
dir_4 = np.full((len(ww_dir), 1), 2)
dir_4[(ww_dir > 45) & (ww_dir < 135)] = 3
dir_4[(ww_dir > 225) & (ww_dir < 315)] = 1
if ship_dir in ("S"):
dir_4 = np.full((len(ww_dir), 1), 2)
dir_4[(ww_dir < 45) | (ww_dir > 315)] = 3
dir_4[(ww_dir > 135) & (ww_dir < 225)] = 1
return dir_4
def concatenate_cmems(cm_wave, cm_phy, ship_param, ship_dir):
"""
concatenate the variables from cmems wave and physics datasets
Parameters
----------
cm_wave : net4CDF dataset
netcdf file cmems wave
cm_phy : net4CDF dataset
netdcf file cmems physics
ship_param : int
ship variable that is used in model later (e.g. draft or length)
ship_dir str, in ("N", "E", "S", "W")
direction the ship is going
"""
array = (np.flipud(cm_wave["VHM0"][0, :, :]).data) # extract data from CMEMS
dim = array.shape
l = np.prod(dim) # get number of "pixel"
# extract parameters from cmems dataset and reshape to array with dimension of 1 x number of pixel
vhm = (np.flipud(cm_wave["VHM0"][0, :, :])).reshape(l, 1)
vtm = (np.flipud(cm_wave["VTPK"][0, :, :])).reshape(l, 1)
temp = (np.flipud(cm_phy["thetao"][0, 1, :, :])).reshape(l, 1)
sal = (np.flipud(cm_phy["so"][0, 1, :, :])).reshape(l, 1)
# create column for ship parameter
ship = np.full((l, 1), ship_param)
# calculate relative direction of wind depending on ship direction
dir = calc_relative_direction(ship_dir, (np.flipud(cm_wave["VMDR_WW"][0, :, :])).reshape(l, 1))
# concatenate parameters
a = np.concatenate((ship, vhm, vtm, temp, sal, dir), axis=1)
# create pd df from array
X_pred = pd.DataFrame(data=a, # values
index=list(range(0, l)), # 1st column as index
columns=["Draft", "VHM0", "VTPK", "thetao", "so", "dir_4"]) # 1st row as the column names
return X_pred
def prepare_grid(cm_wave, cm_phy, ship_param, ship_dir, model):
"""
prepare grid of SOGs
Parameters
----------
cm_wave : net4CDF dataset
netcdf file cmems wave
cm_phy : net4CDF dataset
netdcf file cmems physics
ship_param : int
ship variable that is used in model later (e.g. draft or length)
ship_dir str, in ("N", "E", "S", "W")
direction the ship is going
"""
X_pred = concatenate_cmems(cm_wave, cm_phy, ship_param, ship_dir)
# extract shape from cmems data
input = (np.flipud(cm_wave["VHM0"][0, :, :]))
dim = input.shape
# predict SOG
# model = load('cms_routing/models/DTR_model.joblib') # import model
SOG_pred = model.predict(X_pred)
SOG_pred = SOG_pred.reshape(dim) # reshape to 'coordinates'
SOG_pred[input < -30000] = -5 # -32767.0 # mask data with negative value
return SOG_pred
def calculateTimeGrid(SOG_E, SOG_N, SOG_S, SOG_W, AOI):
kmGridEW = np.load("lengthGridEW.npy")
kmGridEW = kmGridEW[AOI[2]:AOI[3], AOI[0]:AOI[1]]
kmGridNS = np.load("lengthGridNS.npy")
kmGridNS = kmGridNS[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridE = SOG_E
constE = 70 / np.power(timeGridE, 3)
timeGridE80 = np.cbrt(80 / constE)
timeGridE60 = np.cbrt(60 / constE)
timeGridE = timeGridE[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridE80 = timeGridE80[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridE60 = timeGridE60[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridE = np.where(timeGridE < 0, 10000, (kmGridEW * 1000) / (timeGridE * 30.87))
timeGridE80 = np.where(timeGridE80 < 0, 10000, (kmGridEW * 1000) / (timeGridE80 * 30.87))
timeGridE60 = np.where(timeGridE60 < 0, 10000, (kmGridEW * 1000) / (timeGridE60 * 30.87))
timeGridN = SOG_N
constN = 70 / np.power(timeGridN, 3)
timeGridN80 = np.cbrt(80 / constN)
timeGridN60 = np.cbrt(60 / constN)
timeGridN = timeGridN[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridN80 = timeGridN80[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridN60 = timeGridN60[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridN = np.where(timeGridN < 0, 10000, (kmGridNS * 1000) / (timeGridN * 30.87))
timeGridN80 = np.where(timeGridN80 < 0, 10000, (kmGridNS * 1000) / (timeGridN80 * 30.87))
timeGridN60 = np.where(timeGridN60 < 0, 10000, (kmGridNS * 1000) / (timeGridN60 * 30.87))
timeGridS = SOG_S
constS = 70 / np.power(timeGridS, 3)
timeGridS80 = np.cbrt(80 / constS)
timeGridS60 = np.cbrt(60 / constS)
timeGridS = timeGridS[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridS80 = timeGridS80[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridS60 = timeGridS | """
determine relative wind direction for ships going north, east, south or west
Parameters
----------
ship_dir : str, in ("N", "E", "S", "W")
direction the ship is going
ww_dir : array, float
array of relative wind directions [0 - 360]
"""
if ship_dir not in ("N", "E", "S", "W"):
raise Exception("Direction not accepted.")
ww_360 = ww_dir
ww_360[ww_360 < 0] = 360 + ww_dir[0]
if ship_dir in ("N"):
dir_4 = np.full((len(ww_dir), 1), 2)
dir_4[(ww_dir < 45) | (ww_dir > 315)] = 1
dir_4[(ww_dir > 135) & (ww_dir < 225)] = 3
if ship_dir in ("E"):
dir_4 = np.full((len(ww_dir), 1), 2) | identifier_body |
|
prediction.py | (url, user, passwd, ftp_path, filename):
with ftplib.FTP(url) as ftp:
try:
ftp.login(user, passwd)
# Change directory
ftp.cwd(ftp_path)
# Download file (if there is not yet a local copy)
if os.path.isfile(filename):
print("There is already a local copy for this date ({})".format(filename))
else:
with open(filename, 'wb') as fp:
print("Downloading ... ({})".format(filename))
ftp.retrbinary('RETR {}'.format(filename), fp.write)
except ftplib.all_errors as e:
print('FTP error:', e)
# Check contents
"""
with ftplib.FTP('nrt.cmems-du.eu') as ftp:
try:
ftp.login(UN_CMEMS, PW_CMEMS)
# Change directory
ftp.cwd('Core/GLOBAL_ANALYSIS_FORECAST_PHY_001_024/global-analysis-forecast-phy-001-024/2021/07')
# List directory contents with additional information
ftp.retrlines('LIST')
# Get list of directory contents without additional information
files = []
ftp.retrlines('NLST', files.append)
print(files)
# Check file size
print("{} MB".format(ftp.size('mfwamglocep_2020120100_R20201202.nc')/1000000))
except ftplib.all_errors as e:
print('FTP error:', e)
"""
def calc_relative_direction(ship_dir, ww_dir):
"""
determine relative wind direction for ships going north, east, south or west
Parameters
----------
ship_dir : str, in ("N", "E", "S", "W")
direction the ship is going
ww_dir : array, float
array of relative wind directions [0 - 360]
"""
if ship_dir not in ("N", "E", "S", "W"):
raise Exception("Direction not accepted.")
ww_360 = ww_dir
ww_360[ww_360 < 0] = 360 + ww_dir[0]
if ship_dir in ("N"):
dir_4 = np.full((len(ww_dir), 1), 2)
dir_4[(ww_dir < 45) | (ww_dir > 315)] = 1
dir_4[(ww_dir > 135) & (ww_dir < 225)] = 3
if ship_dir in ("E"):
dir_4 = np.full((len(ww_dir), 1), 2)
dir_4[(ww_dir > 45) & (ww_dir < 135)] = 1
dir_4[(ww_dir > 225) & (ww_dir < 315)] = 3
if ship_dir in ("W"):
dir_4 = np.full((len(ww_dir), 1), 2)
dir_4[(ww_dir > 45) & (ww_dir < 135)] = 3
dir_4[(ww_dir > 225) & (ww_dir < 315)] = 1
if ship_dir in ("S"):
dir_4 = np.full((len(ww_dir), 1), 2)
dir_4[(ww_dir < 45) | (ww_dir > 315)] = 3
dir_4[(ww_dir > 135) & (ww_dir < 225)] = 1
return dir_4
def concatenate_cmems(cm_wave, cm_phy, ship_param, ship_dir):
"""
concatenate the variables from cmems wave and physics datasets
Parameters
----------
cm_wave : net4CDF dataset
netcdf file cmems wave
cm_phy : net4CDF dataset
netdcf file cmems physics
ship_param : int
ship variable that is used in model later (e.g. draft or length)
ship_dir str, in ("N", "E", "S", "W")
direction the ship is going
"""
array = (np.flipud(cm_wave["VHM0"][0, :, :]).data) # extract data from CMEMS
dim = array.shape
l = np.prod(dim) # get number of "pixel"
# extract parameters from cmems dataset and reshape to array with dimension of 1 x number of pixel
vhm = (np.flipud(cm_wave["VHM0"][0, :, :])).reshape(l, 1)
vtm = (np.flipud(cm_wave["VTPK"][0, :, :])).reshape(l, 1)
temp = (np.flipud(cm_phy["thetao"][0, 1, :, :])).reshape(l, 1)
sal = (np.flipud(cm_phy["so"][0, 1, :, :])).reshape(l, 1)
# create column for ship parameter
ship = np.full((l, 1), ship_param)
# calculate relative direction of wind depending on ship direction
dir = calc_relative_direction(ship_dir, (np.flipud(cm_wave["VMDR_WW"][0, :, :])).reshape(l, 1))
# concatenate parameters
a = np.concatenate((ship, vhm, vtm, temp, sal, dir), axis=1)
# create pd df from array
X_pred = pd.DataFrame(data=a, # values
index=list(range(0, l)), # 1st column as index
columns=["Draft", "VHM0", "VTPK", "thetao", "so", "dir_4"]) # 1st row as the column names
return X_pred
def prepare_grid(cm_wave, cm_phy, ship_param, ship_dir, model):
"""
prepare grid of SOGs
Parameters
----------
cm_wave : net4CDF dataset
netcdf file cmems wave
cm_phy : net4CDF dataset
netdcf file cmems physics
ship_param : int
ship variable that is used in model later (e.g. draft or length)
ship_dir str, in ("N", "E", "S", "W")
direction the ship is going
"""
X_pred = concatenate_cmems(cm_wave, cm_phy, ship_param, ship_dir)
# extract shape from cmems data
input = (np.flipud(cm_wave["VHM0"][0, :, :]))
dim = input.shape
# predict SOG
# model = load('cms_routing/models/DTR_model.joblib') # import model
SOG_pred = model.predict(X_pred)
SOG_pred = SOG_pred.reshape(dim) # reshape to 'coordinates'
SOG_pred[input < -30000] = -5 # -32767.0 # mask data with negative value
return SOG_pred
def calculateTimeGrid(SOG_E, SOG_N, SOG_S, SOG_W, AOI):
kmGridEW = np.load("lengthGridEW.npy")
kmGridEW = kmGridEW[AOI[2]:AOI[3], AOI[0]:AOI[1]]
kmGridNS = np.load("lengthGridNS.npy")
kmGridNS = kmGridNS[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridE = SOG_E
constE = 70 / np.power(timeGridE, 3)
timeGridE80 = np.cbrt(80 / constE)
timeGridE60 = np.cbrt(60 / constE)
timeGridE = timeGridE[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridE80 = timeGridE80[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridE60 = timeGridE60[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridE = np.where(timeGridE < 0, 10000, (kmGridEW * 1000) / (timeGridE * 30.87))
timeGridE80 = np.where(timeGridE80 < 0, 10000, (kmGridEW * 1000) / (timeGridE80 * 30.87))
timeGridE60 = np.where(timeGridE60 < 0, 10000, (kmGridEW * 1000) / (timeGridE60 * 30.87))
timeGridN = SOG_N
constN = 70 / np.power(timeGridN, 3)
timeGridN80 = np.cbrt(80 / constN)
timeGridN60 = np.cbrt(60 / constN)
timeGridN = timeGridN[AOI[2]:AO | download | identifier_name |
|
prediction.py |
try:
ftp.login(user, passwd)
# Change directory
ftp.cwd(ftp_path)
# Download file (if there is not yet a local copy)
if os.path.isfile(filename):
print("There is already a local copy for this date ({})".format(filename))
else:
with open(filename, 'wb') as fp:
print("Downloading ... ({})".format(filename))
ftp.retrbinary('RETR {}'.format(filename), fp.write)
except ftplib.all_errors as e:
print('FTP error:', e)
# Check contents
"""
with ftplib.FTP('nrt.cmems-du.eu') as ftp:
try:
ftp.login(UN_CMEMS, PW_CMEMS)
# Change directory
ftp.cwd('Core/GLOBAL_ANALYSIS_FORECAST_PHY_001_024/global-analysis-forecast-phy-001-024/2021/07')
# List directory contents with additional information
ftp.retrlines('LIST')
# Get list of directory contents without additional information
files = []
ftp.retrlines('NLST', files.append)
print(files)
# Check file size
print("{} MB".format(ftp.size('mfwamglocep_2020120100_R20201202.nc')/1000000))
except ftplib.all_errors as e:
print('FTP error:', e)
"""
def calc_relative_direction(ship_dir, ww_dir):
"""
determine relative wind direction for ships going north, east, south or west
Parameters
----------
ship_dir : str, in ("N", "E", "S", "W")
direction the ship is going
ww_dir : array, float
array of relative wind directions [0 - 360]
"""
if ship_dir not in ("N", "E", "S", "W"):
raise Exception("Direction not accepted.")
ww_360 = ww_dir
ww_360[ww_360 < 0] = 360 + ww_dir[0]
if ship_dir in ("N"):
dir_4 = np.full((len(ww_dir), 1), 2)
dir_4[(ww_dir < 45) | (ww_dir > 315)] = 1
dir_4[(ww_dir > 135) & (ww_dir < 225)] = 3
if ship_dir in ("E"):
dir_4 = np.full((len(ww_dir), 1), 2)
dir_4[(ww_dir > 45) & (ww_dir < 135)] = 1
dir_4[(ww_dir > 225) & (ww_dir < 315)] = 3
if ship_dir in ("W"):
dir_4 = np.full((len(ww_dir), 1), 2)
dir_4[(ww_dir > 45) & (ww_dir < 135)] = 3
dir_4[(ww_dir > 225) & (ww_dir < 315)] = 1
if ship_dir in ("S"):
dir_4 = np.full((len(ww_dir), 1), 2)
dir_4[(ww_dir < 45) | (ww_dir > 315)] = 3
dir_4[(ww_dir > 135) & (ww_dir < 225)] = 1
return dir_4
def concatenate_cmems(cm_wave, cm_phy, ship_param, ship_dir):
"""
concatenate the variables from cmems wave and physics datasets
Parameters
----------
cm_wave : net4CDF dataset
netcdf file cmems wave
cm_phy : net4CDF dataset
netdcf file cmems physics
ship_param : int
ship variable that is used in model later (e.g. draft or length)
ship_dir str, in ("N", "E", "S", "W")
direction the ship is going
"""
array = (np.flipud(cm_wave["VHM0"][0, :, :]).data) # extract data from CMEMS
dim = array.shape
l = np.prod(dim) # get number of "pixel"
# extract parameters from cmems dataset and reshape to array with dimension of 1 x number of pixel
vhm = (np.flipud(cm_wave["VHM0"][0, :, :])).reshape(l, 1)
vtm = (np.flipud(cm_wave["VTPK"][0, :, :])).reshape(l, 1)
temp = (np.flipud(cm_phy["thetao"][0, 1, :, :])).reshape(l, 1)
sal = (np.flipud(cm_phy["so"][0, 1, :, :])).reshape(l, 1)
# create column for ship parameter
ship = np.full((l, 1), ship_param)
# calculate relative direction of wind depending on ship direction
dir = calc_relative_direction(ship_dir, (np.flipud(cm_wave["VMDR_WW"][0, :, :])).reshape(l, 1))
# concatenate parameters
a = np.concatenate((ship, vhm, vtm, temp, sal, dir), axis=1)
# create pd df from array
X_pred = pd.DataFrame(data=a, # values
index=list(range(0, l)), # 1st column as index
columns=["Draft", "VHM0", "VTPK", "thetao", "so", "dir_4"]) # 1st row as the column names
return X_pred
def prepare_grid(cm_wave, cm_phy, ship_param, ship_dir, model):
"""
prepare grid of SOGs
Parameters
----------
cm_wave : net4CDF dataset
netcdf file cmems wave
cm_phy : net4CDF dataset
netdcf file cmems physics
ship_param : int
ship variable that is used in model later (e.g. draft or length)
ship_dir str, in ("N", "E", "S", "W")
direction the ship is going
"""
X_pred = concatenate_cmems(cm_wave, cm_phy, ship_param, ship_dir)
# extract shape from cmems data
input = (np.flipud(cm_wave["VHM0"][0, :, :]))
dim = input.shape
# predict SOG
# model = load('cms_routing/models/DTR_model.joblib') # import model
SOG_pred = model.predict(X_pred)
SOG_pred = SOG_pred.reshape(dim) # reshape to 'coordinates'
SOG_pred[input < -30000] = -5 # -32767.0 # mask data with negative value
return SOG_pred
def calculateTimeGrid(SOG_E, SOG_N, SOG_S, SOG_W, AOI):
kmGridEW = np.load("lengthGridEW.npy")
kmGridEW = kmGridEW[AOI[2]:AOI[3], AOI[0]:AOI[1]]
kmGridNS = np.load("lengthGridNS.npy")
kmGridNS = kmGridNS[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridE = SOG_E
constE = 70 / np.power(timeGridE, 3)
timeGridE80 = np.cbrt(80 / constE)
timeGridE60 = np.cbrt(60 / constE)
timeGridE = timeGridE[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridE80 = timeGridE80[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridE60 = timeGridE60[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridE = np.where(timeGridE < 0, 10000, (kmGridEW * 1000) / (timeGridE * 30.87))
timeGridE80 = np.where(timeGridE80 < 0, 10000, (kmGridEW * 1000) / (timeGridE80 * 30.87))
timeGridE60 = np.where(timeGridE60 < 0, 10000, (kmGridEW * 1000) / (timeGridE60 * 30.87))
timeGridN = SOG_N
constN = 70 / np.power(timeGridN, 3)
timeGridN80 = np.cbrt(80 / constN)
timeGridN60 = np.cbrt(60 / constN)
timeGridN = timeGridN[AOI[ |
def download(url, user, passwd, ftp_path, filename):
with ftplib.FTP(url) as ftp: | random_line_split |
|
prediction.py | ww_dir : array, float
array of relative wind directions [0 - 360]
"""
if ship_dir not in ("N", "E", "S", "W"):
raise Exception("Direction not accepted.")
ww_360 = ww_dir
ww_360[ww_360 < 0] = 360 + ww_dir[0]
if ship_dir in ("N"):
dir_4 = np.full((len(ww_dir), 1), 2)
dir_4[(ww_dir < 45) | (ww_dir > 315)] = 1
dir_4[(ww_dir > 135) & (ww_dir < 225)] = 3
if ship_dir in ("E"):
dir_4 = np.full((len(ww_dir), 1), 2)
dir_4[(ww_dir > 45) & (ww_dir < 135)] = 1
dir_4[(ww_dir > 225) & (ww_dir < 315)] = 3
if ship_dir in ("W"):
|
if ship_dir in ("S"):
dir_4 = np.full((len(ww_dir), 1), 2)
dir_4[(ww_dir < 45) | (ww_dir > 315)] = 3
dir_4[(ww_dir > 135) & (ww_dir < 225)] = 1
return dir_4
def concatenate_cmems(cm_wave, cm_phy, ship_param, ship_dir):
"""
concatenate the variables from cmems wave and physics datasets
Parameters
----------
cm_wave : net4CDF dataset
netcdf file cmems wave
cm_phy : net4CDF dataset
netdcf file cmems physics
ship_param : int
ship variable that is used in model later (e.g. draft or length)
ship_dir str, in ("N", "E", "S", "W")
direction the ship is going
"""
array = (np.flipud(cm_wave["VHM0"][0, :, :]).data) # extract data from CMEMS
dim = array.shape
l = np.prod(dim) # get number of "pixel"
# extract parameters from cmems dataset and reshape to array with dimension of 1 x number of pixel
vhm = (np.flipud(cm_wave["VHM0"][0, :, :])).reshape(l, 1)
vtm = (np.flipud(cm_wave["VTPK"][0, :, :])).reshape(l, 1)
temp = (np.flipud(cm_phy["thetao"][0, 1, :, :])).reshape(l, 1)
sal = (np.flipud(cm_phy["so"][0, 1, :, :])).reshape(l, 1)
# create column for ship parameter
ship = np.full((l, 1), ship_param)
# calculate relative direction of wind depending on ship direction
dir = calc_relative_direction(ship_dir, (np.flipud(cm_wave["VMDR_WW"][0, :, :])).reshape(l, 1))
# concatenate parameters
a = np.concatenate((ship, vhm, vtm, temp, sal, dir), axis=1)
# create pd df from array
X_pred = pd.DataFrame(data=a, # values
index=list(range(0, l)), # 1st column as index
columns=["Draft", "VHM0", "VTPK", "thetao", "so", "dir_4"]) # 1st row as the column names
return X_pred
def prepare_grid(cm_wave, cm_phy, ship_param, ship_dir, model):
"""
prepare grid of SOGs
Parameters
----------
cm_wave : net4CDF dataset
netcdf file cmems wave
cm_phy : net4CDF dataset
netdcf file cmems physics
ship_param : int
ship variable that is used in model later (e.g. draft or length)
ship_dir str, in ("N", "E", "S", "W")
direction the ship is going
"""
X_pred = concatenate_cmems(cm_wave, cm_phy, ship_param, ship_dir)
# extract shape from cmems data
input = (np.flipud(cm_wave["VHM0"][0, :, :]))
dim = input.shape
# predict SOG
# model = load('cms_routing/models/DTR_model.joblib') # import model
SOG_pred = model.predict(X_pred)
SOG_pred = SOG_pred.reshape(dim) # reshape to 'coordinates'
SOG_pred[input < -30000] = -5 # -32767.0 # mask data with negative value
return SOG_pred
def calculateTimeGrid(SOG_E, SOG_N, SOG_S, SOG_W, AOI):
kmGridEW = np.load("lengthGridEW.npy")
kmGridEW = kmGridEW[AOI[2]:AOI[3], AOI[0]:AOI[1]]
kmGridNS = np.load("lengthGridNS.npy")
kmGridNS = kmGridNS[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridE = SOG_E
constE = 70 / np.power(timeGridE, 3)
timeGridE80 = np.cbrt(80 / constE)
timeGridE60 = np.cbrt(60 / constE)
timeGridE = timeGridE[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridE80 = timeGridE80[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridE60 = timeGridE60[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridE = np.where(timeGridE < 0, 10000, (kmGridEW * 1000) / (timeGridE * 30.87))
timeGridE80 = np.where(timeGridE80 < 0, 10000, (kmGridEW * 1000) / (timeGridE80 * 30.87))
timeGridE60 = np.where(timeGridE60 < 0, 10000, (kmGridEW * 1000) / (timeGridE60 * 30.87))
timeGridN = SOG_N
constN = 70 / np.power(timeGridN, 3)
timeGridN80 = np.cbrt(80 / constN)
timeGridN60 = np.cbrt(60 / constN)
timeGridN = timeGridN[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridN80 = timeGridN80[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridN60 = timeGridN60[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridN = np.where(timeGridN < 0, 10000, (kmGridNS * 1000) / (timeGridN * 30.87))
timeGridN80 = np.where(timeGridN80 < 0, 10000, (kmGridNS * 1000) / (timeGridN80 * 30.87))
timeGridN60 = np.where(timeGridN60 < 0, 10000, (kmGridNS * 1000) / (timeGridN60 * 30.87))
timeGridS = SOG_S
constS = 70 / np.power(timeGridS, 3)
timeGridS80 = np.cbrt(80 / constS)
timeGridS60 = np.cbrt(60 / constS)
timeGridS = timeGridS[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridS80 = timeGridS80[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridS60 = timeGridS60[AOI[2]:AOI[3], AOI[0]:AOI[1]]
timeGridS = np.where(timeGridS < 0, 10000, (kmGridNS | dir_4 = np.full((len(ww_dir), 1), 2)
dir_4[(ww_dir > 45) & (ww_dir < 135)] = 3
dir_4[(ww_dir > 225) & (ww_dir < 315)] = 1 | conditional_block |
iconnect.js | (obj) {
var _key = typeof obj == "number" ? obj : obj[custEventAttribute];
if (_key && custEventCache[_key]) {
if (type) {
type = [].concat(type);
for (var i = 0; i < type.length; i++) {
if (type[i] in custEventCache[_key]) delete custEventCache[_key][type[i]];
}
} else {
delete custEventCache[_key];
}
}
}
},
/**
* 事件添加或绑定
* @method add
* @static
* @param {Object|number} obj 对象引用或获取的(key); 必选
* @param {String} type 自定义事件名称; 必选
* @param {Function} fn 事件处理方法; 必选
* @param {Any} data 扩展数据任意类型; 可选
* @return {number} key 下标
*/
add: function add(obj, type, fn, data) {
return _add(obj, type, fn, data, false);
},
/**
* 单次事件绑定
* @method once
* @static
* @param {Object|number} obj 对象引用或获取的(key); 必选
* @param {String} type 自定义事件名称; 必选
* @param {Function} fn 事件处理方法; 必选
* @param {Any} data 扩展数据任意类型; 可选
* @return {number} key 下标
*/
once: function once(obj, type, fn, data) {
return _add(obj, type, fn, data, true);
},
/**
* 事件删除或解绑
* @method remove
* @static
* @param {Object|number} obj 对象引用或获取的(key); 必选
* @param {String} type 自定义事件名称; 可选; 为空时删除对象下的所有事件绑定
* @param {Function} fn 事件处理方法; 可选; 为空且type不为空时 删除对象下type事件相关的所有处理方法
* @return {number} key 下标
*/
remove: function remove(obj, type, fn) {
if (obj) {
var _cache = findCache(obj, type),
_obj,
index;
if (_cache && (_obj = _cache.obj)) {
if (isArray(_obj)) {
if (fn) {
//for (var i = 0; i < _obj.length && _obj[i].fn !== fn; i++);
var i = 0;
while (_obj[i]) {
if (_obj[i].fn === fn) {
break;
}
i++;
}
_obj.splice(i, 1);
} else {
_obj.splice(0, _obj.length);
}
} else {
for (var i in _obj) {
_obj[i] = [];
}
}
return _cache.key;
}
}
},
/**
* 事件触发
* @method fire
* @static
* @param {Object|number} obj 对象引用或获取的(key); 必选
* @param {String} type 自定义事件名称; 必选
* @param {Any|Array} args 参数数组或单个的其他数据; 可选
* @param {Function} defaultAction 触发事件列表结束后的默认Function; 可选 注:当args不需要时请用undefined/null填充,以保证该参数为第四个参数
* @return {number} key 下标
*/
fire: function fire(obj, type, args, defaultAction) {
return _fire(obj, type, args, defaultAction);
},
/**
* 事件由源对象迁移到目标对象
* @method hook
* @static
* @param {Object} orig 源对象
* @param {Object} dest 目标对象
* @param {Object} typeMap 事件名称对照表
* {
* 源事件名->目标事件名
* }
*/
hook: function hook(orig, dest, typeMap) {
if (!orig || !dest || !typeMap) {
return;
}
var destTypes = [],
origKey = orig[custEventAttribute],
origKeyCache = origKey && custEventCache[origKey],
origTypeCache,
destKey = dest[custEventAttribute] || (dest[custEventAttribute] = custEventKey++),
keyHookCache;
if (origKeyCache) {
keyHookCache = hookCache[origKey + '_' + destKey] || (hookCache[origKey + '_' + destKey] = {});
var fn = function fn(event) {
var preventDefaultFlag = true;
_fire(dest, keyHookCache[event.type].type, Array.prototype.slice.apply(arguments, [1, arguments.length]), function () {
preventDefaultFlag = false;
});
preventDefaultFlag && event.preventDefault();
};
for (var origType in typeMap) {
var destType = typeMap[origType];
if (!keyHookCache[origType]) {
if (origTypeCache = origKeyCache[origType]) {
origTypeCache.push({ fn: fn, data: undefined });
keyHookCache[origType] = {
fn: fn,
type: destType
};
destTypes.push(destType);
}
}
}
that.define(dest, destTypes);
}
},
/**
* 取消事件迁移
* @method unhook
* @static
* @param {Object} orig 源对象
* @param {Object} dest 目标对象
* @param {Object} typeMap 事件名称对照表
* {
* 源事件名->目标事件名
* }
*/
unhook: function unhook(orig, dest, typeMap) {
if (!orig || !dest || !typeMap) {
return;
}
var origKey = orig[custEventAttribute],
destKey = dest[custEventAttribute],
keyHookCache = hookCache[origKey + '_' + destKey];
if (keyHookCache) {
for (var origType in typeMap) {
var destType = typeMap[origType];
if (keyHookCache[origType]) {
that.remove(orig, origType, keyHookCache[origType].fn);
}
}
}
},
/**
* 销毁
* @method destroy
* @static
*/
destroy: function destroy() {
custEventCache = {};
custEventKey = 1;
hookCache = {};
}
};
return that;
}();
var utils = {
count: 0,
getUniqueKey: function getUniqueKey() {
return +new Date() + (Math.random() + '').replace('.', '') + utils.count++;
},
json2str: function () {
function f(n) {
// Format integers to have at least two digits.
return n < 10 ? '0' + n : n;
}
if (typeof Date.prototype.toJSON !== 'function') {
Date.prototype.toJSON = function (key) {
return isFinite(this.valueOf()) ? this.getUTCFullYear() + '-' + f(this.getUTCMonth() + 1) + '-' + f(this.getUTCDate()) + 'T' + f(this.getUTCHours()) + ':' + f(this.getUTCMinutes()) + ':' + f(this.getUTCSeconds()) + 'Z' : null;
};
String.prototype.toJSON = Number.prototype.toJSON = Boolean.prototype.toJSON = function (key) {
return this.valueOf();
};
}
var cx = /[\u0000\u00ad\u0600-\u0604\u070f\u17b4\u17b5\u200c-\u200f\u2028-\u202f\u2060-\u206f\ufeff\ufff0-\uffff]/g,
escapable = /[\\\"\x00-\x1f\x7f-\x9f\u00ad\u0600-\u0604\u070f\u17b4\u17b5\u200c-\u200f\u2028-\u202f\u2060-\u206f\ufeff\ufff0-\uffff]/g,
gap,
indent,
meta = { // table of character substitutions
'\b': '\\b',
'\t': '\\t',
'\n': '\\n',
'\f': '\\f',
'\r': '\\r',
'"': '\\"',
| identifier_name |
||
iconnect.js | }
};
var that = {
/**
* 对象自定义事件的定义 未定义的事件不得绑定
* @method define
* @static
* @param {Object|number} obj 对象引用或获取的下标(key); 必选
* @param {String|Array} type 自定义事件名称; 必选
* @return {number} key 下标
*/
define: function define(obj, type) {
if (obj && type) {
var _key = typeof obj == "number" ? obj : obj[custEventAttribute] || (obj[custEventAttribute] = custEventKey++),
_cache = custEventCache[_key] || (custEventCache[_key] = {});
type = [].concat(type);
for (var i = 0; i < type.length; i++) {
_cache[type[i]] || (_cache[type[i]] = []);
}
return _key;
}
},
/**
* 对象自定义事件的取消定义
* 当对象的所有事件定义都被取消时 删除对对象的引用
* @method define
* @static
* @param {Object|number} obj 对象引用或获取的(key); 必选
* @param {String} type 自定义事件名称; 可选 不填可取消所有事件的定义
*/
undefine: function undefine(obj, type) {
if (obj) {
var _key = typeof obj == "number" ? obj : obj[custEventAttribute];
if (_key && custEventCache[_key]) {
if (type) {
type = [].concat(type);
for (var i = 0; i < type.length; i++) {
if (type[i] in custEventCache[_key]) delete custEventCache[_key][type[i]];
}
} else {
delete custEventCache[_key];
}
}
}
},
/**
* 事件添加或绑定
* @method add
* @static
* @param {Object|number} obj 对象引用或获取的(key); 必选
* @param {String} type 自定义事件名称; 必选
* @param {Function} fn 事件处理方法; 必选
* @param {Any} data 扩展数据任意类型; 可选
* @return {number} key 下标
*/
add: function add(obj, type, fn, data) {
return _add(obj, type, fn, data, false);
},
/**
* 单次事件绑定
* @method once
* @static
* @param {Object|number} obj 对象引用或获取的(key); 必选
* @param {String} type 自定义事件名称; 必选
* @param {Function} fn 事件处理方法; 必选
* @param {Any} data 扩展数据任意类型; 可选
* @return {number} key 下标
*/
once: function once(obj, type, fn, data) {
return _add(obj, type, fn, data, true);
},
/**
* 事件删除或解绑
* @method remove
* @static
* @param {Object|number} obj 对象引用或获取的(key); 必选
* @param {String} type 自定义事件名称; 可选; 为空时删除对象下的所有事件绑定
* @param {Function} fn 事件处理方法; 可选; 为空且type不为空时 删除对象下type事件相关的所有处理方法
* @return {number} key 下标
*/
remove: function remove(obj, type, fn) {
if (obj) {
var _cache = findCache(obj, type),
_obj,
index;
if (_cache && (_obj = _cache.obj)) {
if (isArray(_obj)) {
if (fn) {
//for (var i = 0; i < _obj.length && _obj[i].fn !== fn; i++);
var i = 0;
while (_obj[i]) {
if (_obj[i].fn === fn) {
break;
}
i++;
}
_obj.splice(i, 1);
} else {
_obj.splice(0, _obj.length);
}
} else {
for (var i in _obj) {
_obj[i] = [];
}
}
return _cache.key;
}
}
},
/**
* 事件触发
* @method fire
* @static
* @param {Object|number} obj 对象引用或获取的(key); 必选
* @param {String} type 自定义事件名称; 必选
* @param {Any|Array} args 参数数组或单个的其他数据; 可选
* @param {Function} defaultAction 触发事件列表结束后的默认Function; 可选 注:当args不需要时请用undefined/null填充,以保证该参数为第四个参数
* @return {number} key 下标
*/
fire: function fire(obj, type, args, defaultAction) {
return _fire(obj, type, args, defaultAction);
},
/**
* 事件由源对象迁移到目标对象
* @method hook
* @static
* @param {Object} orig 源对象
* @param {Object} dest 目标对象
* @param {Object} typeMap 事件名称对照表
* {
* 源事件名->目标事件名
* }
*/
hook: function hook(orig, dest, typeMap) {
if (!orig || !dest || !typeMap) {
return;
}
var destTypes = [],
origKey = orig[custEventAttribute],
origKeyCache = origKey && custEventCache[origKey],
origTypeCache,
destKey = dest[custEventAttribute] || (dest[custEventAttribute] = custEventKey++),
keyHookCache;
if (origKeyCache) {
keyHookCache = hookCache[origKey + '_' + destKey] || (hookCache[origKey + '_' + destKey] = {});
var fn = function fn(event) {
var preventDefaultFlag = true;
_fire(dest, keyHookCache[event.type].type, Array.prototype.slice.apply(arguments, [1, arguments.length]), function () {
preventDefaultFlag = false;
});
preventDefaultFlag && event.preventDefault();
};
for (var origType in typeMap) {
var destType = typeMap[origType];
if (!keyHookCache[origType]) {
if (origTypeCache = origKeyCache[origType]) {
origTypeCache.push({ fn: fn, data: undefined });
keyHookCache[origType] = {
fn: fn,
type: destType
};
destTypes.push(destType);
}
}
}
that.define(dest, destTypes);
}
},
/**
* 取消事件迁移
* @method unhook
* @static
* @param {Object} orig 源对象
* @param {Object} dest 目标对象
* @param {Object} typeMap 事件名称对照表
* {
* 源事件名->目标事件名
* }
*/
unhook: function unhook(orig, dest, typeMap) {
if (!orig || !dest || !typeMap) {
return;
}
var origKey = orig[custEventAttribute],
destKey = dest[custEventAttribute],
keyHookCache = hookCache[origKey + '_' + destKey];
if (keyHookCache) {
for (var origType in typeMap) {
var destType = typeMap[origType];
if (keyHookCache[origType]) {
that.remove(orig, origType, keyHookCache[origType].fn);
}
}
}
},
/**
* 销毁
* @method destroy
* @static
*/
destroy: function destroy() {
custEventCache = {};
custEventKey = 1;
hookCache = {};
}
};
return that;
}();
var utils = {
count: 0,
getUniqueKey: function getUniqueKey() {
return +new Date() + (Math.random() + '').replace('.', '') + utils.count++;
},
json2str: function () {
function f(n) {
// Format integers to have at least two digits.
return n < 10 ? '0' + n : n;
}
if (typeof Date.prototype.toJSON !== 'function') {
Date.prototype.toJSON = function (key) {
return isFinite(this.valueOf()) ? this.getUTCFullYear() | }
return _cache.key;
} | random_line_split |
|
iconnect.js | .getUTCHours()) + ':' + f(this.getUTCMinutes()) + ':' + f(this.getUTCSeconds()) + 'Z' : null;
};
String.prototype.toJSON = Number.prototype.toJSON = Boolean.prototype.toJSON = function (key) {
return this.valueOf();
};
}
var cx = /[\u0000\u00ad\u0600-\u0604\u070f\u17b4\u17b5\u200c-\u200f\u2028-\u202f\u2060-\u206f\ufeff\ufff0-\uffff]/g,
escapable = /[\\\"\x00-\x1f\x7f-\x9f\u00ad\u0600-\u0604\u070f\u17b4\u17b5\u200c-\u200f\u2028-\u202f\u2060-\u206f\ufeff\ufff0-\uffff]/g,
gap,
indent,
meta = { // table of character substitutions
'\b': '\\b',
'\t': '\\t',
'\n': '\\n',
'\f': '\\f',
'\r': '\\r',
'"': '\\"',
'\\': '\\\\'
},
rep;
function quote(string) {
// If the string contains no control characters, no quote characters, and no
// backslash characters, then we can safely slap some quotes around it.
// Otherwise we must also replace the offending characters with safe escape
// sequences.
escapable.lastIndex = 0;
return escapable.test(string) ? '"' + string.replace(escapable, function (a) {
var c = meta[a];
return typeof c === 'string' ? c : "\\u" + ('0000' + a.charCodeAt(0).toString(16)).slice(-4);
}) + '"' : '"' + string + '"';
}
function str(key, holder) {
// Produce a string from holder[key].
var i,
// The loop counter.
k,
// The member key.
v,
// The member value.
length,
mind = gap,
partial,
value = holder[key];
// If the value has a toJSON method, call it to obtain a replacement value.
if (value && (typeof value === "undefined" ? "undefined" : _typeof(value)) === 'object' && typeof value.toJSON === 'function') {
value = value.toJSON(key);
}
// If we were called with a replacer function, then call the replacer to
// obtain a replacement value.
if (typeof rep === 'function') {
value = rep.call(holder, key, value);
}
// What happens next depends on the value's type.
switch (typeof value === "undefined" ? "undefined" : _typeof(value)) {
case 'string':
return quote(value);
case 'number':
// JSON numbers must be finite. Encode non-finite numbers as null.
return isFinite(value) ? String(value) : 'null';
case 'boolean':
case 'null':
// If the value is a boolean or null, convert it to a string. Note:
// typeof null does not produce 'null'. The case is included here in
// the remote chance that this gets fixed someday.
return String(value);
// If the type is 'object', we might be dealing with an object or an array or
// null.
case 'object':
// Due to a specification blunder in ECMAScript, typeof null is 'object',
// so watch out for that case.
if (!value) {
return 'null';
}
// Make an array to hold the partial results of stringifying this object value.
gap += indent;
partial = [];
// Is the value an array?
if (Object.prototype.toString.apply(value) === '[object Array]') {
// The value is an array. Stringify every element. Use null as a placeholder
// for non-JSON values.
length = value.length;
for (i = 0; i < length; i += 1) {
partial[i] = str(i, value) || 'null';
}
// Join all of the elements together, separated with commas, and wrap them in
// brackets.
v = partial.length === 0 ? '[]' : gap ? '[\n' + gap + partial.join(',\n' + gap) + '\n' + mind + ']' : '[' + partial.join(',') + ']';
gap = mind;
return v;
}
// If the replacer is an array, use it to select the members to be stringified.
if (rep && (typeof rep === "undefined" ? "undefined" : _typeof(rep)) === 'object') {
length = rep.length;
for (i = 0; i < length; i += 1) {
k = rep[i];
if (typeof k === 'string') {
v = str(k, value);
if (v) {
partial.push(quote(k) + (gap ? ': ' : ':') + v);
}
}
}
} else {
// Otherwise, iterate through all of the keys in the object.
for (k in value) {
if (Object.hasOwnProperty.call(value, k)) {
v = str(k, value);
if (v) {
partial.push(quote(k) + (gap ? ': ' : ':') + v);
}
}
}
}
// Join all of the member texts together, separated with commas,
// and wrap them in braces.
v = partial.length === 0 ? '{}' : gap ? '{\n' + gap + partial.join(',\n' + gap) + '\n' + mind + '}' : '{' + partial.join(',') + '}';
gap = mind;
return v;
}
}
return function (value, replacer, space) {
if (window.JSON && window.JSON.stringify) {
return window.JSON.stringify(value, replacer, space);
}
// The stringify method takes a value and an optional replacer, and an optional
// space parameter, and returns a JSON text. The replacer can be a function
// that can replace values, or an array of strings that will select the keys.
// A default replacer method can be provided. Use of the space parameter can
// produce text that is more easily readable.
var i;
gap = '';
indent = '';
// If the space parameter is a number, make an indent string containing that
// many spaces.
if (typeof space === 'number') {
for (i = 0; i < space; i += 1) {
indent += ' ';
}
// If the space parameter is a string, it will be used as the indent string.
} else if (typeof space === 'string') {
indent = space;
}
// If there is a replacer, it must be a function or an array.
// Otherwise, throw an error.
rep = replacer;
if (replacer && typeof replacer !== 'function' && ((typeof replacer === "undefined" ? "undefined" : _typeof(replacer)) !== 'object' || typeof replacer.length !== 'number')) {
throw new Error('JSON.stringify');
}
// Make a fake root object containing our value under the key of ''.
// Return the result of stringifying the value.
return str('', {
'': value
});
};
}(),
str2json: function str2json(str) {
try {
return eval('(' + str + ')');
} catch (e) {
return null;
}
},
getUrlParam: function getUrlParam(name) {
var reg = new RegExp("(^|&)" + name + "=([^&]*)(&|$)"); //构造一个含有目标参数的正则表达式对象.
var r = window.location.search.substr(1).match(reg); //匹配目标参数
if (r != null) {
return unescape(r[2]);
}
return null;
}
};
var iframeConnect = function iframeConnect() {
var cidList = {};
var event = {};
var iid = window.name;
var post = function post(cid, cmd, param) {
//iid iframe的id
//cid 这个任务的id
//cmd clinet段调用回调函数的方法名
//param 这个任务的参数 数组
var msg = utils.json2str(param === undefined ? { iid: iid, cid: cid, cmd: cmd } : { iid: iid, cid: cid, cmd: cmd, param: param });
if (wind | ow.parent.postMessage) {
window.parent.post | conditional_block |
|
iconnect.js | for (var i = 0; i < _obj.length && _obj[i].fn !== fn; i++);
var i = 0;
while (_obj[i]) {
if (_obj[i].fn === fn) {
break;
}
i++;
}
_obj.splice(i, 1);
} else {
_obj.splice(0, _obj.length);
}
} else {
for (var i in _obj) {
_obj[i] = [];
}
}
return _cache.key;
}
}
},
/**
* 事件触发
* @method fire
* @static
* @param {Object|number} obj 对象引用或获取的(key); 必选
* @param {String} type 自定义事件名称; 必选
* @param {Any|Array} args 参数数组或单个的其他数据; 可选
* @param {Function} defaultAction 触发事件列表结束后的默认Function; 可选 注:当args不需要时请用undefined/null填充,以保证该参数为第四个参数
* @return {number} key 下标
*/
fire: function fire(obj, type, args, defaultAction) {
return _fire(obj, type, args, defaultAction);
},
/**
* 事件由源对象迁移到目标对象
* @method hook
* @static
* @param {Object} orig 源对象
* @param {Object} dest 目标对象
* @param {Object} typeMap 事件名称对照表
* {
* 源事件名->目标事件名
* }
*/
hook: function hook(orig, dest, typeMap) {
if (!orig || !dest || !typeMap) {
return;
}
var destTypes = [],
origKey = orig[custEventAttribute],
origKeyCache = origKey && custEventCache[origKey],
origTypeCache,
destKey = dest[custEventAttribute] || (dest[custEventAttribute] = custEventKey++),
keyHookCache;
if (origKeyCache) {
keyHookCache = hookCache[origKey + '_' + destKey] || (hookCache[origKey + '_' + destKey] = {});
var fn = function fn(event) {
var preventDefaultFlag = true;
_fire(dest, keyHookCache[event.type].type, Array.prototype.slice.apply(arguments, [1, arguments.length]), function () {
preventDefaultFlag = false;
});
preventDefaultFlag && event.preventDefault();
};
for (var origType in typeMap) {
var destType = typeMap[origType];
if (!keyHookCache[origType]) {
if (origTypeCache = origKeyCache[origType]) {
origTypeCache.push({ fn: fn, data: undefined });
keyHookCache[origType] = {
fn: fn,
type: destType
};
destTypes.push(destType);
}
}
}
that.define(dest, destTypes);
}
},
/**
* 取消事件迁移
* @method unhook
* @static
* @param {Object} orig 源对象
* @param {Object} dest 目标对象
* @param {Object} typeMap 事件名称对照表
* {
* 源事件名->目标事件名
* }
*/
unhook: function unhook(orig, dest, typeMap) {
if (!orig || !dest || !typeMap) {
return;
}
var origKey = orig[custEventAttribute],
destKey = dest[custEventAttribute],
keyHookCache = hookCache[origKey + '_' + destKey];
if (keyHookCache) {
for (var origType in typeMap) {
var destType = typeMap[origType];
if (keyHookCache[origType]) {
that.remove(orig, origType, keyHookCache[origType].fn);
}
}
}
},
/**
* 销毁
* @method destroy
* @static
*/
destroy: function destroy() {
custEventCache = {};
custEventKey = 1;
hookCache = {};
}
};
return that;
}();
var utils = {
count: 0,
getUniqueKey: function getUniqueKey() {
return +new Date() + (Math.random() + '').replace('.', '') + utils.count++;
},
json2str: function () {
function f(n) {
// Format integers to have at least two digits.
return n < 10 ? '0' + n : n;
}
if (typeof Date.prototype.toJSON !== 'function') {
Date.prototype.toJSON = function (key) {
return isFinite(this.valueOf()) ? this.getUTCFullYear() + '-' + f(this.getUTCMonth() + 1) + '-' + f(this.getUTCDate()) + 'T' + f(this.getUTCHours()) + ':' + f(this.getUTCMinutes()) + ':' + f(this.getUTCSeconds()) + 'Z' : null;
};
String.prototype.toJSON = Number.prototype.toJSON = Boolean.prototype.toJSON = function (key) {
return this.valueOf();
};
}
var cx = /[\u0000\u00ad\u0600-\u0604\u070f\u17b4\u17b5\u200c-\u200f\u2028-\u202f\u2060-\u206f\ufeff\ufff0-\uffff]/g,
escapable = /[\\\"\x00-\x1f\x7f-\x9f\u00ad\u0600-\u0604\u070f\u17b4\u17b5\u200c-\u200f\u2028-\u202f\u2060-\u206f\ufeff\ufff0-\uffff]/g,
gap,
indent,
meta = { // table of character substitutions
'\b': '\\b',
'\t': '\\t',
'\n': '\\n',
'\f': '\\f',
'\r': '\\r',
'"': '\\"',
'\\': '\\\\'
},
rep;
function quote(string) {
// If the string contains no control characters, no quote characters, and no
// backslash characters, then we can safely slap some quotes around it.
// Otherwise we must also replace the offending characters with safe escape
// sequences.
escapable.lastIndex = 0;
return escapable.test(string) ? '"' + string.replace(escapable, function (a) {
var c = meta[a];
return typeof c === 'string' ? c : "\\u" + ('0000' + a.charCodeAt(0).toString(16)).slice(-4);
}) + '"' : '"' + string + '"';
}
function str(key, holder) {
// Produce a string from holder[key].
var i,
// The loop counter.
k,
// The member key.
v,
// The member value.
length,
mind = gap,
partial,
value = holder[key];
// If the value has a toJSON method, call it to obtain a replacement value.
if (value && (typeof value === "undefined" ? "undefined" : _typeof(value)) === 'object' && typeof value.toJSON === 'function') {
value = value.toJSON(key);
}
// If we were called with a replacer function, then call the replacer to
// obtain a replacement value.
if (typeof rep === 'function') {
value = rep.call(holder, key, value);
}
// What happens next depends on the value's type.
switch (typeof value === "undefined" ? "undefined" : _typeof(value)) {
case 'string':
return quote(value);
case 'number':
// JSON numbers must be finite. Encode non-finite numbers as null.
return isFinite(value) ? String(value) : 'null';
case 'b | oolean':
case 'null':
// If the value is a boolean or null, convert it to a string. Note:
// typeof null does not produce 'null'. The case is included here in
// the remote chance that this gets fixed someday.
return String(value);
// If the type is 'object', we might be dealing with an object or an array or
// null.
case 'object':
// Due to a specification blunder in ECMAScript, typeof null is 'object',
// so watch out for that case.
if (!value) {
return 'null';
}
// Make an array to hold the partial results of stringifying this object value.
gap += indent; | identifier_body |
|
binary.go | returns -1.
func Size(v interface{}) int {
return dataSize(reflect.Indirect(reflect.ValueOf(v)))
}
// dataSize returns the number of bytes the actual data represented by v occupies in memory.
// For compound structures, it sums the sizes of the elements. Thus, for instance, for a slice
// it returns the length of the slice times the element size and does not count the memory
// occupied by the header. If the type of v is not acceptable, dataSize returns -1.
func dataSize(v reflect.Value) int {
if v.Kind() == reflect.Slice {
if s := sizeof(v.Type().Elem()); s >= 0 {
return s * v.Len()
}
return -1
}
return sizeof(v.Type())
}
// sizeof returns the size >= 0 of variables for the given type or -1 if the type is not acceptable.
func sizeof(t reflect.Type) int {
switch t.Kind() {
case reflect.Array:
if s := sizeof(t.Elem()); s >= 0 {
return s * t.Len()
}
case reflect.Struct:
sum := 0
for i, n := 0, t.NumField(); i < n; i++ {
s := sizeof(t.Field(i).Type)
if s < 0 {
return -1
}
sum += s
}
return sum
case reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64,
reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64,
reflect.Float32, reflect.Float64, reflect.Complex64, reflect.Complex128, reflect.Ptr:
return int(t.Size())
}
return -1
}
type coder struct {
order ByteOrder
buf []byte
}
type (
decoder coder
encoder coder
)
func (d *decoder) uint8() uint8 {
x := d.buf[0]
d.buf = d.buf[1:]
return x
}
func (e *encoder) uint8(x uint8) {
e.buf[0] = x
e.buf = e.buf[1:]
}
func (d *decoder) uint16() uint16 {
x := d.order.Uint16(d.buf[0:2])
d.buf = d.buf[2:]
return x
}
func (e *encoder) uint16(x uint16) {
e.order.PutUint16(e.buf[0:2], x)
e.buf = e.buf[2:]
}
func (d *decoder) uint32() uint32 {
x := d.order.Uint32(d.buf[0:4])
d.buf = d.buf[4:]
return x
}
func (e *encoder) uint32(x uint32) {
e.order.PutUint32(e.buf[0:4], x)
e.buf = e.buf[4:]
}
func (d *decoder) uint64() uint64 {
x := d.order.Uint64(d.buf[0:8])
d.buf = d.buf[8:]
return x
}
func (e *encoder) uint64(x uint64) {
e.order.PutUint64(e.buf[0:8], x)
e.buf = e.buf[8:]
}
func (d *decoder) int8() int8 { return int8(d.uint8()) }
func (e *encoder) int8(x int8) { e.uint8(uint8(x)) }
func (d *decoder) int16() int16 { return int16(d.uint16()) }
func (e *encoder) int16(x int16) { e.uint16(uint16(x)) }
func (d *decoder) int32() int32 { return int32(d.uint32()) }
func (e *encoder) int32(x int32) { e.uint32(uint32(x)) }
func (d *decoder) int64() int64 { return int64(d.uint64()) }
func (e *encoder) int64(x int64) { e.uint64(uint64(x)) }
func (d *decoder) value(v reflect.Value) {
switch v.Kind() {
case reflect.Array:
l := v.Len()
for i := 0; i < l; i++ {
d.value(v.Index(i))
}
case reflect.Struct:
t := v.Type()
l := v.NumField()
for i := 0; i < l; i++ {
// Note: Calling v.CanSet() below is an optimization.
// It would be sufficient to check the field name,
// but creating the StructField info for each field is
// costly (run "go test -bench=ReadStruct" and compare
// results when making changes to this code).
if v := v.Field(i); v.CanSet() || t.Field(i).Name != "_" {
d.value(v)
} else {
d.skip(v)
}
}
case reflect.Slice:
l := v.Len()
for i := 0; i < l; i++ {
d.value(v.Index(i))
}
case reflect.Int8:
v.SetInt(int64(d.int8()))
case reflect.Int16:
v.SetInt(int64(d.int16()))
case reflect.Int32:
v.SetInt(int64(d.int32()))
case reflect.Int64:
v.SetInt(d.int64())
case reflect.Uint8:
v.SetUint(uint64(d.uint8()))
case reflect.Uint16:
v.SetUint(uint64(d.uint16()))
case reflect.Uint32:
v.SetUint(uint64(d.uint32()))
case reflect.Uint64:
v.SetUint(d.uint64())
case reflect.Float32:
v.SetFloat(float64(math.Float32frombits(d.uint32())))
case reflect.Float64:
v.SetFloat(math.Float64frombits(d.uint64()))
case reflect.Complex64:
v.SetComplex(complex(
float64(math.Float32frombits(d.uint32())),
float64(math.Float32frombits(d.uint32())),
))
case reflect.Complex128:
v.SetComplex(complex(
math.Float64frombits(d.uint64()),
math.Float64frombits(d.uint64()),
))
}
}
func (e *encoder) value(v reflect.Value) {
switch v.Kind() {
case reflect.Array:
l := v.Len()
for i := 0; i < l; i++ {
e.value(v.Index(i))
}
case reflect.Struct:
t := v.Type()
l := v.NumField()
for i := 0; i < l; i++ {
// see comment for corresponding code in decoder.value()
if v := v.Field(i); v.CanSet() || t.Field(i).Name != "_" {
e.value(v)
} else {
e.skip(v)
}
}
case reflect.Slice:
l := v.Len()
for i := 0; i < l; i++ {
e.value(v.Index(i))
}
case reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64:
switch v.Type().Kind() {
case reflect.Int8:
e.int8(int8(v.Int()))
case reflect.Int16:
e.int16(int16(v.Int()))
case reflect.Int32:
e.int32(int32(v.Int()))
case reflect.Int64:
e.int64(v.Int())
}
case reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr:
switch v.Type().Kind() {
case reflect.Uint8:
e.uint8(uint8(v.Uint()))
case reflect.Uint16:
e.uint16(uint16(v.Uint()))
case reflect.Uint32:
e.uint32(uint32(v.Uint()))
case reflect.Uint64:
e.uint64(v.Uint())
}
case reflect.Float32, reflect.Float64:
switch v.Type().Kind() {
case reflect.Float32:
e.uint32(math.Float32bits(float32(v.Float())))
case reflect.Float64:
e.uint64(math.Float64bits(v.Float()))
}
case reflect.Complex64, reflect.Complex128:
switch v.Type().Kind() {
case reflect.Complex64:
x := v.Complex()
e.uint32(math.Float32bits(float32(real(x))))
e.uint32(math.Float32bits(float32(imag(x))))
case reflect.Complex128:
x := v.Complex()
e.uint64(math.Float64bits(real(x)))
e.uint64(math.Float64bits(imag(x)))
}
}
}
func (d *decoder) skip(v reflect.Value) {
d.buf = d.buf[dataSize(v):]
}
func (e *encoder) skip(v reflect.Value) {
n := dataSize(v)
for i := range e.buf[0:n] {
e.buf[i] = 0
}
e.buf = e.buf[n:]
}
// intDataSize returns the size of the data required to represent the data when encoded.
// It returns zero if the type cannot be implemented by the fast path in Read or Write.
func | intDataSize | identifier_name |
|
binary.go | Endian" }
// Read reads structured binary data from r into data.
// Data must be a pointer to a fixed-size value or a slice
// of fixed-size values.
// Bytes read from r are decoded using the specified byte order
// and written to successive fields of the data.
// When reading into structs, the field data for fields with
// blank (_) field names is skipped; i.e., blank field names
// may be used for padding.
// When reading into a struct, all non-blank fields must be exported.
func Read(r io.Reader, order ByteOrder, data interface{}) error {
// Fast path for basic types and slices.
if n := intDataSize(data); n != 0 {
var b [8]byte
var bs []byte
if n > len(b) {
bs = make([]byte, n)
} else {
bs = b[:n]
}
if _, err := io.ReadFull(r, bs); err != nil {
return err
}
switch data := data.(type) {
case *int8:
*data = int8(b[0])
case *uint8:
*data = b[0]
case *int16:
*data = int16(order.Uint16(bs))
case *uint16:
*data = order.Uint16(bs)
case *int32:
*data = int32(order.Uint32(bs))
case *uint32:
*data = order.Uint32(bs)
case *int64:
*data = int64(order.Uint64(bs))
case *uint64:
*data = order.Uint64(bs)
case []int8:
for i, x := range bs { // Easier to loop over the input for 8-bit values.
data[i] = int8(x)
}
case []uint8:
copy(data, bs)
case []int16:
for i := range data {
data[i] = int16(order.Uint16(bs[2*i:]))
}
case []uint16:
for i := range data {
data[i] = order.Uint16(bs[2*i:])
}
case []int32:
for i := range data {
data[i] = int32(order.Uint32(bs[4*i:]))
}
case []uint32:
for i := range data {
data[i] = order.Uint32(bs[4*i:])
}
case []int64:
for i := range data {
data[i] = int64(order.Uint64(bs[8*i:]))
}
case []uint64:
for i := range data {
data[i] = order.Uint64(bs[8*i:])
}
}
return nil
}
// Fallback to reflect-based decoding.
v := reflect.ValueOf(data)
size := -1
switch v.Kind() {
case reflect.Ptr:
v = v.Elem()
size = dataSize(v)
case reflect.Slice:
size = dataSize(v)
}
if size < 0 {
return errors.New("binary.Read: invalid type " + reflect.TypeOf(data).String())
}
d := &decoder{order: order, buf: make([]byte, size)}
if _, err := io.ReadFull(r, d.buf); err != nil {
return err
}
d.value(v)
return nil
}
// Write writes the binary representation of data into w.
// Data must be a fixed-size value or a slice of fixed-size
// values, or a pointer to such data.
// Bytes written to w are encoded using the specified byte order
// and read from successive fields of the data.
// When writing structs, zero values are written for fields
// with blank (_) field names.
func Write(w io.Writer, order ByteOrder, data interface{}) error {
// Fast path for basic types and slices.
if n := intDataSize(data); n != 0 {
var b [8]byte
var bs []byte
if n > len(b) {
bs = make([]byte, n)
} else {
bs = b[:n]
}
switch v := data.(type) {
case *int8:
bs = b[:1]
b[0] = byte(*v)
case int8:
bs = b[:1]
b[0] = byte(v)
case []int8:
for i, x := range v {
bs[i] = byte(x)
}
case *uint8:
bs = b[:1]
b[0] = *v
case uint8:
bs = b[:1]
b[0] = byte(v)
case []uint8:
bs = v
case *int16:
bs = b[:2]
order.PutUint16(bs, uint16(*v))
case int16:
bs = b[:2]
order.PutUint16(bs, uint16(v))
case []int16:
for i, x := range v {
order.PutUint16(bs[2*i:], uint16(x))
}
case *uint16:
bs = b[:2]
order.PutUint16(bs, *v)
case uint16:
bs = b[:2]
order.PutUint16(bs, v)
case []uint16:
for i, x := range v {
order.PutUint16(bs[2*i:], x)
}
case *int32:
bs = b[:4]
order.PutUint32(bs, uint32(*v))
case int32:
bs = b[:4]
order.PutUint32(bs, uint32(v))
case []int32:
for i, x := range v {
order.PutUint32(bs[4*i:], uint32(x))
}
case *uint32:
bs = b[:4]
order.PutUint32(bs, *v)
case uint32:
bs = b[:4]
order.PutUint32(bs, v)
case []uint32:
for i, x := range v {
order.PutUint32(bs[4*i:], x)
}
case *int64:
bs = b[:8]
order.PutUint64(bs, uint64(*v))
case int64:
bs = b[:8]
order.PutUint64(bs, uint64(v))
case []int64:
for i, x := range v {
order.PutUint64(bs[8*i:], uint64(x))
}
case *uint64:
bs = b[:8]
order.PutUint64(bs, *v)
case uint64:
bs = b[:8]
order.PutUint64(bs, v)
case []uint64:
for i, x := range v {
order.PutUint64(bs[8*i:], x)
}
}
_, err := w.Write(bs)
return err
}
// Fallback to reflect-based encoding.
v := reflect.Indirect(reflect.ValueOf(data))
size := dataSize(v)
if size < 0 {
return errors.New("binary.Write: invalid type " + reflect.TypeOf(data).String())
}
buf := make([]byte, size)
e := &encoder{order: order, buf: buf}
e.value(v)
_, err := w.Write(buf)
return err
}
// Size returns how many bytes Write would generate to encode the value v, which
// must be a fixed-size value or a slice of fixed-size values, or a pointer to such data.
// If v is neither of these, Size returns -1.
func Size(v interface{}) int {
return dataSize(reflect.Indirect(reflect.ValueOf(v)))
}
// dataSize returns the number of bytes the actual data represented by v occupies in memory.
// For compound structures, it sums the sizes of the elements. Thus, for instance, for a slice
// it returns the length of the slice times the element size and does not count the memory
// occupied by the header. If the type of v is not acceptable, dataSize returns -1.
func dataSize(v reflect.Value) int {
if v.Kind() == reflect.Slice {
if s := sizeof(v.Type().Elem()); s >= 0 {
return s * v.Len()
}
return -1
}
return sizeof(v.Type())
}
// sizeof returns the size >= 0 of variables for the given type or -1 if the type is not acceptable.
func sizeof(t reflect.Type) int | {
switch t.Kind() {
case reflect.Array:
if s := sizeof(t.Elem()); s >= 0 {
return s * t.Len()
}
case reflect.Struct:
sum := 0
for i, n := 0, t.NumField(); i < n; i++ {
s := sizeof(t.Field(i).Type)
if s < 0 {
return -1
}
sum += s
}
return sum
case reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64,
reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64, | identifier_body |
|
binary.go | := io.ReadFull(r, d.buf); err != nil {
return err
}
d.value(v)
return nil
}
// Write writes the binary representation of data into w.
// Data must be a fixed-size value or a slice of fixed-size
// values, or a pointer to such data.
// Bytes written to w are encoded using the specified byte order
// and read from successive fields of the data.
// When writing structs, zero values are written for fields
// with blank (_) field names.
func Write(w io.Writer, order ByteOrder, data interface{}) error {
// Fast path for basic types and slices.
if n := intDataSize(data); n != 0 {
var b [8]byte
var bs []byte
if n > len(b) {
bs = make([]byte, n)
} else {
bs = b[:n]
}
switch v := data.(type) {
case *int8:
bs = b[:1]
b[0] = byte(*v)
case int8:
bs = b[:1]
b[0] = byte(v)
case []int8:
for i, x := range v {
bs[i] = byte(x)
}
case *uint8:
bs = b[:1]
b[0] = *v
case uint8:
bs = b[:1]
b[0] = byte(v)
case []uint8:
bs = v
case *int16:
bs = b[:2]
order.PutUint16(bs, uint16(*v))
case int16:
bs = b[:2]
order.PutUint16(bs, uint16(v))
case []int16:
for i, x := range v {
order.PutUint16(bs[2*i:], uint16(x))
}
case *uint16:
bs = b[:2]
order.PutUint16(bs, *v)
case uint16:
bs = b[:2]
order.PutUint16(bs, v)
case []uint16:
for i, x := range v {
order.PutUint16(bs[2*i:], x)
}
case *int32:
bs = b[:4]
order.PutUint32(bs, uint32(*v))
case int32:
bs = b[:4]
order.PutUint32(bs, uint32(v))
case []int32:
for i, x := range v {
order.PutUint32(bs[4*i:], uint32(x))
}
case *uint32:
bs = b[:4]
order.PutUint32(bs, *v)
case uint32:
bs = b[:4]
order.PutUint32(bs, v)
case []uint32:
for i, x := range v {
order.PutUint32(bs[4*i:], x)
}
case *int64:
bs = b[:8]
order.PutUint64(bs, uint64(*v))
case int64:
bs = b[:8]
order.PutUint64(bs, uint64(v))
case []int64:
for i, x := range v {
order.PutUint64(bs[8*i:], uint64(x))
}
case *uint64:
bs = b[:8]
order.PutUint64(bs, *v)
case uint64:
bs = b[:8]
order.PutUint64(bs, v)
case []uint64:
for i, x := range v {
order.PutUint64(bs[8*i:], x)
}
}
_, err := w.Write(bs)
return err
}
// Fallback to reflect-based encoding.
v := reflect.Indirect(reflect.ValueOf(data))
size := dataSize(v)
if size < 0 {
return errors.New("binary.Write: invalid type " + reflect.TypeOf(data).String())
}
buf := make([]byte, size)
e := &encoder{order: order, buf: buf}
e.value(v)
_, err := w.Write(buf)
return err
}
// Size returns how many bytes Write would generate to encode the value v, which
// must be a fixed-size value or a slice of fixed-size values, or a pointer to such data.
// If v is neither of these, Size returns -1.
func Size(v interface{}) int {
return dataSize(reflect.Indirect(reflect.ValueOf(v)))
}
// dataSize returns the number of bytes the actual data represented by v occupies in memory.
// For compound structures, it sums the sizes of the elements. Thus, for instance, for a slice
// it returns the length of the slice times the element size and does not count the memory
// occupied by the header. If the type of v is not acceptable, dataSize returns -1.
func dataSize(v reflect.Value) int {
if v.Kind() == reflect.Slice {
if s := sizeof(v.Type().Elem()); s >= 0 {
return s * v.Len()
}
return -1
}
return sizeof(v.Type())
}
// sizeof returns the size >= 0 of variables for the given type or -1 if the type is not acceptable.
func sizeof(t reflect.Type) int {
switch t.Kind() {
case reflect.Array:
if s := sizeof(t.Elem()); s >= 0 {
return s * t.Len()
}
case reflect.Struct:
sum := 0
for i, n := 0, t.NumField(); i < n; i++ {
s := sizeof(t.Field(i).Type)
if s < 0 {
return -1
}
sum += s
}
return sum
case reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64,
reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64,
reflect.Float32, reflect.Float64, reflect.Complex64, reflect.Complex128, reflect.Ptr:
return int(t.Size())
}
return -1
}
type coder struct {
order ByteOrder
buf []byte
}
type (
decoder coder
encoder coder
)
func (d *decoder) uint8() uint8 {
x := d.buf[0]
d.buf = d.buf[1:]
return x
}
func (e *encoder) uint8(x uint8) {
e.buf[0] = x
e.buf = e.buf[1:]
}
func (d *decoder) uint16() uint16 {
x := d.order.Uint16(d.buf[0:2])
d.buf = d.buf[2:]
return x
}
func (e *encoder) uint16(x uint16) {
e.order.PutUint16(e.buf[0:2], x)
e.buf = e.buf[2:]
}
func (d *decoder) uint32() uint32 {
x := d.order.Uint32(d.buf[0:4])
d.buf = d.buf[4:]
return x
}
func (e *encoder) uint32(x uint32) {
e.order.PutUint32(e.buf[0:4], x)
e.buf = e.buf[4:]
}
func (d *decoder) uint64() uint64 {
x := d.order.Uint64(d.buf[0:8])
d.buf = d.buf[8:]
return x
}
func (e *encoder) uint64(x uint64) {
e.order.PutUint64(e.buf[0:8], x)
e.buf = e.buf[8:]
}
func (d *decoder) int8() int8 { return int8(d.uint8()) }
func (e *encoder) int8(x int8) { e.uint8(uint8(x)) }
func (d *decoder) int16() int16 { return int16(d.uint16()) }
func (e *encoder) int16(x int16) { e.uint16(uint16(x)) }
func (d *decoder) int32() int32 { return int32(d.uint32()) }
func (e *encoder) int32(x int32) { e.uint32(uint32(x)) }
func (d *decoder) int64() int64 { return int64(d.uint64()) }
func (e *encoder) int64(x int64) { e.uint64(uint64(x)) }
func (d *decoder) value(v reflect.Value) {
switch v.Kind() {
case reflect.Array:
l := v.Len()
for i := 0; i < l; i++ {
d.value(v.Index(i))
}
case reflect.Struct:
t := v.Type()
l := v.NumField()
for i := 0; i < l; i++ {
// Note: Calling v.CanSet() below is an optimization.
// It would be sufficient to check the field name,
// but creating the StructField info for each field is
// costly (run "go test -bench=ReadStruct" and compare
// results when making changes to this code).
if v := v.Field(i); v.CanSet() || t.Field(i).Name != "_" | {
d.value(v)
} | conditional_block |
|
binary.go | func (littleEndian) Uint32(b []byte) uint32 {
return uint32(b[0]) | uint32(b[1])<<8 | uint32(b[2])<<16 | uint32(b[3])<<24
}
func (littleEndian) PutUint32(b []byte, v uint32) {
b[0] = byte(v)
b[1] = byte(v >> 8)
b[2] = byte(v >> 16)
b[3] = byte(v >> 24)
}
func (littleEndian) Uint64(b []byte) uint64 {
return uint64(b[0]) | uint64(b[1])<<8 | uint64(b[2])<<16 | uint64(b[3])<<24 |
uint64(b[4])<<32 | uint64(b[5])<<40 | uint64(b[6])<<48 | uint64(b[7])<<56
}
func (littleEndian) PutUint64(b []byte, v uint64) {
b[0] = byte(v)
b[1] = byte(v >> 8)
b[2] = byte(v >> 16)
b[3] = byte(v >> 24)
b[4] = byte(v >> 32)
b[5] = byte(v >> 40)
b[6] = byte(v >> 48)
b[7] = byte(v >> 56)
}
func (littleEndian) String() string { return "LittleEndian" }
func (littleEndian) GoString() string { return "binary.LittleEndian" }
type bigEndian struct{}
func (bigEndian) Uint16(b []byte) uint16 { return uint16(b[1]) | uint16(b[0])<<8 }
func (bigEndian) PutUint16(b []byte, v uint16) {
b[0] = byte(v >> 8)
b[1] = byte(v)
}
func (bigEndian) Uint32(b []byte) uint32 {
return uint32(b[3]) | uint32(b[2])<<8 | uint32(b[1])<<16 | uint32(b[0])<<24
}
func (bigEndian) PutUint32(b []byte, v uint32) {
b[0] = byte(v >> 24)
b[1] = byte(v >> 16)
b[2] = byte(v >> 8)
b[3] = byte(v)
}
func (bigEndian) Uint64(b []byte) uint64 {
return uint64(b[7]) | uint64(b[6])<<8 | uint64(b[5])<<16 | uint64(b[4])<<24 |
uint64(b[3])<<32 | uint64(b[2])<<40 | uint64(b[1])<<48 | uint64(b[0])<<56
}
func (bigEndian) PutUint64(b []byte, v uint64) {
b[0] = byte(v >> 56)
b[1] = byte(v >> 48)
b[2] = byte(v >> 40)
b[3] = byte(v >> 32)
b[4] = byte(v >> 24)
b[5] = byte(v >> 16)
b[6] = byte(v >> 8)
b[7] = byte(v)
}
func (bigEndian) String() string { return "BigEndian" }
func (bigEndian) GoString() string { return "binary.BigEndian" }
// Read reads structured binary data from r into data.
// Data must be a pointer to a fixed-size value or a slice
// of fixed-size values.
// Bytes read from r are decoded using the specified byte order
// and written to successive fields of the data.
// When reading into structs, the field data for fields with
// blank (_) field names is skipped; i.e., blank field names
// may be used for padding.
// When reading into a struct, all non-blank fields must be exported.
func Read(r io.Reader, order ByteOrder, data interface{}) error {
// Fast path for basic types and slices.
if n := intDataSize(data); n != 0 {
var b [8]byte
var bs []byte
if n > len(b) {
bs = make([]byte, n)
} else {
bs = b[:n]
}
if _, err := io.ReadFull(r, bs); err != nil {
return err
}
switch data := data.(type) {
case *int8:
*data = int8(b[0])
case *uint8:
*data = b[0]
case *int16:
*data = int16(order.Uint16(bs))
case *uint16:
*data = order.Uint16(bs)
case *int32:
*data = int32(order.Uint32(bs))
case *uint32:
*data = order.Uint32(bs)
case *int64:
*data = int64(order.Uint64(bs))
case *uint64:
*data = order.Uint64(bs)
case []int8:
for i, x := range bs { // Easier to loop over the input for 8-bit values.
data[i] = int8(x)
}
case []uint8:
copy(data, bs)
case []int16:
for i := range data {
data[i] = int16(order.Uint16(bs[2*i:]))
}
case []uint16:
for i := range data {
data[i] = order.Uint16(bs[2*i:])
}
case []int32:
for i := range data {
data[i] = int32(order.Uint32(bs[4*i:]))
}
case []uint32:
for i := range data {
data[i] = order.Uint32(bs[4*i:])
}
case []int64:
for i := range data {
data[i] = int64(order.Uint64(bs[8*i:]))
}
case []uint64:
for i := range data {
data[i] = order.Uint64(bs[8*i:])
}
}
return nil
}
// Fallback to reflect-based decoding.
v := reflect.ValueOf(data)
size := -1
switch v.Kind() {
case reflect.Ptr:
v = v.Elem()
size = dataSize(v)
case reflect.Slice:
size = dataSize(v)
}
if size < 0 {
return errors.New("binary.Read: invalid type " + reflect.TypeOf(data).String())
}
d := &decoder{order: order, buf: make([]byte, size)}
if _, err := io.ReadFull(r, d.buf); err != nil {
return err
}
d.value(v)
return nil
}
// Write writes the binary representation of data into w.
// Data must be a fixed-size value or a slice of fixed-size
// values, or a pointer to such data.
// Bytes written to w are encoded using the specified byte order
// and read from successive fields of the data.
// When writing structs, zero values are written for fields
// with blank (_) field names.
func Write(w io.Writer, order ByteOrder, data interface{}) error {
// Fast path for basic types and slices.
if n := intDataSize(data); n != 0 {
var b [8]byte
var bs []byte
if n > len(b) {
bs = make([]byte, n)
} else {
bs = b[:n]
}
switch v := data.(type) {
case *int8:
bs = b[:1]
b[0] = byte(*v)
case int8:
bs = b[:1]
b[0] = byte(v)
case []int8:
for i, x := range v {
bs[i] = byte(x)
}
case *uint8:
bs = b[:1]
b[0] = *v
case uint8:
bs = b[:1]
b[0] = byte(v)
case []uint8:
bs = v
case *int16:
bs = b[:2]
order.PutUint16(bs, uint16(*v))
case int16:
bs = b[:2]
order.PutUint16(bs, uint16(v))
case []int16:
for i, x := range v {
order.PutUint16(bs[2*i:], uint16(x))
}
case *uint16:
bs = b[:2]
order.PutUint16(bs, *v)
case uint16:
bs = b[:2]
order.PutUint16(bs, v)
case []uint16:
for i, x := range v {
order.PutUint16(bs[2*i:], x)
}
case *int32:
bs = b[: | }
| random_line_split |
|
tx.rs | (op1, op2) {
(Some(op1), Some(op2)) => {
if op1 != op2 {
panic!("Conflicting prevout information in input.");
}
op1
}
(Some(op), None) => op,
(None, Some(op)) => op,
(None, None) => panic!("No previous output provided in input."),
}
}
fn bytes_32(bytes: &[u8]) -> Option<[u8; 32]> {
if bytes.len() != 32 {
None
} else {
let mut array = [0; 32];
for (x, y) in bytes.iter().zip(array.iter_mut()) {
*y = *x;
}
Some(array)
}
}
fn create_confidential_value(info: ConfidentialValueInfo) -> confidential::Value {
match info.type_ {
ConfidentialType::Null => confidential::Value::Null,
ConfidentialType::Explicit => confidential::Value::Explicit(
info.value.expect("Field \"value\" is required for explicit values."),
),
ConfidentialType::Confidential => {
let comm = PedersenCommitment::from_slice(
&info.commitment
.expect("Field \"commitment\" is required for confidential values.")
.0[..]
).expect("invalid confidential commitment");
confidential::Value::Confidential(comm)
}
}
}
fn create_confidential_asset(info: ConfidentialAssetInfo) -> confidential::Asset {
match info.type_ {
ConfidentialType::Null => confidential::Asset::Null,
ConfidentialType::Explicit => confidential::Asset::Explicit(
info.asset.expect("Field \"asset\" is required for explicit assets."),
),
ConfidentialType::Confidential => {
let gen = Generator::from_slice(
&info.commitment
.expect("Field \"commitment\" is required for confidential values.")
.0[..]
).expect("invalid confidential commitment");
confidential::Asset::Confidential(gen)
}
}
}
fn create_confidential_nonce(info: ConfidentialNonceInfo) -> confidential::Nonce {
match info.type_ {
ConfidentialType::Null => confidential::Nonce::Null,
ConfidentialType::Explicit => confidential::Nonce::Explicit(bytes_32(
&info.nonce
.expect("Field \"nonce\" is required for asset issuances.")
.0[..],
).expect("wrong size of \"nonce\" field")),
ConfidentialType::Confidential => {
let pubkey = PublicKey::from_slice(
&info.commitment
.expect("Field \"commitment\" is required for confidential values.")
.0[..]
).expect("invalid confidential commitment");
confidential::Nonce::Confidential(pubkey)
}
}
}
fn create_asset_issuance(info: AssetIssuanceInfo) -> AssetIssuance {
AssetIssuance {
asset_blinding_nonce: Tweak::from_slice(
&info.asset_blinding_nonce
.expect("Field \"asset_blinding_nonce\" is required for asset issuances.")
.0[..]
).expect("Invalid \"asset_blinding_nonce\"."),
asset_entropy: bytes_32(
&info.asset_entropy
.expect("Field \"asset_entropy\" is required for asset issuances.")
.0[..],
).expect("Invalid size of \"asset_entropy\"."),
amount: create_confidential_value(
info.amount.expect("Field \"amount\" is required for asset issuances."),
),
inflation_keys: create_confidential_value(
info.inflation_keys.expect("Field \"inflation_keys\" is required for asset issuances."),
),
}
}
fn create_script_sig(ss: InputScriptInfo) -> Script {
if let Some(hex) = ss.hex {
if ss.asm.is_some() {
warn!("Field \"asm\" of input is ignored.");
}
hex.0.into()
} else if let Some(_) = ss.asm {
panic!("Decoding script assembly is not yet supported.");
} else {
panic!("No scriptSig info provided.");
}
}
fn create_pegin_witness(pd: PeginDataInfo, prevout: bitcoin::OutPoint) -> Vec<Vec<u8>> {
if prevout != pd.outpoint.parse().expect("Invalid outpoint in field \"pegin_data\".") {
panic!("Outpoint in \"pegin_data\" does not correspond to input value.");
}
let asset = match create_confidential_asset(pd.asset) {
confidential::Asset::Explicit(asset) => asset,
_ => panic!("Asset in \"pegin_data\" should be explicit."),
};
vec![
serialize(&pd.value),
serialize(&asset),
serialize(&pd.genesis_hash),
serialize(&pd.claim_script.0),
serialize(&pd.mainchain_tx_hex.0),
serialize(&pd.merkle_proof.0),
]
}
fn convert_outpoint_to_btc(p: elements::OutPoint) -> bitcoin::OutPoint {
bitcoin::OutPoint {
txid: bitcoin::Txid::from_inner(p.txid.into_inner()),
vout: p.vout,
}
}
fn create_input_witness(
info: Option<InputWitnessInfo>,
pd: Option<PeginDataInfo>,
prevout: OutPoint,
) -> TxInWitness {
let pegin_witness = if info.is_some() && info.as_ref().unwrap().pegin_witness.is_some() {
if pd.is_some() {
warn!("Field \"pegin_data\" of input is ignored.");
}
info.as_ref().unwrap().pegin_witness.clone().unwrap().iter().map(|h| h.clone().0).collect()
} else if let Some(pd) = pd {
create_pegin_witness(pd, convert_outpoint_to_btc(prevout))
} else {
Default::default()
};
if let Some(wi) = info {
TxInWitness {
amount_rangeproof: wi.amount_rangeproof
.map(|b| Box::new(RangeProof::from_slice(&b.0).expect("invalid rangeproof"))),
inflation_keys_rangeproof: wi.inflation_keys_rangeproof
.map(|b| Box::new(RangeProof::from_slice(&b.0).expect("invalid rangeproof"))),
script_witness: match wi.script_witness {
Some(ref w) => w.iter().map(|h| h.clone().0).collect(),
None => Vec::new(),
},
pegin_witness: pegin_witness,
}
} else {
TxInWitness {
pegin_witness: pegin_witness,
..Default::default()
}
}
}
fn create_input(input: InputInfo) -> TxIn | }
fn create_script_pubkey(spk: OutputScriptInfo, used_network: &mut Option<Network>) -> Script {
if spk.type_.is_some() {
warn!("Field \"type\" of output is ignored.");
}
if let Some(hex) = spk.hex {
if spk.asm.is_some() {
warn!("Field \"asm\" of output is ignored.");
}
if spk.address.is_some() {
warn!("Field \"address\" of output is ignored.");
}
//TODO(stevenroose) do script sanity check to avoid blackhole?
hex.0.into()
} else if let Some(_) = spk.asm {
if spk.address.is_some() {
warn!("Field \"address\" of output is ignored.");
}
panic!("Decoding script assembly is not yet supported.");
} else if let Some(address) = spk.address {
// Error if another network had already been used.
if let Some(network) = Network::from_params(address.params) {
if used_network.replace(network).unwrap_or(network) != network {
panic!("Addresses for different networks are used in the output scripts.");
}
}
address.script_pubkey()
} else {
panic!("No scriptPubKey info provided.");
}
}
fn create_bitcoin_script_pubkey(spk: hal::tx::OutputScriptInfo) -> bitcoin::Script {
if spk.type_.is_some() {
warn!("Field \"type\" of output is ignored.");
}
if let Some(hex) | {
let has_issuance = input.has_issuance.unwrap_or(input.asset_issuance.is_some());
let is_pegin = input.is_pegin.unwrap_or(input.pegin_data.is_some());
let prevout = outpoint_from_input_info(&input);
TxIn {
previous_output: prevout,
script_sig: input.script_sig.map(create_script_sig).unwrap_or_default(),
sequence: elements::Sequence::from_height(input.sequence.unwrap_or_default().try_into().unwrap()),
is_pegin: is_pegin,
asset_issuance: if has_issuance {
input.asset_issuance.map(create_asset_issuance).unwrap_or_default()
} else {
if input.asset_issuance.is_some() {
warn!("Field \"asset_issuance\" of input is ignored.");
}
Default::default()
},
witness: create_input_witness(input.witness, input.pegin_data, prevout),
} | identifier_body |
tx.rs | (input: &InputInfo) -> OutPoint {
let op1: Option<OutPoint> =
input.prevout.as_ref().map(|ref op| op.parse().expect("invalid prevout format"));
let op2 = match input.txid {
Some(txid) => match input.vout {
Some(vout) => Some(OutPoint {
txid: txid,
vout: vout,
}),
None => panic!("\"txid\" field given in input without \"vout\" field"),
},
None => None,
};
match (op1, op2) {
(Some(op1), Some(op2)) => {
if op1 != op2 {
panic!("Conflicting prevout information in input.");
}
op1
}
(Some(op), None) => op,
(None, Some(op)) => op,
(None, None) => panic!("No previous output provided in input."),
}
}
fn bytes_32(bytes: &[u8]) -> Option<[u8; 32]> {
if bytes.len() != 32 {
None
} else {
let mut array = [0; 32];
for (x, y) in bytes.iter().zip(array.iter_mut()) {
*y = *x;
}
Some(array)
}
}
fn create_confidential_value(info: ConfidentialValueInfo) -> confidential::Value {
match info.type_ {
ConfidentialType::Null => confidential::Value::Null,
ConfidentialType::Explicit => confidential::Value::Explicit(
info.value.expect("Field \"value\" is required for explicit values."),
),
ConfidentialType::Confidential => {
let comm = PedersenCommitment::from_slice(
&info.commitment
.expect("Field \"commitment\" is required for confidential values.")
.0[..]
).expect("invalid confidential commitment");
confidential::Value::Confidential(comm)
}
}
}
fn create_confidential_asset(info: ConfidentialAssetInfo) -> confidential::Asset {
match info.type_ {
ConfidentialType::Null => confidential::Asset::Null,
ConfidentialType::Explicit => confidential::Asset::Explicit(
info.asset.expect("Field \"asset\" is required for explicit assets."),
),
ConfidentialType::Confidential => {
let gen = Generator::from_slice(
&info.commitment
.expect("Field \"commitment\" is required for confidential values.")
.0[..]
).expect("invalid confidential commitment");
confidential::Asset::Confidential(gen)
}
}
}
fn create_confidential_nonce(info: ConfidentialNonceInfo) -> confidential::Nonce {
match info.type_ {
ConfidentialType::Null => confidential::Nonce::Null,
ConfidentialType::Explicit => confidential::Nonce::Explicit(bytes_32(
&info.nonce
.expect("Field \"nonce\" is required for asset issuances.")
.0[..],
).expect("wrong size of \"nonce\" field")),
ConfidentialType::Confidential => {
let pubkey = PublicKey::from_slice(
&info.commitment
.expect("Field \"commitment\" is required for confidential values.")
.0[..]
).expect("invalid confidential commitment");
confidential::Nonce::Confidential(pubkey)
}
}
}
fn create_asset_issuance(info: AssetIssuanceInfo) -> AssetIssuance {
AssetIssuance {
asset_blinding_nonce: Tweak::from_slice(
&info.asset_blinding_nonce
.expect("Field \"asset_blinding_nonce\" is required for asset issuances.")
.0[..]
).expect("Invalid \"asset_blinding_nonce\"."),
asset_entropy: bytes_32(
&info.asset_entropy
.expect("Field \"asset_entropy\" is required for asset issuances.")
.0[..],
).expect("Invalid size of \"asset_entropy\"."),
amount: create_confidential_value(
info.amount.expect("Field \"amount\" is required for asset issuances."),
),
inflation_keys: create_confidential_value(
info.inflation_keys.expect("Field \"inflation_keys\" is required for asset issuances."),
),
}
}
fn create_script_sig(ss: InputScriptInfo) -> Script {
if let Some(hex) = ss.hex {
if ss.asm.is_some() {
warn!("Field \"asm\" of input is ignored.");
}
hex.0.into()
} else if let Some(_) = ss.asm {
panic!("Decoding script assembly is not yet supported.");
} else {
panic!("No scriptSig info provided.");
}
}
fn create_pegin_witness(pd: PeginDataInfo, prevout: bitcoin::OutPoint) -> Vec<Vec<u8>> {
if prevout != pd.outpoint.parse().expect("Invalid outpoint in field \"pegin_data\".") {
panic!("Outpoint in \"pegin_data\" does not correspond to input value.");
}
let asset = match create_confidential_asset(pd.asset) {
confidential::Asset::Explicit(asset) => asset,
_ => panic!("Asset in \"pegin_data\" should be explicit."),
};
vec![
serialize(&pd.value),
serialize(&asset),
serialize(&pd.genesis_hash),
serialize(&pd.claim_script.0),
serialize(&pd.mainchain_tx_hex.0),
serialize(&pd.merkle_proof.0),
]
}
fn convert_outpoint_to_btc(p: elements::OutPoint) -> bitcoin::OutPoint {
bitcoin::OutPoint {
txid: bitcoin::Txid::from_inner(p.txid.into_inner()),
vout: p.vout,
}
}
fn create_input_witness(
info: Option<InputWitnessInfo>,
pd: Option<PeginDataInfo>,
prevout: OutPoint,
) -> TxInWitness {
let pegin_witness = if info.is_some() && info.as_ref().unwrap().pegin_witness.is_some() {
if pd.is_some() {
warn!("Field \"pegin_data\" of input is ignored.");
}
info.as_ref().unwrap().pegin_witness.clone().unwrap().iter().map(|h| h.clone().0).collect()
} else if let Some(pd) = pd {
create_pegin_witness(pd, convert_outpoint_to_btc(prevout))
} else {
Default::default()
};
if let Some(wi) = info {
TxInWitness {
amount_rangeproof: wi.amount_rangeproof
.map(|b| Box::new(RangeProof::from_slice(&b.0).expect("invalid rangeproof"))),
inflation_keys_rangeproof: wi.inflation_keys_rangeproof
.map(|b| Box::new(RangeProof::from_slice(&b.0).expect("invalid rangeproof"))),
script_witness: match wi.script_witness {
Some(ref w) => w.iter().map(|h| h.clone().0).collect(),
None => Vec::new(),
},
pegin_witness: pegin_witness,
}
} else {
TxInWitness {
pegin_witness: pegin_witness,
..Default::default()
}
}
}
fn create_input(input: InputInfo) -> TxIn {
let has_issuance = input.has_issuance.unwrap_or(input.asset_issuance.is_some());
let is_pegin = input.is_pegin.unwrap_or(input.pegin_data.is_some());
let prevout = outpoint_from_input_info(&input);
TxIn {
previous_output: prevout,
script_sig: input.script_sig.map(create_script_sig).unwrap_or_default(),
sequence: elements::Sequence::from_height(input.sequence.unwrap_or_default().try_into().unwrap()),
is_pegin: is_pegin,
asset_issuance: if has_issuance {
input.asset_issuance.map(create_asset_issuance).unwrap_or_default()
} else {
if input.asset_issuance.is_some() {
warn!("Field \"asset_issuance\" of input is ignored.");
}
Default::default()
},
witness: create_input_witness(input.witness, input.pegin_data, prevout),
}
}
fn create_script_pubkey(spk: OutputScriptInfo, used_network: &mut Option<Network>) -> Script {
if spk.type_.is_some() {
warn!("Field \"type\" of output is ignored.");
}
if let Some(hex) = spk.hex {
if spk.asm.is_some() {
warn!("Field \"asm\" of output is ignored.");
}
if spk.address.is_some() {
warn!("Field \"address\" of output is ignored.");
}
//TODO(stevenroose) do script sanity check to avoid blackhole?
hex.0.into()
} else if let Some(_) = spk.asm {
if spk.address.is_some() {
warn!("Field \"address\" of output is ignored.");
}
panic!("Decoding script assembly is not yet supported.");
} else if let Some(address) = spk.address {
// Error if another network had already | outpoint_from_input_info | identifier_name |
|
tx.rs | serialize(&pd.mainchain_tx_hex.0),
serialize(&pd.merkle_proof.0),
]
}
fn convert_outpoint_to_btc(p: elements::OutPoint) -> bitcoin::OutPoint {
bitcoin::OutPoint {
txid: bitcoin::Txid::from_inner(p.txid.into_inner()),
vout: p.vout,
}
}
fn create_input_witness(
info: Option<InputWitnessInfo>,
pd: Option<PeginDataInfo>,
prevout: OutPoint,
) -> TxInWitness {
let pegin_witness = if info.is_some() && info.as_ref().unwrap().pegin_witness.is_some() {
if pd.is_some() {
warn!("Field \"pegin_data\" of input is ignored.");
}
info.as_ref().unwrap().pegin_witness.clone().unwrap().iter().map(|h| h.clone().0).collect()
} else if let Some(pd) = pd {
create_pegin_witness(pd, convert_outpoint_to_btc(prevout))
} else {
Default::default()
};
if let Some(wi) = info {
TxInWitness {
amount_rangeproof: wi.amount_rangeproof
.map(|b| Box::new(RangeProof::from_slice(&b.0).expect("invalid rangeproof"))),
inflation_keys_rangeproof: wi.inflation_keys_rangeproof
.map(|b| Box::new(RangeProof::from_slice(&b.0).expect("invalid rangeproof"))),
script_witness: match wi.script_witness {
Some(ref w) => w.iter().map(|h| h.clone().0).collect(),
None => Vec::new(),
},
pegin_witness: pegin_witness,
}
} else {
TxInWitness {
pegin_witness: pegin_witness,
..Default::default()
}
}
}
fn create_input(input: InputInfo) -> TxIn {
let has_issuance = input.has_issuance.unwrap_or(input.asset_issuance.is_some());
let is_pegin = input.is_pegin.unwrap_or(input.pegin_data.is_some());
let prevout = outpoint_from_input_info(&input);
TxIn {
previous_output: prevout,
script_sig: input.script_sig.map(create_script_sig).unwrap_or_default(),
sequence: elements::Sequence::from_height(input.sequence.unwrap_or_default().try_into().unwrap()),
is_pegin: is_pegin,
asset_issuance: if has_issuance {
input.asset_issuance.map(create_asset_issuance).unwrap_or_default()
} else {
if input.asset_issuance.is_some() {
warn!("Field \"asset_issuance\" of input is ignored.");
}
Default::default()
},
witness: create_input_witness(input.witness, input.pegin_data, prevout),
}
}
fn create_script_pubkey(spk: OutputScriptInfo, used_network: &mut Option<Network>) -> Script {
if spk.type_.is_some() {
warn!("Field \"type\" of output is ignored.");
}
if let Some(hex) = spk.hex {
if spk.asm.is_some() {
warn!("Field \"asm\" of output is ignored.");
}
if spk.address.is_some() {
warn!("Field \"address\" of output is ignored.");
}
//TODO(stevenroose) do script sanity check to avoid blackhole?
hex.0.into()
} else if let Some(_) = spk.asm {
if spk.address.is_some() {
warn!("Field \"address\" of output is ignored.");
}
panic!("Decoding script assembly is not yet supported.");
} else if let Some(address) = spk.address {
// Error if another network had already been used.
if let Some(network) = Network::from_params(address.params) {
if used_network.replace(network).unwrap_or(network) != network {
panic!("Addresses for different networks are used in the output scripts.");
}
}
address.script_pubkey()
} else {
panic!("No scriptPubKey info provided.");
}
}
fn create_bitcoin_script_pubkey(spk: hal::tx::OutputScriptInfo) -> bitcoin::Script {
if spk.type_.is_some() {
warn!("Field \"type\" of output is ignored.");
}
if let Some(hex) = spk.hex {
if spk.asm.is_some() {
warn!("Field \"asm\" of output is ignored.");
}
if spk.address.is_some() {
warn!("Field \"address\" of output is ignored.");
}
//TODO(stevenroose) do script sanity check to avoid blackhole?
hex.0.into()
} else if let Some(_) = spk.asm {
if spk.address.is_some() {
warn!("Field \"address\" of output is ignored.");
}
panic!("Decoding script assembly is not yet supported.");
} else if let Some(address) = spk.address {
address.script_pubkey()
} else {
panic!("No scriptPubKey info provided.");
}
}
fn create_output_witness(w: OutputWitnessInfo) -> TxOutWitness {
TxOutWitness {
surjection_proof: w.surjection_proof.map(|b| {
Box::new(SurjectionProof::from_slice(&b.0[..]).expect("invalid surjection proof"))
}),
rangeproof: w.rangeproof.map(|b| {
Box::new(RangeProof::from_slice(&b.0[..]).expect("invalid rangeproof"))
}),
}
}
fn create_script_pubkey_from_pegout_data(
pd: PegoutDataInfo,
) -> Script {
let mut builder = elements::script::Builder::new()
.push_opcode(elements::opcodes::all::OP_RETURN)
.push_slice(&pd.genesis_hash.into_inner()[..])
.push_slice(&create_bitcoin_script_pubkey(pd.script_pub_key)[..]);
for d in pd.extra_data {
builder = builder.push_slice(&d.0);
}
builder.into_script()
}
fn create_output(output: OutputInfo) -> TxOut {
// Keep track of which network has been used in addresses and error if two different networks
// are used.
let mut used_network = None;
let value = output
.value
.map(create_confidential_value)
.expect("Field \"value\" is required for outputs.");
let asset = output
.asset
.map(create_confidential_asset)
.expect("Field \"asset\" is required for outputs.");
TxOut {
asset: asset,
value: value,
nonce: output.nonce.map(create_confidential_nonce).unwrap_or(confidential::Nonce::Null),
script_pubkey: if let Some(spk) = output.script_pub_key {
if output.pegout_data.is_some() {
warn!("Field \"pegout_data\" of output is ignored.");
}
create_script_pubkey(spk, &mut used_network)
} else if let Some(pd) = output.pegout_data {
match value {
confidential::Value::Explicit(v) => {
if v != pd.value {
panic!("Value in \"pegout_data\" does not correspond to output value.");
}
}
_ => panic!("Explicit value is required for pegout data."),
}
if asset != create_confidential_asset(pd.asset.clone()) {
panic!("Asset in \"pegout_data\" does not correspond to output value.");
}
create_script_pubkey_from_pegout_data(pd)
} else {
Default::default()
},
witness: output.witness.map(create_output_witness).unwrap_or_default(),
}
}
pub fn create_transaction(info: TransactionInfo) -> Transaction {
// Fields that are ignored.
if info.txid.is_some() {
warn!("Field \"txid\" is ignored.");
}
if info.hash.is_some() {
warn!("Field \"hash\" is ignored.");
}
if info.size.is_some() {
warn!("Field \"size\" is ignored.");
}
if info.weight.is_some() {
warn!("Field \"weight\" is ignored.");
}
if info.vsize.is_some() {
warn!("Field \"vsize\" is ignored.");
}
Transaction {
version: info.version.expect("Field \"version\" is required."),
lock_time: elements::PackedLockTime(info.locktime.expect("Field \"locktime\" is required.")),
input: info
.inputs
.expect("Field \"inputs\" is required.")
.into_iter()
.map(create_input)
.collect(),
output: info
.outputs
.expect("Field \"outputs\" is required.")
.into_iter()
.map(create_output)
.collect(),
}
}
fn exec_create<'a>(matches: &clap::ArgMatches<'a>) {
let info = serde_json::from_str::<TransactionInfo>(&cmd::arg_or_stdin(matches, "tx-info"))
.expect("invalid JSON provided");
let tx = create_transaction(info);
let tx_bytes = serialize(&tx);
if matches.is_present("raw-stdout") { | ::std::io::stdout().write_all(&tx_bytes).unwrap(); | random_line_split |
|
RHD_Load_Filter.py | _data_blocks):
read_one_data_block(data, header, indices, fid)
# Increment indices
indices['amplifier'] += header['num_samples_per_data_block']
indices['aux_input'] += int(header['num_samples_per_data_block'] / 4)
indices['supply_voltage'] += 1
indices['board_adc'] += header['num_samples_per_data_block']
indices['board_dig_in'] += header['num_samples_per_data_block']
indices['board_dig_out'] += header['num_samples_per_data_block']
fraction_done = 100 * (1.0 * i / num_data_blocks)
if fraction_done >= percent_done:
print('{}% done...'.format(percent_done))
percent_done = percent_done + print_increment
# Make sure we have read exactly the right amount of data.
bytes_remaining = filesize - fid.tell()
if bytes_remaining != 0: raise Exception('Error: End of file not reached.')
# Close data file.
fid.close()
if (data_present):
print('Parsing data...')
# Extract digital input channels to separate variables.
for i in range(header['num_board_dig_in_channels']):
data['board_dig_in_data'][i, :] = np.not_equal(np.bitwise_and(data['board_dig_in_raw'], (1 << header['board_dig_in_channels'][i]['native_order'])), 0)
# Extract digital output channels to separate variables.
for i in range(header['num_board_dig_out_channels']):
data['board_dig_out_data'][i, :] = np.not_equal(np.bitwise_and(data['board_dig_out_raw'], (1 << header['board_dig_out_channels'][i]['native_order'])), 0)
# Scale voltage levels appropriately.
data['amplifier_data'] = np.multiply(0.195, (data['amplifier_data'].astype(np.int32) - 32768)) # units = microvolts
data['aux_input_data'] = np.multiply(37.4e-6, data['aux_input_data']) # units = volts
data['supply_voltage_data'] = np.multiply(74.8e-6, data['supply_voltage_data']) # units = volts
if header['eval_board_mode'] == 1:
data['board_adc_data'] = np.multiply(152.59e-6, (data['board_adc_data'].astype(np.int32) - 32768)) # units = volts
elif header['eval_board_mode'] == 13:
data['board_adc_data'] = np.multiply(312.5e-6, (data['board_adc_data'].astype(np.int32) - 32768)) # units = volts
else:
data['board_adc_data'] = np.multiply(50.354e-6, data['board_adc_data']) # units = volts
data['temp_sensor_data'] = np.multiply(0.01, data['temp_sensor_data']) # units = deg C
# Check for gaps in timestamps.
num_gaps = np.sum(np.not_equal(data['t_amplifier'][1:]-data['t_amplifier'][:-1], 1))
if num_gaps == 0:
print('No missing timestamps in data.')
else:
print('Warning: {0} gaps in timestamp data found. Time scale will not be uniform!'.format(num_gaps))
# Scale time steps (units = seconds).
data['t_amplifier'] = data['t_amplifier'] / header['sample_rate']
data['t_aux_input'] = data['t_amplifier'][range(0, len(data['t_amplifier']), 4)]
data['t_supply_voltage'] = data['t_amplifier'][range(0, len(data['t_amplifier']), header['num_samples_per_data_block'])]
data['t_board_adc'] = data['t_amplifier']
data['t_dig'] = data['t_amplifier']
data['t_temp_sensor'] = data['t_supply_voltage']
# If the software notch filter was selected during the recording, apply the
# same notch filter to amplifier data here.
if header['notch_filter_frequency'] > 0 and header['version']['major'] < 3:
print('Applying notch filter...')
print_increment = 10
percent_done = print_increment
for i in range(header['num_amplifier_channels']):
data['amplifier_data'][i,:] = notch_filter(data['amplifier_data'][i,:], header['sample_rate'], header['notch_filter_frequency'], 10)
fraction_done = 100 * (i / header['num_amplifier_channels'])
if fraction_done >= percent_done:
print('{}% done...'.format(percent_done))
percent_done += print_increment
else:
data = [];
# Move variables to result struct.
result = data_to_result(header, data, data_present)
print('Done! Elapsed time: {0:0.1f} seconds'.format(time.time() - tic))
return result
def plural(n):
"""Utility function to optionally pluralize words based on the value of n.
"""
if n == 1:
return ''
else:
return 's'
path="//equipe2-nas1/Gilles.DELBECQ/Data/ePhy/Février2023/Test_Gustave/raw/raw intan/Test_Gustave_15_03_230315_182841/Test_Gustave_15_03_230315_182841.rhd"
reader=read_data(path)
sampling_rate = reader['frequency_parameters']['amplifier_sample_rate']
time_vector=reader['t_amplifier']
signal=reader['amplifier_data']
selected_channels=['2','3','4','9','10','11','12','14','15']
#Filtering parameters
freq_low = 300
freq_high = 3000
order = 4
# Noise parameters
std_threshold = 5 #Times the std
noise_window = 5 #window for the noise calculation in sec
distance = 50 # distance between 2 spikes
#waveform window
waveform_window=5 #ms
Waveforms = True
def extract_spike_waveform(signal, spike_idx, left_width=(waveform_window/1000)*20000/2, right_width=(waveform_window/1000)*20000/2):
'''
Function to extract spikes waveforms in spike2 recordings
INPUTS :
signal (1-d array) : the ephy signal
spike_idx (1-d array or integer list) : array containing the spike indexes (in points)
width (int) = width for spike window
OUTPUTS :
SPIKES (list) : a list containg the waveform of each spike
'''
SPIKES = []
left_width = int(left_width)
right_width = int(right_width)
for i in range(len(spike_idx)):
index = spike_idx[i]
spike_wf = signal[index-left_width : index+right_width]
SPIKES.append(spike_wf)
return SPIKES
def filter_signal(signal, order=order, sample_rate=sampling_rate, freq_low=freq_low, freq_high=freq_high, axis=0):
"""
From Théo G.
Filtering with scipy
inputs raw signal (array)
returns filtered signal (array)
"""
import scipy.signal
Wn = [freq_low / (sample_rate / 2), freq_high / (sample_rate / 2)]
sos_coeff = scipy.signal.iirfilter(order, Wn, btype="band", ftype="butter", output="sos")
filtered_signal = scipy.signal.sosfiltfilt(sos_coeff, signal, axis=axis)
return filtered_signal
# def notch_filter(signal, order=4, sample_rate=20000, freq_low=48, freq_high=52, axis=0):
# import scipy.signal
# Wn = [freq_low / (sample_rate / 2), freq_high / (sample_rate / 2)]
# notch_coeff = scipy.signal.iirfilter(order, Wn, btype="bandstop", ftype="butter", output="sos")
# notch_signal = scipy.signal.sosfiltfilt(notch_coeff, signal, axis=axis)
# return notch_signal
filtered_signals=[]
for i in selected_channels:
filtered_signal=filter_signal(signal[int(i),:])
filtered_signals.append(filtered_signal)
# plt.figure()
# # plt.plot(time_vector,signal[0,:])
# plt.plot(time_vector,filtered_signal)
# plt.title(rf'channel {int(i)}')
filtered_signals = np.array(filtered_signals)
median = np.median(filtered_signals, axis=0)#compute median on all
cmr_signals = filtered_signals-median #compute common ref removal median on all
for i in range(len(cmr_signals)):
pl | t.figure()
plt.plot(time_vector,cmr_signals[i])
plt.title(rf'channel {int(selected_channels[i])}')
| conditional_block |
|
RHD_Load_Filter.py | ['board_dig_out'] += header['num_samples_per_data_block']
fraction_done = 100 * (1.0 * i / num_data_blocks)
if fraction_done >= percent_done:
print('{}% done...'.format(percent_done))
percent_done = percent_done + print_increment
# Make sure we have read exactly the right amount of data.
bytes_remaining = filesize - fid.tell()
if bytes_remaining != 0: raise Exception('Error: End of file not reached.')
# Close data file.
fid.close()
if (data_present):
print('Parsing data...')
# Extract digital input channels to separate variables.
for i in range(header['num_board_dig_in_channels']):
data['board_dig_in_data'][i, :] = np.not_equal(np.bitwise_and(data['board_dig_in_raw'], (1 << header['board_dig_in_channels'][i]['native_order'])), 0)
# Extract digital output channels to separate variables.
for i in range(header['num_board_dig_out_channels']):
data['board_dig_out_data'][i, :] = np.not_equal(np.bitwise_and(data['board_dig_out_raw'], (1 << header['board_dig_out_channels'][i]['native_order'])), 0)
# Scale voltage levels appropriately.
data['amplifier_data'] = np.multiply(0.195, (data['amplifier_data'].astype(np.int32) - 32768)) # units = microvolts
data['aux_input_data'] = np.multiply(37.4e-6, data['aux_input_data']) # units = volts
data['supply_voltage_data'] = np.multiply(74.8e-6, data['supply_voltage_data']) # units = volts
if header['eval_board_mode'] == 1:
data['board_adc_data'] = np.multiply(152.59e-6, (data['board_adc_data'].astype(np.int32) - 32768)) # units = volts
elif header['eval_board_mode'] == 13:
data['board_adc_data'] = np.multiply(312.5e-6, (data['board_adc_data'].astype(np.int32) - 32768)) # units = volts
else:
data['board_adc_data'] = np.multiply(50.354e-6, data['board_adc_data']) # units = volts
data['temp_sensor_data'] = np.multiply(0.01, data['temp_sensor_data']) # units = deg C
# Check for gaps in timestamps.
num_gaps = np.sum(np.not_equal(data['t_amplifier'][1:]-data['t_amplifier'][:-1], 1))
if num_gaps == 0:
print('No missing timestamps in data.')
else:
print('Warning: {0} gaps in timestamp data found. Time scale will not be uniform!'.format(num_gaps))
# Scale time steps (units = seconds).
data['t_amplifier'] = data['t_amplifier'] / header['sample_rate']
data['t_aux_input'] = data['t_amplifier'][range(0, len(data['t_amplifier']), 4)]
data['t_supply_voltage'] = data['t_amplifier'][range(0, len(data['t_amplifier']), header['num_samples_per_data_block'])]
data['t_board_adc'] = data['t_amplifier']
data['t_dig'] = data['t_amplifier']
data['t_temp_sensor'] = data['t_supply_voltage']
# If the software notch filter was selected during the recording, apply the
# same notch filter to amplifier data here.
if header['notch_filter_frequency'] > 0 and header['version']['major'] < 3:
print('Applying notch filter...')
print_increment = 10
percent_done = print_increment
for i in range(header['num_amplifier_channels']):
data['amplifier_data'][i,:] = notch_filter(data['amplifier_data'][i,:], header['sample_rate'], header['notch_filter_frequency'], 10)
fraction_done = 100 * (i / header['num_amplifier_channels'])
if fraction_done >= percent_done:
print('{}% done...'.format(percent_done))
percent_done += print_increment
else:
data = [];
# Move variables to result struct.
result = data_to_result(header, data, data_present)
print('Done! Elapsed time: {0:0.1f} seconds'.format(time.time() - tic))
return result
def plural(n):
"""Utility function to optionally pluralize words based on the value of n.
"""
if n == 1:
return ''
else:
return 's'
path="//equipe2-nas1/Gilles.DELBECQ/Data/ePhy/Février2023/Test_Gustave/raw/raw intan/Test_Gustave_15_03_230315_182841/Test_Gustave_15_03_230315_182841.rhd"
reader=read_data(path)
sampling_rate = reader['frequency_parameters']['amplifier_sample_rate']
time_vector=reader['t_amplifier']
signal=reader['amplifier_data']
selected_channels=['2','3','4','9','10','11','12','14','15']
#Filtering parameters
freq_low = 300
freq_high = 3000
order = 4
# Noise parameters
std_threshold = 5 #Times the std
noise_window = 5 #window for the noise calculation in sec
distance = 50 # distance between 2 spikes
#waveform window
waveform_window=5 #ms
Waveforms = True
def extract_spike_waveform(signal, spike_idx, left_width=(waveform_window/1000)*20000/2, right_width=(waveform_window/1000)*20000/2):
'''
Function to extract spikes waveforms in spike2 recordings
INPUTS :
signal (1-d array) : the ephy signal
spike_idx (1-d array or integer list) : array containing the spike indexes (in points)
width (int) = width for spike window
OUTPUTS :
SPIKES (list) : a list containg the waveform of each spike
'''
SPIKES = []
left_width = int(left_width)
right_width = int(right_width)
for i in range(len(spike_idx)):
index = spike_idx[i]
spike_wf = signal[index-left_width : index+right_width]
SPIKES.append(spike_wf)
return SPIKES
def filter_signal(signal, order=order, sample_rate=sampling_rate, freq_low=freq_low, freq_high=freq_high, axis=0):
"""
From Théo G.
Filtering with scipy
inputs raw signal (array)
returns filtered signal (array)
"""
import scipy.signal
Wn = [freq_low / (sample_rate / 2), freq_high / (sample_rate / 2)]
sos_coeff = scipy.signal.iirfilter(order, Wn, btype="band", ftype="butter", output="sos")
filtered_signal = scipy.signal.sosfiltfilt(sos_coeff, signal, axis=axis)
return filtered_signal
# def notch_filter(signal, order=4, sample_rate=20000, freq_low=48, freq_high=52, axis=0):
# import scipy.signal
# Wn = [freq_low / (sample_rate / 2), freq_high / (sample_rate / 2)]
# notch_coeff = scipy.signal.iirfilter(order, Wn, btype="bandstop", ftype="butter", output="sos")
# notch_signal = scipy.signal.sosfiltfilt(notch_coeff, signal, axis=axis)
# return notch_signal
filtered_signals=[]
for i in selected_channels:
filtered_signal=filter_signal(signal[int(i),:])
filtered_signals.append(filtered_signal)
# plt.figure()
# # plt.plot(time_vector,signal[0,:])
# plt.plot(time_vector,filtered_signal)
# plt.title(rf'channel {int(i)}')
filtered_signals = np.array(filtered_signals)
median = np.median(filtered_signals, axis=0)#compute median on all
cmr_signals = filtered_signals-median #compute common ref removal median on all
for i in range(len(cmr_signals)):
plt.figure()
plt.plot(time_vector,cmr_signals[i])
plt.title(rf'channel {int(selected_channels[i])}')
"""
Spike detection
"""
thresholds=[]
spikes_list=[]
spikes_list_y=[]
wfs=[]
waveforms=[]
for signal in cmr_signals:
# Threshold calculation
noise = signal[0:int(noise_window*sampling_rate)] #noise window taken from individual channel signal
threshold = np.median(noise)+std_threshold*np.std(noise) #threshold calculation for the channel
thresholds.append(threshold) #append it to the list regrouping threshold for each channel | random_line_split |
||
RHD_Load_Filter.py | of data blocks')
num_data_blocks = int(bytes_remaining / bytes_per_block)
num_amplifier_samples = header['num_samples_per_data_block'] * num_data_blocks
num_aux_input_samples = int((header['num_samples_per_data_block'] / 4) * num_data_blocks)
num_supply_voltage_samples = 1 * num_data_blocks
num_board_adc_samples = header['num_samples_per_data_block'] * num_data_blocks
num_board_dig_in_samples = header['num_samples_per_data_block'] * num_data_blocks
num_board_dig_out_samples = header['num_samples_per_data_block'] * num_data_blocks
record_time = num_amplifier_samples / header['sample_rate']
if data_present:
print('File contains {:0.3f} seconds of data. Amplifiers were sampled at {:0.2f} kS/s.'.format(record_time, header['sample_rate'] / 1000))
else:
print('Header file contains no data. Amplifiers were sampled at {:0.2f} kS/s.'.format(header['sample_rate'] / 1000))
if data_present:
# Pre-allocate memory for data.
print('')
print('Allocating memory for data...')
data = {}
if (header['version']['major'] == 1 and header['version']['minor'] >= 2) or (header['version']['major'] > 1):
data['t_amplifier'] = np.zeros(num_amplifier_samples, dtype=np.int_)
else:
data['t_amplifier'] = np.zeros(num_amplifier_samples, dtype=np.uint)
data['amplifier_data'] = np.zeros([header['num_amplifier_channels'], num_amplifier_samples], dtype=np.uint)
data['aux_input_data'] = np.zeros([header['num_aux_input_channels'], num_aux_input_samples], dtype=np.uint)
data['supply_voltage_data'] = np.zeros([header['num_supply_voltage_channels'], num_supply_voltage_samples], dtype=np.uint)
data['temp_sensor_data'] = np.zeros([header['num_temp_sensor_channels'], num_supply_voltage_samples], dtype=np.uint)
data['board_adc_data'] = np.zeros([header['num_board_adc_channels'], num_board_adc_samples], dtype=np.uint)
# by default, this script interprets digital events (digital inputs and outputs) as booleans
# if unsigned int values are preferred(0 for False, 1 for True), replace the 'dtype=np.bool_' argument with 'dtype=np.uint' as shown
# the commented line below illustrates this for digital input data; the same can be done for digital out
#data['board_dig_in_data'] = np.zeros([header['num_board_dig_in_channels'], num_board_dig_in_samples], dtype=np.uint)
data['board_dig_in_data'] = np.zeros([header['num_board_dig_in_channels'], num_board_dig_in_samples], dtype=np.bool_)
data['board_dig_in_raw'] = np.zeros(num_board_dig_in_samples, dtype=np.uint)
data['board_dig_out_data'] = np.zeros([header['num_board_dig_out_channels'], num_board_dig_out_samples], dtype=np.bool_)
data['board_dig_out_raw'] = np.zeros(num_board_dig_out_samples, dtype=np.uint)
# Read sampled data from file.
print('Reading data from file...')
# Initialize indices used in looping
indices = {}
indices['amplifier'] = 0
indices['aux_input'] = 0
indices['supply_voltage'] = 0
indices['board_adc'] = 0
indices['board_dig_in'] = 0
indices['board_dig_out'] = 0
print_increment = 10
percent_done = print_increment
for i in range(num_data_blocks):
read_one_data_block(data, header, indices, fid)
# Increment indices
indices['amplifier'] += header['num_samples_per_data_block']
indices['aux_input'] += int(header['num_samples_per_data_block'] / 4)
indices['supply_voltage'] += 1
indices['board_adc'] += header['num_samples_per_data_block']
indices['board_dig_in'] += header['num_samples_per_data_block']
indices['board_dig_out'] += header['num_samples_per_data_block']
fraction_done = 100 * (1.0 * i / num_data_blocks)
if fraction_done >= percent_done:
print('{}% done...'.format(percent_done))
percent_done = percent_done + print_increment
# Make sure we have read exactly the right amount of data.
bytes_remaining = filesize - fid.tell()
if bytes_remaining != 0: raise Exception('Error: End of file not reached.')
# Close data file.
fid.close()
if (data_present):
print('Parsing data...')
# Extract digital input channels to separate variables.
for i in range(header['num_board_dig_in_channels']):
data['board_dig_in_data'][i, :] = np.not_equal(np.bitwise_and(data['board_dig_in_raw'], (1 << header['board_dig_in_channels'][i]['native_order'])), 0)
# Extract digital output channels to separate variables.
for i in range(header['num_board_dig_out_channels']):
data['board_dig_out_data'][i, :] = np.not_equal(np.bitwise_and(data['board_dig_out_raw'], (1 << header['board_dig_out_channels'][i]['native_order'])), 0)
# Scale voltage levels appropriately.
data['amplifier_data'] = np.multiply(0.195, (data['amplifier_data'].astype(np.int32) - 32768)) # units = microvolts
data['aux_input_data'] = np.multiply(37.4e-6, data['aux_input_data']) # units = volts
data['supply_voltage_data'] = np.multiply(74.8e-6, data['supply_voltage_data']) # units = volts
if header['eval_board_mode'] == 1:
data['board_adc_data'] = np.multiply(152.59e-6, (data['board_adc_data'].astype(np.int32) - 32768)) # units = volts
elif header['eval_board_mode'] == 13:
data['board_adc_data'] = np.multiply(312.5e-6, (data['board_adc_data'].astype(np.int32) - 32768)) # units = volts
else:
data['board_adc_data'] = np.multiply(50.354e-6, data['board_adc_data']) # units = volts
data['temp_sensor_data'] = np.multiply(0.01, data['temp_sensor_data']) # units = deg C
# Check for gaps in timestamps.
num_gaps = np.sum(np.not_equal(data['t_amplifier'][1:]-data['t_amplifier'][:-1], 1))
if num_gaps == 0:
print('No missing timestamps in data.')
else:
print('Warning: {0} gaps in timestamp data found. Time scale will not be uniform!'.format(num_gaps))
# Scale time steps (units = seconds).
data['t_amplifier'] = data['t_amplifier'] / header['sample_rate']
data['t_aux_input'] = data['t_amplifier'][range(0, len(data['t_amplifier']), 4)]
data['t_supply_voltage'] = data['t_amplifier'][range(0, len(data['t_amplifier']), header['num_samples_per_data_block'])]
data['t_board_adc'] = data['t_amplifier']
data['t_dig'] = data['t_amplifier']
data['t_temp_sensor'] = data['t_supply_voltage']
# If the software notch filter was selected during the recording, apply the
# same notch filter to amplifier data here.
if header['notch_filter_frequency'] > 0 and header['version']['major'] < 3:
print('Applying notch filter...')
print_increment = 10
percent_done = print_increment
for i in range(header['num_amplifier_channels']):
data['amplifier_data'][i,:] = notch_filter(data['amplifier_data'][i,:], header['sample_rate'], header['notch_filter_frequency'], 10)
fraction_done = 100 * (i / header['num_amplifier_channels'])
if fraction_done >= percent_done:
print('{}% done...'.format(percent_done))
percent_done += print_increment
else:
data = [];
# Move variables to result struct.
result = data_to_result(header, data, data_present)
print('Done! Elapsed time: {0:0.1f} seconds'.format(time.time() - tic))
return result
def | (n):
"""Utility function to optionally pluralize words based on the value of n.
"""
if n == 1:
return ''
else:
return 's'
path="//equipe2-nas1/Gilles.DELBECQ/Data/ePhy/Février2023/Test_Gustave/raw/raw intan/Test | plural | identifier_name |
RHD_Load_Filter.py | of data blocks')
num_data_blocks = int(bytes_remaining / bytes_per_block)
num_amplifier_samples = header['num_samples_per_data_block'] * num_data_blocks
num_aux_input_samples = int((header['num_samples_per_data_block'] / 4) * num_data_blocks)
num_supply_voltage_samples = 1 * num_data_blocks
num_board_adc_samples = header['num_samples_per_data_block'] * num_data_blocks
num_board_dig_in_samples = header['num_samples_per_data_block'] * num_data_blocks
num_board_dig_out_samples = header['num_samples_per_data_block'] * num_data_blocks
record_time = num_amplifier_samples / header['sample_rate']
if data_present:
print('File contains {:0.3f} seconds of data. Amplifiers were sampled at {:0.2f} kS/s.'.format(record_time, header['sample_rate'] / 1000))
else:
print('Header file contains no data. Amplifiers were sampled at {:0.2f} kS/s.'.format(header['sample_rate'] / 1000))
if data_present:
# Pre-allocate memory for data.
print('')
print('Allocating memory for data...')
data = {}
if (header['version']['major'] == 1 and header['version']['minor'] >= 2) or (header['version']['major'] > 1):
data['t_amplifier'] = np.zeros(num_amplifier_samples, dtype=np.int_)
else:
data['t_amplifier'] = np.zeros(num_amplifier_samples, dtype=np.uint)
data['amplifier_data'] = np.zeros([header['num_amplifier_channels'], num_amplifier_samples], dtype=np.uint)
data['aux_input_data'] = np.zeros([header['num_aux_input_channels'], num_aux_input_samples], dtype=np.uint)
data['supply_voltage_data'] = np.zeros([header['num_supply_voltage_channels'], num_supply_voltage_samples], dtype=np.uint)
data['temp_sensor_data'] = np.zeros([header['num_temp_sensor_channels'], num_supply_voltage_samples], dtype=np.uint)
data['board_adc_data'] = np.zeros([header['num_board_adc_channels'], num_board_adc_samples], dtype=np.uint)
# by default, this script interprets digital events (digital inputs and outputs) as booleans
# if unsigned int values are preferred(0 for False, 1 for True), replace the 'dtype=np.bool_' argument with 'dtype=np.uint' as shown
# the commented line below illustrates this for digital input data; the same can be done for digital out
#data['board_dig_in_data'] = np.zeros([header['num_board_dig_in_channels'], num_board_dig_in_samples], dtype=np.uint)
data['board_dig_in_data'] = np.zeros([header['num_board_dig_in_channels'], num_board_dig_in_samples], dtype=np.bool_)
data['board_dig_in_raw'] = np.zeros(num_board_dig_in_samples, dtype=np.uint)
data['board_dig_out_data'] = np.zeros([header['num_board_dig_out_channels'], num_board_dig_out_samples], dtype=np.bool_)
data['board_dig_out_raw'] = np.zeros(num_board_dig_out_samples, dtype=np.uint)
# Read sampled data from file.
print('Reading data from file...')
# Initialize indices used in looping
indices = {}
indices['amplifier'] = 0
indices['aux_input'] = 0
indices['supply_voltage'] = 0
indices['board_adc'] = 0
indices['board_dig_in'] = 0
indices['board_dig_out'] = 0
print_increment = 10
percent_done = print_increment
for i in range(num_data_blocks):
read_one_data_block(data, header, indices, fid)
# Increment indices
indices['amplifier'] += header['num_samples_per_data_block']
indices['aux_input'] += int(header['num_samples_per_data_block'] / 4)
indices['supply_voltage'] += 1
indices['board_adc'] += header['num_samples_per_data_block']
indices['board_dig_in'] += header['num_samples_per_data_block']
indices['board_dig_out'] += header['num_samples_per_data_block']
fraction_done = 100 * (1.0 * i / num_data_blocks)
if fraction_done >= percent_done:
print('{}% done...'.format(percent_done))
percent_done = percent_done + print_increment
# Make sure we have read exactly the right amount of data.
bytes_remaining = filesize - fid.tell()
if bytes_remaining != 0: raise Exception('Error: End of file not reached.')
# Close data file.
fid.close()
if (data_present):
print('Parsing data...')
# Extract digital input channels to separate variables.
for i in range(header['num_board_dig_in_channels']):
data['board_dig_in_data'][i, :] = np.not_equal(np.bitwise_and(data['board_dig_in_raw'], (1 << header['board_dig_in_channels'][i]['native_order'])), 0)
# Extract digital output channels to separate variables.
for i in range(header['num_board_dig_out_channels']):
data['board_dig_out_data'][i, :] = np.not_equal(np.bitwise_and(data['board_dig_out_raw'], (1 << header['board_dig_out_channels'][i]['native_order'])), 0)
# Scale voltage levels appropriately.
data['amplifier_data'] = np.multiply(0.195, (data['amplifier_data'].astype(np.int32) - 32768)) # units = microvolts
data['aux_input_data'] = np.multiply(37.4e-6, data['aux_input_data']) # units = volts
data['supply_voltage_data'] = np.multiply(74.8e-6, data['supply_voltage_data']) # units = volts
if header['eval_board_mode'] == 1:
data['board_adc_data'] = np.multiply(152.59e-6, (data['board_adc_data'].astype(np.int32) - 32768)) # units = volts
elif header['eval_board_mode'] == 13:
data['board_adc_data'] = np.multiply(312.5e-6, (data['board_adc_data'].astype(np.int32) - 32768)) # units = volts
else:
data['board_adc_data'] = np.multiply(50.354e-6, data['board_adc_data']) # units = volts
data['temp_sensor_data'] = np.multiply(0.01, data['temp_sensor_data']) # units = deg C
# Check for gaps in timestamps.
num_gaps = np.sum(np.not_equal(data['t_amplifier'][1:]-data['t_amplifier'][:-1], 1))
if num_gaps == 0:
print('No missing timestamps in data.')
else:
print('Warning: {0} gaps in timestamp data found. Time scale will not be uniform!'.format(num_gaps))
# Scale time steps (units = seconds).
data['t_amplifier'] = data['t_amplifier'] / header['sample_rate']
data['t_aux_input'] = data['t_amplifier'][range(0, len(data['t_amplifier']), 4)]
data['t_supply_voltage'] = data['t_amplifier'][range(0, len(data['t_amplifier']), header['num_samples_per_data_block'])]
data['t_board_adc'] = data['t_amplifier']
data['t_dig'] = data['t_amplifier']
data['t_temp_sensor'] = data['t_supply_voltage']
# If the software notch filter was selected during the recording, apply the
# same notch filter to amplifier data here.
if header['notch_filter_frequency'] > 0 and header['version']['major'] < 3:
print('Applying notch filter...')
print_increment = 10
percent_done = print_increment
for i in range(header['num_amplifier_channels']):
data['amplifier_data'][i,:] = notch_filter(data['amplifier_data'][i,:], header['sample_rate'], header['notch_filter_frequency'], 10)
fraction_done = 100 * (i / header['num_amplifier_channels'])
if fraction_done >= percent_done:
print('{}% done...'.format(percent_done))
percent_done += print_increment
else:
data = [];
# Move variables to result struct.
result = data_to_result(header, data, data_present)
print('Done! Elapsed time: {0:0.1f} seconds'.format(time.time() - tic))
return result
def plural(n):
|
path="//equipe2-nas1/Gilles.DELBECQ/Data/ePhy/Février2023/Test_Gustave/raw/raw intan/Test | """Utility function to optionally pluralize words based on the value of n.
"""
if n == 1:
return ''
else:
return 's' | identifier_body |
term_gui.rs | of the server, since the server
// might run on a different machine than the client - and certainly in a different
// directory.
let current_dir = env::current_dir().unwrap();
let rpc = try!(client.call("list_files", &swiboe::plugin::list_files::ListFilesRequest {
directory: current_dir.to_string_lossy().into_owned(),
}));
Ok(CompleterWidget {
candidates: subsequence_match::CandidateSet::new(),
rpc: Some(rpc),
query: "".into(),
results: Vec::new(),
selection_index: 0,
})
}
fn on_key(&mut self, key: rustbox::Key) -> CompleterState {
match key {
rustbox::Key::Char(c) => {
self.query.push(c);
self.results.clear();
CompleterState::Running
},
rustbox::Key::Backspace => {
self.query.pop();
self.results.clear();
CompleterState::Running
},
rustbox::Key::Down => {
self.selection_index += 1;
CompleterState::Running
},
rustbox::Key::Up => {
self.selection_index -= 1;
CompleterState::Running
},
rustbox::Key::Esc => {
self.rpc.take().unwrap().cancel().unwrap();
CompleterState::Canceled
},
rustbox::Key::Enter => {
self.rpc.take().unwrap().cancel().unwrap();
if self.results.is_empty() {
CompleterState::Canceled
} else {
clamp(0, self.results.len() as isize - 1, &mut self.selection_index);
CompleterState::Selected(self.results[self.selection_index as usize].text.clone())
}
}
_ => CompleterState::Running,
}
}
fn draw(&mut self, rustbox: &rustbox::RustBox) {
while let Some(b) = self.rpc.as_mut().unwrap().try_recv().unwrap() {
self.results.clear();
let b: swiboe::plugin::list_files::ListFilesUpdate = serde_json::from_value(b).unwrap();
for file in &b.files {
self.candidates.insert(file);
}
}
if self.results.is_empty() {
let query_to_use: String = self.query.chars().filter(|c| !c.is_whitespace()).collect();
self.candidates.query(&query_to_use, subsequence_match::MatchCase::No, &mut self.results);
}
if !self.results.is_empty() {
clamp(0, self.results.len() as isize - 1, &mut self.selection_index);
}
rustbox.print(0, 0, rustbox::RB_BOLD, Color::Yellow, Color::Default, &self.query);
let len_string = format!("{}/{} matching ({})", self.results.len(), self.candidates.len(),
if self.rpc.as_ref().unwrap().done() { "done" } else { "scanning" } );
rustbox.print(rustbox.width() - len_string.len() - 1, 0, rustbox::RB_BOLD, Color::Blue, Color::Default, &len_string);
let mut row = 1usize;
for result in &self.results {
let mut matching_indices = result.matching_indices.iter().peekable();
for (col, c) in result.text.chars().enumerate() {
let matches = match matching_indices.peek() {
Some(val) if **val == col => true,
_ => false,
};
let mut style = if matches {
matching_indices.next();
rustbox::RB_BOLD
} else {
rustbox::RB_NORMAL
};
if row as isize == self.selection_index + 1 {
style = style | rustbox::RB_REVERSE;
}
rustbox.print_char(col, row, style, Color::Default, Color::Default, c);
}
row += 1;
if row > rustbox.height() {
break;
}
}
}
}
struct BufferViewWidget {
view_id: String,
client: client::ThinClient,
cursor_id: String,
}
impl BufferViewWidget {
pub fn new(view_id: String, client: client::ThinClient) -> Self {
BufferViewWidget {
view_id: view_id,
client: client, |
fn draw(&mut self, buffer_view: &buffer_views::BufferView, rustbox: &rustbox::RustBox) {
let mut row = 0;
let top_line_index = buffer_view.top_line_index as usize;
self.cursor_id = buffer_view.cursor.id().to_string();
let mut cursor_drawn = false;
while row < rustbox.height() {
let line_index = top_line_index + row;
if let Some(line) = buffer_view.lines.get(line_index) {
for (col, c) in line.chars().enumerate() {
if col >= rustbox.width() {
break;
}
let bg = if buffer_view.cursor.position.line_index == line_index as isize &&
buffer_view.cursor.position.column_index as usize == col {
cursor_drawn = true;
Color::Red
} else {
Color::Default
};
rustbox.print_char(col, row, rustbox::RB_NORMAL, Color::Default, bg, c);
}
}
row += 1;
}
if !cursor_drawn {
let row = buffer_view.cursor.position.line_index - top_line_index as isize;
rustbox.print_char(buffer_view.cursor.position.column_index as usize,
row as usize, rustbox::RB_NORMAL,
Color::Default, Color::Red, ' ');
}
}
}
#[derive(Debug)]
struct Options {
socket: String,
config_file: path::PathBuf,
}
struct TerminalGui {
config_file_runner: Box<gui::config_file::ConfigFileRunner>,
client: client::Client,
rustbox: rustbox::RustBox,
buffer_views: Arc<RwLock<gui::buffer_views::BufferViews>>,
last_key_down_event: time::PreciseTime,
completer: Option<CompleterWidget>,
buffer_view_widget: Option<BufferViewWidget>,
// NOCOM(#sirver): GuiCommand in namespace gui is very duplicated
gui_commands: mpsc::Receiver<gui::command::GuiCommand>,
}
impl TerminalGui {
fn new(options: &Options) -> swiboe::Result<Self> {
let mut client = match net::SocketAddr::from_str(&options.socket) {
Ok(value) => {
client::Client::connect_tcp(&value).unwrap()
}
Err(_) => {
let socket_path = path::PathBuf::from(&options.socket);
client::Client::connect_unix(&socket_path).unwrap()
}
};
let mut config_file_runner = gui::config_file::ConfigFileRunner::new(
try!(client.clone()));
config_file_runner.run(&options.config_file);
let rustbox = match RustBox::init(rustbox::InitOptions {
input_mode: rustbox::InputMode::Current,
buffer_stderr: true,
}) {
Result::Ok(v) => v,
Result::Err(e) => panic!("{}", e),
};
let gui_id: String = Uuid::new_v4().to_hyphenated_string();
let (gui_commands_tx, gui_commands_rx) = mpsc::channel();
let buffer_views = try!(gui::buffer_views::BufferViews::new(&gui_id, gui_commands_tx, &mut client));
Ok(TerminalGui {
config_file_runner: config_file_runner,
client: client,
rustbox: rustbox,
buffer_views: buffer_views,
last_key_down_event: time::PreciseTime::now(),
completer: None,
buffer_view_widget: None,
gui_commands: gui_commands_rx,
})
}
fn handle_events(&mut self) -> swiboe::Result<bool> {
match self.rustbox.peek_event(std::time::Duration::from_millis(5), false) {
Ok(rustbox::Event::KeyEvent(key)) => {
if self.completer.is_some() {
let rv = self.completer.as_mut().unwrap().on_key(key);
match rv {
CompleterState::Running => (),
CompleterState::Canceled => {
self.completer = None;
},
CompleterState::Selected(result) => {
self.completer = None;
let mut rpc = try!(self.client.call("buffer.open", &swiboe::plugin::buffer::open::Request {
uri: format!("file://{}", result),
}));
let response: swiboe::plugin::buffer::open::Response = rpc.wait_for().unwrap();
let mut buffer_views = self.buffer_views.write().unwrap();
let view_id = buffer_views.new_view(response.buffer_index, self.rustbox.width(), self.rustbox.height());
self.buffer_view_widget = Some(BufferViewWidget::new(view_id, try!(self.client.clone())));
},
}
} else {
if !try!(self.handle_key(key)) {
return Ok(false);
}
}
},
Err(e) => panic!("{}", e),
_ => { }
}
while let Ok(command) = self.gui_commands.try_recv() {
match command {
gui::command::GuiCommand::Quit => return Ok(false),
gui::command::GuiCommand::Redraw => (),
| cursor_id: String::new(),
}
} | random_line_split |
term_gui.rs | {
self.rpc.take().unwrap().cancel().unwrap();
if self.results.is_empty() {
CompleterState::Canceled
} else {
clamp(0, self.results.len() as isize - 1, &mut self.selection_index);
CompleterState::Selected(self.results[self.selection_index as usize].text.clone())
}
}
_ => CompleterState::Running,
}
}
fn draw(&mut self, rustbox: &rustbox::RustBox) {
while let Some(b) = self.rpc.as_mut().unwrap().try_recv().unwrap() {
self.results.clear();
let b: swiboe::plugin::list_files::ListFilesUpdate = serde_json::from_value(b).unwrap();
for file in &b.files {
self.candidates.insert(file);
}
}
if self.results.is_empty() {
let query_to_use: String = self.query.chars().filter(|c| !c.is_whitespace()).collect();
self.candidates.query(&query_to_use, subsequence_match::MatchCase::No, &mut self.results);
}
if !self.results.is_empty() {
clamp(0, self.results.len() as isize - 1, &mut self.selection_index);
}
rustbox.print(0, 0, rustbox::RB_BOLD, Color::Yellow, Color::Default, &self.query);
let len_string = format!("{}/{} matching ({})", self.results.len(), self.candidates.len(),
if self.rpc.as_ref().unwrap().done() { "done" } else { "scanning" } );
rustbox.print(rustbox.width() - len_string.len() - 1, 0, rustbox::RB_BOLD, Color::Blue, Color::Default, &len_string);
let mut row = 1usize;
for result in &self.results {
let mut matching_indices = result.matching_indices.iter().peekable();
for (col, c) in result.text.chars().enumerate() {
let matches = match matching_indices.peek() {
Some(val) if **val == col => true,
_ => false,
};
let mut style = if matches {
matching_indices.next();
rustbox::RB_BOLD
} else {
rustbox::RB_NORMAL
};
if row as isize == self.selection_index + 1 {
style = style | rustbox::RB_REVERSE;
}
rustbox.print_char(col, row, style, Color::Default, Color::Default, c);
}
row += 1;
if row > rustbox.height() {
break;
}
}
}
}
struct BufferViewWidget {
view_id: String,
client: client::ThinClient,
cursor_id: String,
}
impl BufferViewWidget {
pub fn new(view_id: String, client: client::ThinClient) -> Self {
BufferViewWidget {
view_id: view_id,
client: client,
cursor_id: String::new(),
}
}
fn draw(&mut self, buffer_view: &buffer_views::BufferView, rustbox: &rustbox::RustBox) {
let mut row = 0;
let top_line_index = buffer_view.top_line_index as usize;
self.cursor_id = buffer_view.cursor.id().to_string();
let mut cursor_drawn = false;
while row < rustbox.height() {
let line_index = top_line_index + row;
if let Some(line) = buffer_view.lines.get(line_index) {
for (col, c) in line.chars().enumerate() {
if col >= rustbox.width() {
break;
}
let bg = if buffer_view.cursor.position.line_index == line_index as isize &&
buffer_view.cursor.position.column_index as usize == col {
cursor_drawn = true;
Color::Red
} else {
Color::Default
};
rustbox.print_char(col, row, rustbox::RB_NORMAL, Color::Default, bg, c);
}
}
row += 1;
}
if !cursor_drawn {
let row = buffer_view.cursor.position.line_index - top_line_index as isize;
rustbox.print_char(buffer_view.cursor.position.column_index as usize,
row as usize, rustbox::RB_NORMAL,
Color::Default, Color::Red, ' ');
}
}
}
#[derive(Debug)]
struct Options {
socket: String,
config_file: path::PathBuf,
}
struct TerminalGui {
config_file_runner: Box<gui::config_file::ConfigFileRunner>,
client: client::Client,
rustbox: rustbox::RustBox,
buffer_views: Arc<RwLock<gui::buffer_views::BufferViews>>,
last_key_down_event: time::PreciseTime,
completer: Option<CompleterWidget>,
buffer_view_widget: Option<BufferViewWidget>,
// NOCOM(#sirver): GuiCommand in namespace gui is very duplicated
gui_commands: mpsc::Receiver<gui::command::GuiCommand>,
}
impl TerminalGui {
fn new(options: &Options) -> swiboe::Result<Self> {
let mut client = match net::SocketAddr::from_str(&options.socket) {
Ok(value) => {
client::Client::connect_tcp(&value).unwrap()
}
Err(_) => {
let socket_path = path::PathBuf::from(&options.socket);
client::Client::connect_unix(&socket_path).unwrap()
}
};
let mut config_file_runner = gui::config_file::ConfigFileRunner::new(
try!(client.clone()));
config_file_runner.run(&options.config_file);
let rustbox = match RustBox::init(rustbox::InitOptions {
input_mode: rustbox::InputMode::Current,
buffer_stderr: true,
}) {
Result::Ok(v) => v,
Result::Err(e) => panic!("{}", e),
};
let gui_id: String = Uuid::new_v4().to_hyphenated_string();
let (gui_commands_tx, gui_commands_rx) = mpsc::channel();
let buffer_views = try!(gui::buffer_views::BufferViews::new(&gui_id, gui_commands_tx, &mut client));
Ok(TerminalGui {
config_file_runner: config_file_runner,
client: client,
rustbox: rustbox,
buffer_views: buffer_views,
last_key_down_event: time::PreciseTime::now(),
completer: None,
buffer_view_widget: None,
gui_commands: gui_commands_rx,
})
}
fn handle_events(&mut self) -> swiboe::Result<bool> {
match self.rustbox.peek_event(std::time::Duration::from_millis(5), false) {
Ok(rustbox::Event::KeyEvent(key)) => {
if self.completer.is_some() {
let rv = self.completer.as_mut().unwrap().on_key(key);
match rv {
CompleterState::Running => (),
CompleterState::Canceled => {
self.completer = None;
},
CompleterState::Selected(result) => {
self.completer = None;
let mut rpc = try!(self.client.call("buffer.open", &swiboe::plugin::buffer::open::Request {
uri: format!("file://{}", result),
}));
let response: swiboe::plugin::buffer::open::Response = rpc.wait_for().unwrap();
let mut buffer_views = self.buffer_views.write().unwrap();
let view_id = buffer_views.new_view(response.buffer_index, self.rustbox.width(), self.rustbox.height());
self.buffer_view_widget = Some(BufferViewWidget::new(view_id, try!(self.client.clone())));
},
}
} else {
if !try!(self.handle_key(key)) {
return Ok(false);
}
}
},
Err(e) => panic!("{}", e),
_ => { }
}
while let Ok(command) = self.gui_commands.try_recv() {
match command {
gui::command::GuiCommand::Quit => return Ok(false),
gui::command::GuiCommand::Redraw => (),
}
}
return Ok(true);
}
fn handle_key(&mut self, key: rustbox::Key) -> swiboe::Result<bool> {
let delta_t = {
let now = time::PreciseTime::now();
let delta_t = self.last_key_down_event.to(now);
self.last_key_down_event = now;
delta_t
};
let delta_t_in_seconds = delta_t.num_nanoseconds().unwrap() as f64 / 1e9;
match key {
// NOCOM(#sirver): should be handled through plugins.
rustbox::Key::Char('q') => return Ok(false),
rustbox::Key::Ctrl('t') => {
self.completer = Some(try!(CompleterWidget::new(&mut self.client)))
},
rustbox::Key::Esc => {
self.config_file_runner.keymap_handler.timeout();
},
rustbox::Key::Char(a) => {
self.config_file_runner.keymap_handler.key_down(
delta_t_in_seconds, keymap_handler::Key::Char(a));
},
rustbox::Key::Up => {
self.config_file_runner.keymap_handler.key_down(
delta_t_in_seconds, keymap_handler::Key::Up);
},
rustbox::Key::Down => | {
self.config_file_runner.keymap_handler.key_down(
delta_t_in_seconds, keymap_handler::Key::Down);
} | conditional_block |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.