file_name
large_stringlengths 4
140
| prefix
large_stringlengths 0
12.1k
| suffix
large_stringlengths 0
12k
| middle
large_stringlengths 0
7.51k
| fim_type
large_stringclasses 4
values |
---|---|---|---|---|
feature_extract.py | numtaps = 255
fil = firwin(numtaps, norm_cutoff)
x_pad = np.pad(x, (numtaps, numtaps), 'edge')
lpf_x = lfilter(fil, 1, x_pad)
lpf_x = lpf_x[numtaps + numtaps // 2: -numtaps // 2]
return lpf_x
def convert_to_continuos_f0(f0):
"""CONVERT F0 TO CONTINUOUS F0.
Args:
f0 (ndarray): original f0 sequence with the shape (T,).
Returns:
ndarray: continuous f0 with the shape (T,).
"""
# get uv information as binary
uv = np.float32(f0 != 0)
# get start and end of f0
if (f0 == 0).all():
logging.warning("all of the f0 values are 0.")
return uv, f0
start_f0 = f0[f0 != 0][0]
end_f0 = f0[f0 != 0][-1]
# padding start and end of f0 sequence
start_idx = np.where(f0 == start_f0)[0][0]
end_idx = np.where(f0 == end_f0)[0][-1]
f0[:start_idx] = start_f0
f0[end_idx:] = end_f0
|
# perform linear interpolation
f = interp1d(nz_frames, f0[nz_frames])
cont_f0 = f(np.arange(0, f0.shape[0]))
return uv, cont_f0
def stft_mcep(x, fftl=512, shiftl=256, dim=25, alpha=0.41, window="hamming", is_padding=False):
"""EXTRACT STFT-BASED MEL-CEPSTRUM.
Args:
x (ndarray): Numpy double array with the size (T,).
fftl (int): FFT length in point (default=512).
shiftl (int): Shift length in point (default=256).
dim (int): Dimension of mel-cepstrum (default=25).
alpha (float): All pass filter coefficient (default=0.41).
window (str): Analysis window type (default="hamming").
is_padding (bool): Whether to pad the end of signal (default=False).
Returns:
ndarray: Mel-cepstrum with the size (N, n_fft).
"""
# perform padding
if is_padding:
n_pad = fftl - (len(x) - fftl) % shiftl
x = np.pad(x, (0, n_pad), 'reflect')
# get number of frames
n_frame = (len(x) - fftl) // shiftl + 1
# get window function
win = get_window(window, fftl)
# calculate spectrogram
mcep = [pysptk.mcep(x[shiftl * i: shiftl * i + fftl] * win,
dim, alpha, eps=EPS, etype=1)
for i in range(n_frame)]
return np.stack(mcep)
def world_feature_extract(wav_list, args):
"""EXTRACT WORLD FEATURE VECTOR."""
# define feature extractor
feature_extractor = FeatureExtractor(
analyzer="world",
fs=args.fs,
shiftms=args.shiftms,
minf0=args.minf0,
maxf0=args.maxf0,
fftl=args.fftl)
for i, wav_name in enumerate(wav_list):
logging.info("now processing %s (%d/%d)" % (wav_name, i + 1, len(wav_list)))
# load wavfile and apply low cut filter
fs, x = wavfile.read(wav_name)
if x.dtype != np.int16:
logging.warning("wav file format is not 16 bit PCM.")
x = np.array(x, dtype=np.float64)
if args.highpass_cutoff != 0:
x = low_cut_filter(x, fs, cutoff=args.highpass_cutoff)
# check sampling frequency
if not fs == args.fs:
logging.error("sampling frequency is not matched.")
sys.exit(1)
# extract features
f0, _, _ = feature_extractor.analyze(x)
uv, cont_f0 = convert_to_continuos_f0(f0)
cont_f0_lpf = low_pass_filter(cont_f0, int(1.0 / (args.shiftms * 0.001)), cutoff=20)
codeap = feature_extractor.codeap()
mcep = feature_extractor.mcep(dim=args.mcep_dim, alpha=args.mcep_alpha)
# concatenate
cont_f0_lpf = np.expand_dims(cont_f0_lpf, axis=-1)
uv = np.expand_dims(uv, axis=-1)
feats = np.concatenate([uv, cont_f0_lpf, mcep, codeap], axis=1)
# save to hdf5
hdf5name = args.hdf5dir + "/" + os.path.basename(wav_name).replace(".wav", ".h5")
write_hdf5(hdf5name, "/world", feats)
# overwrite wav file
if args.highpass_cutoff != 0 and args.save_wav:
wavfile.write(args.wavdir + "/" + os.path.basename(wav_name), fs, np.int16(x))
def melspectrogram_extract(wav_list, args):
"""EXTRACT MEL SPECTROGRAM."""
# define feature extractor
for i, wav_name in enumerate(wav_list):
logging.info("now processing %s (%d/%d)" % (wav_name, i + 1, len(wav_list)))
# load wavfile and apply low cut filter
fs, x = wavfile.read(wav_name)
if x.dtype != np.int16:
logging.warning("wav file format is not 16 bit PCM.")
x = np.array(x, dtype=np.float64)
if args.highpass_cutoff != 0:
x = low_cut_filter(x, fs, cutoff=args.highpass_cutoff)
# check sampling frequency
if not fs == args.fs:
logging.error("sampling frequency is not matched.")
sys.exit(1)
# extract features
x_norm = x / (np.iinfo(np.int16).max + 1)
shiftl = int(args.shiftms * fs * 0.001)
mspc = librosa.feature.melspectrogram(
x_norm, fs,
n_fft=args.fftl,
hop_length=shiftl,
n_mels=args.mspc_dim,
fmin=args.fmin if args.fmin is not None else 0,
fmax=args.fmax if args.fmax is not None else fs // 2,
power=1.0)
mspc = np.log10(np.maximum(EPS, mspc.T))
# save to hdf5
hdf5name = args.hdf5dir + "/" + os.path.basename(wav_name).replace(".wav", ".h5")
write_hdf5(hdf5name, "/melspc", np.float32(mspc))
# overwrite wav file
if args.highpass_cutoff != 0 and args.save_wav:
wavfile.write(args.wavdir + "/" + os.path.basename(wav_name), fs, np.int16(x))
def melcepstrum_extract(wav_list, args):
"""EXTRACT MEL CEPSTRUM."""
# define feature extractor
for i, wav_name in enumerate(wav_list):
logging.info("now processing %s (%d/%d)" % (wav_name, i + 1, len(wav_list)))
# load wavfile and apply low cut filter
fs, x = wavfile.read(wav_name)
if x.dtype != np.int16:
logging.warning("wav file format is not 16 bit PCM.")
x = np.array(x, dtype=np.float64)
if args.highpass_cutoff != 0:
x = low_cut_filter(x, fs, cutoff=args.highpass_cutoff)
# check sampling frequency
if not fs == args.fs:
logging.error("sampling frequency is not matched.")
sys.exit(1)
# extract features
shiftl = int(args.shiftms * fs * 0.001)
mcep = stft_mcep(x, args.fftl, shiftl, args.mcep_dim, args.mcep_alpha)
# save to hdf5
hdf5name = args.hdf5dir + "/" + os.path.basename(wav_name).replace(".wav", ".h5")
write_hdf5(hdf5name, "/mcep", np.float32(mcep))
# overwrite wav file
if args.highpass_cutoff != 0 and args.save_wav:
wavfile.write(args.wavdir + "/" + os.path.basename(wav_name), fs, np.int16(x))
def main():
"""RUN FEATURE EXTRACTION IN PARALLEL."""
parser = argparse.ArgumentParser(
description="making feature file argsurations.")
parser.add_argument(
"--waveforms", default=None,
help="directory or list of filename of input wav | # get non-zero frame index
nz_frames = np.where(f0 != 0)[0] | random_line_split |
feature_extract.py | numtaps = 255
fil = firwin(numtaps, norm_cutoff)
x_pad = np.pad(x, (numtaps, numtaps), 'edge')
lpf_x = lfilter(fil, 1, x_pad)
lpf_x = lpf_x[numtaps + numtaps // 2: -numtaps // 2]
return lpf_x
def convert_to_continuos_f0(f0):
"""CONVERT F0 TO CONTINUOUS F0.
Args:
f0 (ndarray): original f0 sequence with the shape (T,).
Returns:
ndarray: continuous f0 with the shape (T,).
"""
# get uv information as binary
uv = np.float32(f0 != 0)
# get start and end of f0
if (f0 == 0).all():
logging.warning("all of the f0 values are 0.")
return uv, f0
start_f0 = f0[f0 != 0][0]
end_f0 = f0[f0 != 0][-1]
# padding start and end of f0 sequence
start_idx = np.where(f0 == start_f0)[0][0]
end_idx = np.where(f0 == end_f0)[0][-1]
f0[:start_idx] = start_f0
f0[end_idx:] = end_f0
# get non-zero frame index
nz_frames = np.where(f0 != 0)[0]
# perform linear interpolation
f = interp1d(nz_frames, f0[nz_frames])
cont_f0 = f(np.arange(0, f0.shape[0]))
return uv, cont_f0
def | (x, fftl=512, shiftl=256, dim=25, alpha=0.41, window="hamming", is_padding=False):
"""EXTRACT STFT-BASED MEL-CEPSTRUM.
Args:
x (ndarray): Numpy double array with the size (T,).
fftl (int): FFT length in point (default=512).
shiftl (int): Shift length in point (default=256).
dim (int): Dimension of mel-cepstrum (default=25).
alpha (float): All pass filter coefficient (default=0.41).
window (str): Analysis window type (default="hamming").
is_padding (bool): Whether to pad the end of signal (default=False).
Returns:
ndarray: Mel-cepstrum with the size (N, n_fft).
"""
# perform padding
if is_padding:
n_pad = fftl - (len(x) - fftl) % shiftl
x = np.pad(x, (0, n_pad), 'reflect')
# get number of frames
n_frame = (len(x) - fftl) // shiftl + 1
# get window function
win = get_window(window, fftl)
# calculate spectrogram
mcep = [pysptk.mcep(x[shiftl * i: shiftl * i + fftl] * win,
dim, alpha, eps=EPS, etype=1)
for i in range(n_frame)]
return np.stack(mcep)
def world_feature_extract(wav_list, args):
"""EXTRACT WORLD FEATURE VECTOR."""
# define feature extractor
feature_extractor = FeatureExtractor(
analyzer="world",
fs=args.fs,
shiftms=args.shiftms,
minf0=args.minf0,
maxf0=args.maxf0,
fftl=args.fftl)
for i, wav_name in enumerate(wav_list):
logging.info("now processing %s (%d/%d)" % (wav_name, i + 1, len(wav_list)))
# load wavfile and apply low cut filter
fs, x = wavfile.read(wav_name)
if x.dtype != np.int16:
logging.warning("wav file format is not 16 bit PCM.")
x = np.array(x, dtype=np.float64)
if args.highpass_cutoff != 0:
x = low_cut_filter(x, fs, cutoff=args.highpass_cutoff)
# check sampling frequency
if not fs == args.fs:
logging.error("sampling frequency is not matched.")
sys.exit(1)
# extract features
f0, _, _ = feature_extractor.analyze(x)
uv, cont_f0 = convert_to_continuos_f0(f0)
cont_f0_lpf = low_pass_filter(cont_f0, int(1.0 / (args.shiftms * 0.001)), cutoff=20)
codeap = feature_extractor.codeap()
mcep = feature_extractor.mcep(dim=args.mcep_dim, alpha=args.mcep_alpha)
# concatenate
cont_f0_lpf = np.expand_dims(cont_f0_lpf, axis=-1)
uv = np.expand_dims(uv, axis=-1)
feats = np.concatenate([uv, cont_f0_lpf, mcep, codeap], axis=1)
# save to hdf5
hdf5name = args.hdf5dir + "/" + os.path.basename(wav_name).replace(".wav", ".h5")
write_hdf5(hdf5name, "/world", feats)
# overwrite wav file
if args.highpass_cutoff != 0 and args.save_wav:
wavfile.write(args.wavdir + "/" + os.path.basename(wav_name), fs, np.int16(x))
def melspectrogram_extract(wav_list, args):
"""EXTRACT MEL SPECTROGRAM."""
# define feature extractor
for i, wav_name in enumerate(wav_list):
logging.info("now processing %s (%d/%d)" % (wav_name, i + 1, len(wav_list)))
# load wavfile and apply low cut filter
fs, x = wavfile.read(wav_name)
if x.dtype != np.int16:
logging.warning("wav file format is not 16 bit PCM.")
x = np.array(x, dtype=np.float64)
if args.highpass_cutoff != 0:
x = low_cut_filter(x, fs, cutoff=args.highpass_cutoff)
# check sampling frequency
if not fs == args.fs:
logging.error("sampling frequency is not matched.")
sys.exit(1)
# extract features
x_norm = x / (np.iinfo(np.int16).max + 1)
shiftl = int(args.shiftms * fs * 0.001)
mspc = librosa.feature.melspectrogram(
x_norm, fs,
n_fft=args.fftl,
hop_length=shiftl,
n_mels=args.mspc_dim,
fmin=args.fmin if args.fmin is not None else 0,
fmax=args.fmax if args.fmax is not None else fs // 2,
power=1.0)
mspc = np.log10(np.maximum(EPS, mspc.T))
# save to hdf5
hdf5name = args.hdf5dir + "/" + os.path.basename(wav_name).replace(".wav", ".h5")
write_hdf5(hdf5name, "/melspc", np.float32(mspc))
# overwrite wav file
if args.highpass_cutoff != 0 and args.save_wav:
wavfile.write(args.wavdir + "/" + os.path.basename(wav_name), fs, np.int16(x))
def melcepstrum_extract(wav_list, args):
"""EXTRACT MEL CEPSTRUM."""
# define feature extractor
for i, wav_name in enumerate(wav_list):
logging.info("now processing %s (%d/%d)" % (wav_name, i + 1, len(wav_list)))
# load wavfile and apply low cut filter
fs, x = wavfile.read(wav_name)
if x.dtype != np.int16:
logging.warning("wav file format is not 16 bit PCM.")
x = np.array(x, dtype=np.float64)
if args.highpass_cutoff != 0:
x = low_cut_filter(x, fs, cutoff=args.highpass_cutoff)
# check sampling frequency
if not fs == args.fs:
logging.error("sampling frequency is not matched.")
sys.exit(1)
# extract features
shiftl = int(args.shiftms * fs * 0.001)
mcep = stft_mcep(x, args.fftl, shiftl, args.mcep_dim, args.mcep_alpha)
# save to hdf5
hdf5name = args.hdf5dir + "/" + os.path.basename(wav_name).replace(".wav", ".h5")
write_hdf5(hdf5name, "/mcep", np.float32(mcep))
# overwrite wav file
if args.highpass_cutoff != 0 and args.save_wav:
wavfile.write(args.wavdir + "/" + os.path.basename(wav_name), fs, np.int16(x))
def main():
"""RUN FEATURE EXTRACTION IN PARALLEL."""
parser = argparse.ArgumentParser(
description="making feature file argsurations.")
parser.add_argument(
"--waveforms", default=None,
help="directory or list of filename of input wav | stft_mcep | identifier_name |
feature_extract.py | numtaps = 255
fil = firwin(numtaps, norm_cutoff)
x_pad = np.pad(x, (numtaps, numtaps), 'edge')
lpf_x = lfilter(fil, 1, x_pad)
lpf_x = lpf_x[numtaps + numtaps // 2: -numtaps // 2]
return lpf_x
def convert_to_continuos_f0(f0):
"""CONVERT F0 TO CONTINUOUS F0.
Args:
f0 (ndarray): original f0 sequence with the shape (T,).
Returns:
ndarray: continuous f0 with the shape (T,).
"""
# get uv information as binary
uv = np.float32(f0 != 0)
# get start and end of f0
if (f0 == 0).all():
logging.warning("all of the f0 values are 0.")
return uv, f0
start_f0 = f0[f0 != 0][0]
end_f0 = f0[f0 != 0][-1]
# padding start and end of f0 sequence
start_idx = np.where(f0 == start_f0)[0][0]
end_idx = np.where(f0 == end_f0)[0][-1]
f0[:start_idx] = start_f0
f0[end_idx:] = end_f0
# get non-zero frame index
nz_frames = np.where(f0 != 0)[0]
# perform linear interpolation
f = interp1d(nz_frames, f0[nz_frames])
cont_f0 = f(np.arange(0, f0.shape[0]))
return uv, cont_f0
def stft_mcep(x, fftl=512, shiftl=256, dim=25, alpha=0.41, window="hamming", is_padding=False):
"""EXTRACT STFT-BASED MEL-CEPSTRUM.
Args:
x (ndarray): Numpy double array with the size (T,).
fftl (int): FFT length in point (default=512).
shiftl (int): Shift length in point (default=256).
dim (int): Dimension of mel-cepstrum (default=25).
alpha (float): All pass filter coefficient (default=0.41).
window (str): Analysis window type (default="hamming").
is_padding (bool): Whether to pad the end of signal (default=False).
Returns:
ndarray: Mel-cepstrum with the size (N, n_fft).
"""
# perform padding
if is_padding:
n_pad = fftl - (len(x) - fftl) % shiftl
x = np.pad(x, (0, n_pad), 'reflect')
# get number of frames
n_frame = (len(x) - fftl) // shiftl + 1
# get window function
win = get_window(window, fftl)
# calculate spectrogram
mcep = [pysptk.mcep(x[shiftl * i: shiftl * i + fftl] * win,
dim, alpha, eps=EPS, etype=1)
for i in range(n_frame)]
return np.stack(mcep)
def world_feature_extract(wav_list, args):
"""EXTRACT WORLD FEATURE VECTOR."""
# define feature extractor
feature_extractor = FeatureExtractor(
analyzer="world",
fs=args.fs,
shiftms=args.shiftms,
minf0=args.minf0,
maxf0=args.maxf0,
fftl=args.fftl)
for i, wav_name in enumerate(wav_list):
logging.info("now processing %s (%d/%d)" % (wav_name, i + 1, len(wav_list)))
# load wavfile and apply low cut filter
fs, x = wavfile.read(wav_name)
if x.dtype != np.int16:
logging.warning("wav file format is not 16 bit PCM.")
x = np.array(x, dtype=np.float64)
if args.highpass_cutoff != 0:
x = low_cut_filter(x, fs, cutoff=args.highpass_cutoff)
# check sampling frequency
if not fs == args.fs:
logging.error("sampling frequency is not matched.")
sys.exit(1)
# extract features
f0, _, _ = feature_extractor.analyze(x)
uv, cont_f0 = convert_to_continuos_f0(f0)
cont_f0_lpf = low_pass_filter(cont_f0, int(1.0 / (args.shiftms * 0.001)), cutoff=20)
codeap = feature_extractor.codeap()
mcep = feature_extractor.mcep(dim=args.mcep_dim, alpha=args.mcep_alpha)
# concatenate
cont_f0_lpf = np.expand_dims(cont_f0_lpf, axis=-1)
uv = np.expand_dims(uv, axis=-1)
feats = np.concatenate([uv, cont_f0_lpf, mcep, codeap], axis=1)
# save to hdf5
hdf5name = args.hdf5dir + "/" + os.path.basename(wav_name).replace(".wav", ".h5")
write_hdf5(hdf5name, "/world", feats)
# overwrite wav file
if args.highpass_cutoff != 0 and args.save_wav:
wavfile.write(args.wavdir + "/" + os.path.basename(wav_name), fs, np.int16(x))
def melspectrogram_extract(wav_list, args):
| shiftl = int(args.shiftms * fs * 0.001)
mspc = librosa.feature.melspectrogram(
x_norm, fs,
n_fft=args.fftl,
hop_length=shiftl,
n_mels=args.mspc_dim,
fmin=args.fmin if args.fmin is not None else 0,
fmax=args.fmax if args.fmax is not None else fs // 2,
power=1.0)
mspc = np.log10(np.maximum(EPS, mspc.T))
# save to hdf5
hdf5name = args.hdf5dir + "/" + os.path.basename(wav_name).replace(".wav", ".h5")
write_hdf5(hdf5name, "/melspc", np.float32(mspc))
# overwrite wav file
if args.highpass_cutoff != 0 and args.save_wav:
wavfile.write(args.wavdir + "/" + os.path.basename(wav_name), fs, np.int16(x))
def melcepstrum_extract(wav_list, args):
"""EXTRACT MEL CEPSTRUM."""
# define feature extractor
for i, wav_name in enumerate(wav_list):
logging.info("now processing %s (%d/%d)" % (wav_name, i + 1, len(wav_list)))
# load wavfile and apply low cut filter
fs, x = wavfile.read(wav_name)
if x.dtype != np.int16:
logging.warning("wav file format is not 16 bit PCM.")
x = np.array(x, dtype=np.float64)
if args.highpass_cutoff != 0:
x = low_cut_filter(x, fs, cutoff=args.highpass_cutoff)
# check sampling frequency
if not fs == args.fs:
logging.error("sampling frequency is not matched.")
sys.exit(1)
# extract features
shiftl = int(args.shiftms * fs * 0.001)
mcep = stft_mcep(x, args.fftl, shiftl, args.mcep_dim, args.mcep_alpha)
# save to hdf5
hdf5name = args.hdf5dir + "/" + os.path.basename(wav_name).replace(".wav", ".h5")
write_hdf5(hdf5name, "/mcep", np.float32(mcep))
# overwrite wav file
if args.highpass_cutoff != 0 and args.save_wav:
wavfile.write(args.wavdir + "/" + os.path.basename(wav_name), fs, np.int16(x))
def main():
"""RUN FEATURE EXTRACTION IN PARALLEL."""
parser = argparse.ArgumentParser(
description="making feature file argsurations.")
parser.add_argument(
"--waveforms", default=None,
help="directory or list of filename of input wav | """EXTRACT MEL SPECTROGRAM."""
# define feature extractor
for i, wav_name in enumerate(wav_list):
logging.info("now processing %s (%d/%d)" % (wav_name, i + 1, len(wav_list)))
# load wavfile and apply low cut filter
fs, x = wavfile.read(wav_name)
if x.dtype != np.int16:
logging.warning("wav file format is not 16 bit PCM.")
x = np.array(x, dtype=np.float64)
if args.highpass_cutoff != 0:
x = low_cut_filter(x, fs, cutoff=args.highpass_cutoff)
# check sampling frequency
if not fs == args.fs:
logging.error("sampling frequency is not matched.")
sys.exit(1)
# extract features
x_norm = x / (np.iinfo(np.int16).max + 1) | identifier_body |
feature_extract.py | numtaps = 255
fil = firwin(numtaps, norm_cutoff)
x_pad = np.pad(x, (numtaps, numtaps), 'edge')
lpf_x = lfilter(fil, 1, x_pad)
lpf_x = lpf_x[numtaps + numtaps // 2: -numtaps // 2]
return lpf_x
def convert_to_continuos_f0(f0):
"""CONVERT F0 TO CONTINUOUS F0.
Args:
f0 (ndarray): original f0 sequence with the shape (T,).
Returns:
ndarray: continuous f0 with the shape (T,).
"""
# get uv information as binary
uv = np.float32(f0 != 0)
# get start and end of f0
if (f0 == 0).all():
logging.warning("all of the f0 values are 0.")
return uv, f0
start_f0 = f0[f0 != 0][0]
end_f0 = f0[f0 != 0][-1]
# padding start and end of f0 sequence
start_idx = np.where(f0 == start_f0)[0][0]
end_idx = np.where(f0 == end_f0)[0][-1]
f0[:start_idx] = start_f0
f0[end_idx:] = end_f0
# get non-zero frame index
nz_frames = np.where(f0 != 0)[0]
# perform linear interpolation
f = interp1d(nz_frames, f0[nz_frames])
cont_f0 = f(np.arange(0, f0.shape[0]))
return uv, cont_f0
def stft_mcep(x, fftl=512, shiftl=256, dim=25, alpha=0.41, window="hamming", is_padding=False):
"""EXTRACT STFT-BASED MEL-CEPSTRUM.
Args:
x (ndarray): Numpy double array with the size (T,).
fftl (int): FFT length in point (default=512).
shiftl (int): Shift length in point (default=256).
dim (int): Dimension of mel-cepstrum (default=25).
alpha (float): All pass filter coefficient (default=0.41).
window (str): Analysis window type (default="hamming").
is_padding (bool): Whether to pad the end of signal (default=False).
Returns:
ndarray: Mel-cepstrum with the size (N, n_fft).
"""
# perform padding
if is_padding:
|
# get number of frames
n_frame = (len(x) - fftl) // shiftl + 1
# get window function
win = get_window(window, fftl)
# calculate spectrogram
mcep = [pysptk.mcep(x[shiftl * i: shiftl * i + fftl] * win,
dim, alpha, eps=EPS, etype=1)
for i in range(n_frame)]
return np.stack(mcep)
def world_feature_extract(wav_list, args):
"""EXTRACT WORLD FEATURE VECTOR."""
# define feature extractor
feature_extractor = FeatureExtractor(
analyzer="world",
fs=args.fs,
shiftms=args.shiftms,
minf0=args.minf0,
maxf0=args.maxf0,
fftl=args.fftl)
for i, wav_name in enumerate(wav_list):
logging.info("now processing %s (%d/%d)" % (wav_name, i + 1, len(wav_list)))
# load wavfile and apply low cut filter
fs, x = wavfile.read(wav_name)
if x.dtype != np.int16:
logging.warning("wav file format is not 16 bit PCM.")
x = np.array(x, dtype=np.float64)
if args.highpass_cutoff != 0:
x = low_cut_filter(x, fs, cutoff=args.highpass_cutoff)
# check sampling frequency
if not fs == args.fs:
logging.error("sampling frequency is not matched.")
sys.exit(1)
# extract features
f0, _, _ = feature_extractor.analyze(x)
uv, cont_f0 = convert_to_continuos_f0(f0)
cont_f0_lpf = low_pass_filter(cont_f0, int(1.0 / (args.shiftms * 0.001)), cutoff=20)
codeap = feature_extractor.codeap()
mcep = feature_extractor.mcep(dim=args.mcep_dim, alpha=args.mcep_alpha)
# concatenate
cont_f0_lpf = np.expand_dims(cont_f0_lpf, axis=-1)
uv = np.expand_dims(uv, axis=-1)
feats = np.concatenate([uv, cont_f0_lpf, mcep, codeap], axis=1)
# save to hdf5
hdf5name = args.hdf5dir + "/" + os.path.basename(wav_name).replace(".wav", ".h5")
write_hdf5(hdf5name, "/world", feats)
# overwrite wav file
if args.highpass_cutoff != 0 and args.save_wav:
wavfile.write(args.wavdir + "/" + os.path.basename(wav_name), fs, np.int16(x))
def melspectrogram_extract(wav_list, args):
"""EXTRACT MEL SPECTROGRAM."""
# define feature extractor
for i, wav_name in enumerate(wav_list):
logging.info("now processing %s (%d/%d)" % (wav_name, i + 1, len(wav_list)))
# load wavfile and apply low cut filter
fs, x = wavfile.read(wav_name)
if x.dtype != np.int16:
logging.warning("wav file format is not 16 bit PCM.")
x = np.array(x, dtype=np.float64)
if args.highpass_cutoff != 0:
x = low_cut_filter(x, fs, cutoff=args.highpass_cutoff)
# check sampling frequency
if not fs == args.fs:
logging.error("sampling frequency is not matched.")
sys.exit(1)
# extract features
x_norm = x / (np.iinfo(np.int16).max + 1)
shiftl = int(args.shiftms * fs * 0.001)
mspc = librosa.feature.melspectrogram(
x_norm, fs,
n_fft=args.fftl,
hop_length=shiftl,
n_mels=args.mspc_dim,
fmin=args.fmin if args.fmin is not None else 0,
fmax=args.fmax if args.fmax is not None else fs // 2,
power=1.0)
mspc = np.log10(np.maximum(EPS, mspc.T))
# save to hdf5
hdf5name = args.hdf5dir + "/" + os.path.basename(wav_name).replace(".wav", ".h5")
write_hdf5(hdf5name, "/melspc", np.float32(mspc))
# overwrite wav file
if args.highpass_cutoff != 0 and args.save_wav:
wavfile.write(args.wavdir + "/" + os.path.basename(wav_name), fs, np.int16(x))
def melcepstrum_extract(wav_list, args):
"""EXTRACT MEL CEPSTRUM."""
# define feature extractor
for i, wav_name in enumerate(wav_list):
logging.info("now processing %s (%d/%d)" % (wav_name, i + 1, len(wav_list)))
# load wavfile and apply low cut filter
fs, x = wavfile.read(wav_name)
if x.dtype != np.int16:
logging.warning("wav file format is not 16 bit PCM.")
x = np.array(x, dtype=np.float64)
if args.highpass_cutoff != 0:
x = low_cut_filter(x, fs, cutoff=args.highpass_cutoff)
# check sampling frequency
if not fs == args.fs:
logging.error("sampling frequency is not matched.")
sys.exit(1)
# extract features
shiftl = int(args.shiftms * fs * 0.001)
mcep = stft_mcep(x, args.fftl, shiftl, args.mcep_dim, args.mcep_alpha)
# save to hdf5
hdf5name = args.hdf5dir + "/" + os.path.basename(wav_name).replace(".wav", ".h5")
write_hdf5(hdf5name, "/mcep", np.float32(mcep))
# overwrite wav file
if args.highpass_cutoff != 0 and args.save_wav:
wavfile.write(args.wavdir + "/" + os.path.basename(wav_name), fs, np.int16(x))
def main():
"""RUN FEATURE EXTRACTION IN PARALLEL."""
parser = argparse.ArgumentParser(
description="making feature file argsurations.")
parser.add_argument(
"--waveforms", default=None,
help="directory or list of filename of input wav | n_pad = fftl - (len(x) - fftl) % shiftl
x = np.pad(x, (0, n_pad), 'reflect') | conditional_block |
main.py | ('Set-Cookie', 'user_id=; Path=/')
def initialize(self, *a, **kw):
webapp2.RequestHandler.initialize(self, *a, **kw)
uid = self.read_secure_cookie('user_id')
if uid != None:
uid = check_secure_val(uid)
if uid != None:
self.username = User.by_id(int(uid)).name
else:
self.username = None
else:
self.username = None
#base template end
secret='hiiis'
letters='abcdefghijklmnopqrstuvwxyz'
#hashing and salt stuff
def make_salt(length = 5):
return ''.join(random.choice(letters) for x in xrange(length))
def make_pw_hash(name, pw, salt = None):
if not salt:
salt = make_salt()
h = hashlib.sha256(name + pw + salt).hexdigest()
return '%s,%s' % (salt, h)
def valid_pw(name, password, h):
salt = h.split(',')[0]
return h == make_pw_hash(name, password, salt)
def make_secure_val(val):
return '%s|%s' % (val, hmac.new(secret, val).hexdigest())
def check_secure_val(secure_val):
val = secure_val.split('|')[0]
if secure_val == make_secure_val(val):
return val
#hashing stuff ends
class User(db.Model):
name = db.StringProperty(required=True)
pw_hash = db.StringProperty(required=True)
email = db.StringProperty()
curr_balance = db.IntegerProperty(required=True)
@classmethod
def by_id(cls, uid):
return User.get_by_id(uid)
@classmethod
def by_name(cls, name):
u = User.all().filter('name =', name).get()
return u
@classmethod
def register(cls, name, pw, email = None):
pw_hash = make_pw_hash(name, pw)
return User(name = name,
pw_hash = pw_hash,
email = email,
curr_balance = 0)
@classmethod
def login(cls, name, pw):
u = cls.by_name(name)
# print(u)
if u and valid_pw(name, pw, u.pw_hash):
return u
# account details and related stuff
#
conn = sqlite3.connect('account.db')
# conn.execute('''CREATE TABLE stocks
# (
# uname TEXT NOT NULL,
# stk_symbl TEXT NOT NULL,
# stk_qty INT NOT NULL,
# stk_price INT NOT NULL,
# time_stamp TIMESTAMP NOT NULL )
# ''')
# conn.execute('''
# CREATE TABLE transactionRequests
# (
# uname TEXT NOT NULL,
# stk_qty INT NOT NULL,
# stk_symbl TEXT NOT NULL,
# stk_price INT NOT NULL,
# time_stamp TIMESTAMP NOT NULL,
# status TEXT,
# PRIMARY KEY(uname, time_stamp))
# ''')
t_now = datetime.datetime.now()
c = conn.cursor()
# inst = "INSERT INTO stocks VALUES ('test','TEST',10,100,?)"
# conn.execute("INSERT INTO stocks VALUES ('test','TEST',10,100,?)", (t_now,))
conn.commit()
c.execute('SELECT * FROM stocks')
print c.fetchall()
#conn.commit()
#functions for basic sign-up
USER_RE = re.compile(r"^[a-zA-Z0-9_-]{3,20}$")
def valid_username(username):
return username and USER_RE.match(username)
PASS_RE = re.compile(r"^.{3,20}$")
def valid_password(password):
|
EMAIL_RE = re.compile(r'^[\S]+@[\S]+\.[\S]+$')
def valid_email(email):
return not email or EMAIL_RE.match(email)
class SignUp(BaseHandler):
def get(self):
if self.username != None:
self.redirect('/welcome')
# if valid_username(self.username):
# self.render('welcome.html', username = self.username)
else:
self.render('signup-form.html')
def post(self):
uname = self.request.get('username')
password = self.request.get('password')
verify = self.request.get('verify')
email = self.request.get('email')
# self.username = uname
# self.password = password
# self.email = email
have_error = False
params = dict(username = uname,
email = email)
if not valid_username(uname):
params['error_username'] = "That's not a valid username."
have_error = True
if not valid_password(password):
params['error_password'] = "That wasn't a valid password."
have_error = True
elif password != verify:
params['error_verify'] = "Your passwords didn't match."
have_error = True
if not valid_email(email):
params['error_email'] = "That's not a valid email."
have_error = True
if have_error:
self.render('signup-form.html', **params)
else:
# cooked = 'name=' + uname + ';Path=/'
# self.response.headers.add_header('Set-Cookie',str(cooked))
# self.set_cook('name', uname)
u = User.by_name(uname)
if u:
msg = 'That user already exists.'
self.render('signup-form.html', error_username = msg)
else:
# cooked = 'name=' + uname + ';Path=/'
# self.response.headers.add_header('Set-Cookie',str(cooked))
# self.set_cookie('name',uname)
u = User.register(uname, password, email)
u.put()
self.login(u)
self.redirect('/welcome')
# self.login(u)
# self.redirect('/blog')
class History(BaseHandler):
def get(self):
if self.username != None:
params = {}
sname = self.request.get('sname')
c.execute("SELECT * from transactionRequests where uname=? and stk_symbl=?",(self.username, sname))
params['stk_arr'] = c.fetchall()
params['sname'] = sname
params['username'] = self.username
u = User.by_name(self.username)
print(u.curr_balance)
params['current_balance'] = u.curr_balance
print(params)
self.render('history.html', **params)
else:
self.redirect('/login')
class Welcome(BaseHandler):
def get(self):
if self.username != None:
params = {}
params['username'] = self.username;
c.execute("SELECT * FROM stocks where uname=?",(self.username,))
dat = c.fetchall()
params['stk_arr'] = dat
u = User.by_name(self.username)
print(u.curr_balance)
params['current_balance'] = u.curr_balance
self.render('welcome.html', **params)
else:
self.redirect('/login')
class Login(BaseHandler):
def get(self):
if self.username != None:
self.redirect('/welcome')
self.render('login-form.html')
def post(self):
uname=self.request.get('username')
passwd=self.request.get('password')
u=User.login(uname, passwd)
if u:
self.login(u)
self.redirect('/welcome')
else:
msg = 'Invalid login'
self.render('login-form.html', error = msg)
class LogOut(BaseHandler):
def get(self):
self.logout()
self.redirect('/login')
class SStock(BaseHandler):
def get(self):
sname = self.request.get('sname')
params ={}
params['stock_name']=sname
u = User.by_name(self.username)
params['current_balance'] = u.curr_balance
params['username'] = self.username
# c.execute("SELECT * FROM stocks where uname=?",(self.username,))
# dat = c.fetchall()
# self.render('stock_buy.html',stk_arr=dat)
c.execute("SELECT * FROM stocks where uname=? and stk_symbl=?",(self.username,sname))
global dat
dat = c.fetchone()
params['stk_arr']=dat
print(dat)
if dat:
params['sell_opt'] = True
self.render('stock_page.html',**params)
def post(self):
global dat
req = self.request.get('req')
sname = self.request.get('sname')
stk_qty = self.request.get('qty')
stk_price = self.request.get('stk_valu')
u = User.by_name(self.username)
print(req)
if(req == 'buy'):
print(req)
t_now = datetime.datetime.now()
tot_cost = int(stk_qty) * float(stk_price)
if(u.curr_balance > tot_cost):
u.curr_balance = int(u.curr_balance - tot_cost)
u.put()
c.execute("SELECT * FROM stocks where uname=? and stk_symbl=?",(self.username,sname))
tmp = c.fetchone()
conn.execute("INSERT INTO transactionRequests VALUES (?,?,?,?,?,?)", (self.username,stk_qty,sname,stk_price,t_now,"BUY--SUCCESS"))
if tmp:
print('in if')
tmp2 = int(tmp[2])
tmp3 = float(tmp[3])
avg = ((tmp2 * tmp3) + int(stk_qty) * float(stk_price)) / (tmp2 + int(stk_qty))
print(tmp2 + int(stk_qty))
print("UPDATE stocks SET stk_qty=? | return password and PASS_RE.match(password) | identifier_body |
main.py | _header('Set-Cookie', 'user_id=; Path=/')
def initialize(self, *a, **kw):
webapp2.RequestHandler.initialize(self, *a, **kw)
uid = self.read_secure_cookie('user_id')
if uid != None:
uid = check_secure_val(uid)
if uid != None:
self.username = User.by_id(int(uid)).name
else:
self.username = None
else:
self.username = None
#base template end
secret='hiiis'
letters='abcdefghijklmnopqrstuvwxyz'
#hashing and salt stuff
def make_salt(length = 5):
return ''.join(random.choice(letters) for x in xrange(length))
def make_pw_hash(name, pw, salt = None):
if not salt:
salt = make_salt()
h = hashlib.sha256(name + pw + salt).hexdigest()
return '%s,%s' % (salt, h)
def valid_pw(name, password, h):
salt = h.split(',')[0]
return h == make_pw_hash(name, password, salt)
def make_secure_val(val):
return '%s|%s' % (val, hmac.new(secret, val).hexdigest())
def check_secure_val(secure_val):
val = secure_val.split('|')[0]
if secure_val == make_secure_val(val):
return val
#hashing stuff ends
class User(db.Model):
name = db.StringProperty(required=True)
pw_hash = db.StringProperty(required=True)
email = db.StringProperty()
curr_balance = db.IntegerProperty(required=True)
@classmethod
def by_id(cls, uid):
return User.get_by_id(uid)
@classmethod
def by_name(cls, name):
u = User.all().filter('name =', name).get()
return u
@classmethod
def register(cls, name, pw, email = None):
pw_hash = make_pw_hash(name, pw)
return User(name = name,
pw_hash = pw_hash,
email = email,
curr_balance = 0)
@classmethod
def login(cls, name, pw):
u = cls.by_name(name)
# print(u)
if u and valid_pw(name, pw, u.pw_hash):
return u
# account details and related stuff
#
conn = sqlite3.connect('account.db')
# conn.execute('''CREATE TABLE stocks
# (
# uname TEXT NOT NULL,
# stk_symbl TEXT NOT NULL,
# stk_qty INT NOT NULL,
# stk_price INT NOT NULL,
# time_stamp TIMESTAMP NOT NULL )
# ''')
# conn.execute('''
# CREATE TABLE transactionRequests
# (
# uname TEXT NOT NULL,
# stk_qty INT NOT NULL,
# stk_symbl TEXT NOT NULL,
# stk_price INT NOT NULL,
# time_stamp TIMESTAMP NOT NULL,
# status TEXT,
# PRIMARY KEY(uname, time_stamp))
# ''')
t_now = datetime.datetime.now()
c = conn.cursor()
# inst = "INSERT INTO stocks VALUES ('test','TEST',10,100,?)"
# conn.execute("INSERT INTO stocks VALUES ('test','TEST',10,100,?)", (t_now,))
conn.commit()
c.execute('SELECT * FROM stocks')
print c.fetchall()
#conn.commit()
#functions for basic sign-up
USER_RE = re.compile(r"^[a-zA-Z0-9_-]{3,20}$")
def valid_username(username):
return username and USER_RE.match(username)
PASS_RE = re.compile(r"^.{3,20}$")
def valid_password(password):
return password and PASS_RE.match(password)
EMAIL_RE = re.compile(r'^[\S]+@[\S]+\.[\S]+$')
def valid_email(email):
return not email or EMAIL_RE.match(email)
class SignUp(BaseHandler):
def get(self):
if self.username != None:
self.redirect('/welcome')
# if valid_username(self.username):
# self.render('welcome.html', username = self.username)
else:
self.render('signup-form.html')
def post(self):
uname = self.request.get('username')
password = self.request.get('password')
verify = self.request.get('verify')
email = self.request.get('email')
# self.username = uname
# self.password = password
# self.email = email
have_error = False
params = dict(username = uname,
email = email)
if not valid_username(uname):
params['error_username'] = "That's not a valid username."
have_error = True
if not valid_password(password):
params['error_password'] = "That wasn't a valid password."
have_error = True | if not valid_email(email):
params['error_email'] = "That's not a valid email."
have_error = True
if have_error:
self.render('signup-form.html', **params)
else:
# cooked = 'name=' + uname + ';Path=/'
# self.response.headers.add_header('Set-Cookie',str(cooked))
# self.set_cook('name', uname)
u = User.by_name(uname)
if u:
msg = 'That user already exists.'
self.render('signup-form.html', error_username = msg)
else:
# cooked = 'name=' + uname + ';Path=/'
# self.response.headers.add_header('Set-Cookie',str(cooked))
# self.set_cookie('name',uname)
u = User.register(uname, password, email)
u.put()
self.login(u)
self.redirect('/welcome')
# self.login(u)
# self.redirect('/blog')
class History(BaseHandler):
def get(self):
if self.username != None:
params = {}
sname = self.request.get('sname')
c.execute("SELECT * from transactionRequests where uname=? and stk_symbl=?",(self.username, sname))
params['stk_arr'] = c.fetchall()
params['sname'] = sname
params['username'] = self.username
u = User.by_name(self.username)
print(u.curr_balance)
params['current_balance'] = u.curr_balance
print(params)
self.render('history.html', **params)
else:
self.redirect('/login')
class Welcome(BaseHandler):
def get(self):
if self.username != None:
params = {}
params['username'] = self.username;
c.execute("SELECT * FROM stocks where uname=?",(self.username,))
dat = c.fetchall()
params['stk_arr'] = dat
u = User.by_name(self.username)
print(u.curr_balance)
params['current_balance'] = u.curr_balance
self.render('welcome.html', **params)
else:
self.redirect('/login')
class Login(BaseHandler):
def get(self):
if self.username != None:
self.redirect('/welcome')
self.render('login-form.html')
def post(self):
uname=self.request.get('username')
passwd=self.request.get('password')
u=User.login(uname, passwd)
if u:
self.login(u)
self.redirect('/welcome')
else:
msg = 'Invalid login'
self.render('login-form.html', error = msg)
class LogOut(BaseHandler):
def get(self):
self.logout()
self.redirect('/login')
class SStock(BaseHandler):
def get(self):
sname = self.request.get('sname')
params ={}
params['stock_name']=sname
u = User.by_name(self.username)
params['current_balance'] = u.curr_balance
params['username'] = self.username
# c.execute("SELECT * FROM stocks where uname=?",(self.username,))
# dat = c.fetchall()
# self.render('stock_buy.html',stk_arr=dat)
c.execute("SELECT * FROM stocks where uname=? and stk_symbl=?",(self.username,sname))
global dat
dat = c.fetchone()
params['stk_arr']=dat
print(dat)
if dat:
params['sell_opt'] = True
self.render('stock_page.html',**params)
def post(self):
global dat
req = self.request.get('req')
sname = self.request.get('sname')
stk_qty = self.request.get('qty')
stk_price = self.request.get('stk_valu')
u = User.by_name(self.username)
print(req)
if(req == 'buy'):
print(req)
t_now = datetime.datetime.now()
tot_cost = int(stk_qty) * float(stk_price)
if(u.curr_balance > tot_cost):
u.curr_balance = int(u.curr_balance - tot_cost)
u.put()
c.execute("SELECT * FROM stocks where uname=? and stk_symbl=?",(self.username,sname))
tmp = c.fetchone()
conn.execute("INSERT INTO transactionRequests VALUES (?,?,?,?,?,?)", (self.username,stk_qty,sname,stk_price,t_now,"BUY--SUCCESS"))
if tmp:
print('in if')
tmp2 = int(tmp[2])
tmp3 = float(tmp[3])
avg = ((tmp2 * tmp3) + int(stk_qty) * float(stk_price)) / (tmp2 + int(stk_qty))
print(tmp2 + int(stk_qty))
print("UPDATE stocks SET stk_qty=? and stk | elif password != verify:
params['error_verify'] = "Your passwords didn't match."
have_error = True
| random_line_split |
main.py | .IntegerProperty(required=True)
@classmethod
def by_id(cls, uid):
return User.get_by_id(uid)
@classmethod
def by_name(cls, name):
u = User.all().filter('name =', name).get()
return u
@classmethod
def register(cls, name, pw, email = None):
pw_hash = make_pw_hash(name, pw)
return User(name = name,
pw_hash = pw_hash,
email = email,
curr_balance = 0)
@classmethod
def login(cls, name, pw):
u = cls.by_name(name)
# print(u)
if u and valid_pw(name, pw, u.pw_hash):
return u
# account details and related stuff
#
conn = sqlite3.connect('account.db')
# conn.execute('''CREATE TABLE stocks
# (
# uname TEXT NOT NULL,
# stk_symbl TEXT NOT NULL,
# stk_qty INT NOT NULL,
# stk_price INT NOT NULL,
# time_stamp TIMESTAMP NOT NULL )
# ''')
# conn.execute('''
# CREATE TABLE transactionRequests
# (
# uname TEXT NOT NULL,
# stk_qty INT NOT NULL,
# stk_symbl TEXT NOT NULL,
# stk_price INT NOT NULL,
# time_stamp TIMESTAMP NOT NULL,
# status TEXT,
# PRIMARY KEY(uname, time_stamp))
# ''')
t_now = datetime.datetime.now()
c = conn.cursor()
# inst = "INSERT INTO stocks VALUES ('test','TEST',10,100,?)"
# conn.execute("INSERT INTO stocks VALUES ('test','TEST',10,100,?)", (t_now,))
conn.commit()
c.execute('SELECT * FROM stocks')
print c.fetchall()
#conn.commit()
#functions for basic sign-up
USER_RE = re.compile(r"^[a-zA-Z0-9_-]{3,20}$")
def valid_username(username):
return username and USER_RE.match(username)
PASS_RE = re.compile(r"^.{3,20}$")
def valid_password(password):
return password and PASS_RE.match(password)
EMAIL_RE = re.compile(r'^[\S]+@[\S]+\.[\S]+$')
def valid_email(email):
return not email or EMAIL_RE.match(email)
class SignUp(BaseHandler):
def get(self):
if self.username != None:
self.redirect('/welcome')
# if valid_username(self.username):
# self.render('welcome.html', username = self.username)
else:
self.render('signup-form.html')
def post(self):
uname = self.request.get('username')
password = self.request.get('password')
verify = self.request.get('verify')
email = self.request.get('email')
# self.username = uname
# self.password = password
# self.email = email
have_error = False
params = dict(username = uname,
email = email)
if not valid_username(uname):
params['error_username'] = "That's not a valid username."
have_error = True
if not valid_password(password):
params['error_password'] = "That wasn't a valid password."
have_error = True
elif password != verify:
params['error_verify'] = "Your passwords didn't match."
have_error = True
if not valid_email(email):
params['error_email'] = "That's not a valid email."
have_error = True
if have_error:
self.render('signup-form.html', **params)
else:
# cooked = 'name=' + uname + ';Path=/'
# self.response.headers.add_header('Set-Cookie',str(cooked))
# self.set_cook('name', uname)
u = User.by_name(uname)
if u:
msg = 'That user already exists.'
self.render('signup-form.html', error_username = msg)
else:
# cooked = 'name=' + uname + ';Path=/'
# self.response.headers.add_header('Set-Cookie',str(cooked))
# self.set_cookie('name',uname)
u = User.register(uname, password, email)
u.put()
self.login(u)
self.redirect('/welcome')
# self.login(u)
# self.redirect('/blog')
class History(BaseHandler):
def get(self):
if self.username != None:
params = {}
sname = self.request.get('sname')
c.execute("SELECT * from transactionRequests where uname=? and stk_symbl=?",(self.username, sname))
params['stk_arr'] = c.fetchall()
params['sname'] = sname
params['username'] = self.username
u = User.by_name(self.username)
print(u.curr_balance)
params['current_balance'] = u.curr_balance
print(params)
self.render('history.html', **params)
else:
self.redirect('/login')
class Welcome(BaseHandler):
def get(self):
if self.username != None:
params = {}
params['username'] = self.username;
c.execute("SELECT * FROM stocks where uname=?",(self.username,))
dat = c.fetchall()
params['stk_arr'] = dat
u = User.by_name(self.username)
print(u.curr_balance)
params['current_balance'] = u.curr_balance
self.render('welcome.html', **params)
else:
self.redirect('/login')
class Login(BaseHandler):
def get(self):
if self.username != None:
self.redirect('/welcome')
self.render('login-form.html')
def post(self):
uname=self.request.get('username')
passwd=self.request.get('password')
u=User.login(uname, passwd)
if u:
self.login(u)
self.redirect('/welcome')
else:
msg = 'Invalid login'
self.render('login-form.html', error = msg)
class LogOut(BaseHandler):
def get(self):
self.logout()
self.redirect('/login')
class SStock(BaseHandler):
def get(self):
sname = self.request.get('sname')
params ={}
params['stock_name']=sname
u = User.by_name(self.username)
params['current_balance'] = u.curr_balance
params['username'] = self.username
# c.execute("SELECT * FROM stocks where uname=?",(self.username,))
# dat = c.fetchall()
# self.render('stock_buy.html',stk_arr=dat)
c.execute("SELECT * FROM stocks where uname=? and stk_symbl=?",(self.username,sname))
global dat
dat = c.fetchone()
params['stk_arr']=dat
print(dat)
if dat:
params['sell_opt'] = True
self.render('stock_page.html',**params)
def post(self):
global dat
req = self.request.get('req')
sname = self.request.get('sname')
stk_qty = self.request.get('qty')
stk_price = self.request.get('stk_valu')
u = User.by_name(self.username)
print(req)
if(req == 'buy'):
print(req)
t_now = datetime.datetime.now()
tot_cost = int(stk_qty) * float(stk_price)
if(u.curr_balance > tot_cost):
u.curr_balance = int(u.curr_balance - tot_cost)
u.put()
c.execute("SELECT * FROM stocks where uname=? and stk_symbl=?",(self.username,sname))
tmp = c.fetchone()
conn.execute("INSERT INTO transactionRequests VALUES (?,?,?,?,?,?)", (self.username,stk_qty,sname,stk_price,t_now,"BUY--SUCCESS"))
if tmp:
print('in if')
tmp2 = int(tmp[2])
tmp3 = float(tmp[3])
avg = ((tmp2 * tmp3) + int(stk_qty) * float(stk_price)) / (tmp2 + int(stk_qty))
print(tmp2 + int(stk_qty))
print("UPDATE stocks SET stk_qty=? and stk_price=? where stk_symbl=? and uname=?", ((tmp2 + int(stk_qty)), avg, sname, self.username, ))
conn.execute("UPDATE stocks SET stk_qty=?, stk_price=? where uname=? and stk_symbl=?", ((tmp2 + int(stk_qty)), avg, self.username, sname))
conn.commit()
else:
print('hello world')
conn.execute("INSERT INTO stocks VALUES (?,?,?,?,?)", (self.username,sname,stk_qty,stk_price,t_now))
conn.commit()
print('helo transaction done')
# self.write('transaction complete')
# self.render('success.html')
# time.sleep(5)
self.redirect('/welcome')
else:
# self.render('regret.html')
# time.sleep(5)
# self.redirect('/stock_info?sname='+ sname)
conn.execute("INSERT INTO transactionRequests VALUES (?,?,?,?,?,?)", (self.username,stk_qty,sname,stk_price,t_now,"BUY--FAILED"))
conn.commit()
self.redirect('/regret')
# else
if (req == 'sell'):
t_now = datetime.datetime.now()
tot_cost = int(stk_qty) * float(stk_price)
if(int(stk_qty) > dat[2]):
| conn.execute("INSERT INTO transactionRequests VALUES (?,?,?,?,?,?)", (self.username,stk_qty,sname,stk_price,t_now,"SELL--FAILED"))
print('naaah boy') | conditional_block |
|
main.py | _header('Set-Cookie', 'user_id=; Path=/')
def initialize(self, *a, **kw):
webapp2.RequestHandler.initialize(self, *a, **kw)
uid = self.read_secure_cookie('user_id')
if uid != None:
uid = check_secure_val(uid)
if uid != None:
self.username = User.by_id(int(uid)).name
else:
self.username = None
else:
self.username = None
#base template end
secret='hiiis'
letters='abcdefghijklmnopqrstuvwxyz'
#hashing and salt stuff
def make_salt(length = 5):
return ''.join(random.choice(letters) for x in xrange(length))
def make_pw_hash(name, pw, salt = None):
if not salt:
salt = make_salt()
h = hashlib.sha256(name + pw + salt).hexdigest()
return '%s,%s' % (salt, h)
def valid_pw(name, password, h):
salt = h.split(',')[0]
return h == make_pw_hash(name, password, salt)
def make_secure_val(val):
return '%s|%s' % (val, hmac.new(secret, val).hexdigest())
def check_secure_val(secure_val):
val = secure_val.split('|')[0]
if secure_val == make_secure_val(val):
return val
#hashing stuff ends
class User(db.Model):
name = db.StringProperty(required=True)
pw_hash = db.StringProperty(required=True)
email = db.StringProperty()
curr_balance = db.IntegerProperty(required=True)
@classmethod
def by_id(cls, uid):
return User.get_by_id(uid)
@classmethod
def by_name(cls, name):
u = User.all().filter('name =', name).get()
return u
@classmethod
def register(cls, name, pw, email = None):
pw_hash = make_pw_hash(name, pw)
return User(name = name,
pw_hash = pw_hash,
email = email,
curr_balance = 0)
@classmethod
def login(cls, name, pw):
u = cls.by_name(name)
# print(u)
if u and valid_pw(name, pw, u.pw_hash):
return u
# account details and related stuff
#
conn = sqlite3.connect('account.db')
# conn.execute('''CREATE TABLE stocks
# (
# uname TEXT NOT NULL,
# stk_symbl TEXT NOT NULL,
# stk_qty INT NOT NULL,
# stk_price INT NOT NULL,
# time_stamp TIMESTAMP NOT NULL )
# ''')
# conn.execute('''
# CREATE TABLE transactionRequests
# (
# uname TEXT NOT NULL,
# stk_qty INT NOT NULL,
# stk_symbl TEXT NOT NULL,
# stk_price INT NOT NULL,
# time_stamp TIMESTAMP NOT NULL,
# status TEXT,
# PRIMARY KEY(uname, time_stamp))
# ''')
t_now = datetime.datetime.now()
c = conn.cursor()
# inst = "INSERT INTO stocks VALUES ('test','TEST',10,100,?)"
# conn.execute("INSERT INTO stocks VALUES ('test','TEST',10,100,?)", (t_now,))
conn.commit()
c.execute('SELECT * FROM stocks')
print c.fetchall()
#conn.commit()
#functions for basic sign-up
USER_RE = re.compile(r"^[a-zA-Z0-9_-]{3,20}$")
def valid_username(username):
return username and USER_RE.match(username)
PASS_RE = re.compile(r"^.{3,20}$")
def valid_password(password):
return password and PASS_RE.match(password)
EMAIL_RE = re.compile(r'^[\S]+@[\S]+\.[\S]+$')
def valid_email(email):
return not email or EMAIL_RE.match(email)
class SignUp(BaseHandler):
def get(self):
if self.username != None:
self.redirect('/welcome')
# if valid_username(self.username):
# self.render('welcome.html', username = self.username)
else:
self.render('signup-form.html')
def post(self):
uname = self.request.get('username')
password = self.request.get('password')
verify = self.request.get('verify')
email = self.request.get('email')
# self.username = uname
# self.password = password
# self.email = email
have_error = False
params = dict(username = uname,
email = email)
if not valid_username(uname):
params['error_username'] = "That's not a valid username."
have_error = True
if not valid_password(password):
params['error_password'] = "That wasn't a valid password."
have_error = True
elif password != verify:
params['error_verify'] = "Your passwords didn't match."
have_error = True
if not valid_email(email):
params['error_email'] = "That's not a valid email."
have_error = True
if have_error:
self.render('signup-form.html', **params)
else:
# cooked = 'name=' + uname + ';Path=/'
# self.response.headers.add_header('Set-Cookie',str(cooked))
# self.set_cook('name', uname)
u = User.by_name(uname)
if u:
msg = 'That user already exists.'
self.render('signup-form.html', error_username = msg)
else:
# cooked = 'name=' + uname + ';Path=/'
# self.response.headers.add_header('Set-Cookie',str(cooked))
# self.set_cookie('name',uname)
u = User.register(uname, password, email)
u.put()
self.login(u)
self.redirect('/welcome')
# self.login(u)
# self.redirect('/blog')
class History(BaseHandler):
def get(self):
if self.username != None:
params = {}
sname = self.request.get('sname')
c.execute("SELECT * from transactionRequests where uname=? and stk_symbl=?",(self.username, sname))
params['stk_arr'] = c.fetchall()
params['sname'] = sname
params['username'] = self.username
u = User.by_name(self.username)
print(u.curr_balance)
params['current_balance'] = u.curr_balance
print(params)
self.render('history.html', **params)
else:
self.redirect('/login')
class Welcome(BaseHandler):
def get(self):
if self.username != None:
params = {}
params['username'] = self.username;
c.execute("SELECT * FROM stocks where uname=?",(self.username,))
dat = c.fetchall()
params['stk_arr'] = dat
u = User.by_name(self.username)
print(u.curr_balance)
params['current_balance'] = u.curr_balance
self.render('welcome.html', **params)
else:
self.redirect('/login')
class Login(BaseHandler):
def get(self):
if self.username != None:
self.redirect('/welcome')
self.render('login-form.html')
def | (self):
uname=self.request.get('username')
passwd=self.request.get('password')
u=User.login(uname, passwd)
if u:
self.login(u)
self.redirect('/welcome')
else:
msg = 'Invalid login'
self.render('login-form.html', error = msg)
class LogOut(BaseHandler):
def get(self):
self.logout()
self.redirect('/login')
class SStock(BaseHandler):
def get(self):
sname = self.request.get('sname')
params ={}
params['stock_name']=sname
u = User.by_name(self.username)
params['current_balance'] = u.curr_balance
params['username'] = self.username
# c.execute("SELECT * FROM stocks where uname=?",(self.username,))
# dat = c.fetchall()
# self.render('stock_buy.html',stk_arr=dat)
c.execute("SELECT * FROM stocks where uname=? and stk_symbl=?",(self.username,sname))
global dat
dat = c.fetchone()
params['stk_arr']=dat
print(dat)
if dat:
params['sell_opt'] = True
self.render('stock_page.html',**params)
def post(self):
global dat
req = self.request.get('req')
sname = self.request.get('sname')
stk_qty = self.request.get('qty')
stk_price = self.request.get('stk_valu')
u = User.by_name(self.username)
print(req)
if(req == 'buy'):
print(req)
t_now = datetime.datetime.now()
tot_cost = int(stk_qty) * float(stk_price)
if(u.curr_balance > tot_cost):
u.curr_balance = int(u.curr_balance - tot_cost)
u.put()
c.execute("SELECT * FROM stocks where uname=? and stk_symbl=?",(self.username,sname))
tmp = c.fetchone()
conn.execute("INSERT INTO transactionRequests VALUES (?,?,?,?,?,?)", (self.username,stk_qty,sname,stk_price,t_now,"BUY--SUCCESS"))
if tmp:
print('in if')
tmp2 = int(tmp[2])
tmp3 = float(tmp[3])
avg = ((tmp2 * tmp3) + int(stk_qty) * float(stk_price)) / (tmp2 + int(stk_qty))
print(tmp2 + int(stk_qty))
print("UPDATE stocks SET stk_qty=? | post | identifier_name |
exception.rs | }
paste::item! {
pub fn [< $body:snake _ code >] () -> u16{
$code
}
pub fn [< $body Code >] () -> u16{
$code
}
}
)*
}
}
}
build_exceptions! {
Ok(0),
UnknownTypeOfQuery(1),
UnImplement(2),
UnknownDatabase(3),
UnknownSetting(4),
SyntaxException(5),
BadArguments(6),
IllegalDataType(7),
UnknownFunction(8),
IllegalFunctionState(9),
BadDataValueType(10),
UnknownPlan(11),
IllegalPipelineState(12),
BadTransformType(13),
IllegalTransformConnectionState(14),
LogicalError(15),
EmptyData(16),
DataStructMissMatch(17),
BadDataArrayLength(18),
UnknownContextID(19),
UnknownVariable(20),
UnknownTableFunction(21),
BadOption(22),
CannotReadFile(23),
ParquetError(24),
UnknownTable(25),
IllegalAggregateExp(26),
UnknownAggregateFunction(27),
NumberArgumentsNotMatch(28),
NotFoundStream(29),
EmptyDataFromServer(30),
NotFoundLocalNode(31),
PlanScheduleError(32),
BadPlanInputs(33),
DuplicateClusterNode(34),
NotFoundClusterNode(35),
BadAddressFormat(36),
DnsParseError(37),
CannotConnectNode(38),
DuplicateGetStream(39),
Timeout(40),
TooManyUserConnections(41),
AbortedSession(ABORT_SESSION),
AbortedQuery(ABORT_QUERY),
NotFoundSession(44),
CannotListenerPort(45),
BadBytes(46),
InitPrometheusFailure(47),
ScalarSubqueryBadRows(48),
Overflow(49),
InvalidMetaBinaryFormat(50),
AuthenticateFailure(51),
TLSConfigurationFailure(52),
UnknownSession(53),
UnexpectedError(54),
DateTimeParseError(55),
BadPredicateRows(56),
SHA1CheckFailed(57),
// uncategorized
UnexpectedResponseType(600),
UnknownException(1000),
TokioError(1001),
}
// Store errors
build_exceptions! {
FileMetaNotFound(2001),
FileDamaged(2002),
// dfs node errors
UnknownNode(2101),
// meta service errors
// meta service does not work.
MetaServiceError(2201),
// meta service is shut down.
MetaServiceShutdown(2202),
// meta service is unavailable for now.
MetaServiceUnavailable(2203),
// config errors
InvalidConfig(2301),
// meta store errors
MetaStoreDamaged(2401),
MetaStoreAlreadyExists(2402),
MetaStoreNotFound(2403),
ConcurrentSnapshotInstall(2404),
IllegalSnapshot(2405),
UnknownTableId(2406),
TableVersionMissMatch(2407),
// KVSrv server error
KVSrvError(2501),
// FS error
IllegalFileName(2601),
// Store server error
DatabendStoreError(2701),
// TODO
// We may need to separate front-end errors from API errors (and system errors?)
// That may depend which components are using these error codes, and for what purposes,
// let's figure it out latter.
// user-api error codes
UnknownUser(3000),
UserAlreadyExists(3001),
IllegalUserInfoFormat(3002),
// meta-api error codes
DatabaseAlreadyExists(4001),
TableAlreadyExists(4003),
IllegalMetaOperationArgument(4004),
IllegalSchema(4005),
IllegalMetaState(4006),
MetaNodeInternalError(4007),
TruncateTableFailedError(4008),
CommitTableError(4009),
// namespace error.
NamespaceUnknownNode(4058),
NamespaceNodeAlreadyExists(4059),
NamespaceIllegalNodeFormat(4050),
// storage-api error codes
IllegalScanPlan(5000),
ReadFileError(5001),
BrokenChannel(5002),
// kv-api error codes
UnknownKey(6000),
// DAL error
DALTransportError(7000),
UnknownStorageSchemeName(7001),
SecretKeyNotSet(7002),
// datasource error
DuplicatedTableEngineProvider(8000),
UnknownDatabaseEngine(8001),
UnknownTableEngine(8002),
DuplicatedDatabaseEngineProvider(8003),
}
// General errors
build_exceptions! {
// A task that already stopped and can not stop twice.
AlreadyStarted(7101),
// A task that already started and can not start twice.
AlreadyStopped(7102),
// Trying to cast to a invalid type
InvalidCast(7201),
}
pub type Result<T> = std::result::Result<T, ErrorCode>;
impl Debug for ErrorCode {
fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result {
write!(
f,
"Code: {}, displayText = {}.",
self.code(),
self.message(),
)?;
match self.backtrace.as_ref() {
None => Ok(()), // no backtrace
Some(backtrace) => {
// TODO: Custom stack frame format for print
match backtrace {
ErrorCodeBacktrace::Origin(backtrace) => write!(f, "\n\n{:?}", backtrace),
ErrorCodeBacktrace::Serialized(backtrace) => write!(f, "\n\n{:?}", backtrace),
}
}
}
}
}
impl Display for ErrorCode {
fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result {
write!(
f,
"Code: {}, displayText = {}.",
self.code(),
self.message(),
)
}
}
#[derive(Error)]
enum OtherErrors {
AnyHow { error: anyhow::Error },
}
impl Display for OtherErrors {
fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result {
match self {
OtherErrors::AnyHow { error } => write!(f, "{}", error),
}
}
}
impl Debug for OtherErrors {
fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result {
match self {
OtherErrors::AnyHow { error } => write!(f, "{:?}", error),
}
}
}
impl From<anyhow::Error> for ErrorCode {
fn from(error: anyhow::Error) -> Self {
ErrorCode {
code: 1002,
display_text: format!("{}, source: {:?}", error, error.source()),
cause: Some(Box::new(OtherErrors::AnyHow { error })),
backtrace: Some(ErrorCodeBacktrace::Origin(Arc::new(Backtrace::new()))),
}
}
}
impl From<std::num::ParseIntError> for ErrorCode {
fn from(error: std::num::ParseIntError) -> Self {
ErrorCode::from_std_error(error)
}
}
impl From<std::num::ParseFloatError> for ErrorCode {
fn from(error: std::num::ParseFloatError) -> Self {
ErrorCode::from_std_error(error)
}
}
impl From<common_arrow::arrow::error::ArrowError> for ErrorCode {
fn from(error: common_arrow::arrow::error::ArrowError) -> Self {
ErrorCode::from_std_error(error)
}
}
impl From<serde_json::Error> for ErrorCode {
fn from(error: serde_json::Error) -> Self {
ErrorCode::from_std_error(error)
}
}
impl From<sqlparser::parser::ParserError> for ErrorCode {
fn from(error: sqlparser::parser::ParserError) -> Self {
ErrorCode::from_std_error(error)
}
}
impl From<std::io::Error> for ErrorCode {
fn from(error: std::io::Error) -> Self {
ErrorCode::from_std_error(error)
}
}
impl From<std::net::AddrParseError> for ErrorCode {
fn from(error: AddrParseError) -> Self {
ErrorCode::BadAddressFormat(format!("Bad address format, cause: {}", error))
}
}
impl From<FromUtf8Error> for ErrorCode {
fn from(error: FromUtf8Error) -> Self {
ErrorCode::BadBytes(format!(
"Bad bytes, cannot parse bytes with UTF8, cause: {}",
error
))
}
}
impl From<prost::EncodeError> for ErrorCode {
fn from(error: prost::EncodeError) -> Self {
ErrorCode::BadBytes(format!(
"Bad bytes, cannot parse bytes with prost, cause: {}", | error
))
}
} | random_line_split |
|
exception.rs | self.cause,
backtrace: self.backtrace,
}
}
pub fn add_message_back(self, msg: impl AsRef<str>) -> Self {
Self {
code: self.code(),
display_text: format!("{}{}", self.display_text, msg.as_ref()),
cause: self.cause,
backtrace: self.backtrace,
}
}
pub fn backtrace(&self) -> Option<ErrorCodeBacktrace> {
self.backtrace.clone()
}
pub fn backtrace_str(&self) -> String {
match self.backtrace.as_ref() {
None => "".to_string(),
Some(backtrace) => backtrace.to_string(),
}
}
}
macro_rules! build_exceptions {
($($body:ident($code:expr)),*$(,)*) => {
impl ErrorCode {
$(
pub fn $body(display_text: impl Into<String>) -> ErrorCode {
ErrorCode {
code: $code,
display_text: display_text.into(),
cause: None,
backtrace: Some(ErrorCodeBacktrace::Origin(Arc::new(Backtrace::new()))),
}
}
paste::item! {
pub fn [< $body:snake _ code >] () -> u16{
$code
}
pub fn [< $body Code >] () -> u16{
$code
}
}
)*
}
}
}
build_exceptions! {
Ok(0),
UnknownTypeOfQuery(1),
UnImplement(2),
UnknownDatabase(3),
UnknownSetting(4),
SyntaxException(5),
BadArguments(6),
IllegalDataType(7),
UnknownFunction(8),
IllegalFunctionState(9),
BadDataValueType(10),
UnknownPlan(11),
IllegalPipelineState(12),
BadTransformType(13),
IllegalTransformConnectionState(14),
LogicalError(15),
EmptyData(16),
DataStructMissMatch(17),
BadDataArrayLength(18),
UnknownContextID(19),
UnknownVariable(20),
UnknownTableFunction(21),
BadOption(22),
CannotReadFile(23),
ParquetError(24),
UnknownTable(25),
IllegalAggregateExp(26),
UnknownAggregateFunction(27),
NumberArgumentsNotMatch(28),
NotFoundStream(29),
EmptyDataFromServer(30),
NotFoundLocalNode(31),
PlanScheduleError(32),
BadPlanInputs(33),
DuplicateClusterNode(34),
NotFoundClusterNode(35),
BadAddressFormat(36),
DnsParseError(37),
CannotConnectNode(38),
DuplicateGetStream(39),
Timeout(40),
TooManyUserConnections(41),
AbortedSession(ABORT_SESSION),
AbortedQuery(ABORT_QUERY),
NotFoundSession(44),
CannotListenerPort(45),
BadBytes(46),
InitPrometheusFailure(47),
ScalarSubqueryBadRows(48),
Overflow(49),
InvalidMetaBinaryFormat(50),
AuthenticateFailure(51),
TLSConfigurationFailure(52),
UnknownSession(53),
UnexpectedError(54),
DateTimeParseError(55),
BadPredicateRows(56),
SHA1CheckFailed(57),
// uncategorized
UnexpectedResponseType(600),
UnknownException(1000),
TokioError(1001),
}
// Store errors
build_exceptions! {
FileMetaNotFound(2001),
FileDamaged(2002),
// dfs node errors
UnknownNode(2101),
// meta service errors
// meta service does not work.
MetaServiceError(2201),
// meta service is shut down.
MetaServiceShutdown(2202),
// meta service is unavailable for now.
MetaServiceUnavailable(2203),
// config errors
InvalidConfig(2301),
// meta store errors
MetaStoreDamaged(2401),
MetaStoreAlreadyExists(2402),
MetaStoreNotFound(2403),
ConcurrentSnapshotInstall(2404),
IllegalSnapshot(2405),
UnknownTableId(2406),
TableVersionMissMatch(2407),
// KVSrv server error
KVSrvError(2501),
// FS error
IllegalFileName(2601),
// Store server error
DatabendStoreError(2701),
// TODO
// We may need to separate front-end errors from API errors (and system errors?)
// That may depend which components are using these error codes, and for what purposes,
// let's figure it out latter.
// user-api error codes
UnknownUser(3000),
UserAlreadyExists(3001),
IllegalUserInfoFormat(3002),
// meta-api error codes
DatabaseAlreadyExists(4001),
TableAlreadyExists(4003),
IllegalMetaOperationArgument(4004),
IllegalSchema(4005),
IllegalMetaState(4006),
MetaNodeInternalError(4007),
TruncateTableFailedError(4008),
CommitTableError(4009),
// namespace error.
NamespaceUnknownNode(4058),
NamespaceNodeAlreadyExists(4059),
NamespaceIllegalNodeFormat(4050),
// storage-api error codes
IllegalScanPlan(5000),
ReadFileError(5001),
BrokenChannel(5002),
// kv-api error codes
UnknownKey(6000),
// DAL error
DALTransportError(7000),
UnknownStorageSchemeName(7001),
SecretKeyNotSet(7002),
// datasource error
DuplicatedTableEngineProvider(8000),
UnknownDatabaseEngine(8001),
UnknownTableEngine(8002),
DuplicatedDatabaseEngineProvider(8003),
}
// General errors
build_exceptions! {
// A task that already stopped and can not stop twice.
AlreadyStarted(7101),
// A task that already started and can not start twice.
AlreadyStopped(7102),
// Trying to cast to a invalid type
InvalidCast(7201),
}
pub type Result<T> = std::result::Result<T, ErrorCode>;
impl Debug for ErrorCode {
fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result {
write!(
f,
"Code: {}, displayText = {}.",
self.code(),
self.message(),
)?;
match self.backtrace.as_ref() {
None => Ok(()), // no backtrace
Some(backtrace) => {
// TODO: Custom stack frame format for print
match backtrace {
ErrorCodeBacktrace::Origin(backtrace) => write!(f, "\n\n{:?}", backtrace),
ErrorCodeBacktrace::Serialized(backtrace) => write!(f, "\n\n{:?}", backtrace),
}
}
}
}
}
impl Display for ErrorCode {
fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result {
write!(
f,
"Code: {}, displayText = {}.",
self.code(),
self.message(),
)
}
}
#[derive(Error)]
enum OtherErrors {
AnyHow { error: anyhow::Error },
}
impl Display for OtherErrors {
fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result {
match self {
OtherErrors::AnyHow { error } => write!(f, "{}", error),
}
}
}
impl Debug for OtherErrors {
fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result {
match self {
OtherErrors::AnyHow { error } => write!(f, "{:?}", error),
}
}
}
impl From<anyhow::Error> for ErrorCode {
fn from(error: anyhow::Error) -> Self {
ErrorCode {
code: 1002,
display_text: format!("{}, source: {:?}", error, error.source()),
cause: Some(Box::new(OtherErrors::AnyHow { error })),
backtrace: Some(ErrorCodeBacktrace::Origin(Arc::new(Backtrace::new()))),
}
}
}
impl From<std::num::ParseIntError> for ErrorCode {
fn | (error: std::num::ParseIntError) -> Self {
ErrorCode::from_std_error(error)
}
}
impl From<std::num::ParseFloatError> for ErrorCode {
fn from(error: std::num::ParseFloatError) -> Self {
ErrorCode::from_std_error(error)
}
}
impl From<common_arrow::arrow::error::ArrowError> for ErrorCode {
fn from(error: common_arrow::arrow::error::ArrowError) -> Self {
ErrorCode::from_std_error(error)
}
}
impl From<serde_json::Error> for ErrorCode {
fn from(error: serde_json::Error) -> Self {
ErrorCode:: | from | identifier_name |
exception.rs | .cause,
backtrace: self.backtrace,
}
}
pub fn add_message_back(self, msg: impl AsRef<str>) -> Self {
Self {
code: self.code(),
display_text: format!("{}{}", self.display_text, msg.as_ref()),
cause: self.cause,
backtrace: self.backtrace,
}
}
pub fn backtrace(&self) -> Option<ErrorCodeBacktrace> {
self.backtrace.clone()
}
pub fn backtrace_str(&self) -> String {
match self.backtrace.as_ref() {
None => "".to_string(),
Some(backtrace) => backtrace.to_string(),
}
}
}
macro_rules! build_exceptions {
($($body:ident($code:expr)),*$(,)*) => {
impl ErrorCode {
$(
pub fn $body(display_text: impl Into<String>) -> ErrorCode {
ErrorCode {
code: $code,
display_text: display_text.into(),
cause: None,
backtrace: Some(ErrorCodeBacktrace::Origin(Arc::new(Backtrace::new()))),
}
}
paste::item! {
pub fn [< $body:snake _ code >] () -> u16{
$code
}
pub fn [< $body Code >] () -> u16{
$code
}
}
)*
}
}
}
build_exceptions! {
Ok(0),
UnknownTypeOfQuery(1),
UnImplement(2),
UnknownDatabase(3),
UnknownSetting(4),
SyntaxException(5),
BadArguments(6),
IllegalDataType(7),
UnknownFunction(8),
IllegalFunctionState(9),
BadDataValueType(10),
UnknownPlan(11),
IllegalPipelineState(12),
BadTransformType(13),
IllegalTransformConnectionState(14),
LogicalError(15),
EmptyData(16),
DataStructMissMatch(17),
BadDataArrayLength(18),
UnknownContextID(19),
UnknownVariable(20),
UnknownTableFunction(21),
BadOption(22),
CannotReadFile(23),
ParquetError(24),
UnknownTable(25),
IllegalAggregateExp(26),
UnknownAggregateFunction(27),
NumberArgumentsNotMatch(28),
NotFoundStream(29),
EmptyDataFromServer(30),
NotFoundLocalNode(31),
PlanScheduleError(32),
BadPlanInputs(33),
DuplicateClusterNode(34),
NotFoundClusterNode(35),
BadAddressFormat(36),
DnsParseError(37),
CannotConnectNode(38),
DuplicateGetStream(39),
Timeout(40),
TooManyUserConnections(41),
AbortedSession(ABORT_SESSION),
AbortedQuery(ABORT_QUERY),
NotFoundSession(44),
CannotListenerPort(45),
BadBytes(46),
InitPrometheusFailure(47),
ScalarSubqueryBadRows(48),
Overflow(49),
InvalidMetaBinaryFormat(50),
AuthenticateFailure(51),
TLSConfigurationFailure(52),
UnknownSession(53),
UnexpectedError(54),
DateTimeParseError(55),
BadPredicateRows(56),
SHA1CheckFailed(57),
// uncategorized
UnexpectedResponseType(600),
UnknownException(1000),
TokioError(1001),
}
// Store errors
build_exceptions! {
FileMetaNotFound(2001),
FileDamaged(2002),
// dfs node errors
UnknownNode(2101),
// meta service errors
// meta service does not work.
MetaServiceError(2201),
// meta service is shut down.
MetaServiceShutdown(2202),
// meta service is unavailable for now.
MetaServiceUnavailable(2203),
// config errors
InvalidConfig(2301),
// meta store errors
MetaStoreDamaged(2401),
MetaStoreAlreadyExists(2402),
MetaStoreNotFound(2403),
ConcurrentSnapshotInstall(2404),
IllegalSnapshot(2405),
UnknownTableId(2406),
TableVersionMissMatch(2407),
// KVSrv server error
KVSrvError(2501),
// FS error
IllegalFileName(2601),
// Store server error
DatabendStoreError(2701),
// TODO
// We may need to separate front-end errors from API errors (and system errors?)
// That may depend which components are using these error codes, and for what purposes,
// let's figure it out latter.
// user-api error codes
UnknownUser(3000),
UserAlreadyExists(3001),
IllegalUserInfoFormat(3002),
// meta-api error codes
DatabaseAlreadyExists(4001),
TableAlreadyExists(4003),
IllegalMetaOperationArgument(4004),
IllegalSchema(4005),
IllegalMetaState(4006),
MetaNodeInternalError(4007),
TruncateTableFailedError(4008),
CommitTableError(4009),
// namespace error.
NamespaceUnknownNode(4058),
NamespaceNodeAlreadyExists(4059),
NamespaceIllegalNodeFormat(4050),
// storage-api error codes
IllegalScanPlan(5000),
ReadFileError(5001),
BrokenChannel(5002),
// kv-api error codes
UnknownKey(6000),
// DAL error
DALTransportError(7000),
UnknownStorageSchemeName(7001),
SecretKeyNotSet(7002),
// datasource error
DuplicatedTableEngineProvider(8000),
UnknownDatabaseEngine(8001),
UnknownTableEngine(8002),
DuplicatedDatabaseEngineProvider(8003),
}
// General errors
build_exceptions! {
// A task that already stopped and can not stop twice.
AlreadyStarted(7101),
// A task that already started and can not start twice.
AlreadyStopped(7102),
// Trying to cast to a invalid type
InvalidCast(7201),
}
pub type Result<T> = std::result::Result<T, ErrorCode>;
impl Debug for ErrorCode {
fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result {
write!(
f,
"Code: {}, displayText = {}.",
self.code(),
self.message(),
)?;
match self.backtrace.as_ref() {
None => Ok(()), // no backtrace
Some(backtrace) => {
// TODO: Custom stack frame format for print
match backtrace {
ErrorCodeBacktrace::Origin(backtrace) => write!(f, "\n\n{:?}", backtrace),
ErrorCodeBacktrace::Serialized(backtrace) => write!(f, "\n\n{:?}", backtrace),
}
}
}
}
}
impl Display for ErrorCode {
fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result {
write!(
f,
"Code: {}, displayText = {}.",
self.code(),
self.message(),
)
}
}
#[derive(Error)]
enum OtherErrors {
AnyHow { error: anyhow::Error },
}
impl Display for OtherErrors {
fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result {
match self {
OtherErrors::AnyHow { error } => write!(f, "{}", error),
}
}
}
impl Debug for OtherErrors {
fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result {
match self {
OtherErrors::AnyHow { error } => write!(f, "{:?}", error),
}
}
}
impl From<anyhow::Error> for ErrorCode {
fn from(error: anyhow::Error) -> Self {
ErrorCode {
code: 1002,
display_text: format!("{}, source: {:?}", error, error.source()),
cause: Some(Box::new(OtherErrors::AnyHow { error })),
backtrace: Some(ErrorCodeBacktrace::Origin(Arc::new(Backtrace::new()))),
}
}
}
impl From<std::num::ParseIntError> for ErrorCode {
fn from(error: std::num::ParseIntError) -> Self |
}
impl From<std::num::ParseFloatError> for ErrorCode {
fn from(error: std::num::ParseFloatError) -> Self {
ErrorCode::from_std_error(error)
}
}
impl From<common_arrow::arrow::error::ArrowError> for ErrorCode {
fn from(error: common_arrow::arrow::error::ArrowError) -> Self {
ErrorCode::from_std_error(error)
}
}
impl From<serde_json::Error> for ErrorCode {
fn from(error: serde_json::Error) -> Self {
ErrorCode:: | {
ErrorCode::from_std_error(error)
} | identifier_body |
mailbox.rs | 18,
config: 0x1c,
write: 0x20,
};
// MailboxRegisterOffsets {
// read: 0x20,
// peek: 0x30,
// sender: 0x34,
// status: 0x38,
// config: 0x3c,
// write: 0x40,
// },
// ];
#[inline]
unsafe fn read_reg(base: usize, offset: u8) -> u32 {
((MAPPED_REGISTERS_BASE + base + offset as usize) as *const u32).read_volatile()
}
#[inline]
unsafe fn | (base: usize, offset: u8, value: u32) {
((MAPPED_REGISTERS_BASE + base + offset as usize) as *mut u32).write_volatile(value)
}
unsafe fn read_mailbox(channel: u8) -> u32 {
// 1. Read the status register until the empty flag is not set.
// 2. Read data from the read register.
// 3. If the lower four bits do not match the channel number desired repeat
// from 1.
// 4. The upper 28 bits are the returned data.
// Wait for the mailbox to be non-empty
// Execute a memory barrier
// Read MAIL0_STATUS
// Goto step 1 if MAIL_EMPTY bit is set
// Execute a memory barrier
// Read from MAIL0_READ
// Check the channel (lowest 4 bits) of the read value for the correct channel
// If the channel is not the one we wish to read from (i.e: 1), go to step 1
// Return the data (i.e: the read value >> 4)
// println!("Reading mailbox (want channel {})", channel);
let mut limit = 10;
loop {
let mut empty_limit = 10;
loop {
fence(Ordering::SeqCst);
if read_reg(MAIL_BASE, MAILBOX_OFFFSETS.status) & MAIL_EMPTY == 0 {
break;
}
if empty_limit == 0 {
panic!(
"Gave up waiting for mail when reading from mailbox (channel {})",
channel
);
}
empty_limit -= 1;
}
fence(Ordering::SeqCst);
let data: u32 = read_reg(MAIL_BASE, MAILBOX_OFFFSETS.read);
let read_channel = (data & 0x0F) as u8;
let data = data >> 4;
// println!(
// "Got data from mailbox: {:#8x} (from channel {})",
// data, read_channel
// );
if read_channel != channel {
// println!("Wrong channel, trying again...");
if limit == 0 {
panic!(
"Got trampled too many times when reading from mailbox (channel {})",
channel
);
}
limit -= 1;
continue;
}
return data;
}
}
unsafe fn write_mailbox(channel: u8, data: u32) {
// 1. Read the status register until the full flag is not set.
// 2. Write the data (shifted into the upper 28 bits) combined with the
// channel (in the lower four bits) to the write register.
// println!("Writing {:#8x} to mailbox channel {}", data, channel);
let mut limit = 10;
loop {
// Wait for space
fence(Ordering::SeqCst);
if read_reg(MAIL_BASE, MAILBOX_OFFFSETS.status + 0x20) & MAIL_FULL == 0 {
break;
}
if limit == 0 {
panic!(
"Gave up waiting for space to write to mailbox (channel {}, data: 0x{:08x})",
channel, data
);
}
limit -= 1;
}
write_reg(MAIL_BASE, MAILBOX_OFFFSETS.write, data | (channel as u32));
fence(Ordering::SeqCst);
// println!("Finished writing to mailbox");
}
pub trait PropertyTagList: Sized {
fn prepare(self) -> PropertyMessageWrapper<Self> {
PropertyMessageWrapper::new(self)
}
}
macro_rules! impl_ptl {
( $( $t:ident ),+ ) => {
impl< $($t),+ > PropertyTagList for ( $(PropertyMessage< $t >, )+ )
where $(
$t: Sized,
)+ {}
};
}
impl<T: Sized> PropertyTagList for PropertyMessage<T> {}
// impl_ptl!(T1);
// impl_ptl!(T1, T2);
// impl_ptl!(T1, T2, T3);
// impl_ptl!(T1, T2, T3, T4);
// impl_ptl!(T1, T2, T3, T4, T5);
// impl_ptl!(T1, T2, T3, T4, T5, T6);
// impl_ptl!(T1, T2, T3, T4, T5, T6, T7);
// impl_ptl!(T1, T2, T3, T4, T5, T6, T7, T8);
// impl_ptl!(T1, T2, T3, T4, T5, T6, T7, T8, T9);
// impl_ptl!(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10);
#[repr(C, align(16))]
#[derive(Debug)]
pub struct PropertyMessageWrapper<TL: PropertyTagList> {
buffer_size: u32,
code: u32,
tags: TL,
// extra: [u32; 10],
}
impl<TL: PropertyTagList> PropertyMessageWrapper<TL> {
#[inline]
fn new(tags: TL) -> Self {
// let extra_offset = offset_of!(Self => extra).get_byte_offset();
// assert!(extra_offset % 4 == 0);
let buffer_size = core::mem::size_of::<Self>();
PropertyMessageWrapper {
buffer_size: buffer_size
.try_into()
.expect("Property message list size in bytes is too big to fit in a u32"),
code: 0x0000_0000,
tags,
// extra: [0; 10],
}
}
fn as_quads(&self) -> &[u32] {
let size_bytes = mem::size_of::<Self>();
debug_assert_eq!(size_bytes % 4, 0);
let u32_size: usize = size_bytes / 4;
unsafe { slice::from_raw_parts((self as *const Self) as *const u32, u32_size) }
}
pub fn send<'a>(&'a mut self) -> Option<&'a TL>
where
TL: fmt::Debug,
{
// println!("Property message before sending over mailbox: {:#x?}", self);
// println!(
// "Property message quads before sending over mailbox: {:#x?}",
// self.as_quads()
// );
const CHANNEL: u8 = Channel::PropertyTagsSend as u8;
println!("sending message {:x?}", self);
unsafe {
let ptr = self as *const Self;
let addr = ptr as usize;
write_mailbox(CHANNEL, addr.try_into().ok()?);
let resp_addr = read_mailbox(CHANNEL);
}
// let resp_ptr = resp_addr as *const u32;
// println!("Got response from mailbox: {:#?}", &*resp_ptr);
// let resp_code: u32 = *resp_ptr.offset(1);
// println!(
// "Property message after response {:#8x}: {:#x?}",
// resp_addr, self
// );
// {
// let message_quads = self.as_quads();
// println!("Property message words: {:#x?}", message_quads);
// }
if self.code != 0x8000_0000 {
return None;
}
// let msg_ptr = resp_ptr.offset(2);
// let value_buffer_size_ptr = msg_ptr.offset(1);
// let value_buffer_size = (*value_buffer_size_ptr) as usize;
// let value_buffer_ptr = msg_ptr.offset(3) as *const T;
// assert_eq!(value_buffer_size, mem::size_of::<T>());
// let value_ref = &*(value_buffer_ptr as *const T);
// Some(value_ref)
println!("received message: {:#x?}", self);
Some(&self.tags)
}
}
impl<TL: PropertyTagList> ops::Deref for PropertyMessageWrapper<TL> {
type Target = TL;
fn deref(&self) -> &TL {
&self.tags
}
}
impl<TL: PropertyTagList> ops::DerefMut for PropertyMessageWrapper<TL> {
fn deref_mut(&mut self) -> &mut TL {
&mut self.tags
}
}
#[repr(C, align(4))]
#[derive(Debug)]
| write_reg | identifier_name |
mailbox.rs | 18,
config: 0x1c,
write: 0x20,
};
// MailboxRegisterOffsets {
// read: 0x20,
// peek: 0x30,
// sender: 0x34,
// status: 0x38,
// config: 0x3c,
// write: 0x40,
// },
// ];
#[inline]
unsafe fn read_reg(base: usize, offset: u8) -> u32 {
((MAPPED_REGISTERS_BASE + base + offset as usize) as *const u32).read_volatile()
}
#[inline]
unsafe fn write_reg(base: usize, offset: u8, value: u32) {
((MAPPED_REGISTERS_BASE + base + offset as usize) as *mut u32).write_volatile(value)
}
unsafe fn read_mailbox(channel: u8) -> u32 {
// 1. Read the status register until the empty flag is not set.
// 2. Read data from the read register.
// 3. If the lower four bits do not match the channel number desired repeat
// from 1.
// 4. The upper 28 bits are the returned data.
// Wait for the mailbox to be non-empty
// Execute a memory barrier
// Read MAIL0_STATUS
// Goto step 1 if MAIL_EMPTY bit is set
// Execute a memory barrier
// Read from MAIL0_READ
// Check the channel (lowest 4 bits) of the read value for the correct channel
// If the channel is not the one we wish to read from (i.e: 1), go to step 1
// Return the data (i.e: the read value >> 4)
// println!("Reading mailbox (want channel {})", channel);
let mut limit = 10;
loop {
let mut empty_limit = 10;
loop {
fence(Ordering::SeqCst);
if read_reg(MAIL_BASE, MAILBOX_OFFFSETS.status) & MAIL_EMPTY == 0 {
break;
}
if empty_limit == 0 {
panic!(
"Gave up waiting for mail when reading from mailbox (channel {})",
channel
);
}
empty_limit -= 1;
}
fence(Ordering::SeqCst);
let data: u32 = read_reg(MAIL_BASE, MAILBOX_OFFFSETS.read);
let read_channel = (data & 0x0F) as u8;
let data = data >> 4;
// println!(
// "Got data from mailbox: {:#8x} (from channel {})",
// data, read_channel
// );
if read_channel != channel {
// println!("Wrong channel, trying again...");
if limit == 0 {
panic!(
"Got trampled too many times when reading from mailbox (channel {})",
channel
);
}
limit -= 1;
continue;
}
return data;
}
}
unsafe fn write_mailbox(channel: u8, data: u32) {
// 1. Read the status register until the full flag is not set.
// 2. Write the data (shifted into the upper 28 bits) combined with the
// channel (in the lower four bits) to the write register.
// println!("Writing {:#8x} to mailbox channel {}", data, channel);
let mut limit = 10;
loop {
// Wait for space
fence(Ordering::SeqCst);
if read_reg(MAIL_BASE, MAILBOX_OFFFSETS.status + 0x20) & MAIL_FULL == 0 {
break;
}
if limit == 0 |
limit -= 1;
}
write_reg(MAIL_BASE, MAILBOX_OFFFSETS.write, data | (channel as u32));
fence(Ordering::SeqCst);
// println!("Finished writing to mailbox");
}
pub trait PropertyTagList: Sized {
fn prepare(self) -> PropertyMessageWrapper<Self> {
PropertyMessageWrapper::new(self)
}
}
macro_rules! impl_ptl {
( $( $t:ident ),+ ) => {
impl< $($t),+ > PropertyTagList for ( $(PropertyMessage< $t >, )+ )
where $(
$t: Sized,
)+ {}
};
}
impl<T: Sized> PropertyTagList for PropertyMessage<T> {}
// impl_ptl!(T1);
// impl_ptl!(T1, T2);
// impl_ptl!(T1, T2, T3);
// impl_ptl!(T1, T2, T3, T4);
// impl_ptl!(T1, T2, T3, T4, T5);
// impl_ptl!(T1, T2, T3, T4, T5, T6);
// impl_ptl!(T1, T2, T3, T4, T5, T6, T7);
// impl_ptl!(T1, T2, T3, T4, T5, T6, T7, T8);
// impl_ptl!(T1, T2, T3, T4, T5, T6, T7, T8, T9);
// impl_ptl!(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10);
#[repr(C, align(16))]
#[derive(Debug)]
pub struct PropertyMessageWrapper<TL: PropertyTagList> {
buffer_size: u32,
code: u32,
tags: TL,
// extra: [u32; 10],
}
impl<TL: PropertyTagList> PropertyMessageWrapper<TL> {
#[inline]
fn new(tags: TL) -> Self {
// let extra_offset = offset_of!(Self => extra).get_byte_offset();
// assert!(extra_offset % 4 == 0);
let buffer_size = core::mem::size_of::<Self>();
PropertyMessageWrapper {
buffer_size: buffer_size
.try_into()
.expect("Property message list size in bytes is too big to fit in a u32"),
code: 0x0000_0000,
tags,
// extra: [0; 10],
}
}
fn as_quads(&self) -> &[u32] {
let size_bytes = mem::size_of::<Self>();
debug_assert_eq!(size_bytes % 4, 0);
let u32_size: usize = size_bytes / 4;
unsafe { slice::from_raw_parts((self as *const Self) as *const u32, u32_size) }
}
pub fn send<'a>(&'a mut self) -> Option<&'a TL>
where
TL: fmt::Debug,
{
// println!("Property message before sending over mailbox: {:#x?}", self);
// println!(
// "Property message quads before sending over mailbox: {:#x?}",
// self.as_quads()
// );
const CHANNEL: u8 = Channel::PropertyTagsSend as u8;
println!("sending message {:x?}", self);
unsafe {
let ptr = self as *const Self;
let addr = ptr as usize;
write_mailbox(CHANNEL, addr.try_into().ok()?);
let resp_addr = read_mailbox(CHANNEL);
}
// let resp_ptr = resp_addr as *const u32;
// println!("Got response from mailbox: {:#?}", &*resp_ptr);
// let resp_code: u32 = *resp_ptr.offset(1);
// println!(
// "Property message after response {:#8x}: {:#x?}",
// resp_addr, self
// );
// {
// let message_quads = self.as_quads();
// println!("Property message words: {:#x?}", message_quads);
// }
if self.code != 0x8000_0000 {
return None;
}
// let msg_ptr = resp_ptr.offset(2);
// let value_buffer_size_ptr = msg_ptr.offset(1);
// let value_buffer_size = (*value_buffer_size_ptr) as usize;
// let value_buffer_ptr = msg_ptr.offset(3) as *const T;
// assert_eq!(value_buffer_size, mem::size_of::<T>());
// let value_ref = &*(value_buffer_ptr as *const T);
// Some(value_ref)
println!("received message: {:#x?}", self);
Some(&self.tags)
}
}
impl<TL: PropertyTagList> ops::Deref for PropertyMessageWrapper<TL> {
type Target = TL;
fn deref(&self) -> &TL {
&self.tags
}
}
impl<TL: PropertyTagList> ops::DerefMut for PropertyMessageWrapper<TL> {
fn deref_mut(&mut self) -> &mut TL {
&mut self.tags
}
}
#[repr(C, align(4))]
#[derive(Debug | {
panic!(
"Gave up waiting for space to write to mailbox (channel {}, data: 0x{:08x})",
channel, data
);
} | conditional_block |
mailbox.rs | println!(
// "Got data from mailbox: {:#8x} (from channel {})",
// data, read_channel
// );
if read_channel != channel {
// println!("Wrong channel, trying again...");
if limit == 0 {
panic!(
"Got trampled too many times when reading from mailbox (channel {})",
channel
);
}
limit -= 1;
continue;
}
return data;
}
}
unsafe fn write_mailbox(channel: u8, data: u32) {
// 1. Read the status register until the full flag is not set.
// 2. Write the data (shifted into the upper 28 bits) combined with the
// channel (in the lower four bits) to the write register.
// println!("Writing {:#8x} to mailbox channel {}", data, channel);
let mut limit = 10;
loop {
// Wait for space
fence(Ordering::SeqCst);
if read_reg(MAIL_BASE, MAILBOX_OFFFSETS.status + 0x20) & MAIL_FULL == 0 {
break;
}
if limit == 0 {
panic!(
"Gave up waiting for space to write to mailbox (channel {}, data: 0x{:08x})",
channel, data
);
}
limit -= 1;
}
write_reg(MAIL_BASE, MAILBOX_OFFFSETS.write, data | (channel as u32));
fence(Ordering::SeqCst);
// println!("Finished writing to mailbox");
}
pub trait PropertyTagList: Sized {
fn prepare(self) -> PropertyMessageWrapper<Self> {
PropertyMessageWrapper::new(self)
}
}
macro_rules! impl_ptl {
( $( $t:ident ),+ ) => {
impl< $($t),+ > PropertyTagList for ( $(PropertyMessage< $t >, )+ )
where $(
$t: Sized,
)+ {}
};
}
impl<T: Sized> PropertyTagList for PropertyMessage<T> {}
// impl_ptl!(T1);
// impl_ptl!(T1, T2);
// impl_ptl!(T1, T2, T3);
// impl_ptl!(T1, T2, T3, T4);
// impl_ptl!(T1, T2, T3, T4, T5);
// impl_ptl!(T1, T2, T3, T4, T5, T6);
// impl_ptl!(T1, T2, T3, T4, T5, T6, T7);
// impl_ptl!(T1, T2, T3, T4, T5, T6, T7, T8);
// impl_ptl!(T1, T2, T3, T4, T5, T6, T7, T8, T9);
// impl_ptl!(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10);
#[repr(C, align(16))]
#[derive(Debug)]
pub struct PropertyMessageWrapper<TL: PropertyTagList> {
buffer_size: u32,
code: u32,
tags: TL,
// extra: [u32; 10],
}
impl<TL: PropertyTagList> PropertyMessageWrapper<TL> {
#[inline]
fn new(tags: TL) -> Self {
// let extra_offset = offset_of!(Self => extra).get_byte_offset();
// assert!(extra_offset % 4 == 0);
let buffer_size = core::mem::size_of::<Self>();
PropertyMessageWrapper {
buffer_size: buffer_size
.try_into()
.expect("Property message list size in bytes is too big to fit in a u32"),
code: 0x0000_0000,
tags,
// extra: [0; 10],
}
}
fn as_quads(&self) -> &[u32] {
let size_bytes = mem::size_of::<Self>();
debug_assert_eq!(size_bytes % 4, 0);
let u32_size: usize = size_bytes / 4;
unsafe { slice::from_raw_parts((self as *const Self) as *const u32, u32_size) }
}
pub fn send<'a>(&'a mut self) -> Option<&'a TL>
where
TL: fmt::Debug,
{
// println!("Property message before sending over mailbox: {:#x?}", self);
// println!(
// "Property message quads before sending over mailbox: {:#x?}",
// self.as_quads()
// );
const CHANNEL: u8 = Channel::PropertyTagsSend as u8;
println!("sending message {:x?}", self);
unsafe {
let ptr = self as *const Self;
let addr = ptr as usize;
write_mailbox(CHANNEL, addr.try_into().ok()?);
let resp_addr = read_mailbox(CHANNEL);
}
// let resp_ptr = resp_addr as *const u32;
// println!("Got response from mailbox: {:#?}", &*resp_ptr);
// let resp_code: u32 = *resp_ptr.offset(1);
// println!(
// "Property message after response {:#8x}: {:#x?}",
// resp_addr, self
// );
// {
// let message_quads = self.as_quads();
// println!("Property message words: {:#x?}", message_quads);
// }
if self.code != 0x8000_0000 {
return None;
}
// let msg_ptr = resp_ptr.offset(2);
// let value_buffer_size_ptr = msg_ptr.offset(1);
// let value_buffer_size = (*value_buffer_size_ptr) as usize;
// let value_buffer_ptr = msg_ptr.offset(3) as *const T;
// assert_eq!(value_buffer_size, mem::size_of::<T>());
// let value_ref = &*(value_buffer_ptr as *const T);
// Some(value_ref)
println!("received message: {:#x?}", self);
Some(&self.tags)
}
}
impl<TL: PropertyTagList> ops::Deref for PropertyMessageWrapper<TL> {
type Target = TL;
fn deref(&self) -> &TL {
&self.tags
}
}
impl<TL: PropertyTagList> ops::DerefMut for PropertyMessageWrapper<TL> {
fn deref_mut(&mut self) -> &mut TL {
&mut self.tags
}
}
#[repr(C, align(4))]
#[derive(Debug)]
pub struct PropertyMessage<T> {
tag: u32,
buffer_size: u32,
code: u32,
buffer: T,
}
impl<T> ops::Deref for PropertyMessage<T> {
type Target = T;
fn deref(&self) -> &T {
&self.buffer
}
}
impl<T> ops::DerefMut for PropertyMessage<T> {
fn deref_mut(&mut self) -> &mut T {
&mut self.buffer
}
}
impl<T: Sized> From<(u32, T)> for PropertyMessage<T> {
fn from((tag, buffer): (u32, T)) -> PropertyMessage<T> {
PropertyMessage::new(tag, buffer)
}
}
impl<T> PropertyMessage<T> {
pub fn new(tag: u32, buffer: T) -> Self {
let buffer_size = align_up(mem::size_of::<T>(), 2)
.try_into()
.expect("Property message size is too big to fit in a u32");
let msg = PropertyMessage {
tag,
buffer_size,
code: 0,
buffer,
};
let tag_offset = offset_of!(Self => tag);
assert_eq!(tag_offset.get_byte_offset(), 0);
let size_offset = offset_of!(Self => buffer_size);
assert_eq!(size_offset.get_byte_offset(), 4);
let code_offset = offset_of!(Self => code);
assert_eq!(code_offset.get_byte_offset(), 8);
let buffer_offset = offset_of!(Self => buffer);
assert_eq!(buffer_offset.get_byte_offset(), 12);
msg
}
pub fn value(&self) -> &T {
&self.buffer
}
}
// impl<T: fmt::Debug> PropertyMessage<T> {
// pub fn new(tag: u32, buffer: T) -> Self {
// PropertyMessage {
// }
// }
// }
pub fn send_raw_message<T: fmt::Debug>(channel: Channel, msg: &mut T) -> Result<u32, ()> | {
let resp: u32;
let msg_ptr = msg as *mut T;
let msg_addr_usize = msg_ptr as usize;
let msg_addr_u32 = msg_addr_usize.try_into().map_err(|_| ())?;
unsafe {
write_mailbox(channel as u8, msg_addr_u32);
resp = read_mailbox(channel as u8);
}
// println!(
// "Got response {:#8x} after raw message send: {:#x?}",
// resp, msg
// );
Ok(resp)
} | identifier_body |
|
mailbox.rs | x18,
config: 0x1c,
write: 0x20,
};
// MailboxRegisterOffsets {
// read: 0x20,
// peek: 0x30,
// sender: 0x34,
// status: 0x38,
// config: 0x3c,
// write: 0x40,
// },
// ];
#[inline]
unsafe fn read_reg(base: usize, offset: u8) -> u32 {
((MAPPED_REGISTERS_BASE + base + offset as usize) as *const u32).read_volatile()
}
#[inline]
unsafe fn write_reg(base: usize, offset: u8, value: u32) {
((MAPPED_REGISTERS_BASE + base + offset as usize) as *mut u32).write_volatile(value)
}
unsafe fn read_mailbox(channel: u8) -> u32 {
// 1. Read the status register until the empty flag is not set.
// 2. Read data from the read register.
// 3. If the lower four bits do not match the channel number desired repeat
// from 1.
// 4. The upper 28 bits are the returned data.
// Wait for the mailbox to be non-empty
// Execute a memory barrier
// Read MAIL0_STATUS
// Goto step 1 if MAIL_EMPTY bit is set | // Check the channel (lowest 4 bits) of the read value for the correct channel
// If the channel is not the one we wish to read from (i.e: 1), go to step 1
// Return the data (i.e: the read value >> 4)
// println!("Reading mailbox (want channel {})", channel);
let mut limit = 10;
loop {
let mut empty_limit = 10;
loop {
fence(Ordering::SeqCst);
if read_reg(MAIL_BASE, MAILBOX_OFFFSETS.status) & MAIL_EMPTY == 0 {
break;
}
if empty_limit == 0 {
panic!(
"Gave up waiting for mail when reading from mailbox (channel {})",
channel
);
}
empty_limit -= 1;
}
fence(Ordering::SeqCst);
let data: u32 = read_reg(MAIL_BASE, MAILBOX_OFFFSETS.read);
let read_channel = (data & 0x0F) as u8;
let data = data >> 4;
// println!(
// "Got data from mailbox: {:#8x} (from channel {})",
// data, read_channel
// );
if read_channel != channel {
// println!("Wrong channel, trying again...");
if limit == 0 {
panic!(
"Got trampled too many times when reading from mailbox (channel {})",
channel
);
}
limit -= 1;
continue;
}
return data;
}
}
unsafe fn write_mailbox(channel: u8, data: u32) {
// 1. Read the status register until the full flag is not set.
// 2. Write the data (shifted into the upper 28 bits) combined with the
// channel (in the lower four bits) to the write register.
// println!("Writing {:#8x} to mailbox channel {}", data, channel);
let mut limit = 10;
loop {
// Wait for space
fence(Ordering::SeqCst);
if read_reg(MAIL_BASE, MAILBOX_OFFFSETS.status + 0x20) & MAIL_FULL == 0 {
break;
}
if limit == 0 {
panic!(
"Gave up waiting for space to write to mailbox (channel {}, data: 0x{:08x})",
channel, data
);
}
limit -= 1;
}
write_reg(MAIL_BASE, MAILBOX_OFFFSETS.write, data | (channel as u32));
fence(Ordering::SeqCst);
// println!("Finished writing to mailbox");
}
pub trait PropertyTagList: Sized {
fn prepare(self) -> PropertyMessageWrapper<Self> {
PropertyMessageWrapper::new(self)
}
}
macro_rules! impl_ptl {
( $( $t:ident ),+ ) => {
impl< $($t),+ > PropertyTagList for ( $(PropertyMessage< $t >, )+ )
where $(
$t: Sized,
)+ {}
};
}
impl<T: Sized> PropertyTagList for PropertyMessage<T> {}
// impl_ptl!(T1);
// impl_ptl!(T1, T2);
// impl_ptl!(T1, T2, T3);
// impl_ptl!(T1, T2, T3, T4);
// impl_ptl!(T1, T2, T3, T4, T5);
// impl_ptl!(T1, T2, T3, T4, T5, T6);
// impl_ptl!(T1, T2, T3, T4, T5, T6, T7);
// impl_ptl!(T1, T2, T3, T4, T5, T6, T7, T8);
// impl_ptl!(T1, T2, T3, T4, T5, T6, T7, T8, T9);
// impl_ptl!(T1, T2, T3, T4, T5, T6, T7, T8, T9, T10);
#[repr(C, align(16))]
#[derive(Debug)]
pub struct PropertyMessageWrapper<TL: PropertyTagList> {
buffer_size: u32,
code: u32,
tags: TL,
// extra: [u32; 10],
}
impl<TL: PropertyTagList> PropertyMessageWrapper<TL> {
#[inline]
fn new(tags: TL) -> Self {
// let extra_offset = offset_of!(Self => extra).get_byte_offset();
// assert!(extra_offset % 4 == 0);
let buffer_size = core::mem::size_of::<Self>();
PropertyMessageWrapper {
buffer_size: buffer_size
.try_into()
.expect("Property message list size in bytes is too big to fit in a u32"),
code: 0x0000_0000,
tags,
// extra: [0; 10],
}
}
fn as_quads(&self) -> &[u32] {
let size_bytes = mem::size_of::<Self>();
debug_assert_eq!(size_bytes % 4, 0);
let u32_size: usize = size_bytes / 4;
unsafe { slice::from_raw_parts((self as *const Self) as *const u32, u32_size) }
}
pub fn send<'a>(&'a mut self) -> Option<&'a TL>
where
TL: fmt::Debug,
{
// println!("Property message before sending over mailbox: {:#x?}", self);
// println!(
// "Property message quads before sending over mailbox: {:#x?}",
// self.as_quads()
// );
const CHANNEL: u8 = Channel::PropertyTagsSend as u8;
println!("sending message {:x?}", self);
unsafe {
let ptr = self as *const Self;
let addr = ptr as usize;
write_mailbox(CHANNEL, addr.try_into().ok()?);
let resp_addr = read_mailbox(CHANNEL);
}
// let resp_ptr = resp_addr as *const u32;
// println!("Got response from mailbox: {:#?}", &*resp_ptr);
// let resp_code: u32 = *resp_ptr.offset(1);
// println!(
// "Property message after response {:#8x}: {:#x?}",
// resp_addr, self
// );
// {
// let message_quads = self.as_quads();
// println!("Property message words: {:#x?}", message_quads);
// }
if self.code != 0x8000_0000 {
return None;
}
// let msg_ptr = resp_ptr.offset(2);
// let value_buffer_size_ptr = msg_ptr.offset(1);
// let value_buffer_size = (*value_buffer_size_ptr) as usize;
// let value_buffer_ptr = msg_ptr.offset(3) as *const T;
// assert_eq!(value_buffer_size, mem::size_of::<T>());
// let value_ref = &*(value_buffer_ptr as *const T);
// Some(value_ref)
println!("received message: {:#x?}", self);
Some(&self.tags)
}
}
impl<TL: PropertyTagList> ops::Deref for PropertyMessageWrapper<TL> {
type Target = TL;
fn deref(&self) -> &TL {
&self.tags
}
}
impl<TL: PropertyTagList> ops::DerefMut for PropertyMessageWrapper<TL> {
fn deref_mut(&mut self) -> &mut TL {
&mut self.tags
}
}
#[repr(C, align(4))]
#[derive(Debug)]
pub struct | // Execute a memory barrier
// Read from MAIL0_READ | random_line_split |
internal.rs | ![,]>,
}
// Like Path::parse_mod_style but accepts keywords in the path.
fn parse_meta_path(input: ParseStream) -> syn::Result<syn::Path> {
Ok(syn::Path {
leading_colon: input.parse()?,
segments: {
let mut segments = syn::punctuated::Punctuated::new();
while input.peek(syn::Ident::peek_any) {
let ident = syn::Ident::parse_any(input)?;
segments.push_value(syn::PathSegment::from(ident));
if !input.peek(syn::Token![::]) {
break
}
let punct = input.parse()?;
segments.push_punct(punct);
}
if segments.is_empty() {
return Err(input.error("expected path"))
} else if segments.trailing_punct() {
return Err(input.error("expected path segment"))
}
segments
},
})
}
fn parse_meta_list_after_path(path: syn::Path, input: ParseStream) -> syn::Result<MetaList> {
let content;
Ok(MetaList {
path,
_paren_token: parenthesized!(content in input),
nested: content.parse_terminated(TokenStream2::parse)?,
})
}
fn parse_meta_after_path(path: syn::Path, input: ParseStream) -> syn::Result<NestedMeta> {
if input.peek(syn::token::Paren) {
parse_meta_list_after_path(path, input).map(NestedMeta::List)
} else {
Ok(NestedMeta::Path(path))
}
}
impl Parse for MetaList {
fn parse(input: ParseStream) -> syn::Result<Self> {
let path = input.call(parse_meta_path)?;
parse_meta_list_after_path(path, input)
}
}
pub(crate) enum NestedMeta {
Path(syn::Path),
List(MetaList),
}
impl Parse for NestedMeta {
fn parse(input: ParseStream) -> syn::Result<Self> {
let path = input.call(parse_meta_path)?;
parse_meta_after_path(path, input)
}
}
pub(crate) struct AttributeArgs(Vec<NestedMeta>);
impl Parse for AttributeArgs {
fn parse(input: ParseStream) -> syn::Result<Self> {
let mut attrs = Vec::new();
while input.peek(syn::Ident::peek_any) {
attrs.push(input.parse()?);
if input.is_empty() {
break
}
let _: syn::token::Comma = input.parse()?;
}
Ok(AttributeArgs { 0: attrs })
}
}
impl std::ops::Deref for AttributeArgs {
type Target = Vec<NestedMeta>;
fn deref(&self) -> &Self::Target {
&self.0
}
}
impl std::ops::DerefMut for AttributeArgs {
fn deref_mut(&mut self) -> &mut Vec<NestedMeta> {
&mut self.0
}
}
pub(crate) struct Attributes(Vec<syn::Attribute>);
impl Parse for Attributes {
fn parse(input: ParseStream) -> syn::Result<Self> {
Ok(Self(syn::Attribute::parse_outer(input)?))
}
}
impl Attributes {
pub(crate) fn attr(&self) -> &Vec<syn::Attribute> {
&self.0
}
}
// Returns "ink-as-dependency" and not("ink-as-dependency") impls
pub(crate) fn impl_external_trait(
mut impl_item: syn::ItemImpl,
trait_ident: &syn::Ident,
metadata: &Metadata,
) -> (Vec<syn::Item>, Vec<syn::Item>) {
let impl_ink_attrs = extract_attr(&mut impl_item.attrs, "ink");
let mut ink_methods: HashMap<String, syn::TraitItemMethod> = HashMap::new();
metadata
.external_traits
.get(&trait_ident.to_string())
.methods()
.iter()
.for_each(|method| {
if is_attr(&method.attrs, "ink") {
let mut empty_method = method.clone();
empty_method.default = Some(
syn::parse2(quote! {
{
unimplemented!()
}
})
.unwrap(),
);
let mut attrs = empty_method.attrs.clone();
empty_method.attrs = extract_attr(&mut attrs, "doc");
empty_method.attrs.append(&mut extract_attr(&mut attrs, "ink"));
ink_methods.insert(method.sig.ident.to_string(), empty_method);
}
});
// Move ink! attrs from internal trait to external
impl_item.items.iter_mut().for_each(|mut item| {
if let syn::ImplItem::Method(method) = &mut item {
let method_key = method.sig.ident.to_string();
if ink_methods.contains_key(&method_key) {
// Internal attrs will override external, so user must include full declaration with ink(message) and etc.
ink_methods.get_mut(&method_key).unwrap().attrs = extract_attr(&mut method.attrs, "doc");
ink_methods
.get_mut(&method_key)
.unwrap()
.attrs
.append(&mut extract_attr(&mut method.attrs, "ink"));
}
}
});
let ink_methods_iter = ink_methods.iter().map(|(_, value)| value);
let self_ty = impl_item.self_ty.clone().as_ref().clone();
let draft_impl: ItemImpl = syn::parse2(quote! {
#(#impl_ink_attrs)*
impl #trait_ident for #self_ty {
#(#ink_methods_iter)*
}
})
.unwrap();
// Evaluate selector and metadata_name for each method based on rules in ink!
let ink_impl = ::ink_lang_ir::ItemImpl::try_from(draft_impl).unwrap();
ink_impl.iter_messages().for_each(|message| {
let method = ink_methods.get_mut(&message.ident().to_string()).unwrap();
if message.user_provided_selector().is_none() {
let selector_u32 = u32::from_be_bytes(message.composed_selector().as_bytes().clone()) as usize;
let selector = format!("{:#010x}", selector_u32);
method.attrs.push(new_attribute(quote! {#[ink(selector = #selector)]}));
}
if message.metadata_name() == message.ident().to_string() {
let selector = format!("{}", message.metadata_name());
method
.attrs
.push(new_attribute(quote! {#[ink(metadata_name = #selector)]}));
}
let original_name = message.ident();
let inputs_params = message.inputs().map(|pat_type| &pat_type.pat);
method.default = Some(
syn::parse2(quote! {
{
#trait_ident::#original_name(self #(, #inputs_params )* )
}
})
.unwrap(),
);
});
let ink_methods_iter = ink_methods.iter().map(|(_, value)| value);
let wrapper_trait_ident = format_ident!("{}_{}{}", BRUSH_PREFIX, trait_ident, WRAPPER_TRAIT_SUFFIX);
// We only want to use this implementation in case when ink-as-dependency for wrapper.
// It will provide methods with the same name like in initial trait.
let wrapper_impl: ItemImpl = syn::parse2(quote! {
#(#impl_ink_attrs)*
impl #wrapper_trait_ident for #self_ty {
#(#ink_methods_iter)*
}
})
.unwrap();
let trait_name = ink_impl
.trait_path()
.map(|path| path.segments.last().unwrap().ident.to_string());
let mut metadata_name_attr = quote! {};
if trait_name == ink_impl.trait_metadata_name() |
let external_ink_methods_iter = ink_methods.iter_mut().map(|(_, value)| {
value.sig.ident = format_ident!("{}_{}{}", BRUSH_PREFIX, value.sig.ident, EXTERNAL_METHOD_SUFFIX);
value
});
let external_trait_ident = format_ident!("{}_{}{}", BRUSH_PREFIX, trait_ident, EXTERNAL_TRAIT_SUFFIX);
// It is implementation of "external" trait(trait where all method marked with ink!)
// This trait has another name with external suffix. And all methods have external signature.
// But ABI generated by this impl section is the same as ABI generated by original trait.
let external_impl: ItemImpl = syn::parse2(quote! {
#metadata_name_attr
#(#impl_ink_attrs)*
impl #external_trait_ident for #self_ty {
#(#external_ink_methods_iter)*
}
})
.unwrap();
// Internal implementation must be disable during "ink-as-dependency"
let internal_impl = impl_item;
(
vec![syn::Item::from(wrapper_impl)],
vec![syn::Item::from(internal_impl), syn::Item::from(external_impl)],
)
}
#[inline]
pub(crate) fn is_attr(attrs: &Vec<syn::Attribute>, ident: &str) -> bool {
if let None = attrs
.iter()
.find(|attr| attr.path.segments.last().expect("No segments in path").ident == ident)
{
false
} else {
true
}
}
#[inline]
#[allow(dead_code)]
pub(crate) fn get_attr(attrs: &Vec<syn::Attribute>, ident: &str) -> Option<syn::Attribute> {
for attr in attrs.iter() {
if is_attr(&vec![attr.clone()], ident) {
return Some(attr.clone())
}
}
| {
let name = format!("{}", trait_name.unwrap());
metadata_name_attr = quote! { #[ink(metadata_name = #name)] }
} | conditional_block |
internal.rs | Parse,
ParseStream,
},
ItemImpl,
};
use crate::{
metadata::Metadata,
trait_definition::{
EXTERNAL_METHOD_SUFFIX,
EXTERNAL_TRAIT_SUFFIX,
WRAPPER_TRAIT_SUFFIX,
},
};
pub(crate) const BRUSH_PREFIX: &'static str = "__brush";
pub(crate) struct MetaList {
pub path: syn::Path,
pub _paren_token: syn::token::Paren,
pub nested: syn::punctuated::Punctuated<TokenStream2, syn::Token![,]>,
}
// Like Path::parse_mod_style but accepts keywords in the path.
fn parse_meta_path(input: ParseStream) -> syn::Result<syn::Path> {
Ok(syn::Path {
leading_colon: input.parse()?,
segments: {
let mut segments = syn::punctuated::Punctuated::new();
while input.peek(syn::Ident::peek_any) {
let ident = syn::Ident::parse_any(input)?;
segments.push_value(syn::PathSegment::from(ident));
if !input.peek(syn::Token![::]) {
break
}
let punct = input.parse()?;
segments.push_punct(punct);
}
if segments.is_empty() {
return Err(input.error("expected path"))
} else if segments.trailing_punct() {
return Err(input.error("expected path segment"))
}
segments
},
})
}
fn parse_meta_list_after_path(path: syn::Path, input: ParseStream) -> syn::Result<MetaList> {
let content;
Ok(MetaList {
path,
_paren_token: parenthesized!(content in input),
nested: content.parse_terminated(TokenStream2::parse)?,
})
}
fn parse_meta_after_path(path: syn::Path, input: ParseStream) -> syn::Result<NestedMeta> {
if input.peek(syn::token::Paren) {
parse_meta_list_after_path(path, input).map(NestedMeta::List)
} else {
Ok(NestedMeta::Path(path))
}
}
impl Parse for MetaList {
fn parse(input: ParseStream) -> syn::Result<Self> {
let path = input.call(parse_meta_path)?;
parse_meta_list_after_path(path, input)
}
}
pub(crate) enum NestedMeta {
Path(syn::Path),
List(MetaList),
}
impl Parse for NestedMeta {
fn parse(input: ParseStream) -> syn::Result<Self> {
let path = input.call(parse_meta_path)?;
parse_meta_after_path(path, input)
}
}
pub(crate) struct AttributeArgs(Vec<NestedMeta>);
impl Parse for AttributeArgs {
fn parse(input: ParseStream) -> syn::Result<Self> {
let mut attrs = Vec::new();
while input.peek(syn::Ident::peek_any) {
attrs.push(input.parse()?);
if input.is_empty() {
break
}
let _: syn::token::Comma = input.parse()?;
}
Ok(AttributeArgs { 0: attrs })
}
}
impl std::ops::Deref for AttributeArgs {
type Target = Vec<NestedMeta>;
fn deref(&self) -> &Self::Target {
&self.0
}
}
impl std::ops::DerefMut for AttributeArgs {
fn deref_mut(&mut self) -> &mut Vec<NestedMeta> {
&mut self.0
}
}
pub(crate) struct Attributes(Vec<syn::Attribute>);
impl Parse for Attributes {
fn parse(input: ParseStream) -> syn::Result<Self> {
Ok(Self(syn::Attribute::parse_outer(input)?))
}
}
impl Attributes {
pub(crate) fn attr(&self) -> &Vec<syn::Attribute> {
&self.0
}
}
// Returns "ink-as-dependency" and not("ink-as-dependency") impls
pub(crate) fn impl_external_trait(
mut impl_item: syn::ItemImpl,
trait_ident: &syn::Ident,
metadata: &Metadata,
) -> (Vec<syn::Item>, Vec<syn::Item>) {
let impl_ink_attrs = extract_attr(&mut impl_item.attrs, "ink");
let mut ink_methods: HashMap<String, syn::TraitItemMethod> = HashMap::new();
metadata
.external_traits
.get(&trait_ident.to_string())
.methods()
.iter()
.for_each(|method| {
if is_attr(&method.attrs, "ink") {
let mut empty_method = method.clone();
empty_method.default = Some(
syn::parse2(quote! {
{
unimplemented!()
}
})
.unwrap(),
);
let mut attrs = empty_method.attrs.clone();
empty_method.attrs = extract_attr(&mut attrs, "doc");
empty_method.attrs.append(&mut extract_attr(&mut attrs, "ink"));
ink_methods.insert(method.sig.ident.to_string(), empty_method);
}
});
// Move ink! attrs from internal trait to external
impl_item.items.iter_mut().for_each(|mut item| {
if let syn::ImplItem::Method(method) = &mut item {
let method_key = method.sig.ident.to_string();
if ink_methods.contains_key(&method_key) {
// Internal attrs will override external, so user must include full declaration with ink(message) and etc.
ink_methods.get_mut(&method_key).unwrap().attrs = extract_attr(&mut method.attrs, "doc");
ink_methods
.get_mut(&method_key)
.unwrap()
.attrs
.append(&mut extract_attr(&mut method.attrs, "ink"));
}
}
});
let ink_methods_iter = ink_methods.iter().map(|(_, value)| value);
let self_ty = impl_item.self_ty.clone().as_ref().clone();
let draft_impl: ItemImpl = syn::parse2(quote! {
#(#impl_ink_attrs)*
impl #trait_ident for #self_ty {
#(#ink_methods_iter)*
}
})
.unwrap();
// Evaluate selector and metadata_name for each method based on rules in ink!
let ink_impl = ::ink_lang_ir::ItemImpl::try_from(draft_impl).unwrap();
ink_impl.iter_messages().for_each(|message| {
let method = ink_methods.get_mut(&message.ident().to_string()).unwrap();
if message.user_provided_selector().is_none() {
let selector_u32 = u32::from_be_bytes(message.composed_selector().as_bytes().clone()) as usize;
let selector = format!("{:#010x}", selector_u32);
method.attrs.push(new_attribute(quote! {#[ink(selector = #selector)]}));
}
if message.metadata_name() == message.ident().to_string() {
let selector = format!("{}", message.metadata_name());
method
.attrs
.push(new_attribute(quote! {#[ink(metadata_name = #selector)]}));
}
let original_name = message.ident();
let inputs_params = message.inputs().map(|pat_type| &pat_type.pat);
method.default = Some(
syn::parse2(quote! {
{
#trait_ident::#original_name(self #(, #inputs_params )* )
}
})
.unwrap(),
);
});
let ink_methods_iter = ink_methods.iter().map(|(_, value)| value);
let wrapper_trait_ident = format_ident!("{}_{}{}", BRUSH_PREFIX, trait_ident, WRAPPER_TRAIT_SUFFIX);
// We only want to use this implementation in case when ink-as-dependency for wrapper.
// It will provide methods with the same name like in initial trait.
let wrapper_impl: ItemImpl = syn::parse2(quote! {
#(#impl_ink_attrs)*
impl #wrapper_trait_ident for #self_ty {
#(#ink_methods_iter)*
}
})
.unwrap();
let trait_name = ink_impl
.trait_path()
.map(|path| path.segments.last().unwrap().ident.to_string());
let mut metadata_name_attr = quote! {};
if trait_name == ink_impl.trait_metadata_name() {
let name = format!("{}", trait_name.unwrap());
metadata_name_attr = quote! { #[ink(metadata_name = #name)] }
}
let external_ink_methods_iter = ink_methods.iter_mut().map(|(_, value)| {
value.sig.ident = format_ident!("{}_{}{}", BRUSH_PREFIX, value.sig.ident, EXTERNAL_METHOD_SUFFIX);
value
});
let external_trait_ident = format_ident!("{}_{}{}", BRUSH_PREFIX, trait_ident, EXTERNAL_TRAIT_SUFFIX);
// It is implementation of "external" trait(trait where all method marked with ink!)
// This trait has another name with external suffix. And all methods have external signature.
// But ABI generated by this impl section is the same as ABI generated by original trait.
let external_impl: ItemImpl = syn::parse2(quote! {
#metadata_name_attr
#(#impl_ink_attrs)*
impl #external_trait_ident for #self_ty {
#(#external_ink_methods_iter)*
}
})
.unwrap();
// Internal implementation must be disable during "ink-as-dependency"
let internal_impl = impl_item;
(
vec![syn::Item::from(wrapper_impl)],
vec![syn::Item::from(internal_impl), syn::Item::from(external_impl)],
)
}
#[inline]
pub(crate) fn is_attr(attrs: &Vec<syn::Attribute>, ident: &str) -> bool {
if let | parse::{ | random_line_split |
|
internal.rs | Token![,]>,
}
// Like Path::parse_mod_style but accepts keywords in the path.
fn parse_meta_path(input: ParseStream) -> syn::Result<syn::Path> {
Ok(syn::Path {
leading_colon: input.parse()?,
segments: {
let mut segments = syn::punctuated::Punctuated::new();
while input.peek(syn::Ident::peek_any) {
let ident = syn::Ident::parse_any(input)?;
segments.push_value(syn::PathSegment::from(ident));
if !input.peek(syn::Token![::]) {
break
}
let punct = input.parse()?;
segments.push_punct(punct);
}
if segments.is_empty() {
return Err(input.error("expected path"))
} else if segments.trailing_punct() {
return Err(input.error("expected path segment"))
}
segments
},
})
}
fn parse_meta_list_after_path(path: syn::Path, input: ParseStream) -> syn::Result<MetaList> {
let content;
Ok(MetaList {
path,
_paren_token: parenthesized!(content in input),
nested: content.parse_terminated(TokenStream2::parse)?,
})
}
fn parse_meta_after_path(path: syn::Path, input: ParseStream) -> syn::Result<NestedMeta> {
if input.peek(syn::token::Paren) {
parse_meta_list_after_path(path, input).map(NestedMeta::List)
} else {
Ok(NestedMeta::Path(path))
}
}
impl Parse for MetaList {
fn | (input: ParseStream) -> syn::Result<Self> {
let path = input.call(parse_meta_path)?;
parse_meta_list_after_path(path, input)
}
}
pub(crate) enum NestedMeta {
Path(syn::Path),
List(MetaList),
}
impl Parse for NestedMeta {
fn parse(input: ParseStream) -> syn::Result<Self> {
let path = input.call(parse_meta_path)?;
parse_meta_after_path(path, input)
}
}
pub(crate) struct AttributeArgs(Vec<NestedMeta>);
impl Parse for AttributeArgs {
fn parse(input: ParseStream) -> syn::Result<Self> {
let mut attrs = Vec::new();
while input.peek(syn::Ident::peek_any) {
attrs.push(input.parse()?);
if input.is_empty() {
break
}
let _: syn::token::Comma = input.parse()?;
}
Ok(AttributeArgs { 0: attrs })
}
}
impl std::ops::Deref for AttributeArgs {
type Target = Vec<NestedMeta>;
fn deref(&self) -> &Self::Target {
&self.0
}
}
impl std::ops::DerefMut for AttributeArgs {
fn deref_mut(&mut self) -> &mut Vec<NestedMeta> {
&mut self.0
}
}
pub(crate) struct Attributes(Vec<syn::Attribute>);
impl Parse for Attributes {
fn parse(input: ParseStream) -> syn::Result<Self> {
Ok(Self(syn::Attribute::parse_outer(input)?))
}
}
impl Attributes {
pub(crate) fn attr(&self) -> &Vec<syn::Attribute> {
&self.0
}
}
// Returns "ink-as-dependency" and not("ink-as-dependency") impls
pub(crate) fn impl_external_trait(
mut impl_item: syn::ItemImpl,
trait_ident: &syn::Ident,
metadata: &Metadata,
) -> (Vec<syn::Item>, Vec<syn::Item>) {
let impl_ink_attrs = extract_attr(&mut impl_item.attrs, "ink");
let mut ink_methods: HashMap<String, syn::TraitItemMethod> = HashMap::new();
metadata
.external_traits
.get(&trait_ident.to_string())
.methods()
.iter()
.for_each(|method| {
if is_attr(&method.attrs, "ink") {
let mut empty_method = method.clone();
empty_method.default = Some(
syn::parse2(quote! {
{
unimplemented!()
}
})
.unwrap(),
);
let mut attrs = empty_method.attrs.clone();
empty_method.attrs = extract_attr(&mut attrs, "doc");
empty_method.attrs.append(&mut extract_attr(&mut attrs, "ink"));
ink_methods.insert(method.sig.ident.to_string(), empty_method);
}
});
// Move ink! attrs from internal trait to external
impl_item.items.iter_mut().for_each(|mut item| {
if let syn::ImplItem::Method(method) = &mut item {
let method_key = method.sig.ident.to_string();
if ink_methods.contains_key(&method_key) {
// Internal attrs will override external, so user must include full declaration with ink(message) and etc.
ink_methods.get_mut(&method_key).unwrap().attrs = extract_attr(&mut method.attrs, "doc");
ink_methods
.get_mut(&method_key)
.unwrap()
.attrs
.append(&mut extract_attr(&mut method.attrs, "ink"));
}
}
});
let ink_methods_iter = ink_methods.iter().map(|(_, value)| value);
let self_ty = impl_item.self_ty.clone().as_ref().clone();
let draft_impl: ItemImpl = syn::parse2(quote! {
#(#impl_ink_attrs)*
impl #trait_ident for #self_ty {
#(#ink_methods_iter)*
}
})
.unwrap();
// Evaluate selector and metadata_name for each method based on rules in ink!
let ink_impl = ::ink_lang_ir::ItemImpl::try_from(draft_impl).unwrap();
ink_impl.iter_messages().for_each(|message| {
let method = ink_methods.get_mut(&message.ident().to_string()).unwrap();
if message.user_provided_selector().is_none() {
let selector_u32 = u32::from_be_bytes(message.composed_selector().as_bytes().clone()) as usize;
let selector = format!("{:#010x}", selector_u32);
method.attrs.push(new_attribute(quote! {#[ink(selector = #selector)]}));
}
if message.metadata_name() == message.ident().to_string() {
let selector = format!("{}", message.metadata_name());
method
.attrs
.push(new_attribute(quote! {#[ink(metadata_name = #selector)]}));
}
let original_name = message.ident();
let inputs_params = message.inputs().map(|pat_type| &pat_type.pat);
method.default = Some(
syn::parse2(quote! {
{
#trait_ident::#original_name(self #(, #inputs_params )* )
}
})
.unwrap(),
);
});
let ink_methods_iter = ink_methods.iter().map(|(_, value)| value);
let wrapper_trait_ident = format_ident!("{}_{}{}", BRUSH_PREFIX, trait_ident, WRAPPER_TRAIT_SUFFIX);
// We only want to use this implementation in case when ink-as-dependency for wrapper.
// It will provide methods with the same name like in initial trait.
let wrapper_impl: ItemImpl = syn::parse2(quote! {
#(#impl_ink_attrs)*
impl #wrapper_trait_ident for #self_ty {
#(#ink_methods_iter)*
}
})
.unwrap();
let trait_name = ink_impl
.trait_path()
.map(|path| path.segments.last().unwrap().ident.to_string());
let mut metadata_name_attr = quote! {};
if trait_name == ink_impl.trait_metadata_name() {
let name = format!("{}", trait_name.unwrap());
metadata_name_attr = quote! { #[ink(metadata_name = #name)] }
}
let external_ink_methods_iter = ink_methods.iter_mut().map(|(_, value)| {
value.sig.ident = format_ident!("{}_{}{}", BRUSH_PREFIX, value.sig.ident, EXTERNAL_METHOD_SUFFIX);
value
});
let external_trait_ident = format_ident!("{}_{}{}", BRUSH_PREFIX, trait_ident, EXTERNAL_TRAIT_SUFFIX);
// It is implementation of "external" trait(trait where all method marked with ink!)
// This trait has another name with external suffix. And all methods have external signature.
// But ABI generated by this impl section is the same as ABI generated by original trait.
let external_impl: ItemImpl = syn::parse2(quote! {
#metadata_name_attr
#(#impl_ink_attrs)*
impl #external_trait_ident for #self_ty {
#(#external_ink_methods_iter)*
}
})
.unwrap();
// Internal implementation must be disable during "ink-as-dependency"
let internal_impl = impl_item;
(
vec![syn::Item::from(wrapper_impl)],
vec![syn::Item::from(internal_impl), syn::Item::from(external_impl)],
)
}
#[inline]
pub(crate) fn is_attr(attrs: &Vec<syn::Attribute>, ident: &str) -> bool {
if let None = attrs
.iter()
.find(|attr| attr.path.segments.last().expect("No segments in path").ident == ident)
{
false
} else {
true
}
}
#[inline]
#[allow(dead_code)]
pub(crate) fn get_attr(attrs: &Vec<syn::Attribute>, ident: &str) -> Option<syn::Attribute> {
for attr in attrs.iter() {
if is_attr(&vec![attr.clone()], ident) {
return Some(attr.clone())
}
}
| parse | identifier_name |
internal.rs | Token![,]>,
}
// Like Path::parse_mod_style but accepts keywords in the path.
fn parse_meta_path(input: ParseStream) -> syn::Result<syn::Path> | },
})
}
fn parse_meta_list_after_path(path: syn::Path, input: ParseStream) -> syn::Result<MetaList> {
let content;
Ok(MetaList {
path,
_paren_token: parenthesized!(content in input),
nested: content.parse_terminated(TokenStream2::parse)?,
})
}
fn parse_meta_after_path(path: syn::Path, input: ParseStream) -> syn::Result<NestedMeta> {
if input.peek(syn::token::Paren) {
parse_meta_list_after_path(path, input).map(NestedMeta::List)
} else {
Ok(NestedMeta::Path(path))
}
}
impl Parse for MetaList {
fn parse(input: ParseStream) -> syn::Result<Self> {
let path = input.call(parse_meta_path)?;
parse_meta_list_after_path(path, input)
}
}
pub(crate) enum NestedMeta {
Path(syn::Path),
List(MetaList),
}
impl Parse for NestedMeta {
fn parse(input: ParseStream) -> syn::Result<Self> {
let path = input.call(parse_meta_path)?;
parse_meta_after_path(path, input)
}
}
pub(crate) struct AttributeArgs(Vec<NestedMeta>);
impl Parse for AttributeArgs {
fn parse(input: ParseStream) -> syn::Result<Self> {
let mut attrs = Vec::new();
while input.peek(syn::Ident::peek_any) {
attrs.push(input.parse()?);
if input.is_empty() {
break
}
let _: syn::token::Comma = input.parse()?;
}
Ok(AttributeArgs { 0: attrs })
}
}
impl std::ops::Deref for AttributeArgs {
type Target = Vec<NestedMeta>;
fn deref(&self) -> &Self::Target {
&self.0
}
}
impl std::ops::DerefMut for AttributeArgs {
fn deref_mut(&mut self) -> &mut Vec<NestedMeta> {
&mut self.0
}
}
pub(crate) struct Attributes(Vec<syn::Attribute>);
impl Parse for Attributes {
fn parse(input: ParseStream) -> syn::Result<Self> {
Ok(Self(syn::Attribute::parse_outer(input)?))
}
}
impl Attributes {
pub(crate) fn attr(&self) -> &Vec<syn::Attribute> {
&self.0
}
}
// Returns "ink-as-dependency" and not("ink-as-dependency") impls
pub(crate) fn impl_external_trait(
mut impl_item: syn::ItemImpl,
trait_ident: &syn::Ident,
metadata: &Metadata,
) -> (Vec<syn::Item>, Vec<syn::Item>) {
let impl_ink_attrs = extract_attr(&mut impl_item.attrs, "ink");
let mut ink_methods: HashMap<String, syn::TraitItemMethod> = HashMap::new();
metadata
.external_traits
.get(&trait_ident.to_string())
.methods()
.iter()
.for_each(|method| {
if is_attr(&method.attrs, "ink") {
let mut empty_method = method.clone();
empty_method.default = Some(
syn::parse2(quote! {
{
unimplemented!()
}
})
.unwrap(),
);
let mut attrs = empty_method.attrs.clone();
empty_method.attrs = extract_attr(&mut attrs, "doc");
empty_method.attrs.append(&mut extract_attr(&mut attrs, "ink"));
ink_methods.insert(method.sig.ident.to_string(), empty_method);
}
});
// Move ink! attrs from internal trait to external
impl_item.items.iter_mut().for_each(|mut item| {
if let syn::ImplItem::Method(method) = &mut item {
let method_key = method.sig.ident.to_string();
if ink_methods.contains_key(&method_key) {
// Internal attrs will override external, so user must include full declaration with ink(message) and etc.
ink_methods.get_mut(&method_key).unwrap().attrs = extract_attr(&mut method.attrs, "doc");
ink_methods
.get_mut(&method_key)
.unwrap()
.attrs
.append(&mut extract_attr(&mut method.attrs, "ink"));
}
}
});
let ink_methods_iter = ink_methods.iter().map(|(_, value)| value);
let self_ty = impl_item.self_ty.clone().as_ref().clone();
let draft_impl: ItemImpl = syn::parse2(quote! {
#(#impl_ink_attrs)*
impl #trait_ident for #self_ty {
#(#ink_methods_iter)*
}
})
.unwrap();
// Evaluate selector and metadata_name for each method based on rules in ink!
let ink_impl = ::ink_lang_ir::ItemImpl::try_from(draft_impl).unwrap();
ink_impl.iter_messages().for_each(|message| {
let method = ink_methods.get_mut(&message.ident().to_string()).unwrap();
if message.user_provided_selector().is_none() {
let selector_u32 = u32::from_be_bytes(message.composed_selector().as_bytes().clone()) as usize;
let selector = format!("{:#010x}", selector_u32);
method.attrs.push(new_attribute(quote! {#[ink(selector = #selector)]}));
}
if message.metadata_name() == message.ident().to_string() {
let selector = format!("{}", message.metadata_name());
method
.attrs
.push(new_attribute(quote! {#[ink(metadata_name = #selector)]}));
}
let original_name = message.ident();
let inputs_params = message.inputs().map(|pat_type| &pat_type.pat);
method.default = Some(
syn::parse2(quote! {
{
#trait_ident::#original_name(self #(, #inputs_params )* )
}
})
.unwrap(),
);
});
let ink_methods_iter = ink_methods.iter().map(|(_, value)| value);
let wrapper_trait_ident = format_ident!("{}_{}{}", BRUSH_PREFIX, trait_ident, WRAPPER_TRAIT_SUFFIX);
// We only want to use this implementation in case when ink-as-dependency for wrapper.
// It will provide methods with the same name like in initial trait.
let wrapper_impl: ItemImpl = syn::parse2(quote! {
#(#impl_ink_attrs)*
impl #wrapper_trait_ident for #self_ty {
#(#ink_methods_iter)*
}
})
.unwrap();
let trait_name = ink_impl
.trait_path()
.map(|path| path.segments.last().unwrap().ident.to_string());
let mut metadata_name_attr = quote! {};
if trait_name == ink_impl.trait_metadata_name() {
let name = format!("{}", trait_name.unwrap());
metadata_name_attr = quote! { #[ink(metadata_name = #name)] }
}
let external_ink_methods_iter = ink_methods.iter_mut().map(|(_, value)| {
value.sig.ident = format_ident!("{}_{}{}", BRUSH_PREFIX, value.sig.ident, EXTERNAL_METHOD_SUFFIX);
value
});
let external_trait_ident = format_ident!("{}_{}{}", BRUSH_PREFIX, trait_ident, EXTERNAL_TRAIT_SUFFIX);
// It is implementation of "external" trait(trait where all method marked with ink!)
// This trait has another name with external suffix. And all methods have external signature.
// But ABI generated by this impl section is the same as ABI generated by original trait.
let external_impl: ItemImpl = syn::parse2(quote! {
#metadata_name_attr
#(#impl_ink_attrs)*
impl #external_trait_ident for #self_ty {
#(#external_ink_methods_iter)*
}
})
.unwrap();
// Internal implementation must be disable during "ink-as-dependency"
let internal_impl = impl_item;
(
vec![syn::Item::from(wrapper_impl)],
vec![syn::Item::from(internal_impl), syn::Item::from(external_impl)],
)
}
#[inline]
pub(crate) fn is_attr(attrs: &Vec<syn::Attribute>, ident: &str) -> bool {
if let None = attrs
.iter()
.find(|attr| attr.path.segments.last().expect("No segments in path").ident == ident)
{
false
} else {
true
}
}
#[inline]
#[allow(dead_code)]
pub(crate) fn get_attr(attrs: &Vec<syn::Attribute>, ident: &str) -> Option<syn::Attribute> {
for attr in attrs.iter() {
if is_attr(&vec![attr.clone()], ident) {
return Some(attr.clone())
}
}
| {
Ok(syn::Path {
leading_colon: input.parse()?,
segments: {
let mut segments = syn::punctuated::Punctuated::new();
while input.peek(syn::Ident::peek_any) {
let ident = syn::Ident::parse_any(input)?;
segments.push_value(syn::PathSegment::from(ident));
if !input.peek(syn::Token![::]) {
break
}
let punct = input.parse()?;
segments.push_punct(punct);
}
if segments.is_empty() {
return Err(input.error("expected path"))
} else if segments.trailing_punct() {
return Err(input.error("expected path segment"))
}
segments | identifier_body |
data.py | Args:
filename: string, path to crf file
pserver: jsonrpc.ServerProxy, stanford corenlp server for parsing
Returns:
tuple, (list_stanford_sent_parses, dict_file_corefs, dict_file_synsets)
"""
parses = []
try:
with open(filename) as f:
vprint('OPEN: %s' % filename)
xml = f.read()
except IOError:
print strerror(EIO)
print("ERROR: Could not open %s" % filename)
return (parses, get_tagged_corefs(''), get_synsets({}))
# remove unwanted characters from xml
vprint('\tPARSE: Parsing file: %s' % filename)
# parse_tries = 0
# while parse_tries < 5:
# try:
# t = loads(pserver.parse(_normalize_sentence(_remove_tags(xml))))
# parse_tries = 0
# break
# except jsonrpc.RPCTimeoutError:
# vprint('\tERROR: RPCTimeoutError - retrying')
# parse_tries += 3
# except jsonrpc.RPCTransportError:
# vprint('\tERROR: RPCTransportError - retrying')
# data = _normalize_sentence(_remove_tags(xml))
# sentences = [sent for part in data.split('\n\n')
# for sent in sent_tokenize(part)]
# try:
# xml1 = data[:data.find(sentences[len(sentences)/3])]
# xml2 = data[data.find(sentences[len(sentences)/3+1]):data.find(sentences[2*len(sentences)/3])]
# xml3 = data[data.find(sentences[2*len(sentences)/3+1]):]
# t1 = loads(pserver.parse(xml1))
# t2 = loads(pserver.parse(xml2))
# t3 = loads(pserver.parse(xml3))
# t = dict(t1.items() + t2.items() + t3.items())
# parse_tries = 0
# break
# except Exception:
# parse_tries = -1
# break
# parse_tries += 1
# if parse_tries != 0:
# vprint('\tFATAL: RPCTransportError - skipping')
sentences = [sent for part in xml.split('\n\n')
for sent in sent_tokenize(part)]
vprint('\tPARSE: Parsing sentences: %s' % filename)
for sent in sentences:
sent_corefs = get_tagged_corefs(sent, ordered=True)
# remove unwanted characters from xml
sent = _normalize_sentence(_remove_tags(sent))
parse_tries = 0
while parse_tries < 5:
try:
sparse = loads(pserver.parse(sent))
parse_tries = 0
break
except jsonrpc.RPCTransportError:
vprint('\tERROR: RPCTransportError - retrying')
parse_tries += 1
if parse_tries != 0:
vprint('\tFATAL: RPCTransportError - skipping')
pparse = _process_parse(sparse, sent_corefs)
if pparse:
parses.append(pparse)
pos_tags = {}
for parse in parses:
for word, attr in parse[1]:
tags = pos_tags.get(word, set())
tags.add(attr['PartOfSpeech'])
pos_tags[word] = tags
return parses, get_tagged_corefs(xml), get_synsets(pos_tags)
def tag_ptree(ptree, coreflist):
"""Tags given parse tree with coreferences
Args:
ptree: string, parenthesized str represenation of parse tree
coreflist: list of tuples, [('1', {'text': 'dog', 'ref': None})]
Returns:
string, tagged parse tree
>>> ptree = '(S NP( (NN He)) VP( (V ran)))'
>>> coreflist = [('1', {'text': 'He', 'ref': None})]
>>> tag_ptree(ptree, coreflist)
'(S NP( COREF_TAG_1( (NN He))) VP( (V ran)))'
"""
pattern = r"""(?P<lp>\(?\s*) # left parenthesis
(?P<tg>[a-zA-Z$]+)? # POS tag
(?P<data>\s*%s) # subtree of tag
(?P<rp>(?:\s*\))*) # right parenthesis
"""
for cid, coref in coreflist[::-1]:
words = ''.join(word_tokenize(coref['text']))
nltktree = Tree.parse(ptree)
nltktree.reverse() # perform search right to left
data = None
for subtree in nltktree.subtrees(): # BFS
if ''.join(subtree.leaves()) == words: # equal ignoring whitespace
data = subtree.pprint()
break
# If found via breadth-first search of parse tree
if data:
ptree = ptree.replace(data, '( COREF_TAG_%s%s)' % (cid, data))
else: # Try finding via regex matching instead
dpattern = r'\s*'.join([r'\(\s*[a-zA-Z$]+\s+%s\s*\)' % word
for word in word_tokenize(coref['text'])])
found = re.findall(pattern % dpattern, ptree, re.X)
if found:
repl = '%s%s ( COREF_TAG_%s%s) %s' % (found[0][0],
found[0][1],
cid,
found[0][2],
found[0][3])
ptree = re.sub(pattern % dpattern, repl, ptree, 1, re.X)
return ptree
def get_tagged_corefs(xml, ordered=False):
"""Parses xml to find all tagged coreferences contained in COREF tags
Args:
xml: string, xml markedup with COREF tags
ordered: if True, returns an list in the same order that the corefs
appear in the text
Returns:
if ordered
list of tuples, [(coref_id, {coref, ref_id}), ]
if not ordered
dict of dict, {coref_id: (coref, ref_id)
>>> text = "<TXT>John stubbed <COREF ID='1'>his</COREF> toe.</TXT>"
>>> get_tagged_corefs(text)
{u'1': {'text': u'his', 'ref': None}}
>>> get_tagged_corefs(text, ordered=True)
[(u'1', {'text': u'his', 'ref': None})]
>>> text = "<TXT><COREF ID='A'>John</COREF> stubbed " +\
"<COREF ID='1' REF='A'>his</COREF> toe.</TXT>"
>>> get_tagged_corefs(text)
{u'A': {'text': u'John', 'ref': None}, u'1': {'text': u'his', 'ref': u'A'}}
>>> get_tagged_corefs(text, ordered=True) # doctest: +NORMALIZE_WHITESPACE
[(u'A', {'text': u'John', 'ref': None}),
(u'1', {'text': u'his', 'ref': u'A'})]
"""
nps = {}
if ordered:
nps = []
xml = _normalize_malformed_xml(xml)
try:
corefs = parseString(xml).getElementsByTagName('COREF')
except ExpatError:
return nps
for coref in corefs:
try:
cid = coref.attributes['ID'].value
if ordered:
data = {}
for npid, np in nps:
if npid == cid:
data = np
break
else:
data = nps.get(cid, {})
except KeyError:
continue
try:
data['ref'] = coref.attributes['REF'].value
except KeyError:
data['ref'] = None
data['text'] = coref.firstChild.data
if ordered:
nps.append((cid, data))
else:
nps[cid] = data
return nps
def cid_less_than(cid1, cid2):
if cid1.isdigit() and cid2.isdigit():
return int(cid1) < int(cid2)
else:
return True
# elif not (cid1.isdigit() or cid2.isdigit()):
# num1 = int(cid1[:-1])
# num2 = int(cid2[:-1])
# if num1 == num2:
# return cid1[-1] < cid2[-1]
# else:
# return num1 < num2
# elif cid1.isdigit():
# return True
# else:
# return False
def _normalize_sentence(sent):
"""Removes unwanted characters from sentence for parsing
Args:
sent: string, sentence to normalize
Returns
string, normalized sentence |
""" | random_line_split |
|
data.py |
stderr.write("\nERROR: Could not open list file\n")
exit(EIO)
else:
|
def get_id(path):
"""Parses a file path for the filename without extension
Args:
path: string, full (or relative) file path for coreference file.
Must end in .crf
Returns:
string, file id (filename without extension)
>>> path = '/home/user/Desktop/full.crf'
>>> get_id(path)
'full'
>>> path = 'relative.crf'
>>> get_id(path)
'relative'
"""
fid, ext, _ = path.strip().split('/')[-1].partition('.crf')
if not fid or ext != '.crf':
filetype = 'Co-Reference Input file'
error = 'has incorrect file type'
raise FilenameException("Error: %s %s" % (filetype, error))
return fid
def mk_fparse(filename, pserver):
"""Parses input to get list of paragraphs with sentence structure
and a dictionary of noun phrases contained in the COREF tags
Args:
filename: string, path to crf file
pserver: jsonrpc.ServerProxy, stanford corenlp server for parsing
Returns:
tuple, (list_stanford_sent_parses, dict_file_corefs, dict_file_synsets)
"""
parses = []
try:
with open(filename) as f:
vprint('OPEN: %s' % filename)
xml = f.read()
except IOError:
print strerror(EIO)
print("ERROR: Could not open %s" % filename)
return (parses, get_tagged_corefs(''), get_synsets({}))
# remove unwanted characters from xml
vprint('\tPARSE: Parsing file: %s' % filename)
# parse_tries = 0
# while parse_tries < 5:
# try:
# t = loads(pserver.parse(_normalize_sentence(_remove_tags(xml))))
# parse_tries = 0
# break
# except jsonrpc.RPCTimeoutError:
# vprint('\tERROR: RPCTimeoutError - retrying')
# parse_tries += 3
# except jsonrpc.RPCTransportError:
# vprint('\tERROR: RPCTransportError - retrying')
# data = _normalize_sentence(_remove_tags(xml))
# sentences = [sent for part in data.split('\n\n')
# for sent in sent_tokenize(part)]
# try:
# xml1 = data[:data.find(sentences[len(sentences)/3])]
# xml2 = data[data.find(sentences[len(sentences)/3+1]):data.find(sentences[2*len(sentences)/3])]
# xml3 = data[data.find(sentences[2*len(sentences)/3+1]):]
# t1 = loads(pserver.parse(xml1))
# t2 = loads(pserver.parse(xml2))
# t3 = loads(pserver.parse(xml3))
# t = dict(t1.items() + t2.items() + t3.items())
# parse_tries = 0
# break
# except Exception:
# parse_tries = -1
# break
# parse_tries += 1
# if parse_tries != 0:
# vprint('\tFATAL: RPCTransportError - skipping')
sentences = [sent for part in xml.split('\n\n')
for sent in sent_tokenize(part)]
vprint('\tPARSE: Parsing sentences: %s' % filename)
for sent in sentences:
sent_corefs = get_tagged_corefs(sent, ordered=True)
# remove unwanted characters from xml
sent = _normalize_sentence(_remove_tags(sent))
parse_tries = 0
while parse_tries < 5:
try:
sparse = loads(pserver.parse(sent))
parse_tries = 0
break
except jsonrpc.RPCTransportError:
vprint('\tERROR: RPCTransportError - retrying')
parse_tries += 1
if parse_tries != 0:
vprint('\tFATAL: RPCTransportError - skipping')
pparse = _process_parse(sparse, sent_corefs)
if pparse:
parses.append(pparse)
pos_tags = {}
for parse in parses:
for word, attr in parse[1]:
tags = pos_tags.get(word, set())
tags.add(attr['PartOfSpeech'])
pos_tags[word] = tags
return parses, get_tagged_corefs(xml), get_synsets(pos_tags)
def tag_ptree(ptree, coreflist):
"""Tags given parse tree with coreferences
Args:
ptree: string, parenthesized str represenation of parse tree
coreflist: list of tuples, [('1', {'text': 'dog', 'ref': None})]
Returns:
string, tagged parse tree
>>> ptree = '(S NP( (NN He)) VP( (V ran)))'
>>> coreflist = [('1', {'text': 'He', 'ref': None})]
>>> tag_ptree(ptree, coreflist)
'(S NP( COREF_TAG_1( (NN He))) VP( (V ran)))'
"""
pattern = r"""(?P<lp>\(?\s*) # left parenthesis
(?P<tg>[a-zA-Z$]+)? # POS tag
(?P<data>\s*%s) # subtree of tag
(?P<rp>(?:\s*\))*) # right parenthesis
"""
for cid, coref in coreflist[::-1]:
words = ''.join(word_tokenize(coref['text']))
nltktree = Tree.parse(ptree)
nltktree.reverse() # perform search right to left
data = None
for subtree in nltktree.subtrees(): # BFS
if ''.join(subtree.leaves()) == words: # equal ignoring whitespace
data = subtree.pprint()
break
# If found via breadth-first search of parse tree
if data:
ptree = ptree.replace(data, '( COREF_TAG_%s%s)' % (cid, data))
else: # Try finding via regex matching instead
dpattern = r'\s*'.join([r'\(\s*[a-zA-Z$]+\s+%s\s*\)' % word
for word in word_tokenize(coref['text'])])
found = re.findall(pattern % dpattern, ptree, re.X)
if found:
repl = '%s%s ( COREF_TAG_%s%s) %s' % (found[0][0],
found[0][1],
cid,
found[0][2],
found[0][3])
ptree = re.sub(pattern % dpattern, repl, ptree, 1, re.X)
return ptree
def get_tagged_corefs(xml, ordered=False):
"""Parses xml to find all tagged coreferences contained in COREF tags
Args:
xml: string, xml markedup with COREF tags
ordered: if True, returns an list in the same order that the corefs
appear in the text
Returns:
if ordered
list of tuples, [(coref_id, {coref, ref_id}), ]
if not ordered
dict of dict, {coref_id: (coref, ref_id)
>>> text = "<TXT>John stubbed <COREF ID='1'>his</COREF> toe.</TXT>"
>>> get_tagged_corefs(text)
{u'1': {'text': u'his', 'ref': None}}
>>> get_tagged_corefs(text, ordered=True)
[(u'1', {'text': u'his', 'ref': None})]
>>> text = "<TXT><COREF ID='A'>John</COREF> stubbed " +\
"<COREF ID='1' REF='A'>his</COREF> toe.</TXT>"
>>> get_tagged_corefs(text)
{u'A': {'text': u'John', 'ref': None}, u'1': {'text': u'his', 'ref': u'A'}}
>>> get_tagged_corefs(text, ordered=True) # doctest: +NORMALIZE_WHITESPACE
[(u'A', {'text': u'John', 'ref': None}),
(u'1', {'text': u'his', 'ref': u'A'})]
"""
nps = {}
if ordered:
nps = []
xml = _normalize_malformed_xml(xml)
try:
corefs = parseString(xml).getElementsByTagName('COREF')
except ExpatError:
return nps
for coref in corefs:
try:
cid = coref.attributes['ID'].value
if ordered:
data = {}
for npid, np in nps:
if npid == cid:
data = np
break
else:
data = nps.get(cid, {})
except KeyError:
continue
try:
data['ref'] = coref.attributes | return parses | conditional_block |
data.py |
coreflist: list of tuples, [('1', {'text': 'dog', 'ref': None})]
Returns:
string, tagged parse tree
>>> ptree = '(S NP( (NN He)) VP( (V ran)))'
>>> coreflist = [('1', {'text': 'He', 'ref': None})]
>>> tag_ptree(ptree, coreflist)
'(S NP( COREF_TAG_1( (NN He))) VP( (V ran)))'
"""
pattern = r"""(?P<lp>\(?\s*) # left parenthesis
(?P<tg>[a-zA-Z$]+)? # POS tag
(?P<data>\s*%s) # subtree of tag
(?P<rp>(?:\s*\))*) # right parenthesis
"""
for cid, coref in coreflist[::-1]:
words = ''.join(word_tokenize(coref['text']))
nltktree = Tree.parse(ptree)
nltktree.reverse() # perform search right to left
data = None
for subtree in nltktree.subtrees(): # BFS
if ''.join(subtree.leaves()) == words: # equal ignoring whitespace
data = subtree.pprint()
break
# If found via breadth-first search of parse tree
if data:
ptree = ptree.replace(data, '( COREF_TAG_%s%s)' % (cid, data))
else: # Try finding via regex matching instead
dpattern = r'\s*'.join([r'\(\s*[a-zA-Z$]+\s+%s\s*\)' % word
for word in word_tokenize(coref['text'])])
found = re.findall(pattern % dpattern, ptree, re.X)
if found:
repl = '%s%s ( COREF_TAG_%s%s) %s' % (found[0][0],
found[0][1],
cid,
found[0][2],
found[0][3])
ptree = re.sub(pattern % dpattern, repl, ptree, 1, re.X)
return ptree
def get_tagged_corefs(xml, ordered=False):
"""Parses xml to find all tagged coreferences contained in COREF tags
Args:
xml: string, xml markedup with COREF tags
ordered: if True, returns an list in the same order that the corefs
appear in the text
Returns:
if ordered
list of tuples, [(coref_id, {coref, ref_id}), ]
if not ordered
dict of dict, {coref_id: (coref, ref_id)
>>> text = "<TXT>John stubbed <COREF ID='1'>his</COREF> toe.</TXT>"
>>> get_tagged_corefs(text)
{u'1': {'text': u'his', 'ref': None}}
>>> get_tagged_corefs(text, ordered=True)
[(u'1', {'text': u'his', 'ref': None})]
>>> text = "<TXT><COREF ID='A'>John</COREF> stubbed " +\
"<COREF ID='1' REF='A'>his</COREF> toe.</TXT>"
>>> get_tagged_corefs(text)
{u'A': {'text': u'John', 'ref': None}, u'1': {'text': u'his', 'ref': u'A'}}
>>> get_tagged_corefs(text, ordered=True) # doctest: +NORMALIZE_WHITESPACE
[(u'A', {'text': u'John', 'ref': None}),
(u'1', {'text': u'his', 'ref': u'A'})]
"""
nps = {}
if ordered:
nps = []
xml = _normalize_malformed_xml(xml)
try:
corefs = parseString(xml).getElementsByTagName('COREF')
except ExpatError:
return nps
for coref in corefs:
try:
cid = coref.attributes['ID'].value
if ordered:
data = {}
for npid, np in nps:
if npid == cid:
data = np
break
else:
data = nps.get(cid, {})
except KeyError:
continue
try:
data['ref'] = coref.attributes['REF'].value
except KeyError:
data['ref'] = None
data['text'] = coref.firstChild.data
if ordered:
nps.append((cid, data))
else:
nps[cid] = data
return nps
def cid_less_than(cid1, cid2):
if cid1.isdigit() and cid2.isdigit():
return int(cid1) < int(cid2)
else:
return True
# elif not (cid1.isdigit() or cid2.isdigit()):
# num1 = int(cid1[:-1])
# num2 = int(cid2[:-1])
# if num1 == num2:
# return cid1[-1] < cid2[-1]
# else:
# return num1 < num2
# elif cid1.isdigit():
# return True
# else:
# return False
def _normalize_sentence(sent):
"""Removes unwanted characters from sentence for parsing
Args:
sent: string, sentence to normalize
Returns
string, normalized sentence
"""
#sent = sent[sent.find('\n\n\n'):]
removed = r'[\n ]+'
sent = re.sub(removed, ' ', sent)
return sent.strip()
def _normalize_malformed_xml(xml):
"""Ensures that xml begins and ends with <TXT> </TXT> tags
Args:
xml: string, text to be formatted as xml
Returns:
string, formatted xml
>>> _normalize_malformed_xml('The dog.')
'<TXT>The dog.</TXT>'
>>> _normalize_malformed_xml('<TXT>The dog.')
'<TXT>The dog.</TXT>'
>>> _normalize_malformed_xml('The dog.</TXT>')
'<TXT>The dog.</TXT>'
>>> _normalize_malformed_xml('<TXT>The dog.</TXT>')
'<TXT>The dog.</TXT>'
"""
xml = xml.strip()
if not xml.startswith('<TXT>'):
xml = '<TXT>' + xml
if not xml.endswith('</TXT>'):
xml = xml + '</TXT>'
return xml
def _remove_tags(xml):
"""Removes xml tags from string, returning non-markedup text
Args:
xml: string, xml markedup text
Returns:
string, text from xml
>>> xml = "<TXT>John stubbed <COREF ID='1'>his</COREF> toe.</TXT>"
>>> _remove_tags(xml)
'John stubbed his toe.'
>>> xml = "<TXT><COREF ID='A'>John</COREF> stubbed " +\
"<COREF ID='1' REF='A'>his</COREF> toe.</TXT>"
>>> _remove_tags(xml)
'John stubbed his toe.'
"""
chars = list(xml)
i = 0
while i < len(chars):
if chars[i] == '<':
while chars[i] != '>':
chars.pop(i) # pop everything between brackets
chars.pop(i) # pops the right-angle bracket, too
else:
i += 1
return ''.join(chars)
def _process_parse(parse, coreflist):
"""Tags parse tree with corefs and returns the tree, lexicon, dependencies
and raw text as tuple
Args:
parse: list of stanford corenlp parsed sentences
coreflist: list of coreferences from tagged xml
Returns:
tuple, (ptree, lexicon, dependencies, rawtext) if parse contains a
sentence, else returns None
"""
sentence = parse.get('sentences')
if sentence:
ptree = Tree.parse(tag_ptree(sentence[0]['parsetree'], coreflist))
words = [(w[0], w[1]) for w in sentence[0]['words']]
depends = [(d[0], d[1], d[2]) for d in sentence[0]['dependencies']]
text = sentence[0]['text']
return ptree, words, depends, text
else:
return None
def get_synsets(words):
"""Returns sets of cognitive synonyms for each of the input words
Args:
words: dict, {word: (pos1, pos2, ...)}
Returns:
dict, {synset_name: (syn1, syn2, syn3, ...)}
>>> words = {u'apple': (u'NN')}
>>> get_synsets(words) # doctest: +NORMALIZE_WHITESPACE
{'apple.n.01': ('apple',),
'apple.n.02': ('apple', 'orchard_apple_tree', 'Malus_pumila')}
"""
synsets = {}
for word in words:
for syn in wn.synsets(word):
synsets[syn.name] = tuple([lemma.name for lemma in syn.lemmas])
return synsets
@static_var("id", '1A')
def | _mk_coref_id | identifier_name |
|
data.py |
class FilenameException(Exception):
"""Raised when file does not have the correct extension"""
pass
def mk_parses(listfile, corenlp_host):
"""Creates a list of FileParse objects for the files listed in the listfile
Args:
listfile: string, path to input listfile (see assignment description
for listfile details)
Returns:
list of FileParse objects
"""
# if not listfile.endswith('.listfile'):
# filetype = 'Co-Reference List file'
# error = 'has incorrect file type'
# raise FilenameException("Error: %s %s" % (filetype, error))
try:
with open(listfile) as f:
pserver = jsonrpc.ServerProxy(jsonrpc.JsonRpc20(),
jsonrpc.TransportTcpIp(
addr=(corenlp_host, 8080), limit=1000))
parses = dict([(get_id(path), FileParse(path, pserver))
for path in f.readlines()
if path.lstrip()[0] != '#'])
except IOError:
stderr.write(strerror(EIO)) # stderr.write does not have newlines
stderr.write("\nERROR: Could not open list file\n")
exit(EIO)
else:
return parses
def get_id(path):
"""Parses a file path for the filename without extension
Args:
path: string, full (or relative) file path for coreference file.
Must end in .crf
Returns:
string, file id (filename without extension)
>>> path = '/home/user/Desktop/full.crf'
>>> get_id(path)
'full'
>>> path = 'relative.crf'
>>> get_id(path)
'relative'
"""
fid, ext, _ = path.strip().split('/')[-1].partition('.crf')
if not fid or ext != '.crf':
filetype = 'Co-Reference Input file'
error = 'has incorrect file type'
raise FilenameException("Error: %s %s" % (filetype, error))
return fid
def mk_fparse(filename, pserver):
"""Parses input to get list of paragraphs with sentence structure
and a dictionary of noun phrases contained in the COREF tags
Args:
filename: string, path to crf file
pserver: jsonrpc.ServerProxy, stanford corenlp server for parsing
Returns:
tuple, (list_stanford_sent_parses, dict_file_corefs, dict_file_synsets)
"""
parses = []
try:
with open(filename) as f:
vprint('OPEN: %s' % filename)
xml = f.read()
except IOError:
print strerror(EIO)
print("ERROR: Could not open %s" % filename)
return (parses, get_tagged_corefs(''), get_synsets({}))
# remove unwanted characters from xml
vprint('\tPARSE: Parsing file: %s' % filename)
# parse_tries = 0
# while parse_tries < 5:
# try:
# t = loads(pserver.parse(_normalize_sentence(_remove_tags(xml))))
# parse_tries = 0
# break
# except jsonrpc.RPCTimeoutError:
# vprint('\tERROR: RPCTimeoutError - retrying')
# parse_tries += 3
# except jsonrpc.RPCTransportError:
# vprint('\tERROR: RPCTransportError - retrying')
# data = _normalize_sentence(_remove_tags(xml))
# sentences = [sent for part in data.split('\n\n')
# for sent in sent_tokenize(part)]
# try:
# xml1 = data[:data.find(sentences[len(sentences)/3])]
# xml2 = data[data.find(sentences[len(sentences)/3+1]):data.find(sentences[2*len(sentences)/3])]
# xml3 = data[data.find(sentences[2*len(sentences)/3+1]):]
# t1 = loads(pserver.parse(xml1))
# t2 = loads(pserver.parse(xml2))
# t3 = loads(pserver.parse(xml3))
# t = dict(t1.items() + t2.items() + t3.items())
# parse_tries = 0
# break
# except Exception:
# parse_tries = -1
# break
# parse_tries += 1
# if parse_tries != 0:
# vprint('\tFATAL: RPCTransportError - skipping')
sentences = [sent for part in xml.split('\n\n')
for sent in sent_tokenize(part)]
vprint('\tPARSE: Parsing sentences: %s' % filename)
for sent in sentences:
sent_corefs = get_tagged_corefs(sent, ordered=True)
# remove unwanted characters from xml
sent = _normalize_sentence(_remove_tags(sent))
parse_tries = 0
while parse_tries < 5:
try:
sparse = loads(pserver.parse(sent))
parse_tries = 0
break
except jsonrpc.RPCTransportError:
vprint('\tERROR: RPCTransportError - retrying')
parse_tries += 1
if parse_tries != 0:
vprint('\tFATAL: RPCTransportError - skipping')
pparse = _process_parse(sparse, sent_corefs)
if pparse:
parses.append(pparse)
pos_tags = {}
for parse in parses:
for word, attr in parse[1]:
tags = pos_tags.get(word, set())
tags.add(attr['PartOfSpeech'])
pos_tags[word] = tags
return parses, get_tagged_corefs(xml), get_synsets(pos_tags)
def tag_ptree(ptree, coreflist):
"""Tags given parse tree with coreferences
Args:
ptree: string, parenthesized str represenation of parse tree
coreflist: list of tuples, [('1', {'text': 'dog', 'ref': None})]
Returns:
string, tagged parse tree
>>> ptree = '(S NP( (NN He)) VP( (V ran)))'
>>> coreflist = [('1', {'text': 'He', 'ref': None})]
>>> tag_ptree(ptree, coreflist)
'(S NP( COREF_TAG_1( (NN He))) VP( (V ran)))'
"""
pattern = r"""(?P<lp>\(?\s*) # left parenthesis
(?P<tg>[a-zA-Z$]+)? # POS tag
(?P<data>\s*%s) # subtree of tag
(?P<rp>(?:\s*\))*) # right parenthesis
"""
for cid, coref in coreflist[::-1]:
words = ''.join(word_tokenize(coref['text']))
nltktree = Tree.parse(ptree)
nltktree.reverse() # perform search right to left
data = None
for subtree in nltktree.subtrees(): # BFS
if ''.join(subtree.leaves()) == words: # equal ignoring whitespace
data = subtree.pprint()
break
# If found via breadth-first search of parse tree
if data:
ptree = ptree.replace(data, '( COREF_TAG_%s%s)' % (cid, data))
else: # Try finding via regex matching instead
dpattern = r'\s*'.join([r'\(\s*[a-zA-Z$]+\s+%s\s*\)' % word
for word in word_tokenize(coref['text'])])
found = re.findall(pattern % dpattern, ptree, re.X)
if found:
repl = '%s%s ( COREF_TAG_%s%s) %s' % (found[0][0],
found[0][1],
cid,
found[0][2],
found[0][3])
ptree = re.sub(pattern % dpattern, repl, ptree, 1, re.X)
return ptree
def get_tagged_corefs(xml, ordered=False):
"""Parses xml to find all tagged coreferences contained in COREF tags
Args:
xml: string, xml markedup with COREF tags
ordered: if True, returns an list in the same order that the corefs
appear in the text
Returns:
if ordered
list of tuples, [(coref_id, {coref, ref_id}), ]
if not ordered
dict of dict, {coref_id: (coref, ref_id)
>>> text = "<TXT>John stubbed <COREF ID='1'>his</COREF> toe.</TXT>"
>>> get_tagged_corefs(text)
{u'1': {'text': u'his', 'ref': None}}
>>> get_tagged_corefs(text, ordered=True)
[(u'1', {'text': u'his', 'ref': None})]
>>> text | self.ptree = parse[0]
self.words = parse[1]
self.dependencies = parse[2]
self.text = parse[3] | identifier_body |
|
lucidLog-1.0.0.source.js | .pointer;
delete ua.stackLevel;
delete ua.callIdx;
delete ua.callJSON;
ua.timings.startTime = '' + ua.timings.startTime;
var uaJSON = sa.json.encode (ua);
r[uaJSON] = sa.tracer.traced[uaIdx];//sa.l.cleanupTracerData (sa.tracer.traced[uaIdx]);
};
hm (r, 'sa.tracer dump', { htmlID : 'saLucidLog_hm_javascript', opacity : 0.7 });
},
cleanupTracerData : function (ua) {
if (ua.logMessages) {
var
lm = ua.logMessages,
lmn = {};
for (var i=0; i<lm.length; i++) {
lmn[lm[i][0]] = lm[i][1][1];
};
ua.logMessages = lmn;
};
if (ua.calls) {
for (var i=0; i<ua.calls.length; i++) {
/*
if (!sa.l.settings.gottaCleanup) sa.l.settings.gottaCleanup = [];
sa.l.settings.gottaCleanup[sa.l.settings.gottaCleanup.length] = {
ua : ua.calls[i]
};
*/
sa.l.cleanupTracerData (ua.calls[i]);
};
// setTimeout (sa.l.processGottaCleanup, 100);
};
return ua;
},
processGottaCleanup : function () {
var
gt = sa.l.settings.gottaCleanup.shift(),
count = 0;
while (gt && count < 100) {
sa.l.cleanupTracerData (gt.ua);
gt = sa.l.settings.gottaCleanup.shift();
count++;
};
if (sa.l.settings.gottaCleanup.length>0) {
setTimeout (function () {
sa.tracer.processGottaTrace();
}, 10);
}
},
ui : {
resize : function () {
jQuery('#saLucidLog').css({width:jQuery(window).width()-6});
var h = jQuery('#saLucidLog').height();
var w = jQuery('#saLucidLog').width();
jQuery('#saLucidLog_background').css ({ height : h, width : w });
jQuery('#saLucidLog_overlay1').css ({ top : 1, left : 1, height : h-2, width : w-2 });
jQuery('#saLucidLog_overlay2').css ({ top : 2, left : 2, height : h-4, width : w-4 });
jQuery('#saLucidLog_overlay3').css ({ top : 3, left : 3, height : h-6, width : w-6 });
jQuery('#saLucidLog_content_background1').css ({ top : 30, left : 4, height : h - 34, width : w-8 });
jQuery('#saLucidLog_content_background2').css ({ top : 31, left : 5, height : h - 36, width : w-10 });
jQuery('#saLucidLog_content_background3').css ({ top : 32, left : 6, height : h - 38, width : w-12 });
if (jQuery('#saLucidLog_content_holder__container').length==1) {
jQuery('#saLucidLog_content_holder__container').css ({ top : 30, left : 4, height : h - 54, width : w-8 });
} else {
jQuery('#saLucidLog_content_holder').css ({ top : 30, left : 4, height : h - 34, width : w-8 });
};
jQuery('#saLucidLog_content').css ({ top : 3, left : 3, height : jQuery('#saLucidLog_content_holder').height()-6, width : jQuery('#saLucidLog_content_holder').width()-6 });
var jQueryhm = jQuery('#saLucidLog_hm_javascript');
if (
jQueryhm[0]
&& jQueryhm[0].children[0]
&& jQueryhm[0].children[0].children[2]
) sa.sp.containerSizeChanged(jQuery('#saLucidLog_hm_javascript')[0].children[0].children[2], true);
sa.sp.containerSizeChanged (jQuery('#saLucidLog_content_holder')[0], true);
jQuery('.fhl_item_content').each (function (idx,el) {
var id = el.id.replace ('fhl_', '').replace('_content','');
jQuery('#fhl_'+id+'_bg').css ({width : jQuery(el).width()+10});
});
jQuery('.tabsContainer').each ( sa.m.traceFunction ( function (idx) {
jQuery(this).css ({
height : jQuery(this).parent().height() - jQuery(this).prev().height(),
});
}));
sa.l.php.tools.resizeWindow(sa.l.php.cmd.cmdID);
},
hide : function (callback) {
jQuery('#saLucidLog').fadeOut (sa.l.globals.hideShowSpeed, callback);
},
show : function (callback) {
jQuery('#saLucidLog').fadeIn (sa.l.globals.hideShowSpeed, callback);
},
toggleShowHide : function() {
if (jQuery('#saLucidLog').css('display')=='none') {
sa.l.ui.show();
} else {
sa.l.ui.hide();
}
},
click : {
// (all click handlers for this web component) :
btnRefresh : function () {
sa.l.redrawRawLog();
},
btnHide : function () {
sa.l.ui.hide();
},
btnShow : function () {
sa.tracer.disabled = true;
sa.l.ui.show();
sa.l.ui.resize();
sa.l.visualizeTracerData();
},
btnShowPHP : function () {
jQuery('#saLucidLog_page_log').fadeOut(500);
jQuery('#saLucidLog_page_javascript').fadeOut(500);
setTimeout (sa.m.traceFunction(function() {
jQuery('#saLucidLog_page_php').css({display:'none',visibility:'visible'}).fadeIn(500);
}), 510);
},
btnShowJavascript : function () {
jQuery('#saLucidLog_page_log').fadeOut(500);
jQuery('#saLucidLog_page_php').fadeOut(500);
setTimeout (sa.m.traceFunction(function() {
jQuery('#saLucidLog_page_javascript').css({display:'none',visibility:'visible'}).fadeIn(500);
}), 510);
},
btnShowLog : function () {
jQuery('#saLucidLog_page_javascript').fadeOut(500);
jQuery('#saLucidLog_page_php').fadeOut(500);
setTimeout (sa.m.traceFunction(function() {
jQuery('#saLucidLog_page_log').html('<div id="saLucidLog_log" style="width:100%;height:100%;"></div>');
hm (sa.settings.log, 'Console.log', { htmlID : 'saLucidLog_log', opacity :0.65 });
jQuery('#saLucidLog_page_log').css({display:'none',visibility:'visible'}).fadeIn(500);
}), 510);
}
} // sa.l.ui.click
}, // sa.l.ui
tools : {
setupDragAndDrop_forTopBorder : function () {
// ripped with thanks from http://jsfiddle.net/gaby/Bek9L/186/ ;
var i = 0;
var dragging = false;
jQuery('#saLucidLog_dragBar').mousedown(function(e){
e.preventDefault();
dragging = true;
var main = jQuery('#saLucidLog');
var ghostbar = jQuery('<div>', {
id:'ghostbar',
css: {
position:'absolute',
background : 'black',
opacity : 0.7,
width: main.outerWidth(),
height : 3,
zIndex : 99999,
top: main.offset().top,
left: main.offset().left
}
}).appendTo('body');
jQuery(window).mousemove(function(e){
ghostbar.css("top",e.pageY+2);
jQuery('#saLucidLog').css("height", jQuery(window).height()- e.pageY);
sa.l.ui.resize();
});
if (document.getElementById ('iframe-content'))
jQuery(document.getElementById ('iframe-content').contentWindow).mousemove(function(e){
ghostbar.css("top",jQuery('#iframe-content', window.parent.document).offset().top + e.pageY+2);
jQuery('#saLucidLog').css("height", jQuery(window).height()- ghostbar.css("top").replace('px',''));
sa.l.ui.resize();
});
});
|
jQuery(window).mouseup(function(e){
if (dragging) {
jQuery('#ghostbar').remove();
jQuery(window).unbind('mousemove');
| random_line_split |
|
lucidLog-1.0.0.source.js | >'
+'<div id="saLucidLog_btnShowJavascript" class="vividButton vividTheme__menu_001" style="position:absolute;left:250px;z-index:41000300"><a href="javascript:sa.l.ui.click.btnShowJavascript();">Javascript</a></div>'
+'<div id="saLucidLog_btnShowLog" class="vividButton vividTheme__menu_001" style="position:absolute;left:490px;z-index:41000300"><a href="javascript:sa.l.ui.click.btnShowLog();">Console.log</a></div>'
+'<div id="saLucidLog_overlay1" class="fhl_overlay" style="position:absolute;background:black;z-index:41000013;opacity:0.05;filter:Alpha(opacity=5);"> </div>'
+'<div id="saLucidLog_overlay2" style="position:absolute;background:black;z-index:41000014;opacity:0.05;filter:Alpha(opacity=5);"> </div>'
+'<div id="saLucidLog_overlay3" style="position:absolute;background:black;z-index:41000015;opacity:0.05;filter:Alpha(opacity=5);"> </div>'
+'<div id="saLucidLog_content_background1" class="fhl_content_background" style="position:absolute;z-index:41000017;width:100%;height:100%;background:black;opacity:0.15;filter:Alpha(opacity=15);"> </div>'
+'<div id="saLucidLog_content_background2" class="fhl_content_background" style="position:absolute;z-index:41000018;width:100%;height:100%;background:black;opacity:0.18;filter:Alpha(opacity=18);"> </div>'
+'<div id="saLucidLog_content_background3" class="fhl_content_background" style="position:absolute;z-index:41000019;width:100%;height:100%;background:black;opacity:0.20;filter:Alpha(opacity=20);"> </div>'
+'<div id="saLucidLog_content_holder" class="vividScrollpane vividTheme__scroll_black" style="position:absolute;z-index:42000000;">'
+'<div id="saLucidLog_content" style="position:absolute;z-index:50000000;">'
+'<div id="saLucidLog_page_php" class="saLucidLogTabpage" style="position:absolute;width:100%;height:100%;">'
+'</div>'
+'<div id="saLucidLog_page_javascript" class="saLucidLogTabpage" style="position:absolute;width:100%;height:100%;visibility:hidden;">'
+'<div id="saLucidLog_hm_javascript" style="width:100%;height:100%;opacity:0.7"> </div>'
+'</div>'
+'<div id="saLucidLog_page_log" class="saLucidLogTabpage vividScrollpane vividTheme__scroll_black" style="position:absolute;width:100%;height:100%;visibility:hidden;">'
//+ sa.settings.log.entries.join ('<br/>')
//+'<div id="saLucidLog_hm_log" style="width:100%;height:100%;opacity:0.7"> </div>'
+'</div>'
+'</div>'
+'</div>'
+'</div>'
+'</div>'
//+'<img id="saLucidLog_btnRecord" src="'+sa.m.globals.urls.os+'/'+'com/ui/tools/lucidLog/images/btnRecord.png" style="position:absolute;bottom:4px;right:100px;height:21px;z-index:999999990;" title="show LucidLog (PHP + JS trace log)"/>'
+'<img id="saLucidLog_btnShow" class="saBtn_simpleImg" src="'+sa.m.globals.urls.os+'/'+'com/ui/tools/lucidLog/images/btnShow.png" style="position:absolute;width:89px;height:21px;bottom:3px;right:5px;z-index:999999990;" title="show LucidLog (PHP + JS trace log)"/>';
//+'<div id="saLucidLog_btnRecord" class="vividButton vividTheme__playpause_001" style="position:absolute;bottom:3px;right:3px;width:50px;height:50px;" onclick="sa.tracer.disabled=false;" title="record with sa.tracer"> </div>';
jQuery('body').append (html);
if (true)
//setTimeout (sa.m.traceFunction (function () {
sa.vcc.init (jQuery('#saLucidLog')[0], sa.m.traceFunction(function() {sa.l.componentFullyInitialized();}));
//}), 2000);
else {
sa.l.componentFullyInitialized();
};
},
componentFullyInitialized : function(){
jQuery('.fhl_overlay, .fhl_content_background').css ({borderRadius:'.5em'});
jQuery('#saLucidLog').css ({
bottom : 0,
left : 3,
height : ((jQuery(window).height() / 10 ) * 4.5),
width : (jQuery(window).width() - 6)
});
jQuery('#saLucidLog_btnRefresh').click (sa.l.ui.click.btnRefresh);
jQuery('#saLucidLog_btnHide').click (sa.l.ui.click.btnHide);
jQuery('#saLucidLog_btnShow').click (sa.l.ui.click.btnShow);
jQuery(window).resize (sa.l.ui.resize);
sa.l.tools.setupDragAndDrop_forTopBorder();
sa.l.ui.resize();
sa.l.php.initialize();
//sa.l.ui.hide();
jQuery('#saLucidLog').css({opacity:1,display:'none'});
},
log : function () {
if (typeof sa.tracer!=='undefined') {
var ua = sa.tracer.findUA(arguments);
if (ua && ua.pointer) {
var logIdx = ua.pointer.logMessages.length;
//debugger;
ua.pointer.logMessages[logIdx] = arguments;
}// else debugger;
}
},
visualizeTracerData : function () {
var r = {};
for (var uaIdx=0; uaIdx<sa.tracer.traced.length; uaIdx++) {
var ua = sa.tracer.userActions[uaIdx];
delete ua.pointer;
delete ua.stackLevel;
delete ua.callIdx;
delete ua.callJSON;
ua.timings.startTime = '' + ua.timings.startTime;
var uaJSON = sa.json.encode (ua);
r[uaJSON] = sa.tracer.traced[uaIdx];//sa.l.cleanupTracerData (sa.tracer.traced[uaIdx]);
};
hm (r, 'sa.tracer dump', { htmlID : 'saLucidLog_hm_javascript', opacity : 0.7 });
},
cleanupTracerData : function (ua) {
if (ua.logMessages) {
var
lm = ua.logMessages,
lmn = {};
for (var i=0; i<lm.length; i++) | ;
ua.logMessages = lmn;
};
if (ua.calls) {
for (var i=0; i<ua.calls.length; i++) {
/*
if (!sa.l.settings.gottaCleanup) sa.l.settings.gottaCleanup = [];
sa.l.settings.gottaCleanup[sa.l.settings.gottaCleanup.length] = {
ua : ua.calls[i]
};
*/
sa.l.cleanupTracerData (ua.calls[i]);
};
// setTimeout (sa.l.processGottaCleanup, 100);
};
return ua;
},
processGottaCleanup : function () {
var
gt = sa.l.settings.gottaCleanup.shift(),
count = 0;
while (gt && count < 100) {
sa.l.cleanupTracerData (gt.ua);
gt = sa.l.settings.gottaCleanup.shift();
count++;
};
if (sa.l.settings.gottaCleanup.length>0) {
setTimeout (function () {
sa.tracer.processGottaTrace();
}, 10);
}
},
ui : {
resize : function () {
jQuery('#saLucidLog').css({width:jQuery(window).width()-6});
var h = jQuery('#saLucidLog').height();
var w = jQuery('#sa | {
lmn[lm[i][0]] = lm[i][1][1];
} | conditional_block |
index.ts | to hold onto.
vault: Vault;
metadataCache: MetadataCache;
constructor(vault: Vault, metadataCache: MetadataCache, tag: TagIndex, prefix: PrefixIndex) {
this.vault = vault;
this.metadataCache = metadataCache;
this.tag = tag;
this.prefix = prefix;
this.reloadQueue = [];
this.reloadSet = new Set();
this.reloadHandlers = [];
// Background task which regularly checks for reloads.
this.reloadHandle = window.setInterval(() => this.reloadInternal(), FullIndex.RELOAD_INTERVAL);
// TODO: Metadata cache is not updated on modify, but on metadatacache resolve.
vault.on("modify", file => {
if (file instanceof TFile) {
this.queueReload(file);
}
});
}
/** Queue the file for reloading; several fast reloads in a row will be debounced. */
public queueReload(file: TFile) {
if (this.reloadSet.has(file.path)) return;
this.reloadSet.add(file.path);
this.reloadQueue.push(file);
}
public on(event: 'reload', handler: (a: TFile) => Promise<void>) {
this.reloadHandlers.push(handler);
}
/** Utility method which regularly checks the reload queue. */
private async reloadInternal() {
let copy = Array.from(this.reloadQueue);
this.reloadSet.clear();
this.reloadQueue = [];
for (let file of copy) {
await Promise.all([this.tag.reloadFile(file)].concat(this.reloadHandlers.map(f => f(file))));
}
}
}
/** Index which efficiently allows querying by tags / subtags. */
export class TagIndex {
/** Parse all subtags out of the given tag. I.e., #hello/i/am would yield [#hello/i/am, #hello/i, #hello]. */
public static parseSubtags(tag: string): string[] {
let result = [tag];
while (tag.contains("/")) {
tag = tag.substring(0, tag.lastIndexOf("/"));
result.push(tag);
}
return result;
}
/** Parse all of the tags for the given file. */
public static parseTags(cache: MetadataCache, path: string): Set<string> {
let fileCache = cache.getCache(path);
if (!fileCache) return new Set<string>();
let allTags = new Set<string>();
// Parse tags from in the file contents.
let tagCache = fileCache.tags;
if (tagCache) {
for (let tag of tagCache) {
if (!tag.tag || !(typeof tag.tag == 'string')) continue;
this.parseSubtags(tag.tag).forEach(t => allTags.add(t));
}
}
// Parse tags from YAML frontmatter.
let frontCache = fileCache.frontmatter;
// Search for the 'tags' field, since it may have wierd
let tagsName: string | undefined = undefined;
for (let key of Object.keys(frontCache ?? {})) {
if (key.toLowerCase() == "tags" || key.toLowerCase() == "tag")
tagsName = key;
}
if (frontCache && tagsName && frontCache[tagsName]) {
if (Array.isArray(frontCache[tagsName])) {
for (let tag of frontCache[tagsName]) {
if (!(typeof tag == 'string')) continue;
if (!tag.startsWith("#")) tag = "#" + tag;
this.parseSubtags(tag).forEach(t => allTags.add(t));
}
} else if (typeof frontCache[tagsName] === 'string') {
// Assume tags is a comma-separated list.
let tags = (frontCache[tagsName] as string).split(",").map(elem => {
elem = elem.trim();
if (!elem.startsWith("#")) elem = "#" + elem;
return elem;
});
for (let tag of tags) {
this.parseSubtags(tag).forEach(t => allTags.add(t));
}
}
}
return allTags;
}
public static async generate(vault: Vault, cache: MetadataCache): Promise<TagIndex> {
let initialMap = new Map<string, Set<string>>();
let initialInvMap = new Map<string, Set<string>>();
let timeStart = new Date().getTime();
// First time load...
for (let file of vault.getMarkdownFiles()) {
let allTags = TagIndex.parseTags(cache, file.path);
initialInvMap.set(file.path, allTags);
for (let subtag of allTags) {
if (!initialMap.has(subtag)) initialMap.set(subtag, new Set<string>());
initialMap.get(subtag)?.add(file.path);
}
}
let totalTimeMs = new Date().getTime() - timeStart;
console.log(`Dataview: Parsed ${initialMap.size} tags in ${initialInvMap.size} markdown files (${totalTimeMs / 1000.0}s)`);
return new TagIndex(vault, cache, initialMap, initialInvMap);
}
/** Maps tags -> set of files containing that exact tag. */
map: Map<string, Set<string>>;
/** Cached inverse map; maps file -> tags it was last known to contain. */
invMap: Map<string, Set<string>>;
vault: Vault;
cache: MetadataCache;
constructor(vault: Vault, metadataCache: MetadataCache,
map: Map<string, Set<string>>, invMap: Map<string, Set<string>>) {
this.vault = vault;
this.cache = metadataCache;
this.map = map;
this.invMap = invMap;
}
/** Returns all files which have the given tag. */
public get(tag: string): Set<string> {
let result = this.map.get(tag);
if (result) {
return new Set(result);
} else {
return new Set();
}
}
/** Returns all tags the given file has. */
public getInverse(file: string): Set<string> {
let result = this.invMap.get(file);
if (result) {
return new Set(result);
} else {
return new Set();
}
}
async reloadFile(file: TFile) {
this.clearFile(file.path);
let allTags = TagIndex.parseTags(this.cache, file.path);
for (let subtag of allTags) {
if (!this.map.has(subtag)) this.map.set(subtag, new Set<string>());
this.map.get(subtag)?.add(file.path);
}
this.invMap.set(file.path, allTags);
}
/** Clears all tags for the given file so they can be re-added. */
private clearFile(path: string) {
let oldTags = this.invMap.get(path);
if (!oldTags) return;
this.invMap.delete(path);
for (let tag of oldTags) {
this.map.get(tag)?.delete(path);
}
}
}
/** A node in the prefix tree. */
export class PrefixIndexNode {
// TODO: Instead of only storing file paths at the leaf, consider storing them at every level,
// since this will make for faster deletes and gathers in exchange for slightly slower adds and more memory usage.
// since we are optimizing for gather, and file paths tend to be shallow, this should be ok.
files: Set<string>;
element: string;
totalCount: number;
children: Map<string, PrefixIndexNode>;
constructor(element: string) {
this.element = element;
this.files = new Set();
this.totalCount = 0;
this.children = new Map();
}
public static add(root: PrefixIndexNode, path: string) {
let parts = path.split("/");
let node = root;
for (let index = 0; index < parts.length - 1; index++) |
node.totalCount += 1;
node.files.add(path);
}
public static remove(root: PrefixIndexNode, path: string) {
let parts = path.split("/");
let node = root;
let nodes = [];
for (let index = 0; index < parts.length - 1; index++) {
if (!node.children.has(parts[index])) return;
nodes.push(node);
node = node.children.get(parts[index]) as PrefixIndexNode;
}
if (!node.files.has(path)) return;
node.files.delete(path);
node.totalCount -= 1;
for (let p of nodes) p.totalCount -= 1;
}
public static find(root: PrefixIndexNode, prefix: string): PrefixIndexNode | null {
if (prefix.length == 0 || prefix == '/') return root;
let parts = prefix.split("/");
let node = root;
for (let index = 0; index < parts.length; index++) {
if (!node.children.has(parts[index])) return null;
node = node.children.get(parts[index]) as PrefixIndexNode;
}
return node;
}
public static gather(root: PrefixIndexNode): Set<string> {
let result = new Set<string>();
PrefixIndexNode.gatherRec(root, result);
return result;
}
static gatherRec(root: PrefixIndexNode, output: Set<string>) {
for (let file of root.files) output.add(file);
for (let child of root.children.values()) this.gatherRec(child, output);
}
}
/** Indexes files by their full prefix - | {
if (!node.children.has(parts[index])) node.children.set(parts[index], new PrefixIndexNode(parts[index]));
node.totalCount += 1;
node = node.children.get(parts[index]) as PrefixIndexNode;
} | conditional_block |
index.ts | things to hold onto.
vault: Vault;
metadataCache: MetadataCache;
constructor(vault: Vault, metadataCache: MetadataCache, tag: TagIndex, prefix: PrefixIndex) {
this.vault = vault;
this.metadataCache = metadataCache;
this.tag = tag;
this.prefix = prefix;
this.reloadQueue = [];
this.reloadSet = new Set();
this.reloadHandlers = [];
// Background task which regularly checks for reloads.
this.reloadHandle = window.setInterval(() => this.reloadInternal(), FullIndex.RELOAD_INTERVAL);
// TODO: Metadata cache is not updated on modify, but on metadatacache resolve.
vault.on("modify", file => {
if (file instanceof TFile) {
this.queueReload(file);
}
});
}
/** Queue the file for reloading; several fast reloads in a row will be debounced. */
public queueReload(file: TFile) {
if (this.reloadSet.has(file.path)) return;
this.reloadSet.add(file.path);
this.reloadQueue.push(file);
}
public on(event: 'reload', handler: (a: TFile) => Promise<void>) {
this.reloadHandlers.push(handler);
}
/** Utility method which regularly checks the reload queue. */
private async reloadInternal() {
let copy = Array.from(this.reloadQueue);
this.reloadSet.clear();
this.reloadQueue = [];
for (let file of copy) {
await Promise.all([this.tag.reloadFile(file)].concat(this.reloadHandlers.map(f => f(file))));
}
}
}
/** Index which efficiently allows querying by tags / subtags. */
export class TagIndex {
/** Parse all subtags out of the given tag. I.e., #hello/i/am would yield [#hello/i/am, #hello/i, #hello]. */
public static parseSubtags(tag: string): string[] {
let result = [tag];
while (tag.contains("/")) {
tag = tag.substring(0, tag.lastIndexOf("/"));
result.push(tag);
}
return result;
}
/** Parse all of the tags for the given file. */
public static parseTags(cache: MetadataCache, path: string): Set<string> {
let fileCache = cache.getCache(path);
if (!fileCache) return new Set<string>();
let allTags = new Set<string>();
// Parse tags from in the file contents.
let tagCache = fileCache.tags;
if (tagCache) {
for (let tag of tagCache) {
if (!tag.tag || !(typeof tag.tag == 'string')) continue;
this.parseSubtags(tag.tag).forEach(t => allTags.add(t));
}
}
// Parse tags from YAML frontmatter.
let frontCache = fileCache.frontmatter;
// Search for the 'tags' field, since it may have wierd
let tagsName: string | undefined = undefined;
for (let key of Object.keys(frontCache ?? {})) {
if (key.toLowerCase() == "tags" || key.toLowerCase() == "tag")
tagsName = key;
}
if (frontCache && tagsName && frontCache[tagsName]) {
if (Array.isArray(frontCache[tagsName])) {
for (let tag of frontCache[tagsName]) {
if (!(typeof tag == 'string')) continue;
if (!tag.startsWith("#")) tag = "#" + tag;
this.parseSubtags(tag).forEach(t => allTags.add(t));
}
} else if (typeof frontCache[tagsName] === 'string') {
// Assume tags is a comma-separated list.
let tags = (frontCache[tagsName] as string).split(",").map(elem => {
elem = elem.trim();
if (!elem.startsWith("#")) elem = "#" + elem;
return elem;
});
for (let tag of tags) {
this.parseSubtags(tag).forEach(t => allTags.add(t));
}
}
}
return allTags;
}
public static async generate(vault: Vault, cache: MetadataCache): Promise<TagIndex> {
let initialMap = new Map<string, Set<string>>();
let initialInvMap = new Map<string, Set<string>>();
let timeStart = new Date().getTime();
// First time load...
for (let file of vault.getMarkdownFiles()) {
let allTags = TagIndex.parseTags(cache, file.path);
initialInvMap.set(file.path, allTags);
for (let subtag of allTags) {
if (!initialMap.has(subtag)) initialMap.set(subtag, new Set<string>());
initialMap.get(subtag)?.add(file.path);
}
}
let totalTimeMs = new Date().getTime() - timeStart;
console.log(`Dataview: Parsed ${initialMap.size} tags in ${initialInvMap.size} markdown files (${totalTimeMs / 1000.0}s)`);
return new TagIndex(vault, cache, initialMap, initialInvMap);
}
/** Maps tags -> set of files containing that exact tag. */
map: Map<string, Set<string>>;
/** Cached inverse map; maps file -> tags it was last known to contain. */
invMap: Map<string, Set<string>>;
vault: Vault;
cache: MetadataCache;
constructor(vault: Vault, metadataCache: MetadataCache,
map: Map<string, Set<string>>, invMap: Map<string, Set<string>>) {
this.vault = vault;
this.cache = metadataCache;
this.map = map;
this.invMap = invMap;
}
/** Returns all files which have the given tag. */
public get(tag: string): Set<string> {
let result = this.map.get(tag);
if (result) {
return new Set(result);
} else {
return new Set();
}
}
/** Returns all tags the given file has. */
public getInverse(file: string): Set<string> {
let result = this.invMap.get(file);
if (result) {
return new Set(result);
} else {
return new Set();
}
}
async reloadFile(file: TFile) {
this.clearFile(file.path);
let allTags = TagIndex.parseTags(this.cache, file.path);
for (let subtag of allTags) {
if (!this.map.has(subtag)) this.map.set(subtag, new Set<string>());
this.map.get(subtag)?.add(file.path);
}
this.invMap.set(file.path, allTags);
}
/** Clears all tags for the given file so they can be re-added. */
private clearFile(path: string) {
let oldTags = this.invMap.get(path);
if (!oldTags) return;
this.invMap.delete(path);
for (let tag of oldTags) {
this.map.get(tag)?.delete(path);
}
}
}
/** A node in the prefix tree. */
export class | {
// TODO: Instead of only storing file paths at the leaf, consider storing them at every level,
// since this will make for faster deletes and gathers in exchange for slightly slower adds and more memory usage.
// since we are optimizing for gather, and file paths tend to be shallow, this should be ok.
files: Set<string>;
element: string;
totalCount: number;
children: Map<string, PrefixIndexNode>;
constructor(element: string) {
this.element = element;
this.files = new Set();
this.totalCount = 0;
this.children = new Map();
}
public static add(root: PrefixIndexNode, path: string) {
let parts = path.split("/");
let node = root;
for (let index = 0; index < parts.length - 1; index++) {
if (!node.children.has(parts[index])) node.children.set(parts[index], new PrefixIndexNode(parts[index]));
node.totalCount += 1;
node = node.children.get(parts[index]) as PrefixIndexNode;
}
node.totalCount += 1;
node.files.add(path);
}
public static remove(root: PrefixIndexNode, path: string) {
let parts = path.split("/");
let node = root;
let nodes = [];
for (let index = 0; index < parts.length - 1; index++) {
if (!node.children.has(parts[index])) return;
nodes.push(node);
node = node.children.get(parts[index]) as PrefixIndexNode;
}
if (!node.files.has(path)) return;
node.files.delete(path);
node.totalCount -= 1;
for (let p of nodes) p.totalCount -= 1;
}
public static find(root: PrefixIndexNode, prefix: string): PrefixIndexNode | null {
if (prefix.length == 0 || prefix == '/') return root;
let parts = prefix.split("/");
let node = root;
for (let index = 0; index < parts.length; index++) {
if (!node.children.has(parts[index])) return null;
node = node.children.get(parts[index]) as PrefixIndexNode;
}
return node;
}
public static gather(root: PrefixIndexNode): Set<string> {
let result = new Set<string>();
PrefixIndexNode.gatherRec(root, result);
return result;
}
static gatherRec(root: PrefixIndexNode, output: Set<string>) {
for (let file of root.files) output.add(file);
for (let child of root.children.values()) this.gatherRec(child, output);
}
}
/** Indexes files by their full prefix - | PrefixIndexNode | identifier_name |
index.ts | = fileCache.frontmatter;
// Search for the 'tags' field, since it may have wierd
let tagsName: string | undefined = undefined;
for (let key of Object.keys(frontCache ?? {})) {
if (key.toLowerCase() == "tags" || key.toLowerCase() == "tag")
tagsName = key;
}
if (frontCache && tagsName && frontCache[tagsName]) {
if (Array.isArray(frontCache[tagsName])) {
for (let tag of frontCache[tagsName]) {
if (!(typeof tag == 'string')) continue;
if (!tag.startsWith("#")) tag = "#" + tag;
this.parseSubtags(tag).forEach(t => allTags.add(t));
}
} else if (typeof frontCache[tagsName] === 'string') {
// Assume tags is a comma-separated list.
let tags = (frontCache[tagsName] as string).split(",").map(elem => {
elem = elem.trim();
if (!elem.startsWith("#")) elem = "#" + elem;
return elem;
});
for (let tag of tags) {
this.parseSubtags(tag).forEach(t => allTags.add(t));
}
}
}
return allTags;
}
public static async generate(vault: Vault, cache: MetadataCache): Promise<TagIndex> {
let initialMap = new Map<string, Set<string>>();
let initialInvMap = new Map<string, Set<string>>();
let timeStart = new Date().getTime();
// First time load...
for (let file of vault.getMarkdownFiles()) {
let allTags = TagIndex.parseTags(cache, file.path);
initialInvMap.set(file.path, allTags);
for (let subtag of allTags) {
if (!initialMap.has(subtag)) initialMap.set(subtag, new Set<string>());
initialMap.get(subtag)?.add(file.path);
}
}
let totalTimeMs = new Date().getTime() - timeStart;
console.log(`Dataview: Parsed ${initialMap.size} tags in ${initialInvMap.size} markdown files (${totalTimeMs / 1000.0}s)`);
return new TagIndex(vault, cache, initialMap, initialInvMap);
}
/** Maps tags -> set of files containing that exact tag. */
map: Map<string, Set<string>>;
/** Cached inverse map; maps file -> tags it was last known to contain. */
invMap: Map<string, Set<string>>;
vault: Vault;
cache: MetadataCache;
constructor(vault: Vault, metadataCache: MetadataCache,
map: Map<string, Set<string>>, invMap: Map<string, Set<string>>) {
this.vault = vault;
this.cache = metadataCache;
this.map = map;
this.invMap = invMap;
}
/** Returns all files which have the given tag. */
public get(tag: string): Set<string> {
let result = this.map.get(tag);
if (result) {
return new Set(result);
} else {
return new Set();
}
}
/** Returns all tags the given file has. */
public getInverse(file: string): Set<string> {
let result = this.invMap.get(file);
if (result) {
return new Set(result);
} else {
return new Set();
}
}
async reloadFile(file: TFile) {
this.clearFile(file.path);
let allTags = TagIndex.parseTags(this.cache, file.path);
for (let subtag of allTags) {
if (!this.map.has(subtag)) this.map.set(subtag, new Set<string>());
this.map.get(subtag)?.add(file.path);
}
this.invMap.set(file.path, allTags);
}
/** Clears all tags for the given file so they can be re-added. */
private clearFile(path: string) {
let oldTags = this.invMap.get(path);
if (!oldTags) return;
this.invMap.delete(path);
for (let tag of oldTags) {
this.map.get(tag)?.delete(path);
}
}
}
/** A node in the prefix tree. */
export class PrefixIndexNode {
// TODO: Instead of only storing file paths at the leaf, consider storing them at every level,
// since this will make for faster deletes and gathers in exchange for slightly slower adds and more memory usage.
// since we are optimizing for gather, and file paths tend to be shallow, this should be ok.
files: Set<string>;
element: string;
totalCount: number;
children: Map<string, PrefixIndexNode>;
constructor(element: string) {
this.element = element;
this.files = new Set();
this.totalCount = 0;
this.children = new Map();
}
public static add(root: PrefixIndexNode, path: string) {
let parts = path.split("/");
let node = root;
for (let index = 0; index < parts.length - 1; index++) {
if (!node.children.has(parts[index])) node.children.set(parts[index], new PrefixIndexNode(parts[index]));
node.totalCount += 1;
node = node.children.get(parts[index]) as PrefixIndexNode;
}
node.totalCount += 1;
node.files.add(path);
}
public static remove(root: PrefixIndexNode, path: string) {
let parts = path.split("/");
let node = root;
let nodes = [];
for (let index = 0; index < parts.length - 1; index++) {
if (!node.children.has(parts[index])) return;
nodes.push(node);
node = node.children.get(parts[index]) as PrefixIndexNode;
}
if (!node.files.has(path)) return;
node.files.delete(path);
node.totalCount -= 1;
for (let p of nodes) p.totalCount -= 1;
}
public static find(root: PrefixIndexNode, prefix: string): PrefixIndexNode | null {
if (prefix.length == 0 || prefix == '/') return root;
let parts = prefix.split("/");
let node = root;
for (let index = 0; index < parts.length; index++) {
if (!node.children.has(parts[index])) return null;
node = node.children.get(parts[index]) as PrefixIndexNode;
}
return node;
}
public static gather(root: PrefixIndexNode): Set<string> {
let result = new Set<string>();
PrefixIndexNode.gatherRec(root, result);
return result;
}
static gatherRec(root: PrefixIndexNode, output: Set<string>) {
for (let file of root.files) output.add(file);
for (let child of root.children.values()) this.gatherRec(child, output);
}
}
/** Indexes files by their full prefix - essentially a simple prefix tree. */
export class PrefixIndex {
public static async generate(vault: Vault): Promise<PrefixIndex> {
let root = new PrefixIndexNode("");
let timeStart = new Date().getTime();
// First time load...
for (let file of vault.getMarkdownFiles()) {
PrefixIndexNode.add(root, file.path);
}
let totalTimeMs = new Date().getTime() - timeStart;
console.log(`Dataview: Parsed all file prefixes (${totalTimeMs / 1000.0}s)`);
return Promise.resolve(new PrefixIndex(vault, root));
}
root: PrefixIndexNode;
vault: Vault;
constructor(vault: Vault, root: PrefixIndexNode) {
this.vault = vault;
this.root = root;
// TODO: I'm not sure if there is an event for all files in a folder, or just the folder.
// I'm assuming the former naively for now until I inevitably fix it.
this.vault.on("delete", file => {
PrefixIndexNode.remove(this.root, file.path);
});
this.vault.on("create", file => {
PrefixIndexNode.add(this.root, file.path);
});
this.vault.on("rename", (file, old) => {
PrefixIndexNode.remove(this.root, old);
PrefixIndexNode.add(this.root, file.path);
});
}
public get(prefix: string): Set<string> {
let node = PrefixIndexNode.find(this.root, prefix);
if (node == null || node == undefined) return new Set();
return PrefixIndexNode.gather(node);
}
}
/** Caches tasks for each file to avoid repeated re-loading. */
export class TaskCache {
/** Create a task cache for the given vault. */
static async generate(vault: Vault): Promise<TaskCache> {
let initialCache: Record<string, Task[]> = {};
let timeStart = new Date().getTime();
// First time load...
for (let file of vault.getMarkdownFiles()) {
let tasks = await Tasks.findTasksInFile(vault, file);
if (tasks.length == 0) continue;
initialCache[file.path] = tasks;
}
let totalTimeMs = new Date().getTime() - timeStart;
console.log(`Dataview: Parsed tasks in ${Object.keys(initialCache).length} markdown files (${totalTimeMs / 1000.0}s)`);
return new TaskCache(vault, initialCache);
}
cache: Record<string, Task[]>;
vault: Vault;
constructor(vault: Vault, cache: Record<string, Task[]>) | {
this.vault = vault;
this.cache = cache;
} | identifier_body |
|
simple-job-board-public.js | * - Validate Email
* - Initialize TelInput Plugin
* - Validate Phone Number
* - Allowable Uploaded File's Extensions
* - Validate Required Inputs ( Attachment, Phone & Email )
* - Checkbox Group Required Attribute Callbacks
* - Custom Styling of File Upload Button
*/
(function ($) {
'use strict';
$(document).ready(function () {
var jobpost_submit_button = $('.app-submit');
$(".jobpost-form").on("submit", function (event) {
var jobpost_form_status = $('#jobpost_form_status');
var datastring = new FormData(document.getElementById("sjb-application-form"));
/**
* Application Form Submit -> Validate Email & Phone
* @since 2.2.0
*/
var is_valid_email = sjb_is_valid_input(event, "email", "sjb-email-address");
var is_valid_phone = sjb_is_valid_input(event, "phone", "sjb-phone-number");
var is_attachment = sjb_is_attachment(event);
/* Stop Form Submission on Invalid Phone, Email & File Attachement */
if ( !is_valid_email || !is_valid_phone || !is_attachment ) {
return false;
}
$.ajax({
url: application_form.ajaxurl,
type: 'POST',
dataType: 'json',
data: datastring,
async: false,
cache: false,
contentType: false,
processData: false,
beforeSend: function () {
jobpost_form_status.html('Submitting.....');
jobpost_submit_button.attr('disabled', 'diabled');
},
success: function ( response ) {
if ( response['success'] == true ) {
$('.jobpost-form').slideUp();
/* Translation Ready String Through Script Locaization */
jobpost_form_status.html(response['success_alert']);
}
if ( response['success'] == false ) {
/* Translation Ready String Through Script Locaization */
jobpost_form_status.html( response['error'] + ' ' + application_form.jquery_alerts['application_not_submitted'] + '</div>' );
jobpost_submit_button.removeAttr( 'disabled' );
}
}
});
return false;
});
/* Date Picker */
$('.sjb-datepicker').datepicker({
dateFormat: 'dd-mm-yy',
changeMonth: true,
changeYear: true
});
/**
* Application Form -> On Input Email Validation
*
* @since 2.2.0
*/
$('.sjb-email-address').on('input', function () {
var input = $(this);
var re = /^[a-zA-Z0-9.!#$%&'*+/=?^_`{|}~-]+@[a-zA-Z0-9-]+(?:\.[a-zA-Z0-9-]+)*$/;
var is_email = re.test(input.val());
var error_element = $("span", $(this).parent());
if (is_email) {
input.removeClass("invalid").addClass("valid");
error_element.hide();
} else {
input.removeClass("valid").addClass("invalid");
}
});
/**
* Initialize TelInput Plugin
*
* @since 2.2.0
*/
if ($('.sjb-phone-number').length) {
var telInput_id = $('.sjb-phone-number').map(function () {
return this.id;
}).get();
for (var input_ID in telInput_id) {
var telInput = $('#' + telInput_id[input_ID]);
telInput.intlTelInput({
initialCountry: "auto",
geoIpLookup: function (callback) {
$.get('http://ipinfo.io', function () {
}, "jsonp").always(function (resp) {
var countryCode = (resp && resp.country) ? resp.country : "";
callback(countryCode);
});
},
});
}
}
/**
* Application Form -> Phone Number Validation
*
* @since 2.2.0
*/
$('.sjb-phone-number').on('input', function () {
var telInput = $(this);
var telInput_id = $(this).attr('id');
var error_element = $("#" + telInput_id + "-invalid-phone");
error_element.hide();
// Validate Phone Number
if ($.trim(telInput.val())) {
if (telInput.intlTelInput("isValidNumber")) {
telInput.removeClass("invalid").addClass("valid");
error_element.hide();
} else {
telInput.removeClass("valid").addClass("invalid");
}
}
});
/**
* Check for Allowable Extensions of Uploaded File
*
* @since 2.3.0
*/
$('.sjb-attachment').on('change', function () {
var input = $(this);
var file = $("#" + $(this).attr("id"));
var error_element = file.parent().next("span");
error_element.text('');
error_element.hide();
// Validate on File Attachment
if ( 0 != file.get(0).files.length ) {
/**
* Uploded File Extensions Checks
* Get Uploded File Ext
*/
var file_ext = file.val().split('.').pop().toLowerCase();
// All Allowed File Extensions
var allowed_file_exts = application_form.allowed_extensions;
// Settings File Extensions && Getting value From Script Localization
var settings_file_exts = application_form.setting_extensions;
var selected_file_exts = (('yes' === application_form.all_extensions_check) || null == settings_file_exts) ? allowed_file_exts : settings_file_exts;
// File Extension Validation
if ($.inArray(file_ext, selected_file_exts) > -1) {
jobpost_submit_button.attr( 'disabled', false );
input.removeClass("invalid").addClass("valid");
} else {
/* Translation Ready String Through Script Locaization */
error_element.text(application_form.jquery_alerts['invalid_extension']);
error_element.show();
input.removeClass("valid").addClass("invalid");
}
}
});
/**
* Stop Form Submission -> On Required Attachments
*
* @since 2.3.0
*/
function | ( event ) {
var error_free = true;
$(".sjb-attachment").each(function () {
var element = $("#" + $(this).attr("id"));
var valid = element.hasClass("valid");
var is_required_class = element.hasClass("sjb-not-required");
// Set Error Indicator on Invalid Attachment
if (!valid) {
if (!(is_required_class && 0 === element.get(0).files.length)) {
error_free = false;
}
}
// Stop Form Submission
if (!error_free) {
event.preventDefault();
}
});
return error_free;
}
/**
* Stop Form Submission -> On Invalid Email/Phone
*
* @since 2.2.0
*/
function sjb_is_valid_input(event, input_type, input_class) {
var jobpost_form_inputs = $("." + input_class).serializeArray();
var error_free = true;
for (var i in jobpost_form_inputs) {
var element = $("#" + jobpost_form_inputs[i]['name']);
var valid = element.hasClass("valid");
var is_required_class = element.hasClass("sjb-not-required");
if (!(is_required_class && "" === jobpost_form_inputs[i]['value'])) {
if ("email" === input_type) {
var error_element = $("span", element.parent());
} else if ("phone" === input_type) {
var error_element = $("#" + jobpost_form_inputs[i]['name'] + "-invalid-phone");
}
// Set Error Indicator on Invalid Input
if (!valid) {
error_element.show();
error_free = false;
}
else {
error_element.hide();
}
// Stop Form Submission
if (!error_free) {
event.preventDefault();
}
}
}
return error_free;
}
/**
* Remove Required Attribute from Checkbox Group -> When one of the option is selected.
* Add Required Attribute from Checkboxes Group -> When none of the option is selected.
*
* @since 2.3.0
*/
var requiredCheckboxes = $(':checkbox[required]');
requiredCheckboxes.on('change', function () {
var checkboxGroup = requiredCheckboxes.filter('[name="' + $(this).attr('name') + '"]');
var isChecked = checkboxGroup.is(':checked');
checkboxGroup.prop('required', !isChecked);
});
});
/*
* Custom Styling of Upload Field Button
*
* @since 2.4.0
*/
var file = {
maxlength: 20, // maximum length of filename before it's trimmed
convert: function () {
// Convert all file type inputs.
$('input[type=file].sjb-attachment').each(function () {
$(this).wrap('<div class="file" />');
$(this).parent().prepend('<div>'+ application_form.file['browse']+'</div>');
$(this).parent().prepend('<span>'+ application_form.file['no_file_chosen']+'</span>');
$(this).fadeTo(0, 0);
$(this).attr('size', '50'); // Use this to adjust width for FireFox.
});
},
update: function (x) {
// Update the filename | sjb_is_attachment | identifier_name |
simple-job-board-public.js | * - Validate Email
* - Initialize TelInput Plugin
* - Validate Phone Number
* - Allowable Uploaded File's Extensions
* - Validate Required Inputs ( Attachment, Phone & Email )
* - Checkbox Group Required Attribute Callbacks
* - Custom Styling of File Upload Button
*/
(function ($) {
'use strict';
$(document).ready(function () {
var jobpost_submit_button = $('.app-submit');
$(".jobpost-form").on("submit", function (event) {
var jobpost_form_status = $('#jobpost_form_status');
var datastring = new FormData(document.getElementById("sjb-application-form"));
/**
* Application Form Submit -> Validate Email & Phone
* @since 2.2.0
*/
var is_valid_email = sjb_is_valid_input(event, "email", "sjb-email-address");
var is_valid_phone = sjb_is_valid_input(event, "phone", "sjb-phone-number");
var is_attachment = sjb_is_attachment(event);
/* Stop Form Submission on Invalid Phone, Email & File Attachement */
if ( !is_valid_email || !is_valid_phone || !is_attachment ) {
return false;
}
$.ajax({
url: application_form.ajaxurl,
type: 'POST',
dataType: 'json',
data: datastring,
async: false,
cache: false,
contentType: false,
processData: false,
beforeSend: function () {
jobpost_form_status.html('Submitting.....');
jobpost_submit_button.attr('disabled', 'diabled');
},
success: function ( response ) {
if ( response['success'] == true ) {
$('.jobpost-form').slideUp();
/* Translation Ready String Through Script Locaization */
jobpost_form_status.html(response['success_alert']);
}
if ( response['success'] == false ) {
/* Translation Ready String Through Script Locaization */
jobpost_form_status.html( response['error'] + ' ' + application_form.jquery_alerts['application_not_submitted'] + '</div>' );
jobpost_submit_button.removeAttr( 'disabled' );
}
}
});
return false;
});
/* Date Picker */
$('.sjb-datepicker').datepicker({
dateFormat: 'dd-mm-yy',
changeMonth: true,
changeYear: true
});
/**
* Application Form -> On Input Email Validation
*
* @since 2.2.0
*/
$('.sjb-email-address').on('input', function () {
var input = $(this);
var re = /^[a-zA-Z0-9.!#$%&'*+/=?^_`{|}~-]+@[a-zA-Z0-9-]+(?:\.[a-zA-Z0-9-]+)*$/;
var is_email = re.test(input.val());
var error_element = $("span", $(this).parent());
if (is_email) {
input.removeClass("invalid").addClass("valid");
error_element.hide();
} else {
input.removeClass("valid").addClass("invalid"); | }
});
/**
* Initialize TelInput Plugin
*
* @since 2.2.0
*/
if ($('.sjb-phone-number').length) {
var telInput_id = $('.sjb-phone-number').map(function () {
return this.id;
}).get();
for (var input_ID in telInput_id) {
var telInput = $('#' + telInput_id[input_ID]);
telInput.intlTelInput({
initialCountry: "auto",
geoIpLookup: function (callback) {
$.get('http://ipinfo.io', function () {
}, "jsonp").always(function (resp) {
var countryCode = (resp && resp.country) ? resp.country : "";
callback(countryCode);
});
},
});
}
}
/**
* Application Form -> Phone Number Validation
*
* @since 2.2.0
*/
$('.sjb-phone-number').on('input', function () {
var telInput = $(this);
var telInput_id = $(this).attr('id');
var error_element = $("#" + telInput_id + "-invalid-phone");
error_element.hide();
// Validate Phone Number
if ($.trim(telInput.val())) {
if (telInput.intlTelInput("isValidNumber")) {
telInput.removeClass("invalid").addClass("valid");
error_element.hide();
} else {
telInput.removeClass("valid").addClass("invalid");
}
}
});
/**
* Check for Allowable Extensions of Uploaded File
*
* @since 2.3.0
*/
$('.sjb-attachment').on('change', function () {
var input = $(this);
var file = $("#" + $(this).attr("id"));
var error_element = file.parent().next("span");
error_element.text('');
error_element.hide();
// Validate on File Attachment
if ( 0 != file.get(0).files.length ) {
/**
* Uploded File Extensions Checks
* Get Uploded File Ext
*/
var file_ext = file.val().split('.').pop().toLowerCase();
// All Allowed File Extensions
var allowed_file_exts = application_form.allowed_extensions;
// Settings File Extensions && Getting value From Script Localization
var settings_file_exts = application_form.setting_extensions;
var selected_file_exts = (('yes' === application_form.all_extensions_check) || null == settings_file_exts) ? allowed_file_exts : settings_file_exts;
// File Extension Validation
if ($.inArray(file_ext, selected_file_exts) > -1) {
jobpost_submit_button.attr( 'disabled', false );
input.removeClass("invalid").addClass("valid");
} else {
/* Translation Ready String Through Script Locaization */
error_element.text(application_form.jquery_alerts['invalid_extension']);
error_element.show();
input.removeClass("valid").addClass("invalid");
}
}
});
/**
* Stop Form Submission -> On Required Attachments
*
* @since 2.3.0
*/
function sjb_is_attachment( event ) {
var error_free = true;
$(".sjb-attachment").each(function () {
var element = $("#" + $(this).attr("id"));
var valid = element.hasClass("valid");
var is_required_class = element.hasClass("sjb-not-required");
// Set Error Indicator on Invalid Attachment
if (!valid) {
if (!(is_required_class && 0 === element.get(0).files.length)) {
error_free = false;
}
}
// Stop Form Submission
if (!error_free) {
event.preventDefault();
}
});
return error_free;
}
/**
* Stop Form Submission -> On Invalid Email/Phone
*
* @since 2.2.0
*/
function sjb_is_valid_input(event, input_type, input_class) {
var jobpost_form_inputs = $("." + input_class).serializeArray();
var error_free = true;
for (var i in jobpost_form_inputs) {
var element = $("#" + jobpost_form_inputs[i]['name']);
var valid = element.hasClass("valid");
var is_required_class = element.hasClass("sjb-not-required");
if (!(is_required_class && "" === jobpost_form_inputs[i]['value'])) {
if ("email" === input_type) {
var error_element = $("span", element.parent());
} else if ("phone" === input_type) {
var error_element = $("#" + jobpost_form_inputs[i]['name'] + "-invalid-phone");
}
// Set Error Indicator on Invalid Input
if (!valid) {
error_element.show();
error_free = false;
}
else {
error_element.hide();
}
// Stop Form Submission
if (!error_free) {
event.preventDefault();
}
}
}
return error_free;
}
/**
* Remove Required Attribute from Checkbox Group -> When one of the option is selected.
* Add Required Attribute from Checkboxes Group -> When none of the option is selected.
*
* @since 2.3.0
*/
var requiredCheckboxes = $(':checkbox[required]');
requiredCheckboxes.on('change', function () {
var checkboxGroup = requiredCheckboxes.filter('[name="' + $(this).attr('name') + '"]');
var isChecked = checkboxGroup.is(':checked');
checkboxGroup.prop('required', !isChecked);
});
});
/*
* Custom Styling of Upload Field Button
*
* @since 2.4.0
*/
var file = {
maxlength: 20, // maximum length of filename before it's trimmed
convert: function () {
// Convert all file type inputs.
$('input[type=file].sjb-attachment').each(function () {
$(this).wrap('<div class="file" />');
$(this).parent().prepend('<div>'+ application_form.file['browse']+'</div>');
$(this).parent().prepend('<span>'+ application_form.file['no_file_chosen']+'</span>');
$(this).fadeTo(0, 0);
$(this).attr('size', '50'); // Use this to adjust width for FireFox.
});
},
update: function (x) {
// Update the filename display | random_line_split |
|
simple-job-board-public.js | * - Validate Email
* - Initialize TelInput Plugin
* - Validate Phone Number
* - Allowable Uploaded File's Extensions
* - Validate Required Inputs ( Attachment, Phone & Email )
* - Checkbox Group Required Attribute Callbacks
* - Custom Styling of File Upload Button
*/
(function ($) {
'use strict';
$(document).ready(function () {
var jobpost_submit_button = $('.app-submit');
$(".jobpost-form").on("submit", function (event) {
var jobpost_form_status = $('#jobpost_form_status');
var datastring = new FormData(document.getElementById("sjb-application-form"));
/**
* Application Form Submit -> Validate Email & Phone
* @since 2.2.0
*/
var is_valid_email = sjb_is_valid_input(event, "email", "sjb-email-address");
var is_valid_phone = sjb_is_valid_input(event, "phone", "sjb-phone-number");
var is_attachment = sjb_is_attachment(event);
/* Stop Form Submission on Invalid Phone, Email & File Attachement */
if ( !is_valid_email || !is_valid_phone || !is_attachment ) {
return false;
}
$.ajax({
url: application_form.ajaxurl,
type: 'POST',
dataType: 'json',
data: datastring,
async: false,
cache: false,
contentType: false,
processData: false,
beforeSend: function () {
jobpost_form_status.html('Submitting.....');
jobpost_submit_button.attr('disabled', 'diabled');
},
success: function ( response ) {
if ( response['success'] == true ) {
$('.jobpost-form').slideUp();
/* Translation Ready String Through Script Locaization */
jobpost_form_status.html(response['success_alert']);
}
if ( response['success'] == false ) {
/* Translation Ready String Through Script Locaization */
jobpost_form_status.html( response['error'] + ' ' + application_form.jquery_alerts['application_not_submitted'] + '</div>' );
jobpost_submit_button.removeAttr( 'disabled' );
}
}
});
return false;
});
/* Date Picker */
$('.sjb-datepicker').datepicker({
dateFormat: 'dd-mm-yy',
changeMonth: true,
changeYear: true
});
/**
* Application Form -> On Input Email Validation
*
* @since 2.2.0
*/
$('.sjb-email-address').on('input', function () {
var input = $(this);
var re = /^[a-zA-Z0-9.!#$%&'*+/=?^_`{|}~-]+@[a-zA-Z0-9-]+(?:\.[a-zA-Z0-9-]+)*$/;
var is_email = re.test(input.val());
var error_element = $("span", $(this).parent());
if (is_email) {
input.removeClass("invalid").addClass("valid");
error_element.hide();
} else {
input.removeClass("valid").addClass("invalid");
}
});
/**
* Initialize TelInput Plugin
*
* @since 2.2.0
*/
if ($('.sjb-phone-number').length) {
var telInput_id = $('.sjb-phone-number').map(function () {
return this.id;
}).get();
for (var input_ID in telInput_id) {
var telInput = $('#' + telInput_id[input_ID]);
telInput.intlTelInput({
initialCountry: "auto",
geoIpLookup: function (callback) {
$.get('http://ipinfo.io', function () {
}, "jsonp").always(function (resp) {
var countryCode = (resp && resp.country) ? resp.country : "";
callback(countryCode);
});
},
});
}
}
/**
* Application Form -> Phone Number Validation
*
* @since 2.2.0
*/
$('.sjb-phone-number').on('input', function () {
var telInput = $(this);
var telInput_id = $(this).attr('id');
var error_element = $("#" + telInput_id + "-invalid-phone");
error_element.hide();
// Validate Phone Number
if ($.trim(telInput.val())) {
if (telInput.intlTelInput("isValidNumber")) {
telInput.removeClass("invalid").addClass("valid");
error_element.hide();
} else {
telInput.removeClass("valid").addClass("invalid");
}
}
});
/**
* Check for Allowable Extensions of Uploaded File
*
* @since 2.3.0
*/
$('.sjb-attachment').on('change', function () {
var input = $(this);
var file = $("#" + $(this).attr("id"));
var error_element = file.parent().next("span");
error_element.text('');
error_element.hide();
// Validate on File Attachment
if ( 0 != file.get(0).files.length ) {
/**
* Uploded File Extensions Checks
* Get Uploded File Ext
*/
var file_ext = file.val().split('.').pop().toLowerCase();
// All Allowed File Extensions
var allowed_file_exts = application_form.allowed_extensions;
// Settings File Extensions && Getting value From Script Localization
var settings_file_exts = application_form.setting_extensions;
var selected_file_exts = (('yes' === application_form.all_extensions_check) || null == settings_file_exts) ? allowed_file_exts : settings_file_exts;
// File Extension Validation
if ($.inArray(file_ext, selected_file_exts) > -1) {
jobpost_submit_button.attr( 'disabled', false );
input.removeClass("invalid").addClass("valid");
} else {
/* Translation Ready String Through Script Locaization */
error_element.text(application_form.jquery_alerts['invalid_extension']);
error_element.show();
input.removeClass("valid").addClass("invalid");
}
}
});
/**
* Stop Form Submission -> On Required Attachments
*
* @since 2.3.0
*/
function sjb_is_attachment( event ) {
var error_free = true;
$(".sjb-attachment").each(function () {
var element = $("#" + $(this).attr("id"));
var valid = element.hasClass("valid");
var is_required_class = element.hasClass("sjb-not-required");
// Set Error Indicator on Invalid Attachment
if (!valid) {
if (!(is_required_class && 0 === element.get(0).files.length)) {
error_free = false;
}
}
// Stop Form Submission
if (!error_free) {
event.preventDefault();
}
});
return error_free;
}
/**
* Stop Form Submission -> On Invalid Email/Phone
*
* @since 2.2.0
*/
function sjb_is_valid_input(event, input_type, input_class) | else {
error_element.hide();
}
// Stop Form Submission
if (!error_free) {
event.preventDefault();
}
}
}
return error_free;
}
/**
* Remove Required Attribute from Checkbox Group -> When one of the option is selected.
* Add Required Attribute from Checkboxes Group -> When none of the option is selected.
*
* @since 2.3.0
*/
var requiredCheckboxes = $(':checkbox[required]');
requiredCheckboxes.on('change', function () {
var checkboxGroup = requiredCheckboxes.filter('[name="' + $(this).attr('name') + '"]');
var isChecked = checkboxGroup.is(':checked');
checkboxGroup.prop('required', !isChecked);
});
});
/*
* Custom Styling of Upload Field Button
*
* @since 2.4.0
*/
var file = {
maxlength: 20, // maximum length of filename before it's trimmed
convert: function () {
// Convert all file type inputs.
$('input[type=file].sjb-attachment').each(function () {
$(this).wrap('<div class="file" />');
$(this).parent().prepend('<div>'+ application_form.file['browse']+'</div>');
$(this).parent().prepend('<span>'+ application_form.file['no_file_chosen']+'</span>');
$(this).fadeTo(0, 0);
$(this).attr('size', '50'); // Use this to adjust width for FireFox.
});
},
update: function (x) {
// Update the filename | {
var jobpost_form_inputs = $("." + input_class).serializeArray();
var error_free = true;
for (var i in jobpost_form_inputs) {
var element = $("#" + jobpost_form_inputs[i]['name']);
var valid = element.hasClass("valid");
var is_required_class = element.hasClass("sjb-not-required");
if (!(is_required_class && "" === jobpost_form_inputs[i]['value'])) {
if ("email" === input_type) {
var error_element = $("span", element.parent());
} else if ("phone" === input_type) {
var error_element = $("#" + jobpost_form_inputs[i]['name'] + "-invalid-phone");
}
// Set Error Indicator on Invalid Input
if (!valid) {
error_element.show();
error_free = false;
} | identifier_body |
simple-job-board-public.js | * - Validate Email
* - Initialize TelInput Plugin
* - Validate Phone Number
* - Allowable Uploaded File's Extensions
* - Validate Required Inputs ( Attachment, Phone & Email )
* - Checkbox Group Required Attribute Callbacks
* - Custom Styling of File Upload Button
*/
(function ($) {
'use strict';
$(document).ready(function () {
var jobpost_submit_button = $('.app-submit');
$(".jobpost-form").on("submit", function (event) {
var jobpost_form_status = $('#jobpost_form_status');
var datastring = new FormData(document.getElementById("sjb-application-form"));
/**
* Application Form Submit -> Validate Email & Phone
* @since 2.2.0
*/
var is_valid_email = sjb_is_valid_input(event, "email", "sjb-email-address");
var is_valid_phone = sjb_is_valid_input(event, "phone", "sjb-phone-number");
var is_attachment = sjb_is_attachment(event);
/* Stop Form Submission on Invalid Phone, Email & File Attachement */
if ( !is_valid_email || !is_valid_phone || !is_attachment ) {
return false;
}
$.ajax({
url: application_form.ajaxurl,
type: 'POST',
dataType: 'json',
data: datastring,
async: false,
cache: false,
contentType: false,
processData: false,
beforeSend: function () {
jobpost_form_status.html('Submitting.....');
jobpost_submit_button.attr('disabled', 'diabled');
},
success: function ( response ) {
if ( response['success'] == true ) {
$('.jobpost-form').slideUp();
/* Translation Ready String Through Script Locaization */
jobpost_form_status.html(response['success_alert']);
}
if ( response['success'] == false ) {
/* Translation Ready String Through Script Locaization */
jobpost_form_status.html( response['error'] + ' ' + application_form.jquery_alerts['application_not_submitted'] + '</div>' );
jobpost_submit_button.removeAttr( 'disabled' );
}
}
});
return false;
});
/* Date Picker */
$('.sjb-datepicker').datepicker({
dateFormat: 'dd-mm-yy',
changeMonth: true,
changeYear: true
});
/**
* Application Form -> On Input Email Validation
*
* @since 2.2.0
*/
$('.sjb-email-address').on('input', function () {
var input = $(this);
var re = /^[a-zA-Z0-9.!#$%&'*+/=?^_`{|}~-]+@[a-zA-Z0-9-]+(?:\.[a-zA-Z0-9-]+)*$/;
var is_email = re.test(input.val());
var error_element = $("span", $(this).parent());
if (is_email) | else {
input.removeClass("valid").addClass("invalid");
}
});
/**
* Initialize TelInput Plugin
*
* @since 2.2.0
*/
if ($('.sjb-phone-number').length) {
var telInput_id = $('.sjb-phone-number').map(function () {
return this.id;
}).get();
for (var input_ID in telInput_id) {
var telInput = $('#' + telInput_id[input_ID]);
telInput.intlTelInput({
initialCountry: "auto",
geoIpLookup: function (callback) {
$.get('http://ipinfo.io', function () {
}, "jsonp").always(function (resp) {
var countryCode = (resp && resp.country) ? resp.country : "";
callback(countryCode);
});
},
});
}
}
/**
* Application Form -> Phone Number Validation
*
* @since 2.2.0
*/
$('.sjb-phone-number').on('input', function () {
var telInput = $(this);
var telInput_id = $(this).attr('id');
var error_element = $("#" + telInput_id + "-invalid-phone");
error_element.hide();
// Validate Phone Number
if ($.trim(telInput.val())) {
if (telInput.intlTelInput("isValidNumber")) {
telInput.removeClass("invalid").addClass("valid");
error_element.hide();
} else {
telInput.removeClass("valid").addClass("invalid");
}
}
});
/**
* Check for Allowable Extensions of Uploaded File
*
* @since 2.3.0
*/
$('.sjb-attachment').on('change', function () {
var input = $(this);
var file = $("#" + $(this).attr("id"));
var error_element = file.parent().next("span");
error_element.text('');
error_element.hide();
// Validate on File Attachment
if ( 0 != file.get(0).files.length ) {
/**
* Uploded File Extensions Checks
* Get Uploded File Ext
*/
var file_ext = file.val().split('.').pop().toLowerCase();
// All Allowed File Extensions
var allowed_file_exts = application_form.allowed_extensions;
// Settings File Extensions && Getting value From Script Localization
var settings_file_exts = application_form.setting_extensions;
var selected_file_exts = (('yes' === application_form.all_extensions_check) || null == settings_file_exts) ? allowed_file_exts : settings_file_exts;
// File Extension Validation
if ($.inArray(file_ext, selected_file_exts) > -1) {
jobpost_submit_button.attr( 'disabled', false );
input.removeClass("invalid").addClass("valid");
} else {
/* Translation Ready String Through Script Locaization */
error_element.text(application_form.jquery_alerts['invalid_extension']);
error_element.show();
input.removeClass("valid").addClass("invalid");
}
}
});
/**
* Stop Form Submission -> On Required Attachments
*
* @since 2.3.0
*/
function sjb_is_attachment( event ) {
var error_free = true;
$(".sjb-attachment").each(function () {
var element = $("#" + $(this).attr("id"));
var valid = element.hasClass("valid");
var is_required_class = element.hasClass("sjb-not-required");
// Set Error Indicator on Invalid Attachment
if (!valid) {
if (!(is_required_class && 0 === element.get(0).files.length)) {
error_free = false;
}
}
// Stop Form Submission
if (!error_free) {
event.preventDefault();
}
});
return error_free;
}
/**
* Stop Form Submission -> On Invalid Email/Phone
*
* @since 2.2.0
*/
function sjb_is_valid_input(event, input_type, input_class) {
var jobpost_form_inputs = $("." + input_class).serializeArray();
var error_free = true;
for (var i in jobpost_form_inputs) {
var element = $("#" + jobpost_form_inputs[i]['name']);
var valid = element.hasClass("valid");
var is_required_class = element.hasClass("sjb-not-required");
if (!(is_required_class && "" === jobpost_form_inputs[i]['value'])) {
if ("email" === input_type) {
var error_element = $("span", element.parent());
} else if ("phone" === input_type) {
var error_element = $("#" + jobpost_form_inputs[i]['name'] + "-invalid-phone");
}
// Set Error Indicator on Invalid Input
if (!valid) {
error_element.show();
error_free = false;
}
else {
error_element.hide();
}
// Stop Form Submission
if (!error_free) {
event.preventDefault();
}
}
}
return error_free;
}
/**
* Remove Required Attribute from Checkbox Group -> When one of the option is selected.
* Add Required Attribute from Checkboxes Group -> When none of the option is selected.
*
* @since 2.3.0
*/
var requiredCheckboxes = $(':checkbox[required]');
requiredCheckboxes.on('change', function () {
var checkboxGroup = requiredCheckboxes.filter('[name="' + $(this).attr('name') + '"]');
var isChecked = checkboxGroup.is(':checked');
checkboxGroup.prop('required', !isChecked);
});
});
/*
* Custom Styling of Upload Field Button
*
* @since 2.4.0
*/
var file = {
maxlength: 20, // maximum length of filename before it's trimmed
convert: function () {
// Convert all file type inputs.
$('input[type=file].sjb-attachment').each(function () {
$(this).wrap('<div class="file" />');
$(this).parent().prepend('<div>'+ application_form.file['browse']+'</div>');
$(this).parent().prepend('<span>'+ application_form.file['no_file_chosen']+'</span>');
$(this).fadeTo(0, 0);
$(this).attr('size', '50'); // Use this to adjust width for FireFox.
});
},
update: function (x) {
// Update the filename | {
input.removeClass("invalid").addClass("valid");
error_element.hide();
} | conditional_block |
train_timeline.py | Store some git revision info in a text file in the log directory
src_path,_ = os.path.split(os.path.realpath(__file__))
utils.store_revision_info(src_path, log_dir, ' '.join(sys.argv))
np.random.seed(seed=args.seed)
train_set = utils.get_dataset(args.data_dir)
nrof_classes = len(train_set)
print('nrof_classes: ',nrof_classes)
image_list, label_list = utils.get_image_paths_and_labels(train_set)
image_list = np.array(image_list)
print('total images: {}'.format(len(image_list)))
label_list = np.array(label_list,dtype=np.int32)
dataset_size = len(image_list)
data_reader = DataGenerator(image_list,label_list,args.batch_size)
print('Model directory: %s' % model_dir)
print('Log directory: %s' % log_dir)
if args.pretrained_model:
print('Pre-trained model: %s' % os.path.expanduser(args.pretrained_model))
with tf.Graph().as_default():
tf.set_random_seed(args.seed)
global_step = tf.Variable(0, trainable=False,name='global_step')
# Placeholder for the learning rate
learning_rate_placeholder = tf.placeholder(tf.float32, name='learning_rate')
images_placeholder = tf.placeholder(tf.float32,[None,112,96,3], name='images_placeholder')
#images_placeholder = tf.Variable(np.ones([args.batch_size,112,96,3]), dtype=np.float32,name='images_placeholder')
labels_placeholder = tf.placeholder(tf.int32,[None], name='labels_placeholder')
#labels_placeholder = tf.Variable(np.ones([args.batch_size,]),dtype=tf.int32, name='labels_placeholder')
phase_train_placeholder = tf.placeholder(tf.bool, name='phase_train')
learning_rate = tf.train.exponential_decay(learning_rate_placeholder, global_step,
args.learning_rate_decay_epochs*args.epoch_size, args.learning_rate_decay_factor, staircase=True)
tf.summary.scalar('learning_rate', learning_rate)
print('Using optimizer: {}'.format(args.optimizer))
if args.optimizer == 'ADAGRAD':
opt = tf.train.AdagradOptimizer(learning_rate)
elif args.optimizer == 'MOM':
opt = tf.train.MomentumOptimizer(learning_rate,0.9)
if args.network == 'sphere_network':
prelogits = network.infer(images_placeholder)
else:
raise Exception('Not supported network: {}'.format(args.loss_type))
if args.loss_type == 'softmax':
cross_entropy_mean = utils.softmax_loss(prelogits,labels_placeholder, len(train_set),args.weight_decay,False)
regularization_losses = tf.get_collection(tf.GraphKeys.REGULARIZATION_LOSSES)
#loss = cross_entropy_mean + args.weight_decay*tf.add_n(regularization_losses)
loss = cross_entropy_mean + args.weight_decay*tf.add_n(regularization_losses)
#loss = cross_entropy_mean
else:
raise Exception('Not supported loss type: {}'.format(args.loss_type))
#loss = tf.add_n([cross_entropy_mean] + regularization_losses, name='total_loss')
losses = {}
losses['total_loss'] = loss
losses['softmax_loss'] = cross_entropy_mean
debug_info = {}
debug_info ['prelogits'] = prelogits
grads = opt.compute_gradients(loss,tf.trainable_variables())
train_op = opt.apply_gradients(grads,global_step=global_step)
#save_vars = [var for var in tf.global_variables() if 'Adagrad' not in var.name and 'global_step' not in var.name]
save_vars = tf.global_variables()
#saver = tf.train.Saver(tf.trainable_variables(), max_to_keep=3)
saver = tf.train.Saver(save_vars, max_to_keep=3)
# Build the summary operation based on the TF collection of Summaries.
# Start running operations on the Graph.
gpu_options = tf.GPUOptions(per_process_gpu_memory_fraction=args.gpu_memory_fraction)
sess = tf.Session(config=tf.ConfigProto(gpu_options=gpu_options,allow_soft_placement=True))
# Initialize variables
sess.run(tf.global_variables_initializer(), feed_dict={phase_train_placeholder:True})
sess.run(tf.local_variables_initializer(), feed_dict={phase_train_placeholder:True})
run_metadata = tf.RunMetadata()
run_options = tf.RunOptions(trace_level=tf.RunOptions.FULL_TRACE)
with sess.as_default():
#pdb.set_trace()
if args.pretrained_model:
print('Restoring pretrained model: %s' % args.pretrained_model)
saver.restore(sess, os.path.expanduser(args.pretrained_model))
# Training and validation loop
epoch = 0
while epoch < args.max_nrof_epochs:
step = sess.run(global_step, feed_dict=None)
epoch = step // args.epoch_size
# Train for one epoch
train(args, sess, epoch, images_placeholder, labels_placeholder, data_reader,run_metadata,run_options,
learning_rate_placeholder, global_step,
losses, train_op, args.learning_rate_schedule_file)
# Save variables and the metagraph if it doesn't exist already
model_dir = args.models_base_dir
checkpoint_path = os.path.join(model_dir, 'model-%s.ckpt' % 'softmax')
saver.save(sess, checkpoint_path, global_step=step, write_meta_graph=False)
# Evaluate on LFW
return model_dir
def train(args, sess, epoch, images_placeholder, labels_placeholder, data_reader,run_metadata, run_options,
learning_rate_placeholder, global_step,
loss, train_op, learning_rate_schedule_file):
batch_number = 0
if args.learning_rate>0.0:
lr = args.learning_rate
else:
lr = utils.get_learning_rate_from_file(learning_rate_schedule_file, epoch)
while batch_number < args.epoch_size:
start_time = time.time()
#print('Running forward pass on sampled images: ', end='')
start_time = time.time()
images, labels = data_reader.next_batch(args.image_height,args.image_width)
print("Load {} images cost time: {}".format(images.shape[0],time.time()-start_time))
#if epoch < 10:
# continue
#if batch_number > 49:
# pdb.set_trace()
# print(labels)
#print(images.shape,labels)
feed_dict = {learning_rate_placeholder: lr ,images_placeholder:images, labels_placeholder:labels}
#feed_dict = {learning_rate_placeholder: lr}
start_time = time.time()
#total_err, softmax_err, _, step = sess.run([loss['total_loss'], loss['softmax_loss'], train_op, global_step ], feed_dict=feed_dict, options=run_options, run_metadata=run_metadata)
_ = sess.run(train_op, feed_dict=feed_dict, options=run_options, run_metadata=run_metadata)
tl = timeline.Timeline(run_metadata.step_stats)
ctf = tl.generate_chrome_trace_format()
with open('json_placeholder/tl-{}.json'.format(batch_number),'w') as wd:
wd.write(ctf)
duration = time.time() - start_time
#print('Epoch: [%d][%d/%d]\tTime %.3f\tTotal Loss %2.3f\tSoftmax Loss %2.3f, lr %2.5f' %
# (epoch, batch_number+1, args.epoch_size, duration, total_err, softmax_err, lr))
batch_number += 1
return batch_number
def save_variables_and_metagraph(sess, saver, summary_writer, model_dir, model_name, step):
# Save the model checkpoint
print('Saving variables')
start_time = time.time()
checkpoint_path = os.path.join(model_dir, 'model-%s.ckpt' % model_name)
save_time_variables = time.time() - start_time
print('Variables saved in %.2f seconds' % save_time_variables)
metagraph_filename = os.path.join(model_dir, 'model-%s.meta' % model_name)
save_time_metagraph = 0
if not os.path.exists(metagraph_filename):
print('Saving metagraph')
start_time = time.time()
saver.export_meta_graph(metagraph_filename)
save_time_metagraph = time.time() - start_time
print('Metagraph saved in %.2f seconds' % save_time_metagraph)
def | (filename, epoch):
with open(filename, 'r') as f:
for line in f.readlines():
line = line.split('#', 1)[0]
if line:
par = line.strip().split(':')
e = int(par[0])
lr = float(par[1])
if e <= epoch:
learning_rate = lr
else:
return learning_rate
def parse_arguments(argv):
parser = argparse.ArgumentParser()
parser.add_argument('--logs_base_dir', type=str,
help='Directory where to write event logs.', default='logs/facenet_ms_mp')
parser.add_argument('--models_base_dir', type=str,
help='Directory where to write trained models and checkpoints.', default='models/facenet_ms_mp')
parser.add_argument('--gpu_memory_fraction', type=float,
help='Upper bound on the amount of GPU memory that will be used by the process.', default=.9)
parser.add_argument('--pretrained_model', | get_learning_rate_from_file | identifier_name |
train_timeline.py |
def main(args):
#network = importlib.import_module(args.model_def)
subdir = datetime.strftime(datetime.now(), '%Y%m%d-%H%M%S')
log_dir = os.path.join(os.path.expanduser(args.logs_base_dir), subdir)
if not os.path.isdir(log_dir): # Create the log directory if it doesn't exist
os.makedirs(log_dir)
model_dir = os.path.join(os.path.expanduser(args.models_base_dir), subdir)
if not os.path.isdir(model_dir): # Create the model directory if it doesn't exist
os.makedirs(model_dir)
# Write arguments to a text file
utils.write_arguments_to_file(args, os.path.join(log_dir, 'arguments.txt'))
# Store some git revision info in a text file in the log directory
src_path,_ = os.path.split(os.path.realpath(__file__))
utils.store_revision_info(src_path, log_dir, ' '.join(sys.argv))
np.random.seed(seed=args.seed)
train_set = utils.get_dataset(args.data_dir)
nrof_classes = len(train_set)
print('nrof_classes: ',nrof_classes)
image_list, label_list = utils.get_image_paths_and_labels(train_set)
image_list = np.array(image_list)
print('total images: {}'.format(len(image_list)))
label_list = np.array(label_list,dtype=np.int32)
dataset_size = len(image_list)
data_reader = DataGenerator(image_list,label_list,args.batch_size)
print('Model directory: %s' % model_dir)
print('Log directory: %s' % log_dir)
if args.pretrained_model:
print('Pre-trained model: %s' % os.path.expanduser(args.pretrained_model))
with tf.Graph().as_default():
tf.set_random_seed(args.seed)
global_step = tf.Variable(0, trainable=False,name='global_step')
# Placeholder for the learning rate
learning_rate_placeholder = tf.placeholder(tf.float32, name='learning_rate')
images_placeholder = tf.placeholder(tf.float32,[None,112,96,3], name='images_placeholder')
#images_placeholder = tf.Variable(np.ones([args.batch_size,112,96,3]), dtype=np.float32,name='images_placeholder')
labels_placeholder = tf.placeholder(tf.int32,[None], name='labels_placeholder')
#labels_placeholder = tf.Variable(np.ones([args.batch_size,]),dtype=tf.int32, name='labels_placeholder')
phase_train_placeholder = tf.placeholder(tf.bool, name='phase_train')
learning_rate = tf.train.exponential_decay(learning_rate_placeholder, global_step,
args.learning_rate_decay_epochs*args.epoch_size, args.learning_rate_decay_factor, staircase=True)
tf.summary.scalar('learning_rate', learning_rate)
print('Using optimizer: {}'.format(args.optimizer))
if args.optimizer == 'ADAGRAD':
opt = tf.train.AdagradOptimizer(learning_rate)
elif args.optimizer == 'MOM':
opt = tf.train.MomentumOptimizer(learning_rate,0.9)
if args.network == 'sphere_network':
prelogits = network.infer(images_placeholder)
else:
raise Exception('Not supported network: {}'.format(args.loss_type))
if args.loss_type == 'softmax':
cross_entropy_mean = utils.softmax_loss(prelogits,labels_placeholder, len(train_set),args.weight_decay,False)
regularization_losses = tf.get_collection(tf.GraphKeys.REGULARIZATION_LOSSES)
#loss = cross_entropy_mean + args.weight_decay*tf.add_n(regularization_losses)
loss = cross_entropy_mean + args.weight_decay*tf.add_n(regularization_losses)
#loss = cross_entropy_mean
else:
raise Exception('Not supported loss type: {}'.format(args.loss_type))
#loss = tf.add_n([cross_entropy_mean] + regularization_losses, name='total_loss')
losses = {}
losses['total_loss'] = loss
losses['softmax_loss'] = cross_entropy_mean
debug_info = {}
debug_info ['prelogits'] = prelogits
grads = opt.compute_gradients(loss,tf.trainable_variables())
train_op = opt.apply_gradients(grads,global_step=global_step)
#save_vars = [var for var in tf.global_variables() if 'Adagrad' not in var.name and 'global_step' not in var.name]
save_vars = tf.global_variables()
#saver = tf.train.Saver(tf.trainable_variables(), max_to_keep=3)
saver = tf.train.Saver(save_vars, max_to_keep=3)
# Build the summary operation based on the TF collection of Summaries.
# Start running operations on the Graph.
gpu_options = tf.GPUOptions(per_process_gpu_memory_fraction=args.gpu_memory_fraction)
sess = tf.Session(config=tf.ConfigProto(gpu_options=gpu_options,allow_soft_placement=True))
# Initialize variables
sess.run(tf.global_variables_initializer(), feed_dict={phase_train_placeholder:True})
sess.run(tf.local_variables_initializer(), feed_dict={phase_train_placeholder:True})
run_metadata = tf.RunMetadata()
run_options = tf.RunOptions(trace_level=tf.RunOptions.FULL_TRACE)
with sess.as_default():
#pdb.set_trace()
if args.pretrained_model:
print('Restoring pretrained model: %s' % args.pretrained_model)
saver.restore(sess, os.path.expanduser(args.pretrained_model))
# Training and validation loop
epoch = 0
while epoch < args.max_nrof_epochs:
step = sess.run(global_step, feed_dict=None)
epoch = step // args.epoch_size
# Train for one epoch
train(args, sess, epoch, images_placeholder, labels_placeholder, data_reader,run_metadata,run_options,
learning_rate_placeholder, global_step,
losses, train_op, args.learning_rate_schedule_file)
# Save variables and the metagraph if it doesn't exist already
model_dir = args.models_base_dir
checkpoint_path = os.path.join(model_dir, 'model-%s.ckpt' % 'softmax')
saver.save(sess, checkpoint_path, global_step=step, write_meta_graph=False)
# Evaluate on LFW
return model_dir
def train(args, sess, epoch, images_placeholder, labels_placeholder, data_reader,run_metadata, run_options,
learning_rate_placeholder, global_step,
loss, train_op, learning_rate_schedule_file):
batch_number = 0
if args.learning_rate>0.0:
lr = args.learning_rate
else:
lr = utils.get_learning_rate_from_file(learning_rate_schedule_file, epoch)
while batch_number < args.epoch_size:
start_time = time.time()
#print('Running forward pass on sampled images: ', end='')
start_time = time.time()
images, labels = data_reader.next_batch(args.image_height,args.image_width)
print("Load {} images cost time: {}".format(images.shape[0],time.time()-start_time))
#if epoch < 10:
# continue
#if batch_number > 49:
# pdb.set_trace()
# print(labels)
#print(images.shape,labels)
feed_dict = {learning_rate_placeholder: lr ,images_placeholder:images, labels_placeholder:labels}
#feed_dict = {learning_rate_placeholder: lr}
start_time = time.time()
#total_err, softmax_err, _, step = sess.run([loss['total_loss'], loss['softmax_loss'], train_op, global_step ], feed_dict=feed_dict, options=run_options, run_metadata=run_metadata)
_ = sess.run(train_op, feed_dict=feed_dict, options=run_options, run_metadata=run_metadata)
tl = timeline.Timeline(run_metadata.step_stats)
ctf = tl.generate_chrome_trace_format()
with open('json_placeholder/tl-{}.json'.format(batch_number),'w') as wd:
wd.write(ctf)
duration = time.time() - start_time
#print('Epoch: [%d][%d/%d]\tTime %.3f\tTotal Loss %2.3f\tSoftmax Loss %2.3f, lr %2.5f' %
# (epoch, batch_number+1, args.epoch_size, duration, total_err, softmax_err, lr))
batch_number += 1
return batch_number
def save_variables_and_metagraph(sess, saver, summary_writer, model_dir, model_name, step):
# Save the model checkpoint
print('Saving variables')
start_time = time.time()
checkpoint_path = os.path.join(model_dir, 'model-%s.ckpt' % model_name)
save_time_variables = time.time() - start_time
print('Variables saved in %.2f seconds' % save_time_variables)
metagraph_filename = os.path.join(model_dir, 'model-%s.meta' % model_name)
save_time_metagraph = 0
if not os.path.exists(metagraph_filename):
print('Saving metagraph')
start_time = time.time()
saver.export_meta_graph(metagraph_filename)
save_time_metagraph = time.time() - start_time
print('Metagraph saved in %.2f seconds' % save_time_metagraph)
def get_learning_rate_from_file(filename, epoch):
with open(filename, 'r') as f:
for line in f.readlines():
line = line.split('#', 1)[0]
| return tf_data.Dataset.from_tensor_slices((tensors_x,tensors_y)) | identifier_body |
|
train_timeline.py | some git revision info in a text file in the log directory
src_path,_ = os.path.split(os.path.realpath(__file__))
utils.store_revision_info(src_path, log_dir, ' '.join(sys.argv))
np.random.seed(seed=args.seed)
train_set = utils.get_dataset(args.data_dir)
nrof_classes = len(train_set)
print('nrof_classes: ',nrof_classes)
image_list, label_list = utils.get_image_paths_and_labels(train_set)
image_list = np.array(image_list)
print('total images: {}'.format(len(image_list)))
label_list = np.array(label_list,dtype=np.int32)
dataset_size = len(image_list)
data_reader = DataGenerator(image_list,label_list,args.batch_size)
print('Model directory: %s' % model_dir)
print('Log directory: %s' % log_dir)
if args.pretrained_model:
print('Pre-trained model: %s' % os.path.expanduser(args.pretrained_model))
with tf.Graph().as_default():
tf.set_random_seed(args.seed)
global_step = tf.Variable(0, trainable=False,name='global_step')
# Placeholder for the learning rate
learning_rate_placeholder = tf.placeholder(tf.float32, name='learning_rate')
images_placeholder = tf.placeholder(tf.float32,[None,112,96,3], name='images_placeholder')
#images_placeholder = tf.Variable(np.ones([args.batch_size,112,96,3]), dtype=np.float32,name='images_placeholder')
labels_placeholder = tf.placeholder(tf.int32,[None], name='labels_placeholder') |
learning_rate = tf.train.exponential_decay(learning_rate_placeholder, global_step,
args.learning_rate_decay_epochs*args.epoch_size, args.learning_rate_decay_factor, staircase=True)
tf.summary.scalar('learning_rate', learning_rate)
print('Using optimizer: {}'.format(args.optimizer))
if args.optimizer == 'ADAGRAD':
opt = tf.train.AdagradOptimizer(learning_rate)
elif args.optimizer == 'MOM':
opt = tf.train.MomentumOptimizer(learning_rate,0.9)
if args.network == 'sphere_network':
prelogits = network.infer(images_placeholder)
else:
raise Exception('Not supported network: {}'.format(args.loss_type))
if args.loss_type == 'softmax':
cross_entropy_mean = utils.softmax_loss(prelogits,labels_placeholder, len(train_set),args.weight_decay,False)
regularization_losses = tf.get_collection(tf.GraphKeys.REGULARIZATION_LOSSES)
#loss = cross_entropy_mean + args.weight_decay*tf.add_n(regularization_losses)
loss = cross_entropy_mean + args.weight_decay*tf.add_n(regularization_losses)
#loss = cross_entropy_mean
else:
raise Exception('Not supported loss type: {}'.format(args.loss_type))
#loss = tf.add_n([cross_entropy_mean] + regularization_losses, name='total_loss')
losses = {}
losses['total_loss'] = loss
losses['softmax_loss'] = cross_entropy_mean
debug_info = {}
debug_info ['prelogits'] = prelogits
grads = opt.compute_gradients(loss,tf.trainable_variables())
train_op = opt.apply_gradients(grads,global_step=global_step)
#save_vars = [var for var in tf.global_variables() if 'Adagrad' not in var.name and 'global_step' not in var.name]
save_vars = tf.global_variables()
#saver = tf.train.Saver(tf.trainable_variables(), max_to_keep=3)
saver = tf.train.Saver(save_vars, max_to_keep=3)
# Build the summary operation based on the TF collection of Summaries.
# Start running operations on the Graph.
gpu_options = tf.GPUOptions(per_process_gpu_memory_fraction=args.gpu_memory_fraction)
sess = tf.Session(config=tf.ConfigProto(gpu_options=gpu_options,allow_soft_placement=True))
# Initialize variables
sess.run(tf.global_variables_initializer(), feed_dict={phase_train_placeholder:True})
sess.run(tf.local_variables_initializer(), feed_dict={phase_train_placeholder:True})
run_metadata = tf.RunMetadata()
run_options = tf.RunOptions(trace_level=tf.RunOptions.FULL_TRACE)
with sess.as_default():
#pdb.set_trace()
if args.pretrained_model:
print('Restoring pretrained model: %s' % args.pretrained_model)
saver.restore(sess, os.path.expanduser(args.pretrained_model))
# Training and validation loop
epoch = 0
while epoch < args.max_nrof_epochs:
step = sess.run(global_step, feed_dict=None)
epoch = step // args.epoch_size
# Train for one epoch
train(args, sess, epoch, images_placeholder, labels_placeholder, data_reader,run_metadata,run_options,
learning_rate_placeholder, global_step,
losses, train_op, args.learning_rate_schedule_file)
# Save variables and the metagraph if it doesn't exist already
model_dir = args.models_base_dir
checkpoint_path = os.path.join(model_dir, 'model-%s.ckpt' % 'softmax')
saver.save(sess, checkpoint_path, global_step=step, write_meta_graph=False)
# Evaluate on LFW
return model_dir
def train(args, sess, epoch, images_placeholder, labels_placeholder, data_reader,run_metadata, run_options,
learning_rate_placeholder, global_step,
loss, train_op, learning_rate_schedule_file):
batch_number = 0
if args.learning_rate>0.0:
lr = args.learning_rate
else:
lr = utils.get_learning_rate_from_file(learning_rate_schedule_file, epoch)
while batch_number < args.epoch_size:
start_time = time.time()
#print('Running forward pass on sampled images: ', end='')
start_time = time.time()
images, labels = data_reader.next_batch(args.image_height,args.image_width)
print("Load {} images cost time: {}".format(images.shape[0],time.time()-start_time))
#if epoch < 10:
# continue
#if batch_number > 49:
# pdb.set_trace()
# print(labels)
#print(images.shape,labels)
feed_dict = {learning_rate_placeholder: lr ,images_placeholder:images, labels_placeholder:labels}
#feed_dict = {learning_rate_placeholder: lr}
start_time = time.time()
#total_err, softmax_err, _, step = sess.run([loss['total_loss'], loss['softmax_loss'], train_op, global_step ], feed_dict=feed_dict, options=run_options, run_metadata=run_metadata)
_ = sess.run(train_op, feed_dict=feed_dict, options=run_options, run_metadata=run_metadata)
tl = timeline.Timeline(run_metadata.step_stats)
ctf = tl.generate_chrome_trace_format()
with open('json_placeholder/tl-{}.json'.format(batch_number),'w') as wd:
wd.write(ctf)
duration = time.time() - start_time
#print('Epoch: [%d][%d/%d]\tTime %.3f\tTotal Loss %2.3f\tSoftmax Loss %2.3f, lr %2.5f' %
# (epoch, batch_number+1, args.epoch_size, duration, total_err, softmax_err, lr))
batch_number += 1
return batch_number
def save_variables_and_metagraph(sess, saver, summary_writer, model_dir, model_name, step):
# Save the model checkpoint
print('Saving variables')
start_time = time.time()
checkpoint_path = os.path.join(model_dir, 'model-%s.ckpt' % model_name)
save_time_variables = time.time() - start_time
print('Variables saved in %.2f seconds' % save_time_variables)
metagraph_filename = os.path.join(model_dir, 'model-%s.meta' % model_name)
save_time_metagraph = 0
if not os.path.exists(metagraph_filename):
print('Saving metagraph')
start_time = time.time()
saver.export_meta_graph(metagraph_filename)
save_time_metagraph = time.time() - start_time
print('Metagraph saved in %.2f seconds' % save_time_metagraph)
def get_learning_rate_from_file(filename, epoch):
with open(filename, 'r') as f:
for line in f.readlines():
line = line.split('#', 1)[0]
if line:
par = line.strip().split(':')
e = int(par[0])
lr = float(par[1])
if e <= epoch:
learning_rate = lr
else:
return learning_rate
def parse_arguments(argv):
parser = argparse.ArgumentParser()
parser.add_argument('--logs_base_dir', type=str,
help='Directory where to write event logs.', default='logs/facenet_ms_mp')
parser.add_argument('--models_base_dir', type=str,
help='Directory where to write trained models and checkpoints.', default='models/facenet_ms_mp')
parser.add_argument('--gpu_memory_fraction', type=float,
help='Upper bound on the amount of GPU memory that will be used by the process.', default=.9)
parser.add_argument('--pretrained_model', type=str | #labels_placeholder = tf.Variable(np.ones([args.batch_size,]),dtype=tf.int32, name='labels_placeholder')
phase_train_placeholder = tf.placeholder(tf.bool, name='phase_train')
| random_line_split |
train_timeline.py | Store some git revision info in a text file in the log directory
src_path,_ = os.path.split(os.path.realpath(__file__))
utils.store_revision_info(src_path, log_dir, ' '.join(sys.argv))
np.random.seed(seed=args.seed)
train_set = utils.get_dataset(args.data_dir)
nrof_classes = len(train_set)
print('nrof_classes: ',nrof_classes)
image_list, label_list = utils.get_image_paths_and_labels(train_set)
image_list = np.array(image_list)
print('total images: {}'.format(len(image_list)))
label_list = np.array(label_list,dtype=np.int32)
dataset_size = len(image_list)
data_reader = DataGenerator(image_list,label_list,args.batch_size)
print('Model directory: %s' % model_dir)
print('Log directory: %s' % log_dir)
if args.pretrained_model:
print('Pre-trained model: %s' % os.path.expanduser(args.pretrained_model))
with tf.Graph().as_default():
tf.set_random_seed(args.seed)
global_step = tf.Variable(0, trainable=False,name='global_step')
# Placeholder for the learning rate
learning_rate_placeholder = tf.placeholder(tf.float32, name='learning_rate')
images_placeholder = tf.placeholder(tf.float32,[None,112,96,3], name='images_placeholder')
#images_placeholder = tf.Variable(np.ones([args.batch_size,112,96,3]), dtype=np.float32,name='images_placeholder')
labels_placeholder = tf.placeholder(tf.int32,[None], name='labels_placeholder')
#labels_placeholder = tf.Variable(np.ones([args.batch_size,]),dtype=tf.int32, name='labels_placeholder')
phase_train_placeholder = tf.placeholder(tf.bool, name='phase_train')
learning_rate = tf.train.exponential_decay(learning_rate_placeholder, global_step,
args.learning_rate_decay_epochs*args.epoch_size, args.learning_rate_decay_factor, staircase=True)
tf.summary.scalar('learning_rate', learning_rate)
print('Using optimizer: {}'.format(args.optimizer))
if args.optimizer == 'ADAGRAD':
opt = tf.train.AdagradOptimizer(learning_rate)
elif args.optimizer == 'MOM':
opt = tf.train.MomentumOptimizer(learning_rate,0.9)
if args.network == 'sphere_network':
prelogits = network.infer(images_placeholder)
else:
raise Exception('Not supported network: {}'.format(args.loss_type))
if args.loss_type == 'softmax':
cross_entropy_mean = utils.softmax_loss(prelogits,labels_placeholder, len(train_set),args.weight_decay,False)
regularization_losses = tf.get_collection(tf.GraphKeys.REGULARIZATION_LOSSES)
#loss = cross_entropy_mean + args.weight_decay*tf.add_n(regularization_losses)
loss = cross_entropy_mean + args.weight_decay*tf.add_n(regularization_losses)
#loss = cross_entropy_mean
else:
raise Exception('Not supported loss type: {}'.format(args.loss_type))
#loss = tf.add_n([cross_entropy_mean] + regularization_losses, name='total_loss')
losses = {}
losses['total_loss'] = loss
losses['softmax_loss'] = cross_entropy_mean
debug_info = {}
debug_info ['prelogits'] = prelogits
grads = opt.compute_gradients(loss,tf.trainable_variables())
train_op = opt.apply_gradients(grads,global_step=global_step)
#save_vars = [var for var in tf.global_variables() if 'Adagrad' not in var.name and 'global_step' not in var.name]
save_vars = tf.global_variables()
#saver = tf.train.Saver(tf.trainable_variables(), max_to_keep=3)
saver = tf.train.Saver(save_vars, max_to_keep=3)
# Build the summary operation based on the TF collection of Summaries.
# Start running operations on the Graph.
gpu_options = tf.GPUOptions(per_process_gpu_memory_fraction=args.gpu_memory_fraction)
sess = tf.Session(config=tf.ConfigProto(gpu_options=gpu_options,allow_soft_placement=True))
# Initialize variables
sess.run(tf.global_variables_initializer(), feed_dict={phase_train_placeholder:True})
sess.run(tf.local_variables_initializer(), feed_dict={phase_train_placeholder:True})
run_metadata = tf.RunMetadata()
run_options = tf.RunOptions(trace_level=tf.RunOptions.FULL_TRACE)
with sess.as_default():
#pdb.set_trace()
if args.pretrained_model:
print('Restoring pretrained model: %s' % args.pretrained_model)
saver.restore(sess, os.path.expanduser(args.pretrained_model))
# Training and validation loop
epoch = 0
while epoch < args.max_nrof_epochs:
step = sess.run(global_step, feed_dict=None)
epoch = step // args.epoch_size
# Train for one epoch
train(args, sess, epoch, images_placeholder, labels_placeholder, data_reader,run_metadata,run_options,
learning_rate_placeholder, global_step,
losses, train_op, args.learning_rate_schedule_file)
# Save variables and the metagraph if it doesn't exist already
model_dir = args.models_base_dir
checkpoint_path = os.path.join(model_dir, 'model-%s.ckpt' % 'softmax')
saver.save(sess, checkpoint_path, global_step=step, write_meta_graph=False)
# Evaluate on LFW
return model_dir
def train(args, sess, epoch, images_placeholder, labels_placeholder, data_reader,run_metadata, run_options,
learning_rate_placeholder, global_step,
loss, train_op, learning_rate_schedule_file):
batch_number = 0
if args.learning_rate>0.0:
lr = args.learning_rate
else:
lr = utils.get_learning_rate_from_file(learning_rate_schedule_file, epoch)
while batch_number < args.epoch_size:
start_time = time.time()
#print('Running forward pass on sampled images: ', end='')
start_time = time.time()
images, labels = data_reader.next_batch(args.image_height,args.image_width)
print("Load {} images cost time: {}".format(images.shape[0],time.time()-start_time))
#if epoch < 10:
# continue
#if batch_number > 49:
# pdb.set_trace()
# print(labels)
#print(images.shape,labels)
feed_dict = {learning_rate_placeholder: lr ,images_placeholder:images, labels_placeholder:labels}
#feed_dict = {learning_rate_placeholder: lr}
start_time = time.time()
#total_err, softmax_err, _, step = sess.run([loss['total_loss'], loss['softmax_loss'], train_op, global_step ], feed_dict=feed_dict, options=run_options, run_metadata=run_metadata)
_ = sess.run(train_op, feed_dict=feed_dict, options=run_options, run_metadata=run_metadata)
tl = timeline.Timeline(run_metadata.step_stats)
ctf = tl.generate_chrome_trace_format()
with open('json_placeholder/tl-{}.json'.format(batch_number),'w') as wd:
wd.write(ctf)
duration = time.time() - start_time
#print('Epoch: [%d][%d/%d]\tTime %.3f\tTotal Loss %2.3f\tSoftmax Loss %2.3f, lr %2.5f' %
# (epoch, batch_number+1, args.epoch_size, duration, total_err, softmax_err, lr))
batch_number += 1
return batch_number
def save_variables_and_metagraph(sess, saver, summary_writer, model_dir, model_name, step):
# Save the model checkpoint
print('Saving variables')
start_time = time.time()
checkpoint_path = os.path.join(model_dir, 'model-%s.ckpt' % model_name)
save_time_variables = time.time() - start_time
print('Variables saved in %.2f seconds' % save_time_variables)
metagraph_filename = os.path.join(model_dir, 'model-%s.meta' % model_name)
save_time_metagraph = 0
if not os.path.exists(metagraph_filename):
print('Saving metagraph')
start_time = time.time()
saver.export_meta_graph(metagraph_filename)
save_time_metagraph = time.time() - start_time
print('Metagraph saved in %.2f seconds' % save_time_metagraph)
def get_learning_rate_from_file(filename, epoch):
with open(filename, 'r') as f:
for line in f.readlines():
line = line.split('#', 1)[0]
if line:
par = line.strip().split(':')
e = int(par[0])
lr = float(par[1])
if e <= epoch:
learning_rate = lr
else:
|
def parse_arguments(argv):
parser = argparse.ArgumentParser()
parser.add_argument('--logs_base_dir', type=str,
help='Directory where to write event logs.', default='logs/facenet_ms_mp')
parser.add_argument('--models_base_dir', type=str,
help='Directory where to write trained models and checkpoints.', default='models/facenet_ms_mp')
parser.add_argument('--gpu_memory_fraction', type=float,
help='Upper bound on the amount of GPU memory that will be used by the process.', default=.9)
parser.add_argument('--pretrained_model', type | return learning_rate | conditional_block |
index.js | " id
let dateToday = fDate(now);
let today = document.querySelector("#todaysDate");
today.innerHTML = dateToday.toUpperCase();
}
setDateTime(); //Set the current date time for theTime & todaysDate elements
//update the current city name to match what was searched/submitted
function searchYourCity(event) {
event.preventDefault();
let input = document.querySelector("#citySearch");
let cityName = document.querySelector("#searchedCity");
if(input.value.length > 0)
{
cityName.innerHTML = input.value.toUpperCase();
checkWeatherByCity(input.value,checkUnitsSelected());
}
else
{
alert("Please enter a city");
}
}
function checkUnitsSelected(){
let units = "metric";
let unitC = document.querySelector("#tempC");
//If °C is selected
if (unitC.classList.contains("selectedTempUnit")==true)
{
units = "metric";
}
//If °F is selected
else
{
units = "imperial";
}
return units;
}
//Temperature conversion
function convertFtoC(tempF){
return Math.floor(((tempF -32)*5/9)*100)/100;
}
function convertCtoF(tempC){
return Math.floor(((tempC*9/5) + 32)*100)/100;
}
//Unit conversions
function convertUnits(){
let windSpeed = document.querySelector("#wind_Speed");
let feels_like = document.querySelector("#feels_Like");
let rainAmount = document.querySelector("#rain_Amount");
if (checkUnitsSelected()==="metric")
{
windSpeed.innerHTML = Math.floor(windSpeed.innerHTML.substring(0,windSpeed.innerHTML.length-3)*0.621371*100)/100;
feels_like.innerHTML = convertFtoC(feels_like.innerHTML.substring(0,feels_like.innerHTML.length-2));
rainAmount.innerHTML = Math.floor(rainAmount.innerHTML.substring(0,rainAmount.innerHTML.length-1)*25.4*100)/100;
}
//else °F is selected
else
{
windSpeed.innerHTML = Math.floor(windSpeed.innerHTML.substring(0,windSpeed.innerHTML.length-3)*1.609343502101025*100)/100;
feels_like.innerHTML = convertCtoF(feels_like.innerHTML.substring(0,feels_like.innerHTML.length-2));
rainAmount.innerHTML = Math.floor(rainAmount.innerHTML.substring(0,rainAmount.innerHTML.length-2)/25.4*100)/100;
}
addUnits();
}
//Update temperature display
function displayTemp(currentTemp){
let current_temp = document.querySelector("#currentCityTemp");
//If °C is selected
if (checkUnitsSelected()==="metric")
{
current_temp.innerHTML = currentTemp+"°C";
}
//else °F is selected
else
{
current_temp.innerHTML = currentTemp+"°F";
}
}
function changeUnit(){
let unitC = document.querySelector("#tempC");
let unitF = document.querySelector("#tempF");
let temp = document.querySelector("#currentCityTemp");
//If °F is selected
if (unitC.classList.contains("selectedTempUnit")==false)
{
unitF.classList.remove("selectedTempUnit");
unitC.classList.add("selectedTempUnit");
temp.innerHTML = convertFtoC(temp.innerHTML.substring(0,temp.innerHTML.length-2))+"°C";
convertUnits();
}
//else °C is selected
else
{
unitC.classList.remove("selectedTempUnit");
unitF.classList.add("selectedTempUnit");
temp.innerHTML = convertCtoF(temp.innerHTML.substring(0,temp.innerHTML.length-2))+"°F";
convertUnits();
}
getForecast();
}
//Capture user temperature unit selection
let tempC = document.querySelector("#tempC");
tempC.addEventListener("click", changeUnit);
let tempF = document.querySelector("#tempF");
tempF.addEventListener("click", changeUnit);
//create URL
function checkWeatherByCity(city,units){
//api.openweathermap.org/data/2.5/weather?q={city name}&appid={API key}
let weatherCheckUrl = apiCallWeather +"q="+ city + "&appid=" + apiKeyWeather + "&units=" + units;
//Get weather information from URL and then display weather
axios.get(weatherCheckUrl).then(displayWeather);
}
//Display Temperature of URL
function displayWeather(response) {
lastSearchedCityWeather = response;
displayTemp(lastSearchedCityWeather.data.main.temp);
// Update city name
let cityName = document.querySelector("#searchedCity");
cityName.innerHTML = lastSearchedCityWeather.data.name.toUpperCase();
// Update wind speed
document.querySelector("#wind_Speed").innerHTML = lastSearchedCityWeather.data.wind.speed;
// Update Rain
if (lastSearchedCityWeather.data.rain === undefined) {
document.querySelector("#rain_Amount").innerHTML = 0;
} else {
document.querySelector("#rain_Amount").innerHTML =
lastSearchedCityWeather.data.rain["1h"];
}
// Update weather description
document.querySelector("#weather_Description").innerHTML = lastSearchedCityWeather.data.weather[0].description.toUpperCase();
// Update weather icon
let currentWeatherIcon = document.querySelector("#today_Icon");
currentWeatherIcon.setAttribute("src",`https://openweathermap.org/img/wn/${lastSearchedCityWeather.data.weather[0].icon}@2x.png`);
currentWeatherIcon.setAttribute("alt",lastSearchedCityWeather.data.weather[0].description);
// Update feels like
document.querySelector("#feels_Like").innerHTML =lastSearchedCityWeather.data.main.feels_like;
// Update humidity
document.querySelector("#humidity").innerHTML =lastSearchedCityWeather.data.main.humidity + "%";
addUnits();
//Forecast
getForecast();
}
function getForecast() {
let coordinates = lastSearchedCityWeather.data.coord;
let apiUrl = `https://api.openweathermap.org/data/2.5/onecall?lat=${coordinates.lat}&lon=${coordinates.lon}&appid=${apiKeyWeather}&units=${checkUnitsSelected()}`;
axios.get(apiUrl).then(displayForecast);
}
function formatDay(timestamp) {
let date = new Date(timestamp * 1000);
let day = date.getDay();
let days = ["Mon", "Tues", "Wed", "Thurs", "Fri", "Sat", "Sun"];
return days [day];
}
function displayForecast(response) {
let forecast = response.data.daily;
let forecastElement = document.querySelector("#forecast");
let forecastHTML = `<div class="row">`;
forecast.forEach(function(forecastDay, index) {
if (index <6) {
forecastHTML =
forecastHTML +
`
<div class="col-2">
<div class="weather-forecast-dayName">${formatDay(forecastDay.dt)}</div>
<img
src="http://openweathermap.org/img/wn/${forecastDay.weather[0].icon}@2x.png"
alt=""
width="75"
class="forecastIcon"
/>
<div class="weather-forecast-tempertures">
<span class="weather-forecast-temperature-max"> ${Math.round(forecastDay.temp.max)}°
</span>
<span class="weather-forecast-temperature-min"> ${Math.round(forecastDay.temp.min)}°
</span>
</div>
</div>
`;
}
})
forecastHTML = forecastHTML + `</div>`;
forecastElement.innerHTML = `<div class="forecast-title">
<h5>In the next few days, you can expect...</h5>
</div>`
+ forecastHTML;
}
function addUnits(){
if (checkUnitsSelected()==="metric")
{
document.querySelector("#feels_Like").innerHTML = document.querySelector("#feels_Like").innerHTML+"°C";
document.querySelector("#wind_Speed").innerHTML = document.querySelector("#wind_Speed").innerHTML + "kph";
document.querySelector("#rain_Amount").innerHTML = document.querySelector("#rain_Amount").innerHTML + "mm";
}
//If °F is selected
else
{
document.querySelector("#feels_Like").innerHTML = document.querySelector("#feels_Like").innerHTML+"°F";
document.querySelector("#wind_Speed").innerHTML = document.querySelector("#wind_Speed").innerHTML + "mph";
document.querySelector("#rain_Amount").innerHTML = document.querySelector("#rain_Amount").innerHTML + '"';
}
}
//geo location
function showPosition(position) {
let lat = position.coords.latitude;
let lon = position.coords.longitude;
//Get weather information from URL and then display temp
checkWeatherByLocation(lat,lon,checkUnitsSelected());
}
function checkWeatherByLocation(lat,lon,units){
let currentL | at = `lat=${lat}`;
let currentLon = `lon=${lon}`;
//api.openweathermap.org/data/2.5/weather?lat={lat}&lon={lon}&appid={API key}
let weatherCheckUrl = apiCallWeather + currentLat + "&" + currentLon + "&appid=" + apiKeyWeather + "&units=" + units;
//Get weather information from URL and then display weather
axios.get(weatherCheckUrl).then(displayWeather);
}
//Get the posi | identifier_body |
|
index.js | = fDate(now);
let today = document.querySelector("#todaysDate");
today.innerHTML = dateToday.toUpperCase();
}
setDateTime(); //Set the current date time for theTime & todaysDate elements
//update the current city name to match what was searched/submitted
function searchYourCity(event) {
event.preventDefault();
let input = document.querySelector("#citySearch");
let cityName = document.querySelector("#searchedCity");
if(input.value.length > 0)
{
cityName.innerHTML = input.value.toUpperCase();
checkWeatherByCity(input.value,checkUnitsSelected());
}
else
{
alert("Please enter a city");
}
}
function checkUnitsSelected(){
let units = "metric";
let unitC = document.querySelector("#tempC");
//If °C is selected
if (unitC.classList.contains("selectedTempUnit")==true)
{
units = "metric";
}
//If °F is selected
else
{
units = "imperial";
}
return units;
}
//Temperature conversion
function convertFtoC(tempF){
return Math.floor(((tempF -32)*5/9)*100)/100;
}
function convertCtoF(tempC){
return Math.floor(((tempC*9/5) + 32)*100)/100;
}
//Unit conversions
function convertUnits(){
let windSpeed = document.querySelector("#wind_Speed");
let feels_like = document.querySelector("#feels_Like");
let rainAmount = document.querySelector("#rain_Amount");
if (checkUnitsSelected()==="metric")
{
windSpeed.innerHTML = Math.floor(windSpeed.innerHTML.substring(0,windSpeed.innerHTML.length-3)*0.621371*100)/100;
feels_like.innerHTML = convertFtoC(feels_like.innerHTML.substring(0,feels_like.innerHTML.length-2));
rainAmount.innerHTML = Math.floor(rainAmount.innerHTML.substring(0,rainAmount.innerHTML.length-1)*25.4*100)/100;
}
//else °F is selected
else
{
windSpeed.innerHTML = Math.floor(windSpeed.innerHTML.substring(0,windSpeed.innerHTML.length-3)*1.609343502101025*100)/100;
feels_like.innerHTML = convertCtoF(feels_like.innerHTML.substring(0,feels_like.innerHTML.length-2));
rainAmount.innerHTML = Math.floor(rainAmount.innerHTML.substring(0,rainAmount.innerHTML.length-2)/25.4*100)/100;
}
addUnits();
}
//Update temperature display
function displayTemp(currentTemp){
let current_temp = document.querySelector("#currentCityTemp");
//If °C is selected
if (checkUnitsSelected()==="metric")
{
current_temp.innerHTML = currentTemp+"°C";
}
//else °F is selected
else
{
current_temp.innerHTML = currentTemp+"°F";
}
}
function changeUnit(){
let unitC = document.querySelector("#tempC");
let unitF = document.querySelector("#tempF");
let temp = document.querySelector("#currentCityTemp");
//If °F is selected
if (unitC.classList.contains("selectedTempUnit")==false)
{
unitF.classList.remove("selectedTempUnit");
unitC.classList.add("selectedTempUnit");
temp.innerHTML = convertFtoC(temp.innerHTML.substring(0,temp.innerHTML.length-2))+"°C";
convertUnits();
}
//else °C is selected
else
{
unitC.classList.remove("selectedTempUnit");
unitF.classList.add("selectedTempUnit");
temp.innerHTML = convertCtoF(temp.innerHTML.substring(0,temp.innerHTML.length-2))+"°F";
convertUnits();
}
getForecast();
}
//Capture user temperature unit selection
let tempC = document.querySelector("#tempC");
tempC.addEventListener("click", changeUnit);
let tempF = document.querySelector("#tempF");
tempF.addEventListener("click", changeUnit);
//create URL
function checkWeatherByCity(city,units){
//api.openweathermap.org/data/2.5/weather?q={city name}&appid={API key}
let weatherCheckUrl = apiCallWeather +"q="+ city + "&appid=" + apiKeyWeather + "&units=" + units;
//Get weather information from URL and then display weather
axios.get(weatherCheckUrl).then(displayWeather);
}
//Display Temperature of URL
function displayWeather(response) {
lastSearchedCityWeather = response;
displayTemp(lastSearchedCityWeather.data.main.temp);
// Update city name
let cityName = document.querySelector("#searchedCity");
cityName.innerHTML = lastSearchedCityWeather.data.name.toUpperCase();
// Update wind speed
document.querySelector("#wind_Speed").innerHTML = lastSearchedCityWeather.data.wind.speed;
// Update Rain
if (lastSearchedCityWeather.data.rain === undefined) {
document.querySelector("#rain_Amount").innerHTML = 0;
} else {
document.querySelector("#rain_Amount").innerHTML =
lastSearchedCityWeather.data.rain["1h"];
}
// Update weather description
document.querySelector("#weather_Description").innerHTML = lastSearchedCityWeather.data.weather[0].description.toUpperCase();
// Update weather icon
let currentWeatherIcon = document.querySelector("#today_Icon");
currentWeatherIcon.setAttribute("src",`https://openweathermap.org/img/wn/${lastSearchedCityWeather.data.weather[0].icon}@2x.png`);
currentWeatherIcon.setAttribute("alt",lastSearchedCityWeather.data.weather[0].description);
// Update feels like
document.querySelector("#feels_Like").innerHTML =lastSearchedCityWeather.data.main.feels_like;
// Update humidity
document.querySelector("#humidity").innerHTML =lastSearchedCityWeather.data.main.humidity + "%";
addUnits();
//Forecast
getForecast();
}
function getForecast() {
let coordinates = lastSearchedCityWeather.data.coord;
let apiUrl = `https://api.openweathermap.org/data/2.5/onecall?lat=${coordinates.lat}&lon=${coordinates.lon}&appid=${apiKeyWeather}&units=${checkUnitsSelected()}`;
axios.get(apiUrl).then(displayForecast);
}
function formatDay(timestamp) {
let date = new Date(timestamp * 1000);
let day = date.getDay();
let days = ["Mon", "Tues", "Wed", "Thurs", "Fri", "Sat", "Sun"];
return days [day];
}
function displayForecast(response) {
let forecast = response.data.daily;
let forecastElement = document.querySelector("#forecast");
let forecastHTML = `<div class="row">`;
forecast.forEach(function(forecastDay, index) {
if (index <6) {
forecastHTML =
forecastHTML +
`
<div class="col-2">
<div class="weather-forecast-dayName">${formatDay(forecastDay.dt)}</div>
<img
src="http://openweathermap.org/img/wn/${forecastDay.weather[0].icon}@2x.png"
alt=""
width="75"
class="forecastIcon"
/>
<div class="weather-forecast-tempertures">
<span class="weather-forecast-temperature-max"> ${Math.round(forecastDay.temp.max)}°
</span>
<span class="weather-forecast-temperature-min"> ${Math.round(forecastDay.temp.min)}°
</span>
</div>
</div>
`;
}
})
forecastHTML = forecastHTML + `</div>`;
forecastElement.innerHTML = `<div class="forecast-title">
<h5>In the next few days, you can expect...</h5>
</div>`
+ forecastHTML;
}
function addUnits(){
if (checkUnitsSelected()==="metric")
{
document.querySelector("#feels_Like").innerHTML = document.querySelector("#feels_Like").innerHTML+"°C";
document.querySelector("#wind_Speed").innerHTML = document.querySelector("#wind_Speed").innerHTML + "kph";
document.querySelector("#rain_Amount").innerHTML = document.querySelector("#rain_Amount").innerHTML + "mm";
}
//If °F is selected
else
{
document.querySelector("#feels_Like").innerHTML = document.querySelector("#feels_Like").innerHTML+"°F";
document.querySelector("#wind_Speed").innerHTML = document.querySelector("#wind_Speed").innerHTML + "mph";
document.querySelector("#rain_Amount").innerHTML = document.querySelector("#rain_Amount").innerHTML + '"';
}
}
//geo location
function showPosition(position) {
let lat = position.coords.latitude;
let lon = position.coords.longitude;
//Get weather information from URL and then display temp
checkWeatherByLocation(lat,lon,checkUnitsSelected());
}
function checkWeatherByLocation(lat,lon,units){
let currentLat = `lat=${lat}`;
let currentLon = `lon=${lon}`;
//api.openweathermap.org/data/2.5/weather?lat={lat}&lon={lon}&appid={API key}
let weatherCheckUrl = apiCallWeather + currentLat + "&" + currentLon + "&appid=" + apiKeyWeather + "&units=" + units;
//Get weather information from URL and then display weather
axios.get(weatherCheckUrl).then(displayWeather);
}
//Get the position
function findLocation(){
| navigator. | identifier_name |
|
index.js | Months = [
"January",
"February",
"March",
"April",
"May",
"June",
"July",
"August",
"September",
"October",
"November",
"December"
];
let fMonth = fMonths[currentDate.getMonth()];
return (
fDay + ", " +
fMonth +
" " +
currentDate.getDate() +
" " +
currentDate.getFullYear()
);
}
function setDateTime(){
//Get the current date & time
let now = new Date();
//get and set the time using "theTime" id
let time = document.querySelector("#theTime");
let mins = now.getMinutes();
//Add a Zero to keep mins double digits
if (mins < 10) {
mins = "0" + mins;
}
let theCurrentTime = `${now.getHours()}:${mins}`;
time.innerHTML = theCurrentTime;
//get and set the date using "todaysDate" id
let dateToday = fDate(now);
let today = document.querySelector("#todaysDate");
today.innerHTML = dateToday.toUpperCase();
}
setDateTime(); //Set the current date time for theTime & todaysDate elements
//update the current city name to match what was searched/submitted
function searchYourCity(event) {
event.preventDefault();
let input = document.querySelector("#citySearch");
let cityName = document.querySelector("#searchedCity");
if(input.value.length > 0)
{
cityName.innerHTML = input.value.toUpperCase();
checkWeatherByCity(input.value,checkUnitsSelected());
}
else
{
alert("Please enter a city");
}
}
function checkUnitsSelected(){
let units = "metric";
let unitC = document.querySelector("#tempC");
//If °C is selected
if (unitC.classList.contains("selectedTempUnit")==true)
{ | //If °F is selected
else
{
units = "imperial";
}
return units;
}
//Temperature conversion
function convertFtoC(tempF){
return Math.floor(((tempF -32)*5/9)*100)/100;
}
function convertCtoF(tempC){
return Math.floor(((tempC*9/5) + 32)*100)/100;
}
//Unit conversions
function convertUnits(){
let windSpeed = document.querySelector("#wind_Speed");
let feels_like = document.querySelector("#feels_Like");
let rainAmount = document.querySelector("#rain_Amount");
if (checkUnitsSelected()==="metric")
{
windSpeed.innerHTML = Math.floor(windSpeed.innerHTML.substring(0,windSpeed.innerHTML.length-3)*0.621371*100)/100;
feels_like.innerHTML = convertFtoC(feels_like.innerHTML.substring(0,feels_like.innerHTML.length-2));
rainAmount.innerHTML = Math.floor(rainAmount.innerHTML.substring(0,rainAmount.innerHTML.length-1)*25.4*100)/100;
}
//else °F is selected
else
{
windSpeed.innerHTML = Math.floor(windSpeed.innerHTML.substring(0,windSpeed.innerHTML.length-3)*1.609343502101025*100)/100;
feels_like.innerHTML = convertCtoF(feels_like.innerHTML.substring(0,feels_like.innerHTML.length-2));
rainAmount.innerHTML = Math.floor(rainAmount.innerHTML.substring(0,rainAmount.innerHTML.length-2)/25.4*100)/100;
}
addUnits();
}
//Update temperature display
function displayTemp(currentTemp){
let current_temp = document.querySelector("#currentCityTemp");
//If °C is selected
if (checkUnitsSelected()==="metric")
{
current_temp.innerHTML = currentTemp+"°C";
}
//else °F is selected
else
{
current_temp.innerHTML = currentTemp+"°F";
}
}
function changeUnit(){
let unitC = document.querySelector("#tempC");
let unitF = document.querySelector("#tempF");
let temp = document.querySelector("#currentCityTemp");
//If °F is selected
if (unitC.classList.contains("selectedTempUnit")==false)
{
unitF.classList.remove("selectedTempUnit");
unitC.classList.add("selectedTempUnit");
temp.innerHTML = convertFtoC(temp.innerHTML.substring(0,temp.innerHTML.length-2))+"°C";
convertUnits();
}
//else °C is selected
else
{
unitC.classList.remove("selectedTempUnit");
unitF.classList.add("selectedTempUnit");
temp.innerHTML = convertCtoF(temp.innerHTML.substring(0,temp.innerHTML.length-2))+"°F";
convertUnits();
}
getForecast();
}
//Capture user temperature unit selection
let tempC = document.querySelector("#tempC");
tempC.addEventListener("click", changeUnit);
let tempF = document.querySelector("#tempF");
tempF.addEventListener("click", changeUnit);
//create URL
function checkWeatherByCity(city,units){
//api.openweathermap.org/data/2.5/weather?q={city name}&appid={API key}
let weatherCheckUrl = apiCallWeather +"q="+ city + "&appid=" + apiKeyWeather + "&units=" + units;
//Get weather information from URL and then display weather
axios.get(weatherCheckUrl).then(displayWeather);
}
//Display Temperature of URL
function displayWeather(response) {
lastSearchedCityWeather = response;
displayTemp(lastSearchedCityWeather.data.main.temp);
// Update city name
let cityName = document.querySelector("#searchedCity");
cityName.innerHTML = lastSearchedCityWeather.data.name.toUpperCase();
// Update wind speed
document.querySelector("#wind_Speed").innerHTML = lastSearchedCityWeather.data.wind.speed;
// Update Rain
if (lastSearchedCityWeather.data.rain === undefined) {
document.querySelector("#rain_Amount").innerHTML = 0;
} else {
document.querySelector("#rain_Amount").innerHTML =
lastSearchedCityWeather.data.rain["1h"];
}
// Update weather description
document.querySelector("#weather_Description").innerHTML = lastSearchedCityWeather.data.weather[0].description.toUpperCase();
// Update weather icon
let currentWeatherIcon = document.querySelector("#today_Icon");
currentWeatherIcon.setAttribute("src",`https://openweathermap.org/img/wn/${lastSearchedCityWeather.data.weather[0].icon}@2x.png`);
currentWeatherIcon.setAttribute("alt",lastSearchedCityWeather.data.weather[0].description);
// Update feels like
document.querySelector("#feels_Like").innerHTML =lastSearchedCityWeather.data.main.feels_like;
// Update humidity
document.querySelector("#humidity").innerHTML =lastSearchedCityWeather.data.main.humidity + "%";
addUnits();
//Forecast
getForecast();
}
function getForecast() {
let coordinates = lastSearchedCityWeather.data.coord;
let apiUrl = `https://api.openweathermap.org/data/2.5/onecall?lat=${coordinates.lat}&lon=${coordinates.lon}&appid=${apiKeyWeather}&units=${checkUnitsSelected()}`;
axios.get(apiUrl).then(displayForecast);
}
function formatDay(timestamp) {
let date = new Date(timestamp * 1000);
let day = date.getDay();
let days = ["Mon", "Tues", "Wed", "Thurs", "Fri", "Sat", "Sun"];
return days [day];
}
function displayForecast(response) {
let forecast = response.data.daily;
let forecastElement = document.querySelector("#forecast");
let forecastHTML = `<div class="row">`;
forecast.forEach(function(forecastDay, index) {
if (index <6) {
forecastHTML =
forecastHTML +
`
<div class="col-2">
<div class="weather-forecast-dayName">${formatDay(forecastDay.dt)}</div>
<img
src="http://openweathermap.org/img/wn/${forecastDay.weather[0].icon}@2x.png"
alt=""
width="75"
class="forecastIcon"
/>
<div class="weather-forecast-tempertures">
<span class="weather-forecast-temperature-max"> ${Math.round(forecastDay.temp.max)}°
</span>
<span class="weather-forecast-temperature-min"> ${Math.round(forecastDay.temp.min)}°
</span>
</div>
</div>
`;
}
})
forecastHTML = forecastHTML + `</div>`;
forecastElement.innerHTML = `<div class="forecast-title">
<h5>In the next few days, you can expect...</h5>
</div>`
+ forecastHTML;
}
function addUnits(){
if (checkUnitsSelected()==="metric")
{
document.querySelector("#feels_Like").innerHTML = document.querySelector("#feels_Like").innerHTML+"°C";
document.querySelector("#wind_Speed").innerHTML = document.querySelector("#wind_Speed").innerHTML + "kph";
document.querySelector("#rain_Amount").innerHTML = document.querySelector("#rain_Amount").innerHTML + "mm";
}
//If °F is selected
else
{
document.querySelector("#feels_Like").innerHTML = document.querySelector("#feels_Like").innerHTML+"°F";
|
units = "metric";
}
| conditional_block |
index.js | Months = [
"January",
"February",
"March",
"April",
"May",
"June",
"July",
"August",
"September",
"October",
"November",
"December"
];
let fMonth = fMonths[currentDate.getMonth()];
return (
fDay + ", " +
fMonth +
" " +
currentDate.getDate() +
" " +
currentDate.getFullYear()
);
}
function setDateTime(){
//Get the current date & time
let now = new Date();
//get and set the time using "theTime" id
let time = document.querySelector("#theTime");
let mins = now.getMinutes();
//Add a Zero to keep mins double digits
if (mins < 10) {
mins = "0" + mins;
}
let theCurrentTime = `${now.getHours()}:${mins}`;
time.innerHTML = theCurrentTime;
//get and set the date using "todaysDate" id
let dateToday = fDate(now);
let today = document.querySelector("#todaysDate");
today.innerHTML = dateToday.toUpperCase();
}
setDateTime(); //Set the current date time for theTime & todaysDate elements
//update the current city name to match what was searched/submitted
function searchYourCity(event) {
event.preventDefault();
let input = document.querySelector("#citySearch");
let cityName = document.querySelector("#searchedCity");
if(input.value.length > 0)
{
cityName.innerHTML = input.value.toUpperCase();
checkWeatherByCity(input.value,checkUnitsSelected());
}
else
{
alert("Please enter a city");
}
}
function checkUnitsSelected(){
let units = "metric";
let unitC = document.querySelector("#tempC");
//If °C is selected
if (unitC.classList.contains("selectedTempUnit")==true)
{
units = "metric";
}
//If °F is selected
else
{
units = "imperial"; | return units;
}
//Temperature conversion
function convertFtoC(tempF){
return Math.floor(((tempF -32)*5/9)*100)/100;
}
function convertCtoF(tempC){
return Math.floor(((tempC*9/5) + 32)*100)/100;
}
//Unit conversions
function convertUnits(){
let windSpeed = document.querySelector("#wind_Speed");
let feels_like = document.querySelector("#feels_Like");
let rainAmount = document.querySelector("#rain_Amount");
if (checkUnitsSelected()==="metric")
{
windSpeed.innerHTML = Math.floor(windSpeed.innerHTML.substring(0,windSpeed.innerHTML.length-3)*0.621371*100)/100;
feels_like.innerHTML = convertFtoC(feels_like.innerHTML.substring(0,feels_like.innerHTML.length-2));
rainAmount.innerHTML = Math.floor(rainAmount.innerHTML.substring(0,rainAmount.innerHTML.length-1)*25.4*100)/100;
}
//else °F is selected
else
{
windSpeed.innerHTML = Math.floor(windSpeed.innerHTML.substring(0,windSpeed.innerHTML.length-3)*1.609343502101025*100)/100;
feels_like.innerHTML = convertCtoF(feels_like.innerHTML.substring(0,feels_like.innerHTML.length-2));
rainAmount.innerHTML = Math.floor(rainAmount.innerHTML.substring(0,rainAmount.innerHTML.length-2)/25.4*100)/100;
}
addUnits();
}
//Update temperature display
function displayTemp(currentTemp){
let current_temp = document.querySelector("#currentCityTemp");
//If °C is selected
if (checkUnitsSelected()==="metric")
{
current_temp.innerHTML = currentTemp+"°C";
}
//else °F is selected
else
{
current_temp.innerHTML = currentTemp+"°F";
}
}
function changeUnit(){
let unitC = document.querySelector("#tempC");
let unitF = document.querySelector("#tempF");
let temp = document.querySelector("#currentCityTemp");
//If °F is selected
if (unitC.classList.contains("selectedTempUnit")==false)
{
unitF.classList.remove("selectedTempUnit");
unitC.classList.add("selectedTempUnit");
temp.innerHTML = convertFtoC(temp.innerHTML.substring(0,temp.innerHTML.length-2))+"°C";
convertUnits();
}
//else °C is selected
else
{
unitC.classList.remove("selectedTempUnit");
unitF.classList.add("selectedTempUnit");
temp.innerHTML = convertCtoF(temp.innerHTML.substring(0,temp.innerHTML.length-2))+"°F";
convertUnits();
}
getForecast();
}
//Capture user temperature unit selection
let tempC = document.querySelector("#tempC");
tempC.addEventListener("click", changeUnit);
let tempF = document.querySelector("#tempF");
tempF.addEventListener("click", changeUnit);
//create URL
function checkWeatherByCity(city,units){
//api.openweathermap.org/data/2.5/weather?q={city name}&appid={API key}
let weatherCheckUrl = apiCallWeather +"q="+ city + "&appid=" + apiKeyWeather + "&units=" + units;
//Get weather information from URL and then display weather
axios.get(weatherCheckUrl).then(displayWeather);
}
//Display Temperature of URL
function displayWeather(response) {
lastSearchedCityWeather = response;
displayTemp(lastSearchedCityWeather.data.main.temp);
// Update city name
let cityName = document.querySelector("#searchedCity");
cityName.innerHTML = lastSearchedCityWeather.data.name.toUpperCase();
// Update wind speed
document.querySelector("#wind_Speed").innerHTML = lastSearchedCityWeather.data.wind.speed;
// Update Rain
if (lastSearchedCityWeather.data.rain === undefined) {
document.querySelector("#rain_Amount").innerHTML = 0;
} else {
document.querySelector("#rain_Amount").innerHTML =
lastSearchedCityWeather.data.rain["1h"];
}
// Update weather description
document.querySelector("#weather_Description").innerHTML = lastSearchedCityWeather.data.weather[0].description.toUpperCase();
// Update weather icon
let currentWeatherIcon = document.querySelector("#today_Icon");
currentWeatherIcon.setAttribute("src",`https://openweathermap.org/img/wn/${lastSearchedCityWeather.data.weather[0].icon}@2x.png`);
currentWeatherIcon.setAttribute("alt",lastSearchedCityWeather.data.weather[0].description);
// Update feels like
document.querySelector("#feels_Like").innerHTML =lastSearchedCityWeather.data.main.feels_like;
// Update humidity
document.querySelector("#humidity").innerHTML =lastSearchedCityWeather.data.main.humidity + "%";
addUnits();
//Forecast
getForecast();
}
function getForecast() {
let coordinates = lastSearchedCityWeather.data.coord;
let apiUrl = `https://api.openweathermap.org/data/2.5/onecall?lat=${coordinates.lat}&lon=${coordinates.lon}&appid=${apiKeyWeather}&units=${checkUnitsSelected()}`;
axios.get(apiUrl).then(displayForecast);
}
function formatDay(timestamp) {
let date = new Date(timestamp * 1000);
let day = date.getDay();
let days = ["Mon", "Tues", "Wed", "Thurs", "Fri", "Sat", "Sun"];
return days [day];
}
function displayForecast(response) {
let forecast = response.data.daily;
let forecastElement = document.querySelector("#forecast");
let forecastHTML = `<div class="row">`;
forecast.forEach(function(forecastDay, index) {
if (index <6) {
forecastHTML =
forecastHTML +
`
<div class="col-2">
<div class="weather-forecast-dayName">${formatDay(forecastDay.dt)}</div>
<img
src="http://openweathermap.org/img/wn/${forecastDay.weather[0].icon}@2x.png"
alt=""
width="75"
class="forecastIcon"
/>
<div class="weather-forecast-tempertures">
<span class="weather-forecast-temperature-max"> ${Math.round(forecastDay.temp.max)}°
</span>
<span class="weather-forecast-temperature-min"> ${Math.round(forecastDay.temp.min)}°
</span>
</div>
</div>
`;
}
})
forecastHTML = forecastHTML + `</div>`;
forecastElement.innerHTML = `<div class="forecast-title">
<h5>In the next few days, you can expect...</h5>
</div>`
+ forecastHTML;
}
function addUnits(){
if (checkUnitsSelected()==="metric")
{
document.querySelector("#feels_Like").innerHTML = document.querySelector("#feels_Like").innerHTML+"°C";
document.querySelector("#wind_Speed").innerHTML = document.querySelector("#wind_Speed").innerHTML + "kph";
document.querySelector("#rain_Amount").innerHTML = document.querySelector("#rain_Amount").innerHTML + "mm";
}
//If °F is selected
else
{
document.querySelector("#feels_Like").innerHTML = document.querySelector("#feels_Like").innerHTML+"°F";
| } | random_line_split |
web.go | patterns for the inputted fields
VALID_NAME string = "^\\p{L}{2,}(?:\\x20\\p{L}{2,}){1,5}$"
VALID_SENTENCE string = "^[\\p{L}\\d\\x20-\\x2F\\x3A-\\x40\\x5B-\\x60\\x7B-\\x7E\\x{00B4}]*$"
VALID_MESSAGE string = "^[\\p{L}\\d\\x20-\\x2F\\x3A-\\x40\\x5B-\\x60\\x7B-\\x7E\\x{00B4}\\s]+$"
// This email regex has to get better, missing _ and a proper . check before the @
VALID_EMAIL string = "^[a-z\\x2E\\x5F]+\\x2B?[a-z]*[^\\x2B]\\x40(?:[a-z]+[a-z\\x2D\\x2E]?)+[^\\x2D]\\x2E[a-z]{2,5}$"
// Validation errors
EMAIL_ERROR string = "%s must be a valid email address"
NAME_ERROR string = "%s has to have at least one middle or last name"
SENTENCE_ERROR string = "%s must only contain language letters and ascii symbols"
LENGTH_ERROR string = "%s length mustn't be shorter than %d characters or longer than %d characters"
// Validations were successful
SUCCESS_MSG string = "Alright! The email has been sent."
// Amount of bytes to read from the body of a request
PAYLOAD_MAX_SIZE int64 = 8192
)
var (
config *Config
homepage *[]byte
cwd *string
mailmu sync.Mutex
)
// Concatenate a given hostname and port
func GetAddress(h *string, p *int) string {
return *h + ":" + strconv.FormatUint(uint64(*p), 10)
}
// Append an error message to slice if value fails the checks
func validateField(f *Field, hash map[string]string) {
length := 0
if f.Value != "" {
length = len(f.Value)
}
switch {
case length < f.Length.Min || length > f.Length.Max:
hash[f.Name] = fmt.Sprintf(LENGTH_ERROR, f.DisplayName, f.Length.Min, f.Length.Max)
case f.Regex != nil:
matches, _ := regexp.MatchString(f.Regex.Pattern, f.Value)
if !matches {
hash[f.Name] = fmt.Sprintf(f.Regex.Error, f.DisplayName)
}
}
}
// Encode content to JSON and write it to the response
func JSONResponse(w http.ResponseWriter, content interface{}, statusCode int) {
resp, err := json.Marshal(content)
if err != nil |
w.Header().Set("Content-Type", "application/json")
w.Header().Set("Access-Control-Allow-Origin", "*")
w.Header().Set("Access-Control-Allow-Methods", "GET, POST")
w.WriteHeader(statusCode)
w.Write(resp)
}
func (c *Config) Populate() {
content, err := ioutil.ReadFile(path("/config/app.json"))
if err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Fatal("Couldn't read file /config/app.json")
}
if err := json.Unmarshal(content, &c); err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Fatal("Failed to decode config.json")
}
}
func path(p string) string {
return *cwd + p
}
func (m *Mail) Validate() map[string]string {
errors := make(map[string]string)
validateField(&Field{
Name: "subject",
DisplayName: "Subject",
Value: m.Subject,
Length: &FieldLength{
Min: 12,
Max: 128,
},
Regex: &FieldRegex{
Pattern: VALID_SENTENCE,
Error: SENTENCE_ERROR,
},
}, errors)
validateField(&Field{
Name: "fullname",
DisplayName: "Full name",
Value: m.FullName,
Length: &FieldLength{
Min: 5,
Max: 48,
},
Regex: &FieldRegex{
Pattern: VALID_NAME,
Error: NAME_ERROR,
},
}, errors)
validateField(&Field{
Name: "email",
DisplayName: "Email",
Value: m.Email,
Length: &FieldLength{
Min: 12,
Max: 128,
},
Regex: &FieldRegex{
Pattern: VALID_EMAIL,
Error: EMAIL_ERROR,
},
}, errors)
validateField(&Field{
Name: "business",
DisplayName: "Business",
Value: m.Business,
Length: &FieldLength{
Min: 3,
Max: 32,
},
Regex: &FieldRegex{
Pattern: VALID_SENTENCE,
Error: SENTENCE_ERROR,
},
}, errors)
validateField(&Field{
Name: "body",
DisplayName: "Message",
Value: m.Body,
Length: &FieldLength{
Min: 64,
Max: 4096,
},
Regex: &FieldRegex{
Pattern: VALID_MESSAGE,
Error: SENTENCE_ERROR,
},
}, errors)
validateField(&Field{
Name: "details",
DisplayName: "Additional details",
Value: m.AdditionalDetails,
Length: &FieldLength{
Min: 4,
Max: 1024,
},
Regex: &FieldRegex{
Pattern: VALID_MESSAGE,
Error: SENTENCE_ERROR,
},
}, errors)
return errors
}
func (m *Mail) Send() {
err := smtp.SendMail(
GetAddress(&config.Smtp.Hostname, &config.Smtp.Port),
smtp.PlainAuth("", config.Smtp.Email, config.Smtp.Password, config.Smtp.Hostname),
config.Smtp.Email, []string{config.Smtp.To}, []byte(m.Body))
if err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error(fmt.Sprintf("Failed to send email from <%s> to <%s>", config.Smtp.Email, config.Smtp.To))
}
log.WithFields(log.Fields{
"action": "Send email",
}).Info(fmt.Sprintf("Successfully sent email written by %s <%s>", m.FullName, m.Email))
}
func (m *Mail) Save() {
mailmu.Lock()
defer mailmu.Unlock()
info, err := os.Stat(path("/storage.json"))
if err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error("Failed to fetch info of storage.json")
}
file, err := os.OpenFile(path("/storage.json"), os.O_RDWR, 0600)
defer file.Close()
if err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error("Failed to open storage.json")
}
mail, _ := json.Marshal(m)
buffer := make([]byte, 2)
var toWrite []byte
if _, err = file.ReadAt(buffer, info.Size()-2); err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error("Failed to read file storage.json")
}
switch {
case reflect.DeepEqual(buffer, []byte{'[', ']'}):
toWrite = append(toWrite, mail...)
case reflect.DeepEqual(buffer, []byte{'}', ']'}):
toWrite = append(toWrite, append([]byte{','}, mail...)...)
}
if _, err = file.WriteAt(append(toWrite, ']'), info.Size()-1); err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error("Failed to save email in storage.json")
}
log.WithFields(log.Fields{
"action": "Save email",
}).Info(fmt.Sprintf("Successfully saved email written by %s <%s>", m.FullName, m.Email))
}
func (m *Mail) SaveAndSend() {
m.Save()
m.Send()
}
func handleWriteEmail(w http.ResponseWriter, r *http.Request) {
body, err := ioutil.ReadAll(io.LimitReader(r.Body, PAYLOAD_MAX_SIZE))
if err != nil {
http.Error(w, "Failed to read body from HTTP request", http.StatusInternalServerError)
return
}
mail := new(Mail)
err = json.Unmarshal(body, &mail)
if err != nil {
http.Error(w, "Failed to parse payload", http.StatusBadRequest)
return
}
validationErrors := mail.Validate()
if len(validationErrors) != 0 {
JSONResponse(w, map[string](map[string]string){"errors": validationErrors}, http.StatusBadRequest)
return
}
go mail.SaveAndSend()
w.WriteHeader(200)
}
func handleHomepage(w http.ResponseWriter, _ *http.Request) {
w.Write(*homepage)
}
func createIfNot | {
http.Error(w, "Failed to encode content to JSON", http.StatusInternalServerError)
return
} | conditional_block |
web.go | patterns for the inputted fields
VALID_NAME string = "^\\p{L}{2,}(?:\\x20\\p{L}{2,}){1,5}$"
VALID_SENTENCE string = "^[\\p{L}\\d\\x20-\\x2F\\x3A-\\x40\\x5B-\\x60\\x7B-\\x7E\\x{00B4}]*$"
VALID_MESSAGE string = "^[\\p{L}\\d\\x20-\\x2F\\x3A-\\x40\\x5B-\\x60\\x7B-\\x7E\\x{00B4}\\s]+$"
// This email regex has to get better, missing _ and a proper . check before the @
VALID_EMAIL string = "^[a-z\\x2E\\x5F]+\\x2B?[a-z]*[^\\x2B]\\x40(?:[a-z]+[a-z\\x2D\\x2E]?)+[^\\x2D]\\x2E[a-z]{2,5}$"
// Validation errors
EMAIL_ERROR string = "%s must be a valid email address"
NAME_ERROR string = "%s has to have at least one middle or last name"
SENTENCE_ERROR string = "%s must only contain language letters and ascii symbols"
LENGTH_ERROR string = "%s length mustn't be shorter than %d characters or longer than %d characters"
// Validations were successful
SUCCESS_MSG string = "Alright! The email has been sent."
// Amount of bytes to read from the body of a request
PAYLOAD_MAX_SIZE int64 = 8192
)
var (
config *Config
homepage *[]byte
cwd *string
mailmu sync.Mutex
)
// Concatenate a given hostname and port
func GetAddress(h *string, p *int) string {
return *h + ":" + strconv.FormatUint(uint64(*p), 10)
} |
// Append an error message to slice if value fails the checks
func validateField(f *Field, hash map[string]string) {
length := 0
if f.Value != "" {
length = len(f.Value)
}
switch {
case length < f.Length.Min || length > f.Length.Max:
hash[f.Name] = fmt.Sprintf(LENGTH_ERROR, f.DisplayName, f.Length.Min, f.Length.Max)
case f.Regex != nil:
matches, _ := regexp.MatchString(f.Regex.Pattern, f.Value)
if !matches {
hash[f.Name] = fmt.Sprintf(f.Regex.Error, f.DisplayName)
}
}
}
// Encode content to JSON and write it to the response
func JSONResponse(w http.ResponseWriter, content interface{}, statusCode int) {
resp, err := json.Marshal(content)
if err != nil {
http.Error(w, "Failed to encode content to JSON", http.StatusInternalServerError)
return
}
w.Header().Set("Content-Type", "application/json")
w.Header().Set("Access-Control-Allow-Origin", "*")
w.Header().Set("Access-Control-Allow-Methods", "GET, POST")
w.WriteHeader(statusCode)
w.Write(resp)
}
func (c *Config) Populate() {
content, err := ioutil.ReadFile(path("/config/app.json"))
if err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Fatal("Couldn't read file /config/app.json")
}
if err := json.Unmarshal(content, &c); err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Fatal("Failed to decode config.json")
}
}
func path(p string) string {
return *cwd + p
}
func (m *Mail) Validate() map[string]string {
errors := make(map[string]string)
validateField(&Field{
Name: "subject",
DisplayName: "Subject",
Value: m.Subject,
Length: &FieldLength{
Min: 12,
Max: 128,
},
Regex: &FieldRegex{
Pattern: VALID_SENTENCE,
Error: SENTENCE_ERROR,
},
}, errors)
validateField(&Field{
Name: "fullname",
DisplayName: "Full name",
Value: m.FullName,
Length: &FieldLength{
Min: 5,
Max: 48,
},
Regex: &FieldRegex{
Pattern: VALID_NAME,
Error: NAME_ERROR,
},
}, errors)
validateField(&Field{
Name: "email",
DisplayName: "Email",
Value: m.Email,
Length: &FieldLength{
Min: 12,
Max: 128,
},
Regex: &FieldRegex{
Pattern: VALID_EMAIL,
Error: EMAIL_ERROR,
},
}, errors)
validateField(&Field{
Name: "business",
DisplayName: "Business",
Value: m.Business,
Length: &FieldLength{
Min: 3,
Max: 32,
},
Regex: &FieldRegex{
Pattern: VALID_SENTENCE,
Error: SENTENCE_ERROR,
},
}, errors)
validateField(&Field{
Name: "body",
DisplayName: "Message",
Value: m.Body,
Length: &FieldLength{
Min: 64,
Max: 4096,
},
Regex: &FieldRegex{
Pattern: VALID_MESSAGE,
Error: SENTENCE_ERROR,
},
}, errors)
validateField(&Field{
Name: "details",
DisplayName: "Additional details",
Value: m.AdditionalDetails,
Length: &FieldLength{
Min: 4,
Max: 1024,
},
Regex: &FieldRegex{
Pattern: VALID_MESSAGE,
Error: SENTENCE_ERROR,
},
}, errors)
return errors
}
func (m *Mail) Send() {
err := smtp.SendMail(
GetAddress(&config.Smtp.Hostname, &config.Smtp.Port),
smtp.PlainAuth("", config.Smtp.Email, config.Smtp.Password, config.Smtp.Hostname),
config.Smtp.Email, []string{config.Smtp.To}, []byte(m.Body))
if err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error(fmt.Sprintf("Failed to send email from <%s> to <%s>", config.Smtp.Email, config.Smtp.To))
}
log.WithFields(log.Fields{
"action": "Send email",
}).Info(fmt.Sprintf("Successfully sent email written by %s <%s>", m.FullName, m.Email))
}
func (m *Mail) Save() {
mailmu.Lock()
defer mailmu.Unlock()
info, err := os.Stat(path("/storage.json"))
if err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error("Failed to fetch info of storage.json")
}
file, err := os.OpenFile(path("/storage.json"), os.O_RDWR, 0600)
defer file.Close()
if err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error("Failed to open storage.json")
}
mail, _ := json.Marshal(m)
buffer := make([]byte, 2)
var toWrite []byte
if _, err = file.ReadAt(buffer, info.Size()-2); err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error("Failed to read file storage.json")
}
switch {
case reflect.DeepEqual(buffer, []byte{'[', ']'}):
toWrite = append(toWrite, mail...)
case reflect.DeepEqual(buffer, []byte{'}', ']'}):
toWrite = append(toWrite, append([]byte{','}, mail...)...)
}
if _, err = file.WriteAt(append(toWrite, ']'), info.Size()-1); err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error("Failed to save email in storage.json")
}
log.WithFields(log.Fields{
"action": "Save email",
}).Info(fmt.Sprintf("Successfully saved email written by %s <%s>", m.FullName, m.Email))
}
func (m *Mail) SaveAndSend() {
m.Save()
m.Send()
}
func handleWriteEmail(w http.ResponseWriter, r *http.Request) {
body, err := ioutil.ReadAll(io.LimitReader(r.Body, PAYLOAD_MAX_SIZE))
if err != nil {
http.Error(w, "Failed to read body from HTTP request", http.StatusInternalServerError)
return
}
mail := new(Mail)
err = json.Unmarshal(body, &mail)
if err != nil {
http.Error(w, "Failed to parse payload", http.StatusBadRequest)
return
}
validationErrors := mail.Validate()
if len(validationErrors) != 0 {
JSONResponse(w, map[string](map[string]string){"errors": validationErrors}, http.StatusBadRequest)
return
}
go mail.SaveAndSend()
w.WriteHeader(200)
}
func handleHomepage(w http.ResponseWriter, _ *http.Request) {
w.Write(*homepage)
}
func createIfNotExists | random_line_split |
|
web.go | patterns for the inputted fields
VALID_NAME string = "^\\p{L}{2,}(?:\\x20\\p{L}{2,}){1,5}$"
VALID_SENTENCE string = "^[\\p{L}\\d\\x20-\\x2F\\x3A-\\x40\\x5B-\\x60\\x7B-\\x7E\\x{00B4}]*$"
VALID_MESSAGE string = "^[\\p{L}\\d\\x20-\\x2F\\x3A-\\x40\\x5B-\\x60\\x7B-\\x7E\\x{00B4}\\s]+$"
// This email regex has to get better, missing _ and a proper . check before the @
VALID_EMAIL string = "^[a-z\\x2E\\x5F]+\\x2B?[a-z]*[^\\x2B]\\x40(?:[a-z]+[a-z\\x2D\\x2E]?)+[^\\x2D]\\x2E[a-z]{2,5}$"
// Validation errors
EMAIL_ERROR string = "%s must be a valid email address"
NAME_ERROR string = "%s has to have at least one middle or last name"
SENTENCE_ERROR string = "%s must only contain language letters and ascii symbols"
LENGTH_ERROR string = "%s length mustn't be shorter than %d characters or longer than %d characters"
// Validations were successful
SUCCESS_MSG string = "Alright! The email has been sent."
// Amount of bytes to read from the body of a request
PAYLOAD_MAX_SIZE int64 = 8192
)
var (
config *Config
homepage *[]byte
cwd *string
mailmu sync.Mutex
)
// Concatenate a given hostname and port
func GetAddress(h *string, p *int) string |
// Append an error message to slice if value fails the checks
func validateField(f *Field, hash map[string]string) {
length := 0
if f.Value != "" {
length = len(f.Value)
}
switch {
case length < f.Length.Min || length > f.Length.Max:
hash[f.Name] = fmt.Sprintf(LENGTH_ERROR, f.DisplayName, f.Length.Min, f.Length.Max)
case f.Regex != nil:
matches, _ := regexp.MatchString(f.Regex.Pattern, f.Value)
if !matches {
hash[f.Name] = fmt.Sprintf(f.Regex.Error, f.DisplayName)
}
}
}
// Encode content to JSON and write it to the response
func JSONResponse(w http.ResponseWriter, content interface{}, statusCode int) {
resp, err := json.Marshal(content)
if err != nil {
http.Error(w, "Failed to encode content to JSON", http.StatusInternalServerError)
return
}
w.Header().Set("Content-Type", "application/json")
w.Header().Set("Access-Control-Allow-Origin", "*")
w.Header().Set("Access-Control-Allow-Methods", "GET, POST")
w.WriteHeader(statusCode)
w.Write(resp)
}
func (c *Config) Populate() {
content, err := ioutil.ReadFile(path("/config/app.json"))
if err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Fatal("Couldn't read file /config/app.json")
}
if err := json.Unmarshal(content, &c); err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Fatal("Failed to decode config.json")
}
}
func path(p string) string {
return *cwd + p
}
func (m *Mail) Validate() map[string]string {
errors := make(map[string]string)
validateField(&Field{
Name: "subject",
DisplayName: "Subject",
Value: m.Subject,
Length: &FieldLength{
Min: 12,
Max: 128,
},
Regex: &FieldRegex{
Pattern: VALID_SENTENCE,
Error: SENTENCE_ERROR,
},
}, errors)
validateField(&Field{
Name: "fullname",
DisplayName: "Full name",
Value: m.FullName,
Length: &FieldLength{
Min: 5,
Max: 48,
},
Regex: &FieldRegex{
Pattern: VALID_NAME,
Error: NAME_ERROR,
},
}, errors)
validateField(&Field{
Name: "email",
DisplayName: "Email",
Value: m.Email,
Length: &FieldLength{
Min: 12,
Max: 128,
},
Regex: &FieldRegex{
Pattern: VALID_EMAIL,
Error: EMAIL_ERROR,
},
}, errors)
validateField(&Field{
Name: "business",
DisplayName: "Business",
Value: m.Business,
Length: &FieldLength{
Min: 3,
Max: 32,
},
Regex: &FieldRegex{
Pattern: VALID_SENTENCE,
Error: SENTENCE_ERROR,
},
}, errors)
validateField(&Field{
Name: "body",
DisplayName: "Message",
Value: m.Body,
Length: &FieldLength{
Min: 64,
Max: 4096,
},
Regex: &FieldRegex{
Pattern: VALID_MESSAGE,
Error: SENTENCE_ERROR,
},
}, errors)
validateField(&Field{
Name: "details",
DisplayName: "Additional details",
Value: m.AdditionalDetails,
Length: &FieldLength{
Min: 4,
Max: 1024,
},
Regex: &FieldRegex{
Pattern: VALID_MESSAGE,
Error: SENTENCE_ERROR,
},
}, errors)
return errors
}
func (m *Mail) Send() {
err := smtp.SendMail(
GetAddress(&config.Smtp.Hostname, &config.Smtp.Port),
smtp.PlainAuth("", config.Smtp.Email, config.Smtp.Password, config.Smtp.Hostname),
config.Smtp.Email, []string{config.Smtp.To}, []byte(m.Body))
if err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error(fmt.Sprintf("Failed to send email from <%s> to <%s>", config.Smtp.Email, config.Smtp.To))
}
log.WithFields(log.Fields{
"action": "Send email",
}).Info(fmt.Sprintf("Successfully sent email written by %s <%s>", m.FullName, m.Email))
}
func (m *Mail) Save() {
mailmu.Lock()
defer mailmu.Unlock()
info, err := os.Stat(path("/storage.json"))
if err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error("Failed to fetch info of storage.json")
}
file, err := os.OpenFile(path("/storage.json"), os.O_RDWR, 0600)
defer file.Close()
if err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error("Failed to open storage.json")
}
mail, _ := json.Marshal(m)
buffer := make([]byte, 2)
var toWrite []byte
if _, err = file.ReadAt(buffer, info.Size()-2); err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error("Failed to read file storage.json")
}
switch {
case reflect.DeepEqual(buffer, []byte{'[', ']'}):
toWrite = append(toWrite, mail...)
case reflect.DeepEqual(buffer, []byte{'}', ']'}):
toWrite = append(toWrite, append([]byte{','}, mail...)...)
}
if _, err = file.WriteAt(append(toWrite, ']'), info.Size()-1); err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error("Failed to save email in storage.json")
}
log.WithFields(log.Fields{
"action": "Save email",
}).Info(fmt.Sprintf("Successfully saved email written by %s <%s>", m.FullName, m.Email))
}
func (m *Mail) SaveAndSend() {
m.Save()
m.Send()
}
func handleWriteEmail(w http.ResponseWriter, r *http.Request) {
body, err := ioutil.ReadAll(io.LimitReader(r.Body, PAYLOAD_MAX_SIZE))
if err != nil {
http.Error(w, "Failed to read body from HTTP request", http.StatusInternalServerError)
return
}
mail := new(Mail)
err = json.Unmarshal(body, &mail)
if err != nil {
http.Error(w, "Failed to parse payload", http.StatusBadRequest)
return
}
validationErrors := mail.Validate()
if len(validationErrors) != 0 {
JSONResponse(w, map[string](map[string]string){"errors": validationErrors}, http.StatusBadRequest)
return
}
go mail.SaveAndSend()
w.WriteHeader(200)
}
func handleHomepage(w http.ResponseWriter, _ *http.Request) {
w.Write(*homepage)
}
func createIfNot | {
return *h + ":" + strconv.FormatUint(uint64(*p), 10)
} | identifier_body |
web.go | patterns for the inputted fields
VALID_NAME string = "^\\p{L}{2,}(?:\\x20\\p{L}{2,}){1,5}$"
VALID_SENTENCE string = "^[\\p{L}\\d\\x20-\\x2F\\x3A-\\x40\\x5B-\\x60\\x7B-\\x7E\\x{00B4}]*$"
VALID_MESSAGE string = "^[\\p{L}\\d\\x20-\\x2F\\x3A-\\x40\\x5B-\\x60\\x7B-\\x7E\\x{00B4}\\s]+$"
// This email regex has to get better, missing _ and a proper . check before the @
VALID_EMAIL string = "^[a-z\\x2E\\x5F]+\\x2B?[a-z]*[^\\x2B]\\x40(?:[a-z]+[a-z\\x2D\\x2E]?)+[^\\x2D]\\x2E[a-z]{2,5}$"
// Validation errors
EMAIL_ERROR string = "%s must be a valid email address"
NAME_ERROR string = "%s has to have at least one middle or last name"
SENTENCE_ERROR string = "%s must only contain language letters and ascii symbols"
LENGTH_ERROR string = "%s length mustn't be shorter than %d characters or longer than %d characters"
// Validations were successful
SUCCESS_MSG string = "Alright! The email has been sent."
// Amount of bytes to read from the body of a request
PAYLOAD_MAX_SIZE int64 = 8192
)
var (
config *Config
homepage *[]byte
cwd *string
mailmu sync.Mutex
)
// Concatenate a given hostname and port
func GetAddress(h *string, p *int) string {
return *h + ":" + strconv.FormatUint(uint64(*p), 10)
}
// Append an error message to slice if value fails the checks
func validateField(f *Field, hash map[string]string) {
length := 0
if f.Value != "" {
length = len(f.Value)
}
switch {
case length < f.Length.Min || length > f.Length.Max:
hash[f.Name] = fmt.Sprintf(LENGTH_ERROR, f.DisplayName, f.Length.Min, f.Length.Max)
case f.Regex != nil:
matches, _ := regexp.MatchString(f.Regex.Pattern, f.Value)
if !matches {
hash[f.Name] = fmt.Sprintf(f.Regex.Error, f.DisplayName)
}
}
}
// Encode content to JSON and write it to the response
func JSONResponse(w http.ResponseWriter, content interface{}, statusCode int) {
resp, err := json.Marshal(content)
if err != nil {
http.Error(w, "Failed to encode content to JSON", http.StatusInternalServerError)
return
}
w.Header().Set("Content-Type", "application/json")
w.Header().Set("Access-Control-Allow-Origin", "*")
w.Header().Set("Access-Control-Allow-Methods", "GET, POST")
w.WriteHeader(statusCode)
w.Write(resp)
}
func (c *Config) Populate() {
content, err := ioutil.ReadFile(path("/config/app.json"))
if err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Fatal("Couldn't read file /config/app.json")
}
if err := json.Unmarshal(content, &c); err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Fatal("Failed to decode config.json")
}
}
func path(p string) string {
return *cwd + p
}
func (m *Mail) Validate() map[string]string {
errors := make(map[string]string)
validateField(&Field{
Name: "subject",
DisplayName: "Subject",
Value: m.Subject,
Length: &FieldLength{
Min: 12,
Max: 128,
},
Regex: &FieldRegex{
Pattern: VALID_SENTENCE,
Error: SENTENCE_ERROR,
},
}, errors)
validateField(&Field{
Name: "fullname",
DisplayName: "Full name",
Value: m.FullName,
Length: &FieldLength{
Min: 5,
Max: 48,
},
Regex: &FieldRegex{
Pattern: VALID_NAME,
Error: NAME_ERROR,
},
}, errors)
validateField(&Field{
Name: "email",
DisplayName: "Email",
Value: m.Email,
Length: &FieldLength{
Min: 12,
Max: 128,
},
Regex: &FieldRegex{
Pattern: VALID_EMAIL,
Error: EMAIL_ERROR,
},
}, errors)
validateField(&Field{
Name: "business",
DisplayName: "Business",
Value: m.Business,
Length: &FieldLength{
Min: 3,
Max: 32,
},
Regex: &FieldRegex{
Pattern: VALID_SENTENCE,
Error: SENTENCE_ERROR,
},
}, errors)
validateField(&Field{
Name: "body",
DisplayName: "Message",
Value: m.Body,
Length: &FieldLength{
Min: 64,
Max: 4096,
},
Regex: &FieldRegex{
Pattern: VALID_MESSAGE,
Error: SENTENCE_ERROR,
},
}, errors)
validateField(&Field{
Name: "details",
DisplayName: "Additional details",
Value: m.AdditionalDetails,
Length: &FieldLength{
Min: 4,
Max: 1024,
},
Regex: &FieldRegex{
Pattern: VALID_MESSAGE,
Error: SENTENCE_ERROR,
},
}, errors)
return errors
}
func (m *Mail) Send() {
err := smtp.SendMail(
GetAddress(&config.Smtp.Hostname, &config.Smtp.Port),
smtp.PlainAuth("", config.Smtp.Email, config.Smtp.Password, config.Smtp.Hostname),
config.Smtp.Email, []string{config.Smtp.To}, []byte(m.Body))
if err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error(fmt.Sprintf("Failed to send email from <%s> to <%s>", config.Smtp.Email, config.Smtp.To))
}
log.WithFields(log.Fields{
"action": "Send email",
}).Info(fmt.Sprintf("Successfully sent email written by %s <%s>", m.FullName, m.Email))
}
func (m *Mail) Save() {
mailmu.Lock()
defer mailmu.Unlock()
info, err := os.Stat(path("/storage.json"))
if err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error("Failed to fetch info of storage.json")
}
file, err := os.OpenFile(path("/storage.json"), os.O_RDWR, 0600)
defer file.Close()
if err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error("Failed to open storage.json")
}
mail, _ := json.Marshal(m)
buffer := make([]byte, 2)
var toWrite []byte
if _, err = file.ReadAt(buffer, info.Size()-2); err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error("Failed to read file storage.json")
}
switch {
case reflect.DeepEqual(buffer, []byte{'[', ']'}):
toWrite = append(toWrite, mail...)
case reflect.DeepEqual(buffer, []byte{'}', ']'}):
toWrite = append(toWrite, append([]byte{','}, mail...)...)
}
if _, err = file.WriteAt(append(toWrite, ']'), info.Size()-1); err != nil {
log.WithFields(log.Fields{
"error": err.Error(),
}).Error("Failed to save email in storage.json")
}
log.WithFields(log.Fields{
"action": "Save email",
}).Info(fmt.Sprintf("Successfully saved email written by %s <%s>", m.FullName, m.Email))
}
func (m *Mail) SaveAndSend() {
m.Save()
m.Send()
}
func | (w http.ResponseWriter, r *http.Request) {
body, err := ioutil.ReadAll(io.LimitReader(r.Body, PAYLOAD_MAX_SIZE))
if err != nil {
http.Error(w, "Failed to read body from HTTP request", http.StatusInternalServerError)
return
}
mail := new(Mail)
err = json.Unmarshal(body, &mail)
if err != nil {
http.Error(w, "Failed to parse payload", http.StatusBadRequest)
return
}
validationErrors := mail.Validate()
if len(validationErrors) != 0 {
JSONResponse(w, map[string](map[string]string){"errors": validationErrors}, http.StatusBadRequest)
return
}
go mail.SaveAndSend()
w.WriteHeader(200)
}
func handleHomepage(w http.ResponseWriter, _ *http.Request) {
w.Write(*homepage)
}
func createIfNotExists | handleWriteEmail | identifier_name |
scraper.py | (query)['body']
# find username in submission content using brackets
start, stop = comment.index('['), comment.index(']')
self.username = comment[start + 1:stop]
# -------------------------------------------
# PLANS SCRAPEY-I
# get it? like "api" except... oh, never mind
# -------------------------------------------
class PlansConnection(object):
"""
Encapsulates an active login to plans.
"""
def __init__(self, cookiejar=None,
base_url="https://www.grinnellplans.com",
server_tz='US/Central'):
"""
Create a new plans connection.
Optional keyword arguments:
cookiejar -- an existing cookielib.CookieJar to store
credentials in.
base_url -- URL at which to access plans, no trailing slash.
server_tz -- Name of the timezone used by the server.
This class will convert dates to UTC.
"""
self.base_url = base_url
self.server_tz = server_tz
if cookiejar is None:
self.cookiejar = LWPCookieJar()
else:
self.cookiejar = cookiejar
self.session = requests.Session()
self.session.cookies = self.cookiejar
self.parser = PlansPageParser()
self.username = None
def _get_page(self, name, get=None, post=None):
"""
Retrieve an HTML page from plans.
"""
method = 'GET' if post is None else 'POST'
url = '/'.join((self.base_url, name))
req = requests.Request(method, url, params=get, data=post)
prepped = self.session.prepare_request(req)
try:
handle = self.session.send(prepped,
verify=url.startswith('https'))
except requests.exceptions.ConnectionError:
err = "Check your internet connection. Plans could also be down."
raise PlansError(err)
return handle
def _parse_message(self, soup):
"""
Scrape details from an infomessage or alertmessage div.
Returns a dictionary of the message parameters.
"""
kind, = soup.attrs[u'class']
title = soup.findChild().text
body = ''.join(t.text for t in soup.findChildren()[1:])
message = dict(kind=kind, title=title, body=body)
for val in message.values():
assert type(val) == str
return message
def _canonicalize_plantext(self, plan):
"""
Modify reserialized plan text to match what was served.
For consistency, we want to return plan text *exactly* how it
is formatted when served. However, variants of certain tags
(e.g. <br> vs <br/>) are syntactically equivalent in HTML, and
may be interchanged when parsed and reserialized.
This function manually changes the formatting returned by our
parser to more closely match that given by plans.
"""
# Our parser will correct <hr> and <br> to self closing tags
plan = plan.replace('<br/>', '<br>')
plan = plan.replace('<hr/>', '<hr>')
# put attributes in the right order because I have OCD
plan = re.sub(r'<a class="([^\s]*)" href="([^\s]*)">',
r'<a href="\2" class="\1">', plan)
# to avoid playing whack-a-mole, we should configure the
# parser to not do this, or else treat contents of
# <div class="plan_text"> tags as plain text
# (not sure if this is possible)
return plan
@staticmethod
def _html_esc(string):
"""
Replaces certain characters with html escape sequences.
Meant to be passed to the BS4 'decode' method as kwarg 'formatter'.
By default, BS4 only replaces angle brackets and ampersands with
the html escape sequence, but text served by plans also replaces
double quotes. This function as BS4's decoding formatter
reproduces that behavior.
"""
repls = {
'<': 'lt',
'>': 'gt',
'&': 'amp',
'"': 'quot',
}
def repl(matchobj):
return "&%s;" % repls[matchobj.group(0)]
regex = "([%s])" % ''.join(repls.keys())
return re.sub(regex, repl, string)
def plans_login(self, username='', password=''):
"""
Log into plans.
Returns True on success, False on failure. Leave username and
password blank to check an existing login.
"""
# the provided username and password ONLY get checked
# by the plans server if our cookie is expired.
# hence, if we've logged in recently, this will return True even
# if un/pw are not provided or are otherwise bad.
login_info = {'username': username,
'password': password,
'submit': 'Login'}
response = self._get_page('index.php', post=login_info)
# if login is successful, we'll be redirected to home
success = response.url[-9:] == '/home.php'
if success:
self.parser.feed(response.text) # parse out username
self.username = self.parser.username
return success
def get_edit_text(self):
"""
Retrieve contents of the edit plan field.
Returns the edit_text of the plan and its md5 hash,
as computed on the server side.
"""
# grab edit page
response = self._get_page('edit.php')
html = response.text
# parse out existing plan
soup = bs4.BeautifulSoup(html, 'html5lib')
plan = soup.find('textarea')
if plan is None:
raise PlansError("Couldn't get edit text, are we logged in?")
else:
plan = u'' + plan.contents[0]
# prepending the empty string somehow prevents BS from
# escaping all the HTML characters (weird)
assert type(plan) == str
# convert to CRLF line endings
plan = convert_endings(plan, 'CRLF')
# parse out plan md5
md5sum = soup.find('input',
attrs={'name': 'edit_text_md5'}).attrs['value']
# also, explicitly compute the hash, for kicks
assert md5sum == plans_md5(plan)
# verify that username has not changed
assert self.username == self.parser.username
return plan, md5sum
def set_edit_text(self, newtext, md5):
"""
Update plan with new content.
To prevent errors, the server does a hash check on the existing
plan before replacing it with the new one. We provide an
md5 sum to confirm that yes, we really want to update the plan.
Returns info message.
"""
# convert to CRLF line endings
newtext = convert_endings(newtext, 'CRLF')
newtext = newtext.encode('utf8')
edit_info = {'plan': newtext,
'edit_text_md5': md5,
'submit': 'Change Plan'}
response = self._get_page('edit.php', post=edit_info)
soup = bs4.BeautifulSoup(response.text, "html5lib")
alert = soup.find('div', {'class': 'alertmessage'})
info = soup.find('div', {'class': 'infomessage'})
if alert is not None:
# some kind of error
msg = self._parse_message(alert)
raise PlansError(msg['body'])
elif info is None:
raise PlansError('Plans did not verify update')
else:
# probably success
msg = self._parse_message(info)
return msg['body']
def | (self):
"""
Retrieve all levels of the autofinger (autoread) list.
Returns a dictionary where the keys are the group names
"Level 1", "Level 2", etc. and the values are a list of
usernames waiting to be read.
"""
# this actually doesn't scrape; there's a function for it
# in the old JSON API.
get = {'task': 'autofingerlist'}
response = self._get_page('api/1/index.php', get=get)
data = json.loads(response.text)
# the returned JSON is crufty; clean it up
autofinger = {}
for group in data['autofingerList']:
name = "Level %s" % group['level']
autofinger[name] = group['usernames']
return autofinger
def read_plan(self, plan):
"""
Retrieve the contents of the specified plan.
Returns two objects: the plan header (as a python dictionary)
the plan text (in HTML format)
"""
get = {'searchname': plan}
response = self._get_page('read.php', get=get)
soup = bs4.BeautifulSoup(response.text, 'html5lib')
header = soup.find('div', {'id': 'header'})
text = soup.find('div', {'class': 'plan_text'})
if text is None or header is None:
# probably a nonexistent user
alert = soup.find('div', {'class': 'alertmessage'})
msg = self._parse_message(alert)
raise PlansError(msg['title'])
# convert header html into a python dictionary
header_dict = {}
for key in ('username', 'planname'):
content = header.find | get_autofinger | identifier_name |
scraper.py | dict(query)['body']
# find username in submission content using brackets
start, stop = comment.index('['), comment.index(']')
self.username = comment[start + 1:stop]
# -------------------------------------------
# PLANS SCRAPEY-I
# get it? like "api" except... oh, never mind
# -------------------------------------------
class PlansConnection(object):
"""
Encapsulates an active login to plans.
"""
def __init__(self, cookiejar=None,
base_url="https://www.grinnellplans.com",
server_tz='US/Central'):
"""
Create a new plans connection.
Optional keyword arguments:
cookiejar -- an existing cookielib.CookieJar to store
credentials in.
base_url -- URL at which to access plans, no trailing slash.
server_tz -- Name of the timezone used by the server.
This class will convert dates to UTC.
"""
self.base_url = base_url
self.server_tz = server_tz
if cookiejar is None:
self.cookiejar = LWPCookieJar()
else:
self.cookiejar = cookiejar
self.session = requests.Session()
self.session.cookies = self.cookiejar
self.parser = PlansPageParser()
self.username = None
def _get_page(self, name, get=None, post=None):
"""
Retrieve an HTML page from plans.
"""
method = 'GET' if post is None else 'POST' | verify=url.startswith('https'))
except requests.exceptions.ConnectionError:
err = "Check your internet connection. Plans could also be down."
raise PlansError(err)
return handle
def _parse_message(self, soup):
"""
Scrape details from an infomessage or alertmessage div.
Returns a dictionary of the message parameters.
"""
kind, = soup.attrs[u'class']
title = soup.findChild().text
body = ''.join(t.text for t in soup.findChildren()[1:])
message = dict(kind=kind, title=title, body=body)
for val in message.values():
assert type(val) == str
return message
def _canonicalize_plantext(self, plan):
"""
Modify reserialized plan text to match what was served.
For consistency, we want to return plan text *exactly* how it
is formatted when served. However, variants of certain tags
(e.g. <br> vs <br/>) are syntactically equivalent in HTML, and
may be interchanged when parsed and reserialized.
This function manually changes the formatting returned by our
parser to more closely match that given by plans.
"""
# Our parser will correct <hr> and <br> to self closing tags
plan = plan.replace('<br/>', '<br>')
plan = plan.replace('<hr/>', '<hr>')
# put attributes in the right order because I have OCD
plan = re.sub(r'<a class="([^\s]*)" href="([^\s]*)">',
r'<a href="\2" class="\1">', plan)
# to avoid playing whack-a-mole, we should configure the
# parser to not do this, or else treat contents of
# <div class="plan_text"> tags as plain text
# (not sure if this is possible)
return plan
@staticmethod
def _html_esc(string):
"""
Replaces certain characters with html escape sequences.
Meant to be passed to the BS4 'decode' method as kwarg 'formatter'.
By default, BS4 only replaces angle brackets and ampersands with
the html escape sequence, but text served by plans also replaces
double quotes. This function as BS4's decoding formatter
reproduces that behavior.
"""
repls = {
'<': 'lt',
'>': 'gt',
'&': 'amp',
'"': 'quot',
}
def repl(matchobj):
return "&%s;" % repls[matchobj.group(0)]
regex = "([%s])" % ''.join(repls.keys())
return re.sub(regex, repl, string)
def plans_login(self, username='', password=''):
"""
Log into plans.
Returns True on success, False on failure. Leave username and
password blank to check an existing login.
"""
# the provided username and password ONLY get checked
# by the plans server if our cookie is expired.
# hence, if we've logged in recently, this will return True even
# if un/pw are not provided or are otherwise bad.
login_info = {'username': username,
'password': password,
'submit': 'Login'}
response = self._get_page('index.php', post=login_info)
# if login is successful, we'll be redirected to home
success = response.url[-9:] == '/home.php'
if success:
self.parser.feed(response.text) # parse out username
self.username = self.parser.username
return success
def get_edit_text(self):
"""
Retrieve contents of the edit plan field.
Returns the edit_text of the plan and its md5 hash,
as computed on the server side.
"""
# grab edit page
response = self._get_page('edit.php')
html = response.text
# parse out existing plan
soup = bs4.BeautifulSoup(html, 'html5lib')
plan = soup.find('textarea')
if plan is None:
raise PlansError("Couldn't get edit text, are we logged in?")
else:
plan = u'' + plan.contents[0]
# prepending the empty string somehow prevents BS from
# escaping all the HTML characters (weird)
assert type(plan) == str
# convert to CRLF line endings
plan = convert_endings(plan, 'CRLF')
# parse out plan md5
md5sum = soup.find('input',
attrs={'name': 'edit_text_md5'}).attrs['value']
# also, explicitly compute the hash, for kicks
assert md5sum == plans_md5(plan)
# verify that username has not changed
assert self.username == self.parser.username
return plan, md5sum
def set_edit_text(self, newtext, md5):
"""
Update plan with new content.
To prevent errors, the server does a hash check on the existing
plan before replacing it with the new one. We provide an
md5 sum to confirm that yes, we really want to update the plan.
Returns info message.
"""
# convert to CRLF line endings
newtext = convert_endings(newtext, 'CRLF')
newtext = newtext.encode('utf8')
edit_info = {'plan': newtext,
'edit_text_md5': md5,
'submit': 'Change Plan'}
response = self._get_page('edit.php', post=edit_info)
soup = bs4.BeautifulSoup(response.text, "html5lib")
alert = soup.find('div', {'class': 'alertmessage'})
info = soup.find('div', {'class': 'infomessage'})
if alert is not None:
# some kind of error
msg = self._parse_message(alert)
raise PlansError(msg['body'])
elif info is None:
raise PlansError('Plans did not verify update')
else:
# probably success
msg = self._parse_message(info)
return msg['body']
def get_autofinger(self):
"""
Retrieve all levels of the autofinger (autoread) list.
Returns a dictionary where the keys are the group names
"Level 1", "Level 2", etc. and the values are a list of
usernames waiting to be read.
"""
# this actually doesn't scrape; there's a function for it
# in the old JSON API.
get = {'task': 'autofingerlist'}
response = self._get_page('api/1/index.php', get=get)
data = json.loads(response.text)
# the returned JSON is crufty; clean it up
autofinger = {}
for group in data['autofingerList']:
name = "Level %s" % group['level']
autofinger[name] = group['usernames']
return autofinger
def read_plan(self, plan):
"""
Retrieve the contents of the specified plan.
Returns two objects: the plan header (as a python dictionary)
the plan text (in HTML format)
"""
get = {'searchname': plan}
response = self._get_page('read.php', get=get)
soup = bs4.BeautifulSoup(response.text, 'html5lib')
header = soup.find('div', {'id': 'header'})
text = soup.find('div', {'class': 'plan_text'})
if text is None or header is None:
# probably a nonexistent user
alert = soup.find('div', {'class': 'alertmessage'})
msg = self._parse_message(alert)
raise PlansError(msg['title'])
# convert header html into a python dictionary
header_dict = {}
for key in ('username', 'planname'):
content = header.find | url = '/'.join((self.base_url, name))
req = requests.Request(method, url, params=get, data=post)
prepped = self.session.prepare_request(req)
try:
handle = self.session.send(prepped, | random_line_split |
scraper.py | (query)['body']
# find username in submission content using brackets
start, stop = comment.index('['), comment.index(']')
self.username = comment[start + 1:stop]
# -------------------------------------------
# PLANS SCRAPEY-I
# get it? like "api" except... oh, never mind
# -------------------------------------------
class PlansConnection(object):
"""
Encapsulates an active login to plans.
"""
def __init__(self, cookiejar=None,
base_url="https://www.grinnellplans.com",
server_tz='US/Central'):
"""
Create a new plans connection.
Optional keyword arguments:
cookiejar -- an existing cookielib.CookieJar to store
credentials in.
base_url -- URL at which to access plans, no trailing slash.
server_tz -- Name of the timezone used by the server.
This class will convert dates to UTC.
"""
self.base_url = base_url
self.server_tz = server_tz
if cookiejar is None:
self.cookiejar = LWPCookieJar()
else:
|
self.session = requests.Session()
self.session.cookies = self.cookiejar
self.parser = PlansPageParser()
self.username = None
def _get_page(self, name, get=None, post=None):
"""
Retrieve an HTML page from plans.
"""
method = 'GET' if post is None else 'POST'
url = '/'.join((self.base_url, name))
req = requests.Request(method, url, params=get, data=post)
prepped = self.session.prepare_request(req)
try:
handle = self.session.send(prepped,
verify=url.startswith('https'))
except requests.exceptions.ConnectionError:
err = "Check your internet connection. Plans could also be down."
raise PlansError(err)
return handle
def _parse_message(self, soup):
"""
Scrape details from an infomessage or alertmessage div.
Returns a dictionary of the message parameters.
"""
kind, = soup.attrs[u'class']
title = soup.findChild().text
body = ''.join(t.text for t in soup.findChildren()[1:])
message = dict(kind=kind, title=title, body=body)
for val in message.values():
assert type(val) == str
return message
def _canonicalize_plantext(self, plan):
"""
Modify reserialized plan text to match what was served.
For consistency, we want to return plan text *exactly* how it
is formatted when served. However, variants of certain tags
(e.g. <br> vs <br/>) are syntactically equivalent in HTML, and
may be interchanged when parsed and reserialized.
This function manually changes the formatting returned by our
parser to more closely match that given by plans.
"""
# Our parser will correct <hr> and <br> to self closing tags
plan = plan.replace('<br/>', '<br>')
plan = plan.replace('<hr/>', '<hr>')
# put attributes in the right order because I have OCD
plan = re.sub(r'<a class="([^\s]*)" href="([^\s]*)">',
r'<a href="\2" class="\1">', plan)
# to avoid playing whack-a-mole, we should configure the
# parser to not do this, or else treat contents of
# <div class="plan_text"> tags as plain text
# (not sure if this is possible)
return plan
@staticmethod
def _html_esc(string):
"""
Replaces certain characters with html escape sequences.
Meant to be passed to the BS4 'decode' method as kwarg 'formatter'.
By default, BS4 only replaces angle brackets and ampersands with
the html escape sequence, but text served by plans also replaces
double quotes. This function as BS4's decoding formatter
reproduces that behavior.
"""
repls = {
'<': 'lt',
'>': 'gt',
'&': 'amp',
'"': 'quot',
}
def repl(matchobj):
return "&%s;" % repls[matchobj.group(0)]
regex = "([%s])" % ''.join(repls.keys())
return re.sub(regex, repl, string)
def plans_login(self, username='', password=''):
"""
Log into plans.
Returns True on success, False on failure. Leave username and
password blank to check an existing login.
"""
# the provided username and password ONLY get checked
# by the plans server if our cookie is expired.
# hence, if we've logged in recently, this will return True even
# if un/pw are not provided or are otherwise bad.
login_info = {'username': username,
'password': password,
'submit': 'Login'}
response = self._get_page('index.php', post=login_info)
# if login is successful, we'll be redirected to home
success = response.url[-9:] == '/home.php'
if success:
self.parser.feed(response.text) # parse out username
self.username = self.parser.username
return success
def get_edit_text(self):
"""
Retrieve contents of the edit plan field.
Returns the edit_text of the plan and its md5 hash,
as computed on the server side.
"""
# grab edit page
response = self._get_page('edit.php')
html = response.text
# parse out existing plan
soup = bs4.BeautifulSoup(html, 'html5lib')
plan = soup.find('textarea')
if plan is None:
raise PlansError("Couldn't get edit text, are we logged in?")
else:
plan = u'' + plan.contents[0]
# prepending the empty string somehow prevents BS from
# escaping all the HTML characters (weird)
assert type(plan) == str
# convert to CRLF line endings
plan = convert_endings(plan, 'CRLF')
# parse out plan md5
md5sum = soup.find('input',
attrs={'name': 'edit_text_md5'}).attrs['value']
# also, explicitly compute the hash, for kicks
assert md5sum == plans_md5(plan)
# verify that username has not changed
assert self.username == self.parser.username
return plan, md5sum
def set_edit_text(self, newtext, md5):
"""
Update plan with new content.
To prevent errors, the server does a hash check on the existing
plan before replacing it with the new one. We provide an
md5 sum to confirm that yes, we really want to update the plan.
Returns info message.
"""
# convert to CRLF line endings
newtext = convert_endings(newtext, 'CRLF')
newtext = newtext.encode('utf8')
edit_info = {'plan': newtext,
'edit_text_md5': md5,
'submit': 'Change Plan'}
response = self._get_page('edit.php', post=edit_info)
soup = bs4.BeautifulSoup(response.text, "html5lib")
alert = soup.find('div', {'class': 'alertmessage'})
info = soup.find('div', {'class': 'infomessage'})
if alert is not None:
# some kind of error
msg = self._parse_message(alert)
raise PlansError(msg['body'])
elif info is None:
raise PlansError('Plans did not verify update')
else:
# probably success
msg = self._parse_message(info)
return msg['body']
def get_autofinger(self):
"""
Retrieve all levels of the autofinger (autoread) list.
Returns a dictionary where the keys are the group names
"Level 1", "Level 2", etc. and the values are a list of
usernames waiting to be read.
"""
# this actually doesn't scrape; there's a function for it
# in the old JSON API.
get = {'task': 'autofingerlist'}
response = self._get_page('api/1/index.php', get=get)
data = json.loads(response.text)
# the returned JSON is crufty; clean it up
autofinger = {}
for group in data['autofingerList']:
name = "Level %s" % group['level']
autofinger[name] = group['usernames']
return autofinger
def read_plan(self, plan):
"""
Retrieve the contents of the specified plan.
Returns two objects: the plan header (as a python dictionary)
the plan text (in HTML format)
"""
get = {'searchname': plan}
response = self._get_page('read.php', get=get)
soup = bs4.BeautifulSoup(response.text, 'html5lib')
header = soup.find('div', {'id': 'header'})
text = soup.find('div', {'class': 'plan_text'})
if text is None or header is None:
# probably a nonexistent user
alert = soup.find('div', {'class': 'alertmessage'})
msg = self._parse_message(alert)
raise PlansError(msg['title'])
# convert header html into a python dictionary
header_dict = {}
for key in ('username', 'planname'):
content = header.find(
| self.cookiejar = cookiejar | conditional_block |
scraper.py | function as BS4's decoding formatter
reproduces that behavior.
"""
repls = {
'<': 'lt',
'>': 'gt',
'&': 'amp',
'"': 'quot',
}
def repl(matchobj):
return "&%s;" % repls[matchobj.group(0)]
regex = "([%s])" % ''.join(repls.keys())
return re.sub(regex, repl, string)
def plans_login(self, username='', password=''):
"""
Log into plans.
Returns True on success, False on failure. Leave username and
password blank to check an existing login.
"""
# the provided username and password ONLY get checked
# by the plans server if our cookie is expired.
# hence, if we've logged in recently, this will return True even
# if un/pw are not provided or are otherwise bad.
login_info = {'username': username,
'password': password,
'submit': 'Login'}
response = self._get_page('index.php', post=login_info)
# if login is successful, we'll be redirected to home
success = response.url[-9:] == '/home.php'
if success:
self.parser.feed(response.text) # parse out username
self.username = self.parser.username
return success
def get_edit_text(self):
"""
Retrieve contents of the edit plan field.
Returns the edit_text of the plan and its md5 hash,
as computed on the server side.
"""
# grab edit page
response = self._get_page('edit.php')
html = response.text
# parse out existing plan
soup = bs4.BeautifulSoup(html, 'html5lib')
plan = soup.find('textarea')
if plan is None:
raise PlansError("Couldn't get edit text, are we logged in?")
else:
plan = u'' + plan.contents[0]
# prepending the empty string somehow prevents BS from
# escaping all the HTML characters (weird)
assert type(plan) == str
# convert to CRLF line endings
plan = convert_endings(plan, 'CRLF')
# parse out plan md5
md5sum = soup.find('input',
attrs={'name': 'edit_text_md5'}).attrs['value']
# also, explicitly compute the hash, for kicks
assert md5sum == plans_md5(plan)
# verify that username has not changed
assert self.username == self.parser.username
return plan, md5sum
def set_edit_text(self, newtext, md5):
"""
Update plan with new content.
To prevent errors, the server does a hash check on the existing
plan before replacing it with the new one. We provide an
md5 sum to confirm that yes, we really want to update the plan.
Returns info message.
"""
# convert to CRLF line endings
newtext = convert_endings(newtext, 'CRLF')
newtext = newtext.encode('utf8')
edit_info = {'plan': newtext,
'edit_text_md5': md5,
'submit': 'Change Plan'}
response = self._get_page('edit.php', post=edit_info)
soup = bs4.BeautifulSoup(response.text, "html5lib")
alert = soup.find('div', {'class': 'alertmessage'})
info = soup.find('div', {'class': 'infomessage'})
if alert is not None:
# some kind of error
msg = self._parse_message(alert)
raise PlansError(msg['body'])
elif info is None:
raise PlansError('Plans did not verify update')
else:
# probably success
msg = self._parse_message(info)
return msg['body']
def get_autofinger(self):
"""
Retrieve all levels of the autofinger (autoread) list.
Returns a dictionary where the keys are the group names
"Level 1", "Level 2", etc. and the values are a list of
usernames waiting to be read.
"""
# this actually doesn't scrape; there's a function for it
# in the old JSON API.
get = {'task': 'autofingerlist'}
response = self._get_page('api/1/index.php', get=get)
data = json.loads(response.text)
# the returned JSON is crufty; clean it up
autofinger = {}
for group in data['autofingerList']:
name = "Level %s" % group['level']
autofinger[name] = group['usernames']
return autofinger
def read_plan(self, plan):
"""
Retrieve the contents of the specified plan.
Returns two objects: the plan header (as a python dictionary)
the plan text (in HTML format)
"""
get = {'searchname': plan}
response = self._get_page('read.php', get=get)
soup = bs4.BeautifulSoup(response.text, 'html5lib')
header = soup.find('div', {'id': 'header'})
text = soup.find('div', {'class': 'plan_text'})
if text is None or header is None:
# probably a nonexistent user
alert = soup.find('div', {'class': 'alertmessage'})
msg = self._parse_message(alert)
raise PlansError(msg['title'])
# convert header html into a python dictionary
header_dict = {}
for key in ('username', 'planname'):
content = header.find(
'li', {'class': key}
).find(
'span', {'class': 'value'}
).contents
value = str(content[0]) if len(content) > 0 else None
header_dict[key] = value
for key in ('lastupdated', 'lastlogin'):
content = header.find(
'li', {'class': key}
).find(
'span', {'class': 'value'}
).find(
'span', {'class': 'long'}
).contents
if len(content) > 0:
value = str(content[0])
value = parse_plans_date(value, tz_name=self.server_tz)
else:
value = None
header_dict[key] = value
text.hidden = True # prevents BS from wrapping contents in
# <div> upon conversion to unicode string
plan = text.decode(formatter=self._html_esc) # soup to unicode
assert plan[0] == '\n' # drop leading newline
plan = self._canonicalize_plantext(plan[1:])
return header_dict, plan
def search_plans(self, term, planlove=False):
"""
Search plans for the provided ``term``.
If ``planlove`` is ``True``, ``term`` is a username, and the
search will be for incidences of planlove for that user.
returns: list of plans upon which the search term was found.
each list element is a 3-tuple:
- plan name
- number of occurrences of search term on the plan
- list of plan excerpts giving context
the length of the excerpt list may be equal to or less than
the number of occurrences of the search term, since
overlapping excerpts are consolidated.
"""
get = {'mysearch': term,
'planlove': int(bool(planlove))}
response = self._get_page('search.php', get=get)
soup = bs4.BeautifulSoup(response.text, 'html5lib')
results = soup.find('ul', {'id': 'search_results'})
if results is None:
return [] # no results
# results are grouped by the plan
# on which the result was found
user_groups = results.findAll(
'div', {'class': 'result_user_group'})
resultlist = []
for group in user_groups:
user = group.find('a', {'class': 'planlove'}).contents[0]
count = group.find('span').contents[0]
# now extract snippets
snippetlist = group.findAll('li')
snippets = []
for li in snippetlist:
tag = li.find('span')
tag.hidden = True # prevents BS from wrapping contents in
# <span> upon conversion to unicode string
snip = tag.decode(formatter=self._html_esc) # soup to unicode
snip = self._canonicalize_plantext(snip)
snippets.append(snip)
resultlist.append((str(user), int(count), snippets))
return resultlist
def planwatch(self, hours=12):
| """
Return plans updated in the last ``hours`` hours.
The result is a list of (username, timestamp) 2-tuples.
"""
post = {'mytime': str(hours)}
response = self._get_page('planwatch.php', post=post)
soup = bs4.BeautifulSoup(response.text, 'html5lib')
results = soup.find('ul', {'id': 'new_plan_list'})
new_plans = results.findAll('div', {'class': 'newplan'})
resultlist = []
for div in new_plans:
user = div.find('a', {'class': 'planlove'}).contents[0]
time = div.find('span').contents[0]
time = parse_plans_date(time, tz_name=self.server_tz)
resultlist.append((user, time))
return resultlist | identifier_body |
|
AnimePageFetcher.py | return html
def load_html_from_file(file):
with open(file, 'r') as f:
html = f.read()
return html
#these are the stats found on the stat page of an anime
#Example: https://myanimelist.net/anime/30/Neon_Genesis_Evangelion/stats
def getGeneralStatistics(soup, aggregate_dict={}):
return aggregate_dict
# Returns (type, id) where type is str in ["anime", "manga"], and id an int
def getCategoryAndIDFromUrl(url):
url = str(url)
result = url.split("myanimelist.net/")
if(len(result) > 1):
result = result[1].split("/")
content_category = result[0]
content_id = int(result[1])
return (content_category, content_id)
#these are generic info like the producers and the source found in the
#sidebar of an anime page
#Example https://myanimelist.net/anime/30/Neon_Genesis_Evangelion
def getGeneralInformation(html, aggregate_dict={}):
soup = BeautifulSoup(html, 'html.parser')
# Title
title = soup.find("h1", class_="h1").get_text()
aggregate_dict["title"] = title
# Score
score_tag = soup.find("div", class_="fl-l score")
score_users = extract_comma_number(score_tag["data-user"])
score_value = 0.0
try:
score_value = float(score_tag.get_text().strip())
except:
score_value = 0.0
aggregate_dict["score_users"] = score_users
aggregate_dict["score"] = score_value
# Rank
rank_text = soup.find("span", class_="numbers ranked").get_text()
rank_match = rank_regex.search(rank_text)
if rank_match is None:
aggregate_dict["rank"] = None
else:
rank_value = int(rank_match.group(1))
aggregate_dict["rank"] = rank_value
# Popularity
popularity_text = soup.find("span", class_="numbers popularity").get_text()
popularity_value = int(rank_regex.search(popularity_text).group(1))
aggregate_dict["popularity"] = popularity_value
# Members
members_text = soup.find("span", class_="numbers members").get_text().strip()
members_value = int(extract_comma_number(members_text))
aggregate_dict["members"] = members_value
# Synoposis
synopsis_soup = soup.find("span", itemprop="description")
if synopsis_soup is None:
aggregate_dict["synopsis"] = None
else:
synopsis_text = " ".join(synopsis_soup.strings)
aggregate_dict["synopsis"] = synopsis_text
#RelatedAnime
related_table = soup.find_all("table", class_="anime_detail_related_anime")
if len(related_table) == 0:
aggregate_dict["related_ids"] = []
else:
related_entries = [t['href'].strip() for t in related_table[0].find_all("a")]
related_entries = filter(lambda x: "/anime/" in x and len(x.split("/")[2]) > 0, related_entries)
related_titles = map(lambda x: int(x.split("/")[2]), related_entries)
aggregate_dict["related_ids"] = related_titles
# Statistics/Favorites (we have everything else)
favorites_match = favorites_regex.search(html)
favorites_text = favorites_match.group(1)
aggregate_dict["favorites"] = int(favorites_text.replace(",", ""))
# Information section
info_dict = extract_info_section(html)
# Info/Type
aggregate_dict["type"] = info_dict.get("Type")
# Info/Episodes
if "Episodes" in info_dict:
episodes_value = 0
# Some anime pages have "Unknown" for number of episodes.
# 0 represents unknown, because no anime can truly have 0 episodes.
try:
episodes_value = int(info_dict["Episodes"])
except:
episodes_value = 0
aggregate_dict["episodes"] = episodes_value
else:
aggregate_dict["episodes"] = None
# Info/Status
aggregate_dict["status"] = info_dict.get("Status")
# Info/Aired
aired_start = None
aired_end = None
if "Aired" in info_dict:
aired_text = info_dict["Aired"]
# Some animes are aired on one date. Others run for some time period.
# Those that run over a duration have "to" in the text.
if "to" in aired_text:
start_end_split = aired_text.split("to")
aired_start = parse_date(start_end_split[0])
# Some currently running animes have ? for their end date.
if "?" in start_end_split[1]:
aired_end = None
else:
aired_end = parse_date(start_end_split[1])
else:
aired_start = parse_date(aired_text)
aired_end = aired_start
aggregate_dict["aired_start"] = aired_start
aggregate_dict["aired_end"] = aired_end
# Info/Premiered
aggregate_dict["premiered"] = info_dict.get("Premiered")
# Info/Broadcast
aggregate_dict["broadcast"] = info_dict.get("Broadcast")
# Info/Producers
if "Producers" in info_dict:
aggregate_dict["producers"] = parse_info_list(info_dict["Producers"])
else:
aggregate_dict["producers"] = None
# Info/Licensors
if "Licensors" in info_dict:
aggregate_dict["licensors"] = parse_info_list(info_dict["Licensors"])
else:
aggregate_dict["licensors"] = None
# Info/Studios
if "Studios" in info_dict:
aggregate_dict["studios"] = parse_info_list(info_dict["Studios"])
else:
aggregate_dict["studios"] = studios_list
# Info/Source
aggregate_dict["source"] = info_dict.get("Source")
# Info/Genres
if "Genres" in info_dict:
aggregate_dict["genres"] = parse_info_list(info_dict["Genres"])
else:
aggregate_dict["genres"] = None
# Info/Duration
if "Duration" in info_dict:
duration_text = info_dict["Duration"]
mins = 0.0
# Hours
match = hours_regex.search(duration_text)
if match is not None:
mins += float(match.group(1)) * 60.0
# Minutes
match = min_regex.search(duration_text)
if match is not None:
mins += float(match.group(1))
# Seconds
match = sec_regex.search(duration_text)
if match is not None:
mins += float(match.group(1)) / 60.0
aggregate_dict["duration"] = mins
else:
aggregate_dict["duration"] = None
# Info/Rating
if "Rating" in info_dict:
rating_text = info_dict["Rating"]
rating_shorthand = rating_text.split(" - ")[0].strip()
aggregate_dict["rating"] = rating_shorthand
else:
aggregate_dict["rating"] = None
return aggregate_dict
#Fetches the production staff associated with a show
#Example: https://myanimelist.net/anime/30/Neon_Genesis_Evangelion/characters#staff
def getStaff(html):
return
#Fetches the characters and voice actors associted with a show
#Example: https://myanimelist.net/anime/30/Neon_Genesis_Evangelion/characters
def getCharactersAndJapaneseCast(html):
return
#Fetches the list of other related anime, same series etc.
def getRelatedTitles(html):
return
def cooldown():
COOLDOWN_IN_SECONDS = 0.5
time.sleep(COOLDOWN_IN_SECONDS)
def bs_preprocess(html):
"""remove distracting whitespaces and newline characters"""
pat = re.compile('(^[\s]+)|([\s]+$)', re.MULTILINE)
html = re.sub(pat, '', html) # remove leading and trailing whitespaces
html = re.sub('\n', ' ', html) # convert newlines to spaces
# this preserves newline delimiters
html = re.sub('[\s]+<', '<', html) # remove whitespaces before opening tags
html = re.sub('>[\s]+', '>', html) # remove whitespaces after closing tags
return html
# Returns dictionary with data
def scrape_main_page(html, aggregate_data={}):
getGeneralInformation(html, aggregate_data)
return aggregate_data
# Returns dictionary with data
def | (html, aggregate_data={}):
getStatSummary(html, aggregate_data)
getStatDistribution(html, aggregate_data)
return aggregate_data
def getStatSummary(html, aggregate_dict={}):
#soup = BeautifulSoup(html, 'html.parser')
start_index = stats_summary_regex.search(html).end()
end_index = stats_score_regex.search(html).start()
subs = html[start_index:end_index]
stat_soup = BeautifulSoup(html[start_index:end_index], 'html.parser')
field_list = [t.get_text().strip() for t in stat_soup.find_all("div")]
for field in field_list[:6]:
t = "users_" | scrape_stats_page | identifier_name |
AnimePageFetcher.py | with open(out_file, 'w') as f:
soup = BeautifulSoup(html, 'html.parser')
html = soup.prettify()
f.write(html.encode("utf8"))
return html
def load_html_from_file(file):
with open(file, 'r') as f:
html = f.read()
return html
#these are the stats found on the stat page of an anime
#Example: https://myanimelist.net/anime/30/Neon_Genesis_Evangelion/stats
def getGeneralStatistics(soup, aggregate_dict={}):
return aggregate_dict
# Returns (type, id) where type is str in ["anime", "manga"], and id an int
def getCategoryAndIDFromUrl(url):
url = str(url)
result = url.split("myanimelist.net/")
if(len(result) > 1):
result = result[1].split("/")
content_category = result[0]
content_id = int(result[1])
return (content_category, content_id)
#these are generic info like the producers and the source found in the
#sidebar of an anime page
#Example https://myanimelist.net/anime/30/Neon_Genesis_Evangelion
def getGeneralInformation(html, aggregate_dict={}):
soup = BeautifulSoup(html, 'html.parser')
# Title
title = soup.find("h1", class_="h1").get_text()
aggregate_dict["title"] = title
# Score
score_tag = soup.find("div", class_="fl-l score")
score_users = extract_comma_number(score_tag["data-user"])
score_value = 0.0
try:
score_value = float(score_tag.get_text().strip())
except:
score_value = 0.0
aggregate_dict["score_users"] = score_users
aggregate_dict["score"] = score_value
# Rank
rank_text = soup.find("span", class_="numbers ranked").get_text()
rank_match = rank_regex.search(rank_text)
if rank_match is None:
aggregate_dict["rank"] = None
else:
rank_value = int(rank_match.group(1))
aggregate_dict["rank"] = rank_value
# Popularity
popularity_text = soup.find("span", class_="numbers popularity").get_text()
popularity_value = int(rank_regex.search(popularity_text).group(1))
aggregate_dict["popularity"] = popularity_value
# Members
members_text = soup.find("span", class_="numbers members").get_text().strip()
members_value = int(extract_comma_number(members_text))
aggregate_dict["members"] = members_value
# Synoposis
synopsis_soup = soup.find("span", itemprop="description")
if synopsis_soup is None:
aggregate_dict["synopsis"] = None
else:
synopsis_text = " ".join(synopsis_soup.strings)
aggregate_dict["synopsis"] = synopsis_text
#RelatedAnime
related_table = soup.find_all("table", class_="anime_detail_related_anime")
if len(related_table) == 0:
aggregate_dict["related_ids"] = []
else:
related_entries = [t['href'].strip() for t in related_table[0].find_all("a")]
related_entries = filter(lambda x: "/anime/" in x and len(x.split("/")[2]) > 0, related_entries)
related_titles = map(lambda x: int(x.split("/")[2]), related_entries)
aggregate_dict["related_ids"] = related_titles
# Statistics/Favorites (we have everything else)
favorites_match = favorites_regex.search(html)
favorites_text = favorites_match.group(1)
aggregate_dict["favorites"] = int(favorites_text.replace(",", ""))
# Information section
info_dict = extract_info_section(html)
# Info/Type
aggregate_dict["type"] = info_dict.get("Type")
# Info/Episodes
if "Episodes" in info_dict:
episodes_value = 0
# Some anime pages have "Unknown" for number of episodes.
# 0 represents unknown, because no anime can truly have 0 episodes.
try:
episodes_value = int(info_dict["Episodes"])
except:
episodes_value = 0
aggregate_dict["episodes"] = episodes_value
else:
aggregate_dict["episodes"] = None
# Info/Status
aggregate_dict["status"] = info_dict.get("Status")
# Info/Aired
aired_start = None
aired_end = None
if "Aired" in info_dict:
aired_text = info_dict["Aired"]
# Some animes are aired on one date. Others run for some time period.
# Those that run over a duration have "to" in the text.
if "to" in aired_text:
start_end_split = aired_text.split("to")
aired_start = parse_date(start_end_split[0])
# Some currently running animes have ? for their end date.
if "?" in start_end_split[1]:
aired_end = None
else:
aired_end = parse_date(start_end_split[1])
else:
aired_start = parse_date(aired_text)
aired_end = aired_start
aggregate_dict["aired_start"] = aired_start
aggregate_dict["aired_end"] = aired_end
# Info/Premiered
aggregate_dict["premiered"] = info_dict.get("Premiered")
# Info/Broadcast
aggregate_dict["broadcast"] = info_dict.get("Broadcast")
# Info/Producers
if "Producers" in info_dict:
aggregate_dict["producers"] = parse_info_list(info_dict["Producers"])
else:
aggregate_dict["producers"] = None
# Info/Licensors
if "Licensors" in info_dict:
aggregate_dict["licensors"] = parse_info_list(info_dict["Licensors"])
else:
aggregate_dict["licensors"] = None
# Info/Studios
if "Studios" in info_dict:
aggregate_dict["studios"] = parse_info_list(info_dict["Studios"])
else:
aggregate_dict["studios"] = studios_list
# Info/Source
aggregate_dict["source"] = info_dict.get("Source")
# Info/Genres
if "Genres" in info_dict:
aggregate_dict["genres"] = parse_info_list(info_dict["Genres"])
else:
aggregate_dict["genres"] = None
# Info/Duration
if "Duration" in info_dict:
duration_text = info_dict["Duration"]
mins = 0.0
# Hours
match = hours_regex.search(duration_text)
if match is not None:
mins += float(match.group(1)) * 60.0
# Minutes
match = min_regex.search(duration_text)
if match is not None:
mins += float(match.group(1))
# Seconds
match = sec_regex.search(duration_text)
if match is not None:
mins += float(match.group(1)) / 60.0
aggregate_dict["duration"] = mins
else:
aggregate_dict["duration"] = None
# Info/Rating
if "Rating" in info_dict:
rating_text = info_dict["Rating"]
rating_shorthand = rating_text.split(" - ")[0].strip()
aggregate_dict["rating"] = rating_shorthand
else:
aggregate_dict["rating"] = None
return aggregate_dict
#Fetches the production staff associated with a show
#Example: https://myanimelist.net/anime/30/Neon_Genesis_Evangelion/characters#staff
def getStaff(html):
return
#Fetches the characters and voice actors associted with a show
#Example: https://myanimelist.net/anime/30/Neon_Genesis_Evangelion/characters
def getCharactersAndJapaneseCast(html):
return
#Fetches the list of other related anime, same series etc.
def getRelatedTitles(html):
return
def cooldown():
COOLDOWN_IN_SECONDS = 0.5
time.sleep(COOLDOWN_IN_SECONDS)
def bs_preprocess(html):
"""remove distracting whitespaces and newline characters"""
pat = re.compile('(^[\s]+)|([\s]+$)', re.MULTILINE)
html = re.sub(pat, '', html) # remove leading and trailing whitespaces
html = re.sub('\n', ' ', html) # convert newlines to spaces
# this preserves newline delimiters
html = re.sub('[\s]+<', '<', html) # remove whitespaces before opening tags
html = re.sub('>[\s]+', '>', html) # remove whitespaces after closing tags
return html
# Returns dictionary with data
def scrape_main_page(html, aggregate_data={}):
getGeneralInformation(html, aggregate_data)
return aggregate_data
# Returns dictionary with data
def scrape_stats_page(html, aggregate_data={}):
getStatSummary(html, aggregate_data)
getStatDistribution(html, aggregate_data)
return aggregate_data
def getStatSummary(html, aggregate_dict={}):
#soup = BeautifulSoup(html, 'html.parser')
start_index = stats_summary_regex.search(html).end()
end_index = stats_score_regex.search(html).start()
subs = | print "Writing html to", out_file
print "Type is", type(html) | random_line_split |
|
AnimePageFetcher.py | return html
def load_html_from_file(file):
with open(file, 'r') as f:
html = f.read()
return html
#these are the stats found on the stat page of an anime
#Example: https://myanimelist.net/anime/30/Neon_Genesis_Evangelion/stats
def getGeneralStatistics(soup, aggregate_dict={}):
return aggregate_dict
# Returns (type, id) where type is str in ["anime", "manga"], and id an int
def getCategoryAndIDFromUrl(url):
url = str(url)
result = url.split("myanimelist.net/")
if(len(result) > 1):
result = result[1].split("/")
content_category = result[0]
content_id = int(result[1])
return (content_category, content_id)
#these are generic info like the producers and the source found in the
#sidebar of an anime page
#Example https://myanimelist.net/anime/30/Neon_Genesis_Evangelion
def getGeneralInformation(html, aggregate_dict={}):
soup = BeautifulSoup(html, 'html.parser')
# Title
title = soup.find("h1", class_="h1").get_text()
aggregate_dict["title"] = title
# Score
score_tag = soup.find("div", class_="fl-l score")
score_users = extract_comma_number(score_tag["data-user"])
score_value = 0.0
try:
score_value = float(score_tag.get_text().strip())
except:
score_value = 0.0
aggregate_dict["score_users"] = score_users
aggregate_dict["score"] = score_value
# Rank
rank_text = soup.find("span", class_="numbers ranked").get_text()
rank_match = rank_regex.search(rank_text)
if rank_match is None:
aggregate_dict["rank"] = None
else:
rank_value = int(rank_match.group(1))
aggregate_dict["rank"] = rank_value
# Popularity
popularity_text = soup.find("span", class_="numbers popularity").get_text()
popularity_value = int(rank_regex.search(popularity_text).group(1))
aggregate_dict["popularity"] = popularity_value
# Members
members_text = soup.find("span", class_="numbers members").get_text().strip()
members_value = int(extract_comma_number(members_text))
aggregate_dict["members"] = members_value
# Synoposis
synopsis_soup = soup.find("span", itemprop="description")
if synopsis_soup is None:
aggregate_dict["synopsis"] = None
else:
synopsis_text = " ".join(synopsis_soup.strings)
aggregate_dict["synopsis"] = synopsis_text
#RelatedAnime
related_table = soup.find_all("table", class_="anime_detail_related_anime")
if len(related_table) == 0:
aggregate_dict["related_ids"] = []
else:
related_entries = [t['href'].strip() for t in related_table[0].find_all("a")]
related_entries = filter(lambda x: "/anime/" in x and len(x.split("/")[2]) > 0, related_entries)
related_titles = map(lambda x: int(x.split("/")[2]), related_entries)
aggregate_dict["related_ids"] = related_titles
# Statistics/Favorites (we have everything else)
favorites_match = favorites_regex.search(html)
favorites_text = favorites_match.group(1)
aggregate_dict["favorites"] = int(favorites_text.replace(",", ""))
# Information section
info_dict = extract_info_section(html)
# Info/Type
aggregate_dict["type"] = info_dict.get("Type")
# Info/Episodes
if "Episodes" in info_dict:
episodes_value = 0
# Some anime pages have "Unknown" for number of episodes.
# 0 represents unknown, because no anime can truly have 0 episodes.
try:
episodes_value = int(info_dict["Episodes"])
except:
episodes_value = 0
aggregate_dict["episodes"] = episodes_value
else:
aggregate_dict["episodes"] = None
# Info/Status
aggregate_dict["status"] = info_dict.get("Status")
# Info/Aired
aired_start = None
aired_end = None
if "Aired" in info_dict:
aired_text = info_dict["Aired"]
# Some animes are aired on one date. Others run for some time period.
# Those that run over a duration have "to" in the text.
if "to" in aired_text:
start_end_split = aired_text.split("to")
aired_start = parse_date(start_end_split[0])
# Some currently running animes have ? for their end date.
if "?" in start_end_split[1]:
aired_end = None
else:
aired_end = parse_date(start_end_split[1])
else:
aired_start = parse_date(aired_text)
aired_end = aired_start
aggregate_dict["aired_start"] = aired_start
aggregate_dict["aired_end"] = aired_end
# Info/Premiered
aggregate_dict["premiered"] = info_dict.get("Premiered")
# Info/Broadcast
aggregate_dict["broadcast"] = info_dict.get("Broadcast")
# Info/Producers
if "Producers" in info_dict:
aggregate_dict["producers"] = parse_info_list(info_dict["Producers"])
else:
aggregate_dict["producers"] = None
# Info/Licensors
if "Licensors" in info_dict:
aggregate_dict["licensors"] = parse_info_list(info_dict["Licensors"])
else:
aggregate_dict["licensors"] = None
# Info/Studios
if "Studios" in info_dict:
aggregate_dict["studios"] = parse_info_list(info_dict["Studios"])
else:
aggregate_dict["studios"] = studios_list
# Info/Source
aggregate_dict["source"] = info_dict.get("Source")
# Info/Genres
if "Genres" in info_dict:
aggregate_dict["genres"] = parse_info_list(info_dict["Genres"])
else:
aggregate_dict["genres"] = None
# Info/Duration
if "Duration" in info_dict:
|
else:
aggregate_dict["duration"] = None
# Info/Rating
if "Rating" in info_dict:
rating_text = info_dict["Rating"]
rating_shorthand = rating_text.split(" - ")[0].strip()
aggregate_dict["rating"] = rating_shorthand
else:
aggregate_dict["rating"] = None
return aggregate_dict
#Fetches the production staff associated with a show
#Example: https://myanimelist.net/anime/30/Neon_Genesis_Evangelion/characters#staff
def getStaff(html):
return
#Fetches the characters and voice actors associted with a show
#Example: https://myanimelist.net/anime/30/Neon_Genesis_Evangelion/characters
def getCharactersAndJapaneseCast(html):
return
#Fetches the list of other related anime, same series etc.
def getRelatedTitles(html):
return
def cooldown():
COOLDOWN_IN_SECONDS = 0.5
time.sleep(COOLDOWN_IN_SECONDS)
def bs_preprocess(html):
"""remove distracting whitespaces and newline characters"""
pat = re.compile('(^[\s]+)|([\s]+$)', re.MULTILINE)
html = re.sub(pat, '', html) # remove leading and trailing whitespaces
html = re.sub('\n', ' ', html) # convert newlines to spaces
# this preserves newline delimiters
html = re.sub('[\s]+<', '<', html) # remove whitespaces before opening tags
html = re.sub('>[\s]+', '>', html) # remove whitespaces after closing tags
return html
# Returns dictionary with data
def scrape_main_page(html, aggregate_data={}):
getGeneralInformation(html, aggregate_data)
return aggregate_data
# Returns dictionary with data
def scrape_stats_page(html, aggregate_data={}):
getStatSummary(html, aggregate_data)
getStatDistribution(html, aggregate_data)
return aggregate_data
def getStatSummary(html, aggregate_dict={}):
#soup = BeautifulSoup(html, 'html.parser')
start_index = stats_summary_regex.search(html).end()
end_index = stats_score_regex.search(html).start()
subs = html[start_index:end_index]
stat_soup = BeautifulSoup(html[start_index:end_index], 'html.parser')
field_list = [t.get_text().strip() for t in stat_soup.find_all("div")]
for field in field_list[:6]:
t = "users_" + | duration_text = info_dict["Duration"]
mins = 0.0
# Hours
match = hours_regex.search(duration_text)
if match is not None:
mins += float(match.group(1)) * 60.0
# Minutes
match = min_regex.search(duration_text)
if match is not None:
mins += float(match.group(1))
# Seconds
match = sec_regex.search(duration_text)
if match is not None:
mins += float(match.group(1)) / 60.0
aggregate_dict["duration"] = mins | conditional_block |
AnimePageFetcher.py | return html
def load_html_from_file(file):
with open(file, 'r') as f:
html = f.read()
return html
#these are the stats found on the stat page of an anime
#Example: https://myanimelist.net/anime/30/Neon_Genesis_Evangelion/stats
def getGeneralStatistics(soup, aggregate_dict={}):
return aggregate_dict
# Returns (type, id) where type is str in ["anime", "manga"], and id an int
def getCategoryAndIDFromUrl(url):
url = str(url)
result = url.split("myanimelist.net/")
if(len(result) > 1):
result = result[1].split("/")
content_category = result[0]
content_id = int(result[1])
return (content_category, content_id)
#these are generic info like the producers and the source found in the
#sidebar of an anime page
#Example https://myanimelist.net/anime/30/Neon_Genesis_Evangelion
def getGeneralInformation(html, aggregate_dict={}):
soup = BeautifulSoup(html, 'html.parser')
# Title
title = soup.find("h1", class_="h1").get_text()
aggregate_dict["title"] = title
# Score
score_tag = soup.find("div", class_="fl-l score")
score_users = extract_comma_number(score_tag["data-user"])
score_value = 0.0
try:
score_value = float(score_tag.get_text().strip())
except:
score_value = 0.0
aggregate_dict["score_users"] = score_users
aggregate_dict["score"] = score_value
# Rank
rank_text = soup.find("span", class_="numbers ranked").get_text()
rank_match = rank_regex.search(rank_text)
if rank_match is None:
aggregate_dict["rank"] = None
else:
rank_value = int(rank_match.group(1))
aggregate_dict["rank"] = rank_value
# Popularity
popularity_text = soup.find("span", class_="numbers popularity").get_text()
popularity_value = int(rank_regex.search(popularity_text).group(1))
aggregate_dict["popularity"] = popularity_value
# Members
members_text = soup.find("span", class_="numbers members").get_text().strip()
members_value = int(extract_comma_number(members_text))
aggregate_dict["members"] = members_value
# Synoposis
synopsis_soup = soup.find("span", itemprop="description")
if synopsis_soup is None:
aggregate_dict["synopsis"] = None
else:
synopsis_text = " ".join(synopsis_soup.strings)
aggregate_dict["synopsis"] = synopsis_text
#RelatedAnime
related_table = soup.find_all("table", class_="anime_detail_related_anime")
if len(related_table) == 0:
aggregate_dict["related_ids"] = []
else:
related_entries = [t['href'].strip() for t in related_table[0].find_all("a")]
related_entries = filter(lambda x: "/anime/" in x and len(x.split("/")[2]) > 0, related_entries)
related_titles = map(lambda x: int(x.split("/")[2]), related_entries)
aggregate_dict["related_ids"] = related_titles
# Statistics/Favorites (we have everything else)
favorites_match = favorites_regex.search(html)
favorites_text = favorites_match.group(1)
aggregate_dict["favorites"] = int(favorites_text.replace(",", ""))
# Information section
info_dict = extract_info_section(html)
# Info/Type
aggregate_dict["type"] = info_dict.get("Type")
# Info/Episodes
if "Episodes" in info_dict:
episodes_value = 0
# Some anime pages have "Unknown" for number of episodes.
# 0 represents unknown, because no anime can truly have 0 episodes.
try:
episodes_value = int(info_dict["Episodes"])
except:
episodes_value = 0
aggregate_dict["episodes"] = episodes_value
else:
aggregate_dict["episodes"] = None
# Info/Status
aggregate_dict["status"] = info_dict.get("Status")
# Info/Aired
aired_start = None
aired_end = None
if "Aired" in info_dict:
aired_text = info_dict["Aired"]
# Some animes are aired on one date. Others run for some time period.
# Those that run over a duration have "to" in the text.
if "to" in aired_text:
start_end_split = aired_text.split("to")
aired_start = parse_date(start_end_split[0])
# Some currently running animes have ? for their end date.
if "?" in start_end_split[1]:
aired_end = None
else:
aired_end = parse_date(start_end_split[1])
else:
aired_start = parse_date(aired_text)
aired_end = aired_start
aggregate_dict["aired_start"] = aired_start
aggregate_dict["aired_end"] = aired_end
# Info/Premiered
aggregate_dict["premiered"] = info_dict.get("Premiered")
# Info/Broadcast
aggregate_dict["broadcast"] = info_dict.get("Broadcast")
# Info/Producers
if "Producers" in info_dict:
aggregate_dict["producers"] = parse_info_list(info_dict["Producers"])
else:
aggregate_dict["producers"] = None
# Info/Licensors
if "Licensors" in info_dict:
aggregate_dict["licensors"] = parse_info_list(info_dict["Licensors"])
else:
aggregate_dict["licensors"] = None
# Info/Studios
if "Studios" in info_dict:
aggregate_dict["studios"] = parse_info_list(info_dict["Studios"])
else:
aggregate_dict["studios"] = studios_list
# Info/Source
aggregate_dict["source"] = info_dict.get("Source")
# Info/Genres
if "Genres" in info_dict:
aggregate_dict["genres"] = parse_info_list(info_dict["Genres"])
else:
aggregate_dict["genres"] = None
# Info/Duration
if "Duration" in info_dict:
duration_text = info_dict["Duration"]
mins = 0.0
# Hours
match = hours_regex.search(duration_text)
if match is not None:
mins += float(match.group(1)) * 60.0
# Minutes
match = min_regex.search(duration_text)
if match is not None:
mins += float(match.group(1))
# Seconds
match = sec_regex.search(duration_text)
if match is not None:
mins += float(match.group(1)) / 60.0
aggregate_dict["duration"] = mins
else:
aggregate_dict["duration"] = None
# Info/Rating
if "Rating" in info_dict:
rating_text = info_dict["Rating"]
rating_shorthand = rating_text.split(" - ")[0].strip()
aggregate_dict["rating"] = rating_shorthand
else:
aggregate_dict["rating"] = None
return aggregate_dict
#Fetches the production staff associated with a show
#Example: https://myanimelist.net/anime/30/Neon_Genesis_Evangelion/characters#staff
def getStaff(html):
|
#Fetches the characters and voice actors associted with a show
#Example: https://myanimelist.net/anime/30/Neon_Genesis_Evangelion/characters
def getCharactersAndJapaneseCast(html):
return
#Fetches the list of other related anime, same series etc.
def getRelatedTitles(html):
return
def cooldown():
COOLDOWN_IN_SECONDS = 0.5
time.sleep(COOLDOWN_IN_SECONDS)
def bs_preprocess(html):
"""remove distracting whitespaces and newline characters"""
pat = re.compile('(^[\s]+)|([\s]+$)', re.MULTILINE)
html = re.sub(pat, '', html) # remove leading and trailing whitespaces
html = re.sub('\n', ' ', html) # convert newlines to spaces
# this preserves newline delimiters
html = re.sub('[\s]+<', '<', html) # remove whitespaces before opening tags
html = re.sub('>[\s]+', '>', html) # remove whitespaces after closing tags
return html
# Returns dictionary with data
def scrape_main_page(html, aggregate_data={}):
getGeneralInformation(html, aggregate_data)
return aggregate_data
# Returns dictionary with data
def scrape_stats_page(html, aggregate_data={}):
getStatSummary(html, aggregate_data)
getStatDistribution(html, aggregate_data)
return aggregate_data
def getStatSummary(html, aggregate_dict={}):
#soup = BeautifulSoup(html, 'html.parser')
start_index = stats_summary_regex.search(html).end()
end_index = stats_score_regex.search(html).start()
subs = html[start_index:end_index]
stat_soup = BeautifulSoup(html[start_index:end_index], 'html.parser')
field_list = [t.get_text().strip() for t in stat_soup.find_all("div")]
for field in field_list[:6]:
t = "users_" + | return | identifier_body |
spider.js | 101043000,
101043100,
101043200,
101043300,
101043400,
101043600,
101050101,
101050102,
101050103,
101050104,
101050105,
101050106,
101050107,
101050108,
101050109,
101050110,
101050111,
101050112,
101050113,
101050201,
101050202,
101050203
];
// var dump = require("utils").dump;
// script argument
// casper.log("Casper CLI passed args:",'info');
// dump(casper.cli.args);
// filter the png & jpg, to speed up
casper.on('resource.requested', function (request) {
//if (/\.(png|jpg)$/i.test(request.url)) {
// 过滤广告链接
if (/(google|tanx|toruk|tq121|tongji|googlesyndication|taobao|taobaocdn|googlesyndication|doubleclick|baidu)/i.test(request.url)) {
//this.log("Abort resource.requested:" + request.url,'info');
request.abort();
} else {
//this.log("Resource.requested:" + request.url, 'info');
}
});
//casper.on('navigation.requested', function (url, navigationType, navigationLocked, isMainFrame) {
// //this.log("navigation.requested:" + url + " " + navigationType + " " + navigationLocked + " " + isMainFrame, 'info');
//});
//
//casper.on('page.resource.requested', function (requestData, request) {
// //this.log("page.resource.requested:" + requestData.url, 'info');
//
//});
//
//casper.on('load.finished', function (status) {
// //this.log("load.finished:" + status, 'info');
//
//});
casper.getHTML = function getHTML(selector, outer) {
"use strict";
this.checkStarted();
if (!selector) {
return this.page.frameContent;
}
return this.evaluate(function getSelectorHTML(selector, outer) {
var element = __utils__.findOne(selector);
return outer ? element.outerHTML : element.innerHTML;
}, selector, !!outer);
};
function send(host, result) {
try {
return __utils__.sendAJAX(host, 'POST', result, false,
{
contentType: 'application/x-www-form-urlencoded; charset=UTF-8'
});
} catch (e) {
__utils__.log("Server error:" + e, 'error');
}
}
function grab(cityId) {
var start = new Date().getTime();
var result = {};
this.waitForSelector("div.fl h1", (function () {
// ------------- Today weather --------------
result["cityId"] = cityId;
var city = this.getHTML('div.cityName.clearfix div.fl h2');
if (city === null) {
result["city"] = this.getHTML('div.cityName.clearfix div.fl h3');
} else {
result['city'] = city;
}
result["district"] = this.getHTML('div.cityName.clearfix div.fl h1');
result["temp"] = this.getHTML('p.tem span');
//风向
var wd = this.getElementAttr('span.sp2', 'title');
if (wd != null) {
result["wd"] = wd;
}
//风级
var ws = this.getHTML('span.sp2');
if (ws != null) {
result["ws"] = ws.replace(/[^0-9]/ig, "");
}
//湿度
var sd = this.getHTML('span.sp1');
if (sd != null) {
result["sd"] = sd.replace(/[^0-9]/ig, "");
}
var time = this.getHTML('#today span');
if (time != null) {
result["time"] = time.replace(/[^0-9:]/ig, "");
}
var now = new Date();
result["sysdate"] = now.getHours() + ":" + now.getMinutes();
// ------------- Today sunrise & sunset ------
var sunrise = this.getHTML("p.sunUp");
if (sunrise != null) {
result["sunrise"] = sunrise.replace(/[^0-9:]/ig, "");
}
var sunset = this.getHTML("p.sunDown");
if (sunset != null) {
result["sunset"] = sunset.replace(/[^0-9:]/ig, "");
}
// ------------- Today suggestion ------------
//穿衣
result["cyint"] = this.getHTML("section.mask section.ct b");
result["cydes"] = this.getHTML("section.mask section.ct aside").split("</b>")[1];
//感冒
result["gmint"] = this.getHTML("section.mask section.gm b");
result["gmdes"] = this.getHTML("section.mask section.gm aside").split("</b>")[1];
//紫外线
result["uvint"] = this.getHTML("section.mask section.uv b");
result["uvdes"] = this.getHTML("section.mask section.uv aside").split("</b>")[1];
//洗车
result["xcint"] = this.getHTML("section.mask section.xc b");
result["xcdes"] = this.getHTML("section.mask section.xc aside").split("</b>")[1];
//太阳镜
result["tyjint"] = this.getHTML("section.mask section.gl b");
result["tyjdes"] = this.getHTML("section.mask section.gl aside").split("</b>")[1];
//旅游
result["trint"] = this.getHTML("section.mask section.tr b");
result["trdes"] = this.getHTML("section.mask section.tr aside").split("</b>")[1];
//美容
result["mint"] = this.getHTML("section.mask section.gm b");
result["nydes"] = this.getHTML("section.mask section.gm aside").split("</b>")[1];
//晨练
result["clint"] = this.getHTML("section.mask section.cl b");
result["cldes"] = this.getHTML("section.mask section.cl aside").split("</b>")[1];
//过敏
result["agint"] = this.getHTML("section.mask section.ag b");
result["agdes"] = this.getHTML("section.mask section.ag aside").split("</b>")[1];
//雨伞
result["ysint"] = this.getHTML("section.mask section.ys b");
result["ysdes"] = this.getHTML("section.mask section.ys aside").split("</b>")[1];
//运动
result["ydint"] = this.getHTML("section.mask section.yd b");
result["iddes"] = this.getHTML("section.mask section.yd aside").split("</b>")[1];
//化妆
result["hzint"] = this.getHTML("section.mask section.pp b");
result["hzdes"] = this.getHTML("section.mask section.pp aside").split("</b>")[1];
//舒适度
result["ssdint"] = this.getHTML("section.mask section.co b");
result["ssdes"] = this.getHTML("section.mask section.co aside").split("</b>")[1];
this.log('Grab data in ' + (now.getTime() - start) + 'ms', 'info');
// ---------------- 发送给收集端 ---------------
//var response = this.evaluate(send, {host: host, result: result});
//this.log("Server response:" + response, 'debug');
//----------------- 调试输出 -----------
utils.dump(result);
}), (function () {
this.die(this.getCurrentUrl() + " timeout reached.");
}), 12000);
}
//手机版
//function grabFuture(cityId) {
// var result = {};
// result["cityId"] = cityId;
// this.evaluate(function (host, result) {
// var futures = __utils__.querySelectorAll("div.days7 li");
// for (var i = 0, j = futures.length; i < j; i++) {
// //天气图片标题,多云或者晴
// var img = futures[i].querySelectorAll("i img"); | // var before = img[0].getAttribute("alt");
// var after = img[1].getAttribute("alt");
// result["img_title" + (i + 1)] = before;
// result["img_title" + (i + 2)] = after;
// //天气转变,多云转晴 | random_line_split |
|
spider.js | 101031000,
101031100,
101031200,
101031400,
101040100,
101040200,
101040300,
101040400,
101040500,
101040600,
101040700,
101040800,
101040900,
101041000,
101041100,
101041300,
101041400,
101041500,
101041600,
101041700,
101041800,
101041900,
101042000,
101042100,
101042200,
101042300,
101042400,
101042500,
101042600,
101042700,
101042800,
101042900,
101043000,
101043100,
101043200,
101043300,
101043400,
101043600,
101050101,
101050102,
101050103,
101050104,
101050105,
101050106,
101050107,
101050108,
101050109,
101050110,
101050111,
101050112,
101050113,
101050201,
101050202,
101050203
];
// var dump = require("utils").dump;
// script argument
// casper.log("Casper CLI passed args:",'info');
// dump(casper.cli.args);
// filter the png & jpg, to speed up
casper.on('resource.requested', function (request) {
//if (/\.(png|jpg)$/i.test(request.url)) {
// 过滤广告链接
if (/(google|tanx|toruk|tq121|tongji|googlesyndication|taobao|taobaocdn|googlesyndication|doubleclick|baidu)/i.test(request.url)) {
//this.log("Abort resource.requested:" + request.url,'info');
request.abort();
} else {
//this.log("Resource.requested:" + request.url, 'info');
}
});
//casper.on('navigation.requested', function (url, navigationType, navigationLocked, isMainFrame) {
// //this.log("navigation.requested:" + url + " " + navigationType + " " + navigationLocked + " " + isMainFrame, 'info');
//});
//
//casper.on('page.resource.requested', function (requestData, request) {
// //this.log("page.resource.requested:" + requestData.url, 'info');
//
//});
//
//casper.on('load.finished', function (status) {
// //this.log("load.finished:" + status, 'info');
//
//});
casper.getHTML = function getHTML(selector, outer) {
"use strict";
this.checkStarted();
if (!selector) {
return this.page.frameContent;
}
return this.evaluate(function getSelectorHTML(selector, outer) {
var element = __utils__.findOne(selector);
return outer ? element.outerHTML : element.innerHTML;
}, selector, !!outer);
};
function send(host, result) {
try {
return _ | var start = new Date().getTime();
var result = {};
this.waitForSelector("div.fl h1", (function () {
// ------------- Today weather --------------
result["cityId"] = cityId;
var city = this.getHTML('div.cityName.clearfix div.fl h2');
if (city === null) {
result["city"] = this.getHTML('div.cityName.clearfix div.fl h3');
} else {
result['city'] = city;
}
result["district"] = this.getHTML('div.cityName.clearfix div.fl h1');
result["temp"] = this.getHTML('p.tem span');
//风向
var wd = this.getElementAttr('span.sp2', 'title');
if (wd != null) {
result["wd"] = wd;
}
//风级
var ws = this.getHTML('span.sp2');
if (ws != null) {
result["ws"] = ws.replace(/[^0-9]/ig, "");
}
//湿度
var sd = this.getHTML('span.sp1');
if (sd != null) {
result["sd"] = sd.replace(/[^0-9]/ig, "");
}
var time = this.getHTML('#today span');
if (time != null) {
result["time"] = time.replace(/[^0-9:]/ig, "");
}
var now = new Date();
result["sysdate"] = now.getHours() + ":" + now.getMinutes();
// ------------- Today sunrise & sunset ------
var sunrise = this.getHTML("p.sunUp");
if (sunrise != null) {
result["sunrise"] = sunrise.replace(/[^0-9:]/ig, "");
}
var sunset = this.getHTML("p.sunDown");
if (sunset != null) {
result["sunset"] = sunset.replace(/[^0-9:]/ig, "");
}
// ------------- Today suggestion ------------
//穿衣
result["cyint"] = this.getHTML("section.mask section.ct b");
result["cydes"] = this.getHTML("section.mask section.ct aside").split("</b>")[1];
//感冒
result["gmint"] = this.getHTML("section.mask section.gm b");
result["gmdes"] = this.getHTML("section.mask section.gm aside").split("</b>")[1];
//紫外线
result["uvint"] = this.getHTML("section.mask section.uv b");
result["uvdes"] = this.getHTML("section.mask section.uv aside").split("</b>")[1];
//洗车
result["xcint"] = this.getHTML("section.mask section.xc b");
result["xcdes"] = this.getHTML("section.mask section.xc aside").split("</b>")[1];
//太阳镜
result["tyjint"] = this.getHTML("section.mask section.gl b");
result["tyjdes"] = this.getHTML("section.mask section.gl aside").split("</b>")[1];
//旅游
result["trint"] = this.getHTML("section.mask section.tr b");
result["trdes"] = this.getHTML("section.mask section.tr aside").split("</b>")[1];
//美容
result["mint"] = this.getHTML("section.mask section.gm b");
result["nydes"] = this.getHTML("section.mask section.gm aside").split("</b>")[1];
//晨练
result["clint"] = this.getHTML("section.mask section.cl b");
result["cldes"] = this.getHTML("section.mask section.cl aside").split("</b>")[1];
//过敏
result["agint"] = this.getHTML("section.mask section.ag b");
result["agdes"] = this.getHTML("section.mask section.ag aside").split("</b>")[1];
//雨伞
result["ysint"] = this.getHTML("section.mask section.ys b");
result["ysdes"] = this.getHTML("section.mask section.ys aside").split("</b>")[1];
//运动
result["ydint"] = this.getHTML("section.mask section.yd b");
result["iddes"] = this.getHTML("section.mask | _utils__.sendAJAX(host, 'POST', result, false,
{
contentType: 'application/x-www-form-urlencoded; charset=UTF-8'
});
} catch (e) {
__utils__.log("Server error:" + e, 'error');
}
}
function grab(cityId) {
| identifier_body |
spider.js | 101031000,
101031100,
101031200,
101031400,
101040100,
101040200,
101040300,
101040400,
101040500,
101040600,
101040700,
101040800,
101040900,
101041000,
101041100,
101041300,
101041400,
101041500,
101041600,
101041700,
101041800,
101041900,
101042000,
101042100,
101042200,
101042300,
101042400,
101042500,
101042600,
101042700,
101042800,
101042900,
101043000,
101043100,
101043200,
101043300,
101043400,
101043600,
101050101,
101050102,
101050103,
101050104,
101050105,
101050106,
101050107,
101050108,
101050109,
101050110,
101050111,
101050112,
101050113,
101050201,
101050202,
101050203
];
// var dump = require("utils").dump;
// script argument
// casper.log("Casper CLI passed args:",'info');
// dump(casper.cli.args);
// filter the png & jpg, to speed up
casper.on('resource.requested', function (request) {
//if (/\.(png|jpg)$/i.test(request.url)) {
// 过滤广告链接
if (/(google|tanx|toruk|tq121|tongji|googlesyndication|taobao|taobaocdn|googlesyndication|doubleclick|baidu)/i.test(request.url)) {
//this.log("Abort | Resource.requested:" + request.url, 'info');
}
});
//casper.on('navigation.requested', function (url, navigationType, navigationLocked, isMainFrame) {
// //this.log("navigation.requested:" + url + " " + navigationType + " " + navigationLocked + " " + isMainFrame, 'info');
//});
//
//casper.on('page.resource.requested', function (requestData, request) {
// //this.log("page.resource.requested:" + requestData.url, 'info');
//
//});
//
//casper.on('load.finished', function (status) {
// //this.log("load.finished:" + status, 'info');
//
//});
casper.getHTML = function getHTML(selector, outer) {
"use strict";
this.checkStarted();
if (!selector) {
return this.page.frameContent;
}
return this.evaluate(function getSelectorHTML(selector, outer) {
var element = __utils__.findOne(selector);
return outer ? element.outerHTML : element.innerHTML;
}, selector, !!outer);
};
function send(host, result) {
try {
return __utils__.sendAJAX(host, 'POST', result, false,
{
contentType: 'application/x-www-form-urlencoded; charset=UTF-8'
});
} catch (e) {
__utils__.log("Server error:" + e, 'error');
}
}
function grab(cityId) {
var start = new Date().getTime();
var result = {};
this.waitForSelector("div.fl h1", (function () {
// ------------- Today weather --------------
result["cityId"] = cityId;
var city = this.getHTML('div.cityName.clearfix div.fl h2');
if (city === null) {
result["city"] = this.getHTML('div.cityName.clearfix div.fl h3');
} else {
result['city'] = city;
}
result["district"] = this.getHTML('div.cityName.clearfix div.fl h1');
result["temp"] = this.getHTML('p.tem span');
//风向
var wd = this.getElementAttr('span.sp2', 'title');
if (wd != null) {
result["wd"] = wd;
}
//风级
var ws = this.getHTML('span.sp2');
if (ws != null) {
result["ws"] = ws.replace(/[^0-9]/ig, "");
}
//湿度
var sd = this.getHTML('span.sp1');
if (sd != null) {
result["sd"] = sd.replace(/[^0-9]/ig, "");
}
var time = this.getHTML('#today span');
if (time != null) {
result["time"] = time.replace(/[^0-9:]/ig, "");
}
var now = new Date();
result["sysdate"] = now.getHours() + ":" + now.getMinutes();
// ------------- Today sunrise & sunset ------
var sunrise = this.getHTML("p.sunUp");
if (sunrise != null) {
result["sunrise"] = sunrise.replace(/[^0-9:]/ig, "");
}
var sunset = this.getHTML("p.sunDown");
if (sunset != null) {
result["sunset"] = sunset.replace(/[^0-9:]/ig, "");
}
// ------------- Today suggestion ------------
//穿衣
result["cyint"] = this.getHTML("section.mask section.ct b");
result["cydes"] = this.getHTML("section.mask section.ct aside").split("</b>")[1];
//感冒
result["gmint"] = this.getHTML("section.mask section.gm b");
result["gmdes"] = this.getHTML("section.mask section.gm aside").split("</b>")[1];
//紫外线
result["uvint"] = this.getHTML("section.mask section.uv b");
result["uvdes"] = this.getHTML("section.mask section.uv aside").split("</b>")[1];
//洗车
result["xcint"] = this.getHTML("section.mask section.xc b");
result["xcdes"] = this.getHTML("section.mask section.xc aside").split("</b>")[1];
//太阳镜
result["tyjint"] = this.getHTML("section.mask section.gl b");
result["tyjdes"] = this.getHTML("section.mask section.gl aside").split("</b>")[1];
//旅游
result["trint"] = this.getHTML("section.mask section.tr b");
result["trdes"] = this.getHTML("section.mask section.tr aside").split("</b>")[1];
//美容
result["mint"] = this.getHTML("section.mask section.gm b");
result["nydes"] = this.getHTML("section.mask section.gm aside").split("</b>")[1];
//晨练
result["clint"] = this.getHTML("section.mask section.cl b");
result["cldes"] = this.getHTML("section.mask section.cl aside").split("</b>")[1];
//过敏
result["agint"] = this.getHTML("section.mask section.ag b");
result["agdes"] = this.getHTML("section.mask section.ag aside").split("</b>")[1];
//雨伞
result["ysint"] = this.getHTML("section.mask section.ys b");
result["ysdes"] = this.getHTML("section.mask section.ys aside").split("</b>")[1];
//运动
result["ydint"] = this.getHTML("section.mask section.yd b");
result["iddes"] = this.getHTML("section.mask | resource.requested:" + request.url,'info');
request.abort();
} else {
//this.log(" | conditional_block |
spider.js | 01050203
];
// var dump = require("utils").dump;
// script argument
// casper.log("Casper CLI passed args:",'info');
// dump(casper.cli.args);
// filter the png & jpg, to speed up
casper.on('resource.requested', function (request) {
//if (/\.(png|jpg)$/i.test(request.url)) {
// 过滤广告链接
if (/(google|tanx|toruk|tq121|tongji|googlesyndication|taobao|taobaocdn|googlesyndication|doubleclick|baidu)/i.test(request.url)) {
//this.log("Abort resource.requested:" + request.url,'info');
request.abort();
} else {
//this.log("Resource.requested:" + request.url, 'info');
}
});
//casper.on('navigation.requested', function (url, navigationType, navigationLocked, isMainFrame) {
// //this.log("navigation.requested:" + url + " " + navigationType + " " + navigationLocked + " " + isMainFrame, 'info');
//});
//
//casper.on('page.resource.requested', function (requestData, request) {
// //this.log("page.resource.requested:" + requestData.url, 'info');
//
//});
//
//casper.on('load.finished', function (status) {
// //this.log("load.finished:" + status, 'info');
//
//});
casper.getHTML = function getHTML(selector, outer) {
"use strict";
this.checkStarted();
if (!selector) {
return this.page.frameContent;
}
return this.evaluate(function getSelectorHTML(selector, outer) {
var element = __utils__.findOne(selector);
return outer ? element.outerHTML : element.innerHTML;
}, selector, !!outer);
};
function send(host, result) {
try {
return __utils__.sendAJAX(host, 'POST', result, false,
{
contentType: 'application/x-www-form-urlencoded; charset=UTF-8'
});
} catch (e) {
__utils__.log("Server error:" + e, 'error');
}
}
function grab(cityId) {
var start = new Date().getTime();
var result = {};
this.waitForSelector("div.fl h1", (function () {
// ------------- Today weather --------------
result["cityId"] = cityId;
var city = this.getHTML('div.cityName.clearfix div.fl h2');
if (city === null) {
result["city"] = this.getHTML('div.cityName.clearfix div.fl h3');
} else {
result['city'] = city;
}
result["district"] = this.getHTML('div.cityName.clearfix div.fl h1');
result["temp"] = this.getHTML('p.tem span');
//风向
var wd = this.getElementAttr('span.sp2', 'title');
if (wd != null) {
result["wd"] = wd;
}
//风级
var ws = this.getHTML('span.sp2');
if (ws != null) {
result["ws"] = ws.replace(/[^0-9]/ig, "");
}
//湿度
var sd = this.getHTML('span.sp1');
if (sd != null) {
result["sd"] = sd.replace(/[^0-9]/ig, "");
}
var time = this.getHTML('#today span');
if (time != null) {
result["time"] = time.replace(/[^0-9:]/ig, "");
}
var now = new Date();
result["sysdate"] = now.getHours() + ":" + now.getMinutes();
// ------------- Today sunrise & sunset ------
var sunrise = this.getHTML("p.sunUp");
if (sunrise != null) {
result["sunrise"] = sunrise.replace(/[^0-9:]/ig, "");
}
var sunset = this.getHTML("p.sunDown");
if (sunset != null) {
result["sunset"] = sunset.replace(/[^0-9:]/ig, "");
}
// ------------- Today suggestion ------------
//穿衣
result["cyint"] = this.getHTML("section.mask section.ct b");
result["cydes"] = this.getHTML("section.mask section.ct aside").split("</b>")[1];
//感冒
result["gmint"] = this.getHTML("section.mask section.gm b");
result["gmdes"] = this.getHTML("section.mask section.gm aside").split("</b>")[1];
//紫外线
result["uvint"] = this.getHTML("section.mask section.uv b");
result["uvdes"] = this.getHTML("section.mask section.uv aside").split("</b>")[1];
//洗车
result["xcint"] = this.getHTML("section.mask section.xc b");
result["xcdes"] = this.getHTML("section.mask section.xc aside").split("</b>")[1];
//太阳镜
result["tyjint"] = this.getHTML("section.mask section.gl b");
result["tyjdes"] = this.getHTML("section.mask section.gl aside").split("</b>")[1];
//旅游
result["trint"] = this.getHTML("section.mask section.tr b");
result["trdes"] = this.getHTML("section.mask section.tr aside").split("</b>")[1];
//美容
result["mint"] = this.getHTML("section.mask section.gm b");
result["nydes"] = this.getHTML("section.mask section.gm aside").split("</b>")[1];
//晨练
result["clint"] = this.getHTML("section.mask section.cl b");
result["cldes"] = this.getHTML("section.mask section.cl aside").split("</b>")[1];
//过敏
result["agint"] = this.getHTML("section.mask section.ag b");
result["agdes"] = this.getHTML("section.mask section.ag aside").split("</b>")[1];
//雨伞
result["ysint"] = this.getHTML("section.mask section.ys b");
result["ysdes"] = this.getHTML("section.mask section.ys aside").split("</b>")[1];
//运动
result["ydint"] = this.getHTML("section.mask section.yd b");
result["iddes"] = this.getHTML("section.mask section.yd aside").split("</b>")[1];
//化妆
result["hzint"] = this.getHTML("section.mask section.pp b");
result["hzdes"] = this.getHTML("section.mask section.pp aside").split("</b>")[1];
//舒适度
result["ssdint"] = this.getHTML("section.mask section.co b");
result["ssdes"] = this.getHTML("section.mask section.co aside").split("</b>")[1];
this.log('Grab data in ' + (now.getTime() - start) + 'ms', 'info');
// ---------------- 发送给收集端 ---------------
//var response = this.evaluate(send, {host: host, result: result});
//this.log("Server response:" + response, 'debug');
//----------------- 调试输出 -----------
utils.dump(result);
}), (function () {
this.die(this.getCurrentUrl() + " timeout reached.");
}), 12000);
}
//手机版
//function grabFuture(cityId) {
// var result = {};
// result["cityId"] = cityId;
// this.evaluate(function (host, result) {
// var futures = __utils__.querySelectorAll("div.days7 li");
// for (var i = 0, j = futures.length; i < j; i++) {
// //天气图片标题,多云或者晴
// var img = futures[i].querySelectorAll("i img");
// var before = img[0].getAttribute("alt");
// var after = img[1].getAttribute("alt");
// result["img_title" + (i + 1)] = before;
// result["img_title" + (i + 2)] = after;
// //天气转变,多云转晴
// if (before === after) {
// result["weather" + (i + 1)] = before;
// } else {
// result["weather" + (i + 1)] = before + "转" + after;
// }
// //温度,5°/10°
// result["temp" + (i + 1)] = futures[i].querySelector("span").innerHTML;
// //天气图标号码
// var link1 = img[0].getAttribute("src");
// var link2 = img[1].getAttribute("src");
// result["img" + (i + 1)] = link1.charAt(link1.length - 5);
// result["img" + (i + 2)] = link2.charAt(link1.length - 5);
//
// }
//
// }, {host: host, result: result})
//}
//网页版
function grabFuture(cityId) {
var start = new Date().getTime();
var result = {};
result["cityId"] = cityId;
var city = this.getHTML('div.cityName.clearfix div.fl h2');
if (city === | null) {
| identifier_name |
|
real_time_plotting_new.py | generate the constrains of the picture manually)
'''
import datetime
import os
import numpy as np
import matplotlib.pyplot as plt
from matplotlib.lines import Line2D
import matplotlib.image as mpimg
import cartopy.crs as ccrs
from cartopy.mpl.gridliner import LONGITUDE_FORMATTER, LATITUDE_FORMATTER
# **** VARIABLES TO DEFINE BEFORE FLIGHT **************
location_of_base_image = r'C:/Users/kimdu/Documents/ph549/basemap.png'
home_lat = 52.4904018
home_lon = -105.719035
# coordinates of background image
left_lon = -107.7
right_lon = -103.7
bottom_lat = 50.7
top_lat = 54.3
approx_start_time = datetime.datetime(2019, 7, 17, 1, 1, 1)
approx_end_time = datetime.datetime(2019, 7, 17, 23, 1, 1)
# ****************************************************
plt.style.use('plotstyle.mplstyle')
def | (projection=ccrs.PlateCarree()):
"""
Code from https://ocefpaf.github.io/python4oceanographers/blog/2015/06/22/osm/
"""
fig, ax = plt.subplots(figsize=(9, 13),
subplot_kw=dict(projection=projection))
gl = ax.gridlines(draw_labels=True)
gl.xlabels_top = gl.ylabels_right = False
gl.xformatter = LONGITUDE_FORMATTER
gl.yformatter = LATITUDE_FORMATTER
return fig, ax
def save_map_image(loc):
"""
Code adapted from https://ocefpaf.github.io/python4oceanographers/blog/2015/06/22/osm/
Grab google maps image covering geographic area given by 'extent'
and save image to file for use as background map.
Use this function to generate the basemap image. It is also possible to plot data
directly on the cartopy object (map_ax) which provides higher resolution,
however that requires an internet connection onsite.
:param loc: path defining location to save image
:return: nothing
"""
import cartopy.io.img_tiles as cimgt
extent = [left_lon, right_lon, bottom_lat, top_lat]
request = cimgt.GoogleTiles()
map_fig, map_ax = make_map(projection=request.crs)
map_ax.set_extent(extent)
map_ax.add_image(request, 10)
# put star at launch site
map_ax.plot(home_lon, home_lat, marker='*', color='black', markersize=10,
transform=ccrs.Geodetic())
map_ax.savefig(loc)
def set_up_plots():
'''
Set the the axes of the desired plots
Written by Curtis Puetz 2018-07-07
Completely changed by Kimberlee Dube 2019-07-17
:return: None
'''
fig, axes = plt.subplots(2, 2, figsize=(20, 15), num=1,
sharex=False, sharey=False)
ax0 = axes[0, 0]
ax0.set_title('Altitude')
ax0.set_xlabel('Time')
ax0.set_ylabel('Altitude [m]')
ax1 = axes[0, 1]
ax1.set_title('Internal Temperature')
ax1.set_xlabel('Temperature [$\degree$C]')
ax1.set_ylabel('Altitude [m]')
ax2 = axes[1, 1]
ax2.set_title('External Temperature')
ax2.set_xlabel('Temperature [$\degree$C]')
ax2.set_ylabel('Altitude [m]')
ax3 = axes[1, 0]
ax3.set_title('Geiger Counters')
ax3.set_xlabel('Count/Time')
ax3.set_ylabel('Altitude [m]')
ax3.legend([Line2D([0], [0], color='red', lw=4),
Line2D([0], [0], color='blue', lw=4)], ['C1', 'C2'])
plt.tight_layout()
map_fig, map_ax = plt.subplots(figsize=(9, 13))
map_ax.set_title("Where my payload at?")
img = mpimg.imread(location_of_base_image)
imgplot = map_ax.imshow(img)
return axes, map_ax, img
def plot_data(data, header_data, axes, map_ax, img):
'''
Plot a single data point for each of the plots defined in 'set_up_plots()'
This will occur each time a comma separated data list is received
Written by Curtis Puetz 2018-07-07
Rewritten by Kimberlee Dube 2019-07-17
:param data: the list of data generated from the downlinked comma separated data list
:return: None
'''
pi_time = datetime.datetime.strptime(data[0], '%Y%m%d_%X.%f')
# isolate the floats and save them in a dictionary (while checking the units of altitude)
data = data[1:] # removes the time datetime value
header_data = header_data[1:] # removes the time datetime value
data_dict = dict(zip(header_data, data))
if data_dict['Altu'] == "KM":
alt_factor = 1000
else:
alt_factor = 1
data_dict['Alt'] *= alt_factor
del data_dict['Altu']
del data_dict['NS']
del data_dict['EW']
data_float = [[] for i in range(len(data_dict))]
for i, dai in enumerate(list(data_dict.values())):
if dai == "":
data_float[i] = ""
else:
data_float[i] = float(dai)
data_dict = dict(zip(list(data_dict.keys()), data_float))
# Change in altitude over time
if not data_dict['Alt'] == "":
axes[0, 0].scatter(pi_time, data_dict['Alt'], color='green')
# Need to manually set the approximate flight start
# and end times for the plot to look nice
axes[0, 0].set_xlim([approx_start_time, approx_end_time])
# Altitude profile of internal temperature
if not data_dict['TC'] == "" and not data_dict['Alt'] == "":
axes[0, 1].scatter(data_dict['TC'], data_dict['Alt'], color='green')
# Altitude profile of external temperature
if not data_dict['temp'] == "" and not data_dict['Alt'] == "":
axes[1, 1].scatter(data_dict['temp'], data_dict['Alt'], color='green')
# Altitude profiles of Geiger counter measurements
if not data_dict['C1'] == "":
axes[1, 0].scatter(data_dict['C1'], data_dict['Alt'], color='red', label='C1')
if not data_dict['C2'] == "":
axes[1, 0].scatter(data_dict['C2'], data_dict['Alt'], color='blue', label='C2')
# Map of geographic location
if not data_dict['LtDgMn'] == "" and not data_dict['LnDgMn'] == "":
lat = int(data_dict['LtDgMn']/100) + (data_dict['LtDgMn'] - int(data_dict['LtDgMn']/100)*100)/60
lon = -(int(data_dict['LnDgMn']/100) + (data_dict['LnDgMn'] - int(data_dict['LnDgMn']/100)*100)/60)
# change to sask coords for testing
#lat = 52 + (lat % 1)
#lon = -105 + (lon % 1)
index_y = np.interp(lat, np.linspace(bottom_lat, top_lat, len(img)), np.arange(0, len(img))[::-1])
index_x = np.interp(lon, np.linspace(left_lon, right_lon, len(img[0])), np.arange(0, len(img[0])))
# map_ax.plot(lon, lat, marker='o', color='red', markersize=5,
# transform=ccrs.Geodetic())
map_ax.scatter(index_x, index_y, marker='o', color='red')
plt.pause(0.05)
def read_last_line_in_data_log():
"""
This function will read the last line in the data log file and return it
Written by Daniel Letros, 2018-07-03
:return: None
"""
timestamp = datetime.datetime.utcnow().strftime("%Y%m%d")
log_file_path = r'C:/Users/kimdu/Documents/ph549/Telemetry_logs'
log_file_path += os.sep + timestamp
file_name = log_file_path + os.sep + timestamp + "_data.txt"
# file_name = r'C:/Users/kimdu/Documents/ph549/Telemetry_logs/test.txt' # test generated data
try:
with open(file_name, | make_map | identifier_name |
real_time_plotting_new.py | generate the constrains of the picture manually)
'''
import datetime
import os
import numpy as np
import matplotlib.pyplot as plt
from matplotlib.lines import Line2D
import matplotlib.image as mpimg
import cartopy.crs as ccrs
from cartopy.mpl.gridliner import LONGITUDE_FORMATTER, LATITUDE_FORMATTER
# **** VARIABLES TO DEFINE BEFORE FLIGHT **************
location_of_base_image = r'C:/Users/kimdu/Documents/ph549/basemap.png'
home_lat = 52.4904018
home_lon = -105.719035
# coordinates of background image
left_lon = -107.7
right_lon = -103.7
bottom_lat = 50.7
top_lat = 54.3
approx_start_time = datetime.datetime(2019, 7, 17, 1, 1, 1)
approx_end_time = datetime.datetime(2019, 7, 17, 23, 1, 1)
# ****************************************************
plt.style.use('plotstyle.mplstyle')
def make_map(projection=ccrs.PlateCarree()):
"""
Code from https://ocefpaf.github.io/python4oceanographers/blog/2015/06/22/osm/
""" | gl.xlabels_top = gl.ylabels_right = False
gl.xformatter = LONGITUDE_FORMATTER
gl.yformatter = LATITUDE_FORMATTER
return fig, ax
def save_map_image(loc):
"""
Code adapted from https://ocefpaf.github.io/python4oceanographers/blog/2015/06/22/osm/
Grab google maps image covering geographic area given by 'extent'
and save image to file for use as background map.
Use this function to generate the basemap image. It is also possible to plot data
directly on the cartopy object (map_ax) which provides higher resolution,
however that requires an internet connection onsite.
:param loc: path defining location to save image
:return: nothing
"""
import cartopy.io.img_tiles as cimgt
extent = [left_lon, right_lon, bottom_lat, top_lat]
request = cimgt.GoogleTiles()
map_fig, map_ax = make_map(projection=request.crs)
map_ax.set_extent(extent)
map_ax.add_image(request, 10)
# put star at launch site
map_ax.plot(home_lon, home_lat, marker='*', color='black', markersize=10,
transform=ccrs.Geodetic())
map_ax.savefig(loc)
def set_up_plots():
'''
Set the the axes of the desired plots
Written by Curtis Puetz 2018-07-07
Completely changed by Kimberlee Dube 2019-07-17
:return: None
'''
fig, axes = plt.subplots(2, 2, figsize=(20, 15), num=1,
sharex=False, sharey=False)
ax0 = axes[0, 0]
ax0.set_title('Altitude')
ax0.set_xlabel('Time')
ax0.set_ylabel('Altitude [m]')
ax1 = axes[0, 1]
ax1.set_title('Internal Temperature')
ax1.set_xlabel('Temperature [$\degree$C]')
ax1.set_ylabel('Altitude [m]')
ax2 = axes[1, 1]
ax2.set_title('External Temperature')
ax2.set_xlabel('Temperature [$\degree$C]')
ax2.set_ylabel('Altitude [m]')
ax3 = axes[1, 0]
ax3.set_title('Geiger Counters')
ax3.set_xlabel('Count/Time')
ax3.set_ylabel('Altitude [m]')
ax3.legend([Line2D([0], [0], color='red', lw=4),
Line2D([0], [0], color='blue', lw=4)], ['C1', 'C2'])
plt.tight_layout()
map_fig, map_ax = plt.subplots(figsize=(9, 13))
map_ax.set_title("Where my payload at?")
img = mpimg.imread(location_of_base_image)
imgplot = map_ax.imshow(img)
return axes, map_ax, img
def plot_data(data, header_data, axes, map_ax, img):
'''
Plot a single data point for each of the plots defined in 'set_up_plots()'
This will occur each time a comma separated data list is received
Written by Curtis Puetz 2018-07-07
Rewritten by Kimberlee Dube 2019-07-17
:param data: the list of data generated from the downlinked comma separated data list
:return: None
'''
pi_time = datetime.datetime.strptime(data[0], '%Y%m%d_%X.%f')
# isolate the floats and save them in a dictionary (while checking the units of altitude)
data = data[1:] # removes the time datetime value
header_data = header_data[1:] # removes the time datetime value
data_dict = dict(zip(header_data, data))
if data_dict['Altu'] == "KM":
alt_factor = 1000
else:
alt_factor = 1
data_dict['Alt'] *= alt_factor
del data_dict['Altu']
del data_dict['NS']
del data_dict['EW']
data_float = [[] for i in range(len(data_dict))]
for i, dai in enumerate(list(data_dict.values())):
if dai == "":
data_float[i] = ""
else:
data_float[i] = float(dai)
data_dict = dict(zip(list(data_dict.keys()), data_float))
# Change in altitude over time
if not data_dict['Alt'] == "":
axes[0, 0].scatter(pi_time, data_dict['Alt'], color='green')
# Need to manually set the approximate flight start
# and end times for the plot to look nice
axes[0, 0].set_xlim([approx_start_time, approx_end_time])
# Altitude profile of internal temperature
if not data_dict['TC'] == "" and not data_dict['Alt'] == "":
axes[0, 1].scatter(data_dict['TC'], data_dict['Alt'], color='green')
# Altitude profile of external temperature
if not data_dict['temp'] == "" and not data_dict['Alt'] == "":
axes[1, 1].scatter(data_dict['temp'], data_dict['Alt'], color='green')
# Altitude profiles of Geiger counter measurements
if not data_dict['C1'] == "":
axes[1, 0].scatter(data_dict['C1'], data_dict['Alt'], color='red', label='C1')
if not data_dict['C2'] == "":
axes[1, 0].scatter(data_dict['C2'], data_dict['Alt'], color='blue', label='C2')
# Map of geographic location
if not data_dict['LtDgMn'] == "" and not data_dict['LnDgMn'] == "":
lat = int(data_dict['LtDgMn']/100) + (data_dict['LtDgMn'] - int(data_dict['LtDgMn']/100)*100)/60
lon = -(int(data_dict['LnDgMn']/100) + (data_dict['LnDgMn'] - int(data_dict['LnDgMn']/100)*100)/60)
# change to sask coords for testing
#lat = 52 + (lat % 1)
#lon = -105 + (lon % 1)
index_y = np.interp(lat, np.linspace(bottom_lat, top_lat, len(img)), np.arange(0, len(img))[::-1])
index_x = np.interp(lon, np.linspace(left_lon, right_lon, len(img[0])), np.arange(0, len(img[0])))
# map_ax.plot(lon, lat, marker='o', color='red', markersize=5,
# transform=ccrs.Geodetic())
map_ax.scatter(index_x, index_y, marker='o', color='red')
plt.pause(0.05)
def read_last_line_in_data_log():
"""
This function will read the last line in the data log file and return it
Written by Daniel Letros, 2018-07-03
:return: None
"""
timestamp = datetime.datetime.utcnow().strftime("%Y%m%d")
log_file_path = r'C:/Users/kimdu/Documents/ph549/Telemetry_logs'
log_file_path += os.sep + timestamp
file_name = log_file_path + os.sep + timestamp + "_data.txt"
# file_name = r'C:/Users/kimdu/Documents/ph549/Telemetry_logs/test.txt' # test generated data
try:
with open(file_name, ' | fig, ax = plt.subplots(figsize=(9, 13),
subplot_kw=dict(projection=projection))
gl = ax.gridlines(draw_labels=True) | random_line_split |
real_time_plotting_new.py | '''
fig, axes = plt.subplots(2, 2, figsize=(20, 15), num=1,
sharex=False, sharey=False)
ax0 = axes[0, 0]
ax0.set_title('Altitude')
ax0.set_xlabel('Time')
ax0.set_ylabel('Altitude [m]')
ax1 = axes[0, 1]
ax1.set_title('Internal Temperature')
ax1.set_xlabel('Temperature [$\degree$C]')
ax1.set_ylabel('Altitude [m]')
ax2 = axes[1, 1]
ax2.set_title('External Temperature')
ax2.set_xlabel('Temperature [$\degree$C]')
ax2.set_ylabel('Altitude [m]')
ax3 = axes[1, 0]
ax3.set_title('Geiger Counters')
ax3.set_xlabel('Count/Time')
ax3.set_ylabel('Altitude [m]')
ax3.legend([Line2D([0], [0], color='red', lw=4),
Line2D([0], [0], color='blue', lw=4)], ['C1', 'C2'])
plt.tight_layout()
map_fig, map_ax = plt.subplots(figsize=(9, 13))
map_ax.set_title("Where my payload at?")
img = mpimg.imread(location_of_base_image)
imgplot = map_ax.imshow(img)
return axes, map_ax, img
def plot_data(data, header_data, axes, map_ax, img):
'''
Plot a single data point for each of the plots defined in 'set_up_plots()'
This will occur each time a comma separated data list is received
Written by Curtis Puetz 2018-07-07
Rewritten by Kimberlee Dube 2019-07-17
:param data: the list of data generated from the downlinked comma separated data list
:return: None
'''
pi_time = datetime.datetime.strptime(data[0], '%Y%m%d_%X.%f')
# isolate the floats and save them in a dictionary (while checking the units of altitude)
data = data[1:] # removes the time datetime value
header_data = header_data[1:] # removes the time datetime value
data_dict = dict(zip(header_data, data))
if data_dict['Altu'] == "KM":
alt_factor = 1000
else:
alt_factor = 1
data_dict['Alt'] *= alt_factor
del data_dict['Altu']
del data_dict['NS']
del data_dict['EW']
data_float = [[] for i in range(len(data_dict))]
for i, dai in enumerate(list(data_dict.values())):
if dai == "":
data_float[i] = ""
else:
data_float[i] = float(dai)
data_dict = dict(zip(list(data_dict.keys()), data_float))
# Change in altitude over time
if not data_dict['Alt'] == "":
axes[0, 0].scatter(pi_time, data_dict['Alt'], color='green')
# Need to manually set the approximate flight start
# and end times for the plot to look nice
axes[0, 0].set_xlim([approx_start_time, approx_end_time])
# Altitude profile of internal temperature
if not data_dict['TC'] == "" and not data_dict['Alt'] == "":
axes[0, 1].scatter(data_dict['TC'], data_dict['Alt'], color='green')
# Altitude profile of external temperature
if not data_dict['temp'] == "" and not data_dict['Alt'] == "":
axes[1, 1].scatter(data_dict['temp'], data_dict['Alt'], color='green')
# Altitude profiles of Geiger counter measurements
if not data_dict['C1'] == "":
axes[1, 0].scatter(data_dict['C1'], data_dict['Alt'], color='red', label='C1')
if not data_dict['C2'] == "":
axes[1, 0].scatter(data_dict['C2'], data_dict['Alt'], color='blue', label='C2')
# Map of geographic location
if not data_dict['LtDgMn'] == "" and not data_dict['LnDgMn'] == "":
lat = int(data_dict['LtDgMn']/100) + (data_dict['LtDgMn'] - int(data_dict['LtDgMn']/100)*100)/60
lon = -(int(data_dict['LnDgMn']/100) + (data_dict['LnDgMn'] - int(data_dict['LnDgMn']/100)*100)/60)
# change to sask coords for testing
#lat = 52 + (lat % 1)
#lon = -105 + (lon % 1)
index_y = np.interp(lat, np.linspace(bottom_lat, top_lat, len(img)), np.arange(0, len(img))[::-1])
index_x = np.interp(lon, np.linspace(left_lon, right_lon, len(img[0])), np.arange(0, len(img[0])))
# map_ax.plot(lon, lat, marker='o', color='red', markersize=5,
# transform=ccrs.Geodetic())
map_ax.scatter(index_x, index_y, marker='o', color='red')
plt.pause(0.05)
def read_last_line_in_data_log():
"""
This function will read the last line in the data log file and return it
Written by Daniel Letros, 2018-07-03
:return: None
"""
timestamp = datetime.datetime.utcnow().strftime("%Y%m%d")
log_file_path = r'C:/Users/kimdu/Documents/ph549/Telemetry_logs'
log_file_path += os.sep + timestamp
file_name = log_file_path + os.sep + timestamp + "_data.txt"
# file_name = r'C:/Users/kimdu/Documents/ph549/Telemetry_logs/test.txt' # test generated data
try:
with open(file_name, 'rb') as f:
f.seek(-2, os.SEEK_END)
while f.read(1) != b'\n':
f.seek(-2, os.SEEK_CUR)
content = f.readline().decode()
except:
with open(file_name, 'rb') as f:
content = f.readlines()[-1].decode()
return content
if __name__ == '__main__':
header_2018 = ['PiTS', 'ATSms', 'UTC', 'LtDgMn', 'NS', 'LnDgMn', 'EW',
'Nsat', 'Alt', 'Altu', 'Acxms2', 'Acyms2', 'Aczms2', 'Gyxrs',
'Gyyrs', 'Gyzrs', 'MgxuT', 'MgyuT', 'MgzuT', 'Elxdg', 'Elydg',
'Elzdg', 'LAcxms2', 'LAcyms2', 'LAczms2', 'Gvxms2', 'Gvyms2',
'Gvzms2', 'TC', 'SyCl03', 'GyCl03', 'AcCl03', 'MgCl03',
'', 'C1', 'C2', 'SC', 'RSSI']
header = ['PiTS', 'ATSms', 'UTC', 'LtDgMn', 'NS', 'LnDgMn', 'EW',
'Nsat', 'Alt', 'Altu', 'Acxms2', 'Acyms2', 'Aczms2', 'Gyxrs',
'Gyyrs', 'Gyzrs', 'MgxuT', 'MgyuT', 'MgzuT', 'Elxdg', 'Elydg',
'Elzdg', 'LAcxms2', 'LAcyms2', 'LAczms2', 'Gvxms2', 'Gvyms2',
'Gvzms2', 'TC', 'SyCl03', 'GyCl03', 'AcCl03', 'MgCl03',
'C1', 'C2', 'GN', 'BBL1', 'IRL1', 'BBL2', 'IRL2',
'BBL3', 'IRL3', 'temp']
plot_pause_for_interactive = 4
axes, map_ax, img = set_up_plots()
plt.ion()
hold = ""
while True:
data = read_last_line_in_data_log()
if data == hold:
plt.pause(plot_pause_for_interactive)
continue
hold = data
data = data[:-2] # remove newline character
print(data)
if data[0] == "P": # first character of header string
header_data = data.split(',')
elif data[0] == '2': # first character of a row of good data (starts with year)
| data = data.split(',')
plot_data(data, header, axes, map_ax, img) | conditional_block |
|
real_time_plotting_new.py | generate the constrains of the picture manually)
'''
import datetime
import os
import numpy as np
import matplotlib.pyplot as plt
from matplotlib.lines import Line2D
import matplotlib.image as mpimg
import cartopy.crs as ccrs
from cartopy.mpl.gridliner import LONGITUDE_FORMATTER, LATITUDE_FORMATTER
# **** VARIABLES TO DEFINE BEFORE FLIGHT **************
location_of_base_image = r'C:/Users/kimdu/Documents/ph549/basemap.png'
home_lat = 52.4904018
home_lon = -105.719035
# coordinates of background image
left_lon = -107.7
right_lon = -103.7
bottom_lat = 50.7
top_lat = 54.3
approx_start_time = datetime.datetime(2019, 7, 17, 1, 1, 1)
approx_end_time = datetime.datetime(2019, 7, 17, 23, 1, 1)
# ****************************************************
plt.style.use('plotstyle.mplstyle')
def make_map(projection=ccrs.PlateCarree()):
"""
Code from https://ocefpaf.github.io/python4oceanographers/blog/2015/06/22/osm/
"""
fig, ax = plt.subplots(figsize=(9, 13),
subplot_kw=dict(projection=projection))
gl = ax.gridlines(draw_labels=True)
gl.xlabels_top = gl.ylabels_right = False
gl.xformatter = LONGITUDE_FORMATTER
gl.yformatter = LATITUDE_FORMATTER
return fig, ax
def save_map_image(loc):
"""
Code adapted from https://ocefpaf.github.io/python4oceanographers/blog/2015/06/22/osm/
Grab google maps image covering geographic area given by 'extent'
and save image to file for use as background map.
Use this function to generate the basemap image. It is also possible to plot data
directly on the cartopy object (map_ax) which provides higher resolution,
however that requires an internet connection onsite.
:param loc: path defining location to save image
:return: nothing
"""
import cartopy.io.img_tiles as cimgt
extent = [left_lon, right_lon, bottom_lat, top_lat]
request = cimgt.GoogleTiles()
map_fig, map_ax = make_map(projection=request.crs)
map_ax.set_extent(extent)
map_ax.add_image(request, 10)
# put star at launch site
map_ax.plot(home_lon, home_lat, marker='*', color='black', markersize=10,
transform=ccrs.Geodetic())
map_ax.savefig(loc)
def set_up_plots():
| ax2.set_title('External Temperature')
ax2.set_xlabel('Temperature [$\degree$C]')
ax2.set_ylabel('Altitude [m]')
ax3 = axes[1, 0]
ax3.set_title('Geiger Counters')
ax3.set_xlabel('Count/Time')
ax3.set_ylabel('Altitude [m]')
ax3.legend([Line2D([0], [0], color='red', lw=4),
Line2D([0], [0], color='blue', lw=4)], ['C1', 'C2'])
plt.tight_layout()
map_fig, map_ax = plt.subplots(figsize=(9, 13))
map_ax.set_title("Where my payload at?")
img = mpimg.imread(location_of_base_image)
imgplot = map_ax.imshow(img)
return axes, map_ax, img
def plot_data(data, header_data, axes, map_ax, img):
'''
Plot a single data point for each of the plots defined in 'set_up_plots()'
This will occur each time a comma separated data list is received
Written by Curtis Puetz 2018-07-07
Rewritten by Kimberlee Dube 2019-07-17
:param data: the list of data generated from the downlinked comma separated data list
:return: None
'''
pi_time = datetime.datetime.strptime(data[0], '%Y%m%d_%X.%f')
# isolate the floats and save them in a dictionary (while checking the units of altitude)
data = data[1:] # removes the time datetime value
header_data = header_data[1:] # removes the time datetime value
data_dict = dict(zip(header_data, data))
if data_dict['Altu'] == "KM":
alt_factor = 1000
else:
alt_factor = 1
data_dict['Alt'] *= alt_factor
del data_dict['Altu']
del data_dict['NS']
del data_dict['EW']
data_float = [[] for i in range(len(data_dict))]
for i, dai in enumerate(list(data_dict.values())):
if dai == "":
data_float[i] = ""
else:
data_float[i] = float(dai)
data_dict = dict(zip(list(data_dict.keys()), data_float))
# Change in altitude over time
if not data_dict['Alt'] == "":
axes[0, 0].scatter(pi_time, data_dict['Alt'], color='green')
# Need to manually set the approximate flight start
# and end times for the plot to look nice
axes[0, 0].set_xlim([approx_start_time, approx_end_time])
# Altitude profile of internal temperature
if not data_dict['TC'] == "" and not data_dict['Alt'] == "":
axes[0, 1].scatter(data_dict['TC'], data_dict['Alt'], color='green')
# Altitude profile of external temperature
if not data_dict['temp'] == "" and not data_dict['Alt'] == "":
axes[1, 1].scatter(data_dict['temp'], data_dict['Alt'], color='green')
# Altitude profiles of Geiger counter measurements
if not data_dict['C1'] == "":
axes[1, 0].scatter(data_dict['C1'], data_dict['Alt'], color='red', label='C1')
if not data_dict['C2'] == "":
axes[1, 0].scatter(data_dict['C2'], data_dict['Alt'], color='blue', label='C2')
# Map of geographic location
if not data_dict['LtDgMn'] == "" and not data_dict['LnDgMn'] == "":
lat = int(data_dict['LtDgMn']/100) + (data_dict['LtDgMn'] - int(data_dict['LtDgMn']/100)*100)/60
lon = -(int(data_dict['LnDgMn']/100) + (data_dict['LnDgMn'] - int(data_dict['LnDgMn']/100)*100)/60)
# change to sask coords for testing
#lat = 52 + (lat % 1)
#lon = -105 + (lon % 1)
index_y = np.interp(lat, np.linspace(bottom_lat, top_lat, len(img)), np.arange(0, len(img))[::-1])
index_x = np.interp(lon, np.linspace(left_lon, right_lon, len(img[0])), np.arange(0, len(img[0])))
# map_ax.plot(lon, lat, marker='o', color='red', markersize=5,
# transform=ccrs.Geodetic())
map_ax.scatter(index_x, index_y, marker='o', color='red')
plt.pause(0.05)
def read_last_line_in_data_log():
"""
This function will read the last line in the data log file and return it
Written by Daniel Letros, 2018-07-03
:return: None
"""
timestamp = datetime.datetime.utcnow().strftime("%Y%m%d")
log_file_path = r'C:/Users/kimdu/Documents/ph549/Telemetry_logs'
log_file_path += os.sep + timestamp
file_name = log_file_path + os.sep + timestamp + "_data.txt"
# file_name = r'C:/Users/kimdu/Documents/ph549/Telemetry_logs/test.txt' # test generated data
try:
with open(file_name, ' | '''
Set the the axes of the desired plots
Written by Curtis Puetz 2018-07-07
Completely changed by Kimberlee Dube 2019-07-17
:return: None
'''
fig, axes = plt.subplots(2, 2, figsize=(20, 15), num=1,
sharex=False, sharey=False)
ax0 = axes[0, 0]
ax0.set_title('Altitude')
ax0.set_xlabel('Time')
ax0.set_ylabel('Altitude [m]')
ax1 = axes[0, 1]
ax1.set_title('Internal Temperature')
ax1.set_xlabel('Temperature [$\degree$C]')
ax1.set_ylabel('Altitude [m]')
ax2 = axes[1, 1] | identifier_body |
main.go | of the share.
DisplaynameOwner string `json:"displayname_owner" xml:"displayname_owner"`
// Additional info to identify the share owner, eg. the email or username
AdditionalInfoOwner string `json:"additional_info_owner" xml:"additional_info_owner"`
// The permission attribute set on the file.
// TODO(jfd) change the default to read only
Permissions Permissions `json:"permissions" xml:"permissions"`
// The UNIX timestamp when the share was created.
STime uint64 `json:"stime" xml:"stime"`
// ?
Parent string `json:"parent" xml:"parent"`
// The UNIX timestamp when the share expires.
Expiration string `json:"expiration" xml:"expiration"`
// The public link to the item being shared.
Token string `json:"token" xml:"token"`
// The unique id of the user that owns the file or folder being shared.
UIDFileOwner string `json:"uid_file_owner" xml:"uid_file_owner"`
// The display name of the user that owns the file or folder being shared.
DisplaynameFileOwner string `json:"displayname_file_owner" xml:"displayname_file_owner"`
// Additional info to identify the file owner, eg. the email or username
AdditionalInfoFileOwner string `json:"additional_info_file_owner" xml:"additional_info_file_owner"`
// share state, 0 = accepted, 1 = pending, 2 = declined
State int `json:"state" xml:"state"`
// The path to the shared file or folder.
Path string `json:"path" xml:"path"`
// The type of the object being shared. This can be one of 'file' or 'folder'.
ItemType string `json:"item_type" xml:"item_type"`
// The RFC2045-compliant mimetype of the file.
MimeType string `json:"mimetype" xml:"mimetype"`
// The space ID of the original file location
SpaceID string `json:"space_id" xml:"space_id"`
// The space alias of the original file location
SpaceAlias string `json:"space_alias" xml:"space_alias"`
StorageID string `json:"storage_id" xml:"storage_id"`
Storage uint64 `json:"storage" xml:"storage"`
// The unique node id of the item being shared.
ItemSource string `json:"item_source" xml:"item_source"`
// The unique node id of the item being shared. For legacy reasons item_source and file_source attributes have the same value.
FileSource string `json:"file_source" xml:"file_source"`
// The unique node id of the parent node of the item being shared.
FileParent string `json:"file_parent" xml:"file_parent"`
// The basename of the shared file.
FileTarget string `json:"file_target" xml:"file_target"`
// The uid of the share recipient. This is either
// - a GID (group id) if it is being shared with a group or
// - a UID (user id) if the share is shared with a user.
// - a password for public links
ShareWith string `json:"share_with,omitempty" xml:"share_with,omitempty"`
// The type of user
// - 0 = normal user
// - 1 = guest account
ShareWithUserType ShareWithUserType `json:"share_with_user_type" xml:"share_with_user_type"`
// The display name of the share recipient
ShareWithDisplayname string `json:"share_with_displayname,omitempty" xml:"share_with_displayname,omitempty"`
// Additional info to identify the share recipient, eg. the email or username
ShareWithAdditionalInfo string `json:"share_with_additional_info" xml:"share_with_additional_info"`
// Whether the recipient was notified, by mail, about the share being shared with them.
MailSend int `json:"mail_send" xml:"mail_send"`
// Name of the public share
Name string `json:"name" xml:"name"`
// URL of the public share
URL string `json:"url,omitempty" xml:"url,omitempty"`
// Attributes associated
Attributes string `json:"attributes,omitempty" xml:"attributes,omitempty"`
// Quicklink indicates if the link is the quicklink
Quicklink bool `json:"quicklink,omitempty" xml:"quicklink,omitempty"`
// PasswordProtected represents a public share is password protected
// PasswordProtected bool `json:"password_protected,omitempty" xml:"password_protected,omitempty"`
}
// ShareeData holds share recipient search results
type ShareeData struct {
Exact *ExactMatchesData `json:"exact" xml:"exact"`
Users []*MatchData `json:"users" xml:"users>element"`
Groups []*MatchData `json:"groups" xml:"groups>element"`
Remotes []*MatchData `json:"remotes" xml:"remotes>element"`
}
// TokenInfo holds token information
type TokenInfo struct {
// for all callers
Token string `json:"token" xml:"token"`
LinkURL string `json:"link_url" xml:"link_url"`
PasswordProtected bool `json:"password_protected" xml:"password_protected"`
Aliaslink bool `json:"alias_link" xml:"alias_link"`
// if not password protected
ID string `json:"id" xml:"id"`
StorageID string `json:"storage_id" xml:"storage_id"`
SpaceID string `json:"space_id" xml:"space_id"`
OpaqueID string `json:"opaque_id" xml:"opaque_id"`
Path string `json:"path" xml:"path"`
// if native access
SpacePath string `json:"space_path" xml:"space_path"`
SpaceAlias string `json:"space_alias" xml:"space_alias"`
SpaceURL string `json:"space_url" xml:"space_url"`
SpaceType string `json:"space_type" xml:"space_type"`
}
// ExactMatchesData hold exact matches
type ExactMatchesData struct {
Users []*MatchData `json:"users" xml:"users>element"`
Groups []*MatchData `json:"groups" xml:"groups>element"`
Remotes []*MatchData `json:"remotes" xml:"remotes>element"`
}
// MatchData describes a single match
type MatchData struct {
Label string `json:"label" xml:"label,omitempty"`
Value *MatchValueData `json:"value" xml:"value"`
}
// MatchValueData holds the type and actual value
type MatchValueData struct {
ShareType int `json:"shareType" xml:"shareType"`
ShareWith string `json:"shareWith" xml:"shareWith"`
ShareWithAdditionalInfo string `json:"shareWithAdditionalInfo" xml:"shareWithAdditionalInfo,omitempty"`
UserType int `json:"userType" xml:"userType"`
}
// CS3Share2ShareData converts a cs3api user share into shareData data model
func CS3Share2ShareData(ctx context.Context, share *collaboration.Share) (*ShareData, error) {
sd := &ShareData{
// share.permissions are mapped below
// Displaynames are added later
UIDOwner: LocalUserIDToString(share.GetCreator()),
UIDFileOwner: LocalUserIDToString(share.GetOwner()),
}
if share.Grantee.Type == provider.GranteeType_GRANTEE_TYPE_USER {
sd.ShareType = ShareTypeUser
sd.ShareWith = LocalUserIDToString(share.Grantee.GetUserId())
shareType := share.GetGrantee().GetUserId().GetType()
if shareType == userpb.UserType_USER_TYPE_LIGHTWEIGHT || shareType == userpb.UserType_USER_TYPE_GUEST {
sd.ShareWithUserType = ShareWithUserTypeGuest
} else {
sd.ShareWithUserType = ShareWithUserTypeUser
}
} else if share.Grantee.Type == provider.GranteeType_GRANTEE_TYPE_GROUP {
sd.ShareType = ShareTypeGroup
sd.ShareWith = LocalGroupIDToString(share.Grantee.GetGroupId())
}
if share.Id != nil {
sd.ID = share.Id.OpaqueId
}
if share.GetPermissions().GetPermissions() != nil {
sd.Permissions = RoleFromResourcePermissions(share.GetPermissions().GetPermissions(), false).OCSPermissions()
}
if share.Ctime != nil {
sd.STime = share.Ctime.Seconds // TODO CS3 api birth time = btime
}
if share.Expiration != nil {
expiration := time.Unix(int64(share.Expiration.Seconds), int64(share.Expiration.Nanos))
sd.Expiration = expiration.Format(_iso8601)
}
return sd, nil
}
// PublicShare2ShareData converts a cs3api public share into shareData data model
func PublicShare2ShareData(share *link.PublicShare, r *http.Request, publicURL string) *ShareData {
sd := &ShareData{
// share.permissions are mapped below
// Displaynames are added later
ShareType: ShareTypePublicLink,
Token: share.Token,
Name: share.DisplayName,
MailSend: 0,
URL: publicURL + path.Join("/", "s/"+share.Token),
UIDOwner: LocalUserIDToString(share.Creator),
UIDFileOwner: LocalUserIDToString(share.Owner),
Quicklink: share.Quicklink,
} | if share.Id != nil { | random_line_split |
|
main.go | UserTypeUser ShareWithUserType = 0
// ShareWithUserTypeGuest represents a guest user
ShareWithUserTypeGuest ShareWithUserType = 1
// The datetime format of ISO8601
_iso8601 = "2006-01-02T15:04:05Z0700"
)
// ResourceType indicates the OCS type of the resource
type ResourceType int
func (rt ResourceType) String() (s string) {
switch rt {
case 0:
s = "invalid"
case 1:
s = "file"
case 2:
s = "folder"
case 3:
s = "reference"
default:
s = "invalid"
}
return
}
// ShareType denotes a type of share
type ShareType int
// ShareWithUserType denotes a type of user
type ShareWithUserType int
// ShareData represents https://doc.owncloud.com/server/developer_manual/core/ocs-share-api.html#response-attributes-1
type ShareData struct {
// TODO int?
ID string `json:"id" xml:"id"`
// The share’s type
ShareType ShareType `json:"share_type" xml:"share_type"`
// The username of the owner of the share.
UIDOwner string `json:"uid_owner" xml:"uid_owner"`
// The display name of the owner of the share.
DisplaynameOwner string `json:"displayname_owner" xml:"displayname_owner"`
// Additional info to identify the share owner, eg. the email or username
AdditionalInfoOwner string `json:"additional_info_owner" xml:"additional_info_owner"`
// The permission attribute set on the file.
// TODO(jfd) change the default to read only
Permissions Permissions `json:"permissions" xml:"permissions"`
// The UNIX timestamp when the share was created.
STime uint64 `json:"stime" xml:"stime"`
// ?
Parent string `json:"parent" xml:"parent"`
// The UNIX timestamp when the share expires.
Expiration string `json:"expiration" xml:"expiration"`
// The public link to the item being shared.
Token string `json:"token" xml:"token"`
// The unique id of the user that owns the file or folder being shared.
UIDFileOwner string `json:"uid_file_owner" xml:"uid_file_owner"`
// The display name of the user that owns the file or folder being shared.
DisplaynameFileOwner string `json:"displayname_file_owner" xml:"displayname_file_owner"`
// Additional info to identify the file owner, eg. the email or username
AdditionalInfoFileOwner string `json:"additional_info_file_owner" xml:"additional_info_file_owner"`
// share state, 0 = accepted, 1 = pending, 2 = declined
State int `json:"state" xml:"state"`
// The path to the shared file or folder.
Path string `json:"path" xml:"path"`
// The type of the object being shared. This can be one of 'file' or 'folder'.
ItemType string `json:"item_type" xml:"item_type"`
// The RFC2045-compliant mimetype of the file.
MimeType string `json:"mimetype" xml:"mimetype"`
// The space ID of the original file location
SpaceID string `json:"space_id" xml:"space_id"`
// The space alias of the original file location
SpaceAlias string `json:"space_alias" xml:"space_alias"`
StorageID string `json:"storage_id" xml:"storage_id"`
Storage uint64 `json:"storage" xml:"storage"`
// The unique node id of the item being shared.
ItemSource string `json:"item_source" xml:"item_source"`
// The unique node id of the item being shared. For legacy reasons item_source and file_source attributes have the same value.
FileSource string `json:"file_source" xml:"file_source"`
// The unique node id of the parent node of the item being shared.
FileParent string `json:"file_parent" xml:"file_parent"`
// The basename of the shared file.
FileTarget string `json:"file_target" xml:"file_target"`
// The uid of the share recipient. This is either
// - a GID (group id) if it is being shared with a group or
// - a UID (user id) if the share is shared with a user.
// - a password for public links
ShareWith string `json:"share_with,omitempty" xml:"share_with,omitempty"`
// The type of user
// - 0 = normal user
// - 1 = guest account
ShareWithUserType ShareWithUserType `json:"share_with_user_type" xml:"share_with_user_type"`
// The display name of the share recipient
ShareWithDisplayname string `json:"share_with_displayname,omitempty" xml:"share_with_displayname,omitempty"`
// Additional info to identify the share recipient, eg. the email or username
ShareWithAdditionalInfo string `json:"share_with_additional_info" xml:"share_with_additional_info"`
// Whether the recipient was notified, by mail, about the share being shared with them.
MailSend int `json:"mail_send" xml:"mail_send"`
// Name of the public share
Name string `json:"name" xml:"name"`
// URL of the public share
URL string `json:"url,omitempty" xml:"url,omitempty"`
// Attributes associated
Attributes string `json:"attributes,omitempty" xml:"attributes,omitempty"`
// Quicklink indicates if the link is the quicklink
Quicklink bool `json:"quicklink,omitempty" xml:"quicklink,omitempty"`
// PasswordProtected represents a public share is password protected
// PasswordProtected bool `json:"password_protected,omitempty" xml:"password_protected,omitempty"`
}
// ShareeData holds share recipient search results
type ShareeData struct {
Exact *ExactMatchesData `json:"exact" xml:"exact"`
Users []*MatchData `json:"users" xml:"users>element"`
Groups []*MatchData `json:"groups" xml:"groups>element"`
Remotes []*MatchData `json:"remotes" xml:"remotes>element"`
}
// TokenInfo holds token information
type TokenInfo struct {
// for all callers
Token string `json:"token" xml:"token"`
LinkURL string `json:"link_url" xml:"link_url"`
PasswordProtected bool `json:"password_protected" xml:"password_protected"`
Aliaslink bool `json:"alias_link" xml:"alias_link"`
// if not password protected
ID string `json:"id" xml:"id"`
StorageID string `json:"storage_id" xml:"storage_id"`
SpaceID string `json:"space_id" xml:"space_id"`
OpaqueID string `json:"opaque_id" xml:"opaque_id"`
Path string `json:"path" xml:"path"`
// if native access
SpacePath string `json:"space_path" xml:"space_path"`
SpaceAlias string `json:"space_alias" xml:"space_alias"`
SpaceURL string `json:"space_url" xml:"space_url"`
SpaceType string `json:"space_type" xml:"space_type"`
}
// ExactMatchesData hold exact matches
type ExactMatchesData struct {
Users []*MatchData `json:"users" xml:"users>element"`
Groups []*MatchData `json:"groups" xml:"groups>element"`
Remotes []*MatchData `json:"remotes" xml:"remotes>element"`
}
// MatchData describes a single match
type MatchData struct {
Label string `json:"label" xml:"label,omitempty"`
Value *MatchValueData `json:"value" xml:"value"`
}
// MatchValueData holds the type and actual value
type MatchValueData struct {
ShareType int `json:"shareType" xml:"shareType"`
ShareWith string `json:"shareWith" xml:"shareWith"`
ShareWithAdditionalInfo string `json:"shareWithAdditionalInfo" xml:"shareWithAdditionalInfo,omitempty"`
UserType int `json:"userType" xml:"userType"`
}
// CS3Share2ShareData converts a cs3api user share into shareData data model
func CS3Share2ShareData(ctx context.Context, share *collaboration.Share) (*ShareData, error) {
| sd := &ShareData{
// share.permissions are mapped below
// Displaynames are added later
UIDOwner: LocalUserIDToString(share.GetCreator()),
UIDFileOwner: LocalUserIDToString(share.GetOwner()),
}
if share.Grantee.Type == provider.GranteeType_GRANTEE_TYPE_USER {
sd.ShareType = ShareTypeUser
sd.ShareWith = LocalUserIDToString(share.Grantee.GetUserId())
shareType := share.GetGrantee().GetUserId().GetType()
if shareType == userpb.UserType_USER_TYPE_LIGHTWEIGHT || shareType == userpb.UserType_USER_TYPE_GUEST {
sd.ShareWithUserType = ShareWithUserTypeGuest
} else {
sd.ShareWithUserType = ShareWithUserTypeUser
}
} else if share.Grantee.Type == provider.GranteeType_GRANTEE_TYPE_GROUP {
sd.ShareType = ShareTypeGroup
sd.ShareWith = LocalGroupIDToString(share.Grantee.GetGroupId())
} | identifier_body |
|
main.go | = declined
State int `json:"state" xml:"state"`
// The path to the shared file or folder.
Path string `json:"path" xml:"path"`
// The type of the object being shared. This can be one of 'file' or 'folder'.
ItemType string `json:"item_type" xml:"item_type"`
// The RFC2045-compliant mimetype of the file.
MimeType string `json:"mimetype" xml:"mimetype"`
// The space ID of the original file location
SpaceID string `json:"space_id" xml:"space_id"`
// The space alias of the original file location
SpaceAlias string `json:"space_alias" xml:"space_alias"`
StorageID string `json:"storage_id" xml:"storage_id"`
Storage uint64 `json:"storage" xml:"storage"`
// The unique node id of the item being shared.
ItemSource string `json:"item_source" xml:"item_source"`
// The unique node id of the item being shared. For legacy reasons item_source and file_source attributes have the same value.
FileSource string `json:"file_source" xml:"file_source"`
// The unique node id of the parent node of the item being shared.
FileParent string `json:"file_parent" xml:"file_parent"`
// The basename of the shared file.
FileTarget string `json:"file_target" xml:"file_target"`
// The uid of the share recipient. This is either
// - a GID (group id) if it is being shared with a group or
// - a UID (user id) if the share is shared with a user.
// - a password for public links
ShareWith string `json:"share_with,omitempty" xml:"share_with,omitempty"`
// The type of user
// - 0 = normal user
// - 1 = guest account
ShareWithUserType ShareWithUserType `json:"share_with_user_type" xml:"share_with_user_type"`
// The display name of the share recipient
ShareWithDisplayname string `json:"share_with_displayname,omitempty" xml:"share_with_displayname,omitempty"`
// Additional info to identify the share recipient, eg. the email or username
ShareWithAdditionalInfo string `json:"share_with_additional_info" xml:"share_with_additional_info"`
// Whether the recipient was notified, by mail, about the share being shared with them.
MailSend int `json:"mail_send" xml:"mail_send"`
// Name of the public share
Name string `json:"name" xml:"name"`
// URL of the public share
URL string `json:"url,omitempty" xml:"url,omitempty"`
// Attributes associated
Attributes string `json:"attributes,omitempty" xml:"attributes,omitempty"`
// Quicklink indicates if the link is the quicklink
Quicklink bool `json:"quicklink,omitempty" xml:"quicklink,omitempty"`
// PasswordProtected represents a public share is password protected
// PasswordProtected bool `json:"password_protected,omitempty" xml:"password_protected,omitempty"`
}
// ShareeData holds share recipient search results
type ShareeData struct {
Exact *ExactMatchesData `json:"exact" xml:"exact"`
Users []*MatchData `json:"users" xml:"users>element"`
Groups []*MatchData `json:"groups" xml:"groups>element"`
Remotes []*MatchData `json:"remotes" xml:"remotes>element"`
}
// TokenInfo holds token information
type TokenInfo struct {
// for all callers
Token string `json:"token" xml:"token"`
LinkURL string `json:"link_url" xml:"link_url"`
PasswordProtected bool `json:"password_protected" xml:"password_protected"`
Aliaslink bool `json:"alias_link" xml:"alias_link"`
// if not password protected
ID string `json:"id" xml:"id"`
StorageID string `json:"storage_id" xml:"storage_id"`
SpaceID string `json:"space_id" xml:"space_id"`
OpaqueID string `json:"opaque_id" xml:"opaque_id"`
Path string `json:"path" xml:"path"`
// if native access
SpacePath string `json:"space_path" xml:"space_path"`
SpaceAlias string `json:"space_alias" xml:"space_alias"`
SpaceURL string `json:"space_url" xml:"space_url"`
SpaceType string `json:"space_type" xml:"space_type"`
}
// ExactMatchesData hold exact matches
type ExactMatchesData struct {
Users []*MatchData `json:"users" xml:"users>element"`
Groups []*MatchData `json:"groups" xml:"groups>element"`
Remotes []*MatchData `json:"remotes" xml:"remotes>element"`
}
// MatchData describes a single match
type MatchData struct {
Label string `json:"label" xml:"label,omitempty"`
Value *MatchValueData `json:"value" xml:"value"`
}
// MatchValueData holds the type and actual value
type MatchValueData struct {
ShareType int `json:"shareType" xml:"shareType"`
ShareWith string `json:"shareWith" xml:"shareWith"`
ShareWithAdditionalInfo string `json:"shareWithAdditionalInfo" xml:"shareWithAdditionalInfo,omitempty"`
UserType int `json:"userType" xml:"userType"`
}
// CS3Share2ShareData converts a cs3api user share into shareData data model
func CS3Share2ShareData(ctx context.Context, share *collaboration.Share) (*ShareData, error) {
sd := &ShareData{
// share.permissions are mapped below
// Displaynames are added later
UIDOwner: LocalUserIDToString(share.GetCreator()),
UIDFileOwner: LocalUserIDToString(share.GetOwner()),
}
if share.Grantee.Type == provider.GranteeType_GRANTEE_TYPE_USER {
sd.ShareType = ShareTypeUser
sd.ShareWith = LocalUserIDToString(share.Grantee.GetUserId())
shareType := share.GetGrantee().GetUserId().GetType()
if shareType == userpb.UserType_USER_TYPE_LIGHTWEIGHT || shareType == userpb.UserType_USER_TYPE_GUEST {
sd.ShareWithUserType = ShareWithUserTypeGuest
} else {
sd.ShareWithUserType = ShareWithUserTypeUser
}
} else if share.Grantee.Type == provider.GranteeType_GRANTEE_TYPE_GROUP {
sd.ShareType = ShareTypeGroup
sd.ShareWith = LocalGroupIDToString(share.Grantee.GetGroupId())
}
if share.Id != nil {
sd.ID = share.Id.OpaqueId
}
if share.GetPermissions().GetPermissions() != nil {
sd.Permissions = RoleFromResourcePermissions(share.GetPermissions().GetPermissions(), false).OCSPermissions()
}
if share.Ctime != nil {
sd.STime = share.Ctime.Seconds // TODO CS3 api birth time = btime
}
if share.Expiration != nil {
expiration := time.Unix(int64(share.Expiration.Seconds), int64(share.Expiration.Nanos))
sd.Expiration = expiration.Format(_iso8601)
}
return sd, nil
}
// PublicShare2ShareData converts a cs3api public share into shareData data model
func PublicShare2ShareData(share *link.PublicShare, r *http.Request, publicURL string) *ShareData {
sd := &ShareData{
// share.permissions are mapped below
// Displaynames are added later
ShareType: ShareTypePublicLink,
Token: share.Token,
Name: share.DisplayName,
MailSend: 0,
URL: publicURL + path.Join("/", "s/"+share.Token),
UIDOwner: LocalUserIDToString(share.Creator),
UIDFileOwner: LocalUserIDToString(share.Owner),
Quicklink: share.Quicklink,
}
if share.Id != nil {
sd.ID = share.Id.OpaqueId
}
if s := share.GetPermissions().GetPermissions(); s != nil {
sd.Permissions = RoleFromResourcePermissions(share.GetPermissions().GetPermissions(), true).OCSPermissions()
}
if share.Expiration != nil {
sd.Expiration = timestampToExpiration(share.Expiration)
}
if share.Ctime != nil {
sd.STime = share.Ctime.Seconds // TODO CS3 api birth time = btime
}
// hide password
if share.PasswordProtected {
sd.ShareWith = "***redacted***"
sd.ShareWithDisplayname = "***redacted***"
}
return sd
}
// LocalUserIDToString transforms a cs3api user id into an ocs data model without domain name
// TODO ocs uses user names ... so an additional lookup is needed. see mapUserIds()
func LocalUserIDToString(userID *userpb.UserId) string {
if userID == nil || userID.OpaqueId == "" {
return ""
}
return userID.OpaqueId
}
// LocalGroupIDToString transforms a cs3api group id into an ocs data model without domain name
func LocalGroupIDToString(groupID *grouppb.GroupId) string {
if groupID == nil || groupID.OpaqueId == "" {
return ""
}
return groupID.OpaqueId
}
// GetUserManager returns a connection to a user share manager
func Ge | tUserManager(m | identifier_name |
|
main.go | share expires.
Expiration string `json:"expiration" xml:"expiration"`
// The public link to the item being shared.
Token string `json:"token" xml:"token"`
// The unique id of the user that owns the file or folder being shared.
UIDFileOwner string `json:"uid_file_owner" xml:"uid_file_owner"`
// The display name of the user that owns the file or folder being shared.
DisplaynameFileOwner string `json:"displayname_file_owner" xml:"displayname_file_owner"`
// Additional info to identify the file owner, eg. the email or username
AdditionalInfoFileOwner string `json:"additional_info_file_owner" xml:"additional_info_file_owner"`
// share state, 0 = accepted, 1 = pending, 2 = declined
State int `json:"state" xml:"state"`
// The path to the shared file or folder.
Path string `json:"path" xml:"path"`
// The type of the object being shared. This can be one of 'file' or 'folder'.
ItemType string `json:"item_type" xml:"item_type"`
// The RFC2045-compliant mimetype of the file.
MimeType string `json:"mimetype" xml:"mimetype"`
// The space ID of the original file location
SpaceID string `json:"space_id" xml:"space_id"`
// The space alias of the original file location
SpaceAlias string `json:"space_alias" xml:"space_alias"`
StorageID string `json:"storage_id" xml:"storage_id"`
Storage uint64 `json:"storage" xml:"storage"`
// The unique node id of the item being shared.
ItemSource string `json:"item_source" xml:"item_source"`
// The unique node id of the item being shared. For legacy reasons item_source and file_source attributes have the same value.
FileSource string `json:"file_source" xml:"file_source"`
// The unique node id of the parent node of the item being shared.
FileParent string `json:"file_parent" xml:"file_parent"`
// The basename of the shared file.
FileTarget string `json:"file_target" xml:"file_target"`
// The uid of the share recipient. This is either
// - a GID (group id) if it is being shared with a group or
// - a UID (user id) if the share is shared with a user.
// - a password for public links
ShareWith string `json:"share_with,omitempty" xml:"share_with,omitempty"`
// The type of user
// - 0 = normal user
// - 1 = guest account
ShareWithUserType ShareWithUserType `json:"share_with_user_type" xml:"share_with_user_type"`
// The display name of the share recipient
ShareWithDisplayname string `json:"share_with_displayname,omitempty" xml:"share_with_displayname,omitempty"`
// Additional info to identify the share recipient, eg. the email or username
ShareWithAdditionalInfo string `json:"share_with_additional_info" xml:"share_with_additional_info"`
// Whether the recipient was notified, by mail, about the share being shared with them.
MailSend int `json:"mail_send" xml:"mail_send"`
// Name of the public share
Name string `json:"name" xml:"name"`
// URL of the public share
URL string `json:"url,omitempty" xml:"url,omitempty"`
// Attributes associated
Attributes string `json:"attributes,omitempty" xml:"attributes,omitempty"`
// Quicklink indicates if the link is the quicklink
Quicklink bool `json:"quicklink,omitempty" xml:"quicklink,omitempty"`
// PasswordProtected represents a public share is password protected
// PasswordProtected bool `json:"password_protected,omitempty" xml:"password_protected,omitempty"`
}
// ShareeData holds share recipient search results
type ShareeData struct {
Exact *ExactMatchesData `json:"exact" xml:"exact"`
Users []*MatchData `json:"users" xml:"users>element"`
Groups []*MatchData `json:"groups" xml:"groups>element"`
Remotes []*MatchData `json:"remotes" xml:"remotes>element"`
}
// TokenInfo holds token information
type TokenInfo struct {
// for all callers
Token string `json:"token" xml:"token"`
LinkURL string `json:"link_url" xml:"link_url"`
PasswordProtected bool `json:"password_protected" xml:"password_protected"`
Aliaslink bool `json:"alias_link" xml:"alias_link"`
// if not password protected
ID string `json:"id" xml:"id"`
StorageID string `json:"storage_id" xml:"storage_id"`
SpaceID string `json:"space_id" xml:"space_id"`
OpaqueID string `json:"opaque_id" xml:"opaque_id"`
Path string `json:"path" xml:"path"`
// if native access
SpacePath string `json:"space_path" xml:"space_path"`
SpaceAlias string `json:"space_alias" xml:"space_alias"`
SpaceURL string `json:"space_url" xml:"space_url"`
SpaceType string `json:"space_type" xml:"space_type"`
}
// ExactMatchesData hold exact matches
type ExactMatchesData struct {
Users []*MatchData `json:"users" xml:"users>element"`
Groups []*MatchData `json:"groups" xml:"groups>element"`
Remotes []*MatchData `json:"remotes" xml:"remotes>element"`
}
// MatchData describes a single match
type MatchData struct {
Label string `json:"label" xml:"label,omitempty"`
Value *MatchValueData `json:"value" xml:"value"`
}
// MatchValueData holds the type and actual value
type MatchValueData struct {
ShareType int `json:"shareType" xml:"shareType"`
ShareWith string `json:"shareWith" xml:"shareWith"`
ShareWithAdditionalInfo string `json:"shareWithAdditionalInfo" xml:"shareWithAdditionalInfo,omitempty"`
UserType int `json:"userType" xml:"userType"`
}
// CS3Share2ShareData converts a cs3api user share into shareData data model
func CS3Share2ShareData(ctx context.Context, share *collaboration.Share) (*ShareData, error) {
sd := &ShareData{
// share.permissions are mapped below
// Displaynames are added later
UIDOwner: LocalUserIDToString(share.GetCreator()),
UIDFileOwner: LocalUserIDToString(share.GetOwner()),
}
if share.Grantee.Type == provider.GranteeType_GRANTEE_TYPE_USER {
sd.ShareType = ShareTypeUser
sd.ShareWith = LocalUserIDToString(share.Grantee.GetUserId())
shareType := share.GetGrantee().GetUserId().GetType()
if shareType == userpb.UserType_USER_TYPE_LIGHTWEIGHT || shareType == userpb.UserType_USER_TYPE_GUEST {
sd.ShareWithUserType = ShareWithUserTypeGuest
} else {
sd.ShareWithUserType = ShareWithUserTypeUser
}
} else if share.Grantee.Type == provider.GranteeType_GRANTEE_TYPE_GROUP {
sd.ShareType = ShareTypeGroup
sd.ShareWith = LocalGroupIDToString(share.Grantee.GetGroupId())
}
if share.Id != nil {
sd.ID = share.Id.OpaqueId
}
if share.GetPermissions().GetPermissions() != nil {
sd.Permissions = RoleFromResourcePermissions(share.GetPermissions().GetPermissions(), false).OCSPermissions()
}
if share.Ctime != nil {
sd.STime = share.Ctime.Seconds // TODO CS3 api birth time = btime
}
if share.Expiration != nil {
expiration := time.Unix(int64(share.Expiration.Seconds), int64(share.Expiration.Nanos))
sd.Expiration = expiration.Format(_iso8601)
}
return sd, nil
}
// PublicShare2ShareData converts a cs3api public share into shareData data model
func PublicShare2ShareData(share *link.PublicShare, r *http.Request, publicURL string) *ShareData {
sd := &ShareData{
// share.permissions are mapped below
// Displaynames are added later
ShareType: ShareTypePublicLink,
Token: share.Token,
Name: share.DisplayName,
MailSend: 0,
URL: publicURL + path.Join("/", "s/"+share.Token),
UIDOwner: LocalUserIDToString(share.Creator),
UIDFileOwner: LocalUserIDToString(share.Owner),
Quicklink: share.Quicklink,
}
if share.Id != nil {
sd.ID = share.Id.OpaqueId
}
if s := share.GetPermissions().GetPermissions(); s != nil {
sd.Permissions = RoleFromResourcePermissions(share.GetPermissions().GetPermissions(), true).OCSPermissions()
}
if share.Expiration != nil {
sd.Expiration = timestampToExpiration(share.Expiration)
}
if share.Ctime != nil {
sd.STime = share.Ctime.Seconds // TODO CS3 api birth time = btime
}
// hide password
if share.PasswordProtected {
| sd.ShareWith = "***redacted***"
sd.ShareWithDisplayname = "***redacted***"
}
| conditional_block |
|
main.rs | }
total_welfare
}
// assumes that `guest_list` has already been sorted
// just doles out memory in sorted order such that we
// don't give anyone any more memory than they asked for
// it's still possible that we give people too little memory
// which should be checked after getting the result of this
// function
// I think what I really want to return from this is a vector of allocation amounts
// it'll even take up less space than (reference, allocation) pairs and will be much
// less problematic
fn naive_allocation
(
guest_list: &Vec<&Guest>,
available_memory: u64
)
-> Vec<(u64)>
{
let mut remaining_memory: u64 = available_memory;
guest_list.iter().map
(
|guest|
{
// if there's no memory left to hand out our job is simple
if remaining_memory == 0
{
0
}
// otherwise get the maximum amount memory this guest
// wants to pay for and give them that
else
{
// if the last forbidden range goes to infinity we want
//to use the minimum of that forbidden range
let upper_bound =
guest
.forbidden_ranges
.last()
.filter(|range| {range.max == u64::max_value()})
.map(|range| {range.min})
.unwrap_or(u64::max_value());
let mem_to_alloc = min(remaining_memory, upper_bound);
remaining_memory -= mem_to_alloc;
mem_to_alloc
}
}
).collect()
}
// this guy already returns indices instead of references
fn | (proposed_allocations: &Vec<(&Guest, u64)>) -> Vec<(usize, Range)>
{
let mut violations: Vec<(usize, Range)> = Vec::new();
let mut index: usize = 0;
for &(guest, amount) in proposed_allocations.iter()
{
// we want to get the guest's forbidden range with the greatest
// min value that is less than amount. This would probably best
// be done with a binary search, but for now iterative is fine
for range in guest.forbidden_ranges.clone()
{
if range.min < amount && range.max > amount
{
violations.push((index, range));
}
}
index = index + 1;
}
violations
}
fn public_auction_function
(
guest_list: &Vec<&Guest>,
available_memory: u64
)
-> Vec<u64>
{
auction_with_pinned_allocations(guest_list, available_memory, Vec::new())
}
// returns the list of allocations of the provided memory to the list of
// provided guests which results in the maximal social welfare possible without
// changing the provided pinned allocations
fn auction_with_pinned_allocations
(
guest_list: &Vec<&Guest>,
available_memory: u64,
pinned_allocations: &HashMap<u64, u64> //list of (index, allocation) pairs
)
-> Vec<u64>
{
// so I think the idea is we filter the pinned allocations out in an
// enumerate into a filter into a map that discards the index into a
// collect
// then we put them back into the result of the recursive case with an
// enumerate into a flatmap that returns element +
// contiguous-succesive-elements. So we just need a mutable `index_correction`
// variable that gets incremented every time we insert an element so we
// can keep track of each elements place in the new list so we can map positions in
// the enumerate to positions in the original list
let my_copy =
guest_list
.iter()
.enumerate()
.filter(|(index, &guest)| !pinned_allocations.contains_key(index))
.map(|(index, &guest)| guest.clone())
.collect()
// let mut my_copy = guest_list.clone();
let invalid = invalid_allocations(my_copy.iter().zip(naive_allocation(my_copy, available_memory))
}
// Given the applicable range for an invalid allocation and the amount of available_memory
// this function returns the one to two allocation options available to replace the
// invalid allocation
fn two_options(applicable_range: Range, available_memory: u64) -> Vec<u64>
{
match applicable_range.max.cmp(&available_memory)
{
Ordering::Less | Ordering::Equal =>
vec![applicable_range.min, applicable_range.max],
Ordering::Greater => vec![applicable_range.min],
}
}
//I think I want an immutable list slice that I jut progressively slice more off of
// or present different iterators of maybe?
// the payment rule should be implemented in a different function
// so that we can use this function recursively
// I think what we should really return from this is a list of (index, allocation) pairs
// and take as input a list of inidices of invalid allocations (we can have a helper function
//without that argument that passes an empty list into this function)
// fn auction<'a>(guest_list: &'a mut Vec<&'a Guest>, available_memory: u64) -> Vec<(&'a Guest, u64)>
// {
// if guest_list.is_empty()
// {
// return Vec::new();
// }
//
// // so first we try just try giving as much as
// // possible of the remaining_memory
// // to each guest
// let applicable_range: Range;
// let invalid_guest_number;
// {
// //then we first attempt a naive allocation based solely on the sorting
// //and any upper bounds the guests may have set
// // naive_allocation should maintain the same ordering of geusts as guest_list
// // so indices of proposed_allocation can be used to index into guest_list
// let proposed_allocations = naive_allocation(guest_list, available_memory);
//
// let mut invalid = invalid_allocations(&proposed_allocations);
//
// // yup, commenting out this early return got rid of two of the 3 compile errors
// // and the third is pretty obviously a good thing to error out on because it let me know
// //I was adding two guests for case two instead of just one
// // if invalid.is_empty()
// // {
// // return proposed_allocations;
// // }
//
// // so with this first attempt we want to first check and see if we're
// // assigning someone an amount of memory in one of their forbidden ranges
// // and for each case in which someone was allocated an invalid amount, we
// // need to try two cases.
// // so we just need to try removing the first invalid allocation, which means
// // we can just mutate the guest_list instead of cloning every time
// let (_invalid_guest_number, _applicable_range) = invalid.remove(0);
// invalid_guest_number = _invalid_guest_number;
// applicable_range = _applicable_range;
//
// }
// //so we remove the first invalid allcoation
// let badly_served_guest = guest_list.remove(invalid_guest_number);
//
// // and then we try the two cases with that guest
//
// // So I think the idea is that we try the minimum and maximum of the
// // forbidden range that the invalid value fell into
//
// //case one is no more than the minimum of the forbidden range
// let allocation_amount_one = applicable_range.min;
//
// let mut case_one_proposal = auction(guest_list, available_memory - allocation_amount_one);
//
// case_one_proposal.push((badly_served_guest, allocation_amount_one));
//
// let case_one_welfare = social_welfare(&case_one_proposal);
//
// //case two is at least as much as the maximum of the forbidden range
// let allocation_amount_two = applicable_range.max;
//
// let (case_two_welfare, case_two_proposal) =
// if allocation_amount_two <= available_memory
// {
// let mut inner_case_two_proposal =
// auction(guest_list, available_memory - allocation_amount_two);
//
// inner_case_two_proposal.push((badly_served_guest, allocation_amount_two));
//
// (social_welfare(&inner_case_two_proposal), inner_case_two_proposal)
// }
// else
// {
// (0, Vec::new())
// };
//
//
//
// //return the one with greater welfare, or if equal, the one that allocates less memory
// match case_one_welfare.cmp(&case_two_welfare)
// {
// Ordering::Less => case_two_proposal,
//
// Ordering::Greater => case_one_proposal,
//
// Ordering::Equal => case_one_proposal,
// }
// }
// fn registerGuest(baseMemory: i64)
// {
//
// }
// fn makeBid(mem_unit_price: f64, guest: Guest)
// {
//
// }
fn main()
{
let guest1 =
Guest
{
mem_unit_price: 2,
current_holdings: 1,
forbidden_ranges:
vec!
[
Range{min: 0, max: 3},
Range{min: 4, max: u64::max_value()}
],
base_memory: 10
};
let guest2 | invalid_allocations | identifier_name |
main.rs | }
total_welfare
}
// assumes that `guest_list` has already been sorted
// just doles out memory in sorted order such that we
// don't give anyone any more memory than they asked for
// it's still possible that we give people too little memory
// which should be checked after getting the result of this
// function
// I think what I really want to return from this is a vector of allocation amounts
// it'll even take up less space than (reference, allocation) pairs and will be much
// less problematic
fn naive_allocation
(
guest_list: &Vec<&Guest>,
available_memory: u64
)
-> Vec<(u64)>
| .forbidden_ranges
.last()
.filter(|range| {range.max == u64::max_value()})
.map(|range| {range.min})
.unwrap_or(u64::max_value());
let mem_to_alloc = min(remaining_memory, upper_bound);
remaining_memory -= mem_to_alloc;
mem_to_alloc
}
}
).collect()
}
// this guy already returns indices instead of references
fn invalid_allocations(proposed_allocations: &Vec<(&Guest, u64)>) -> Vec<(usize, Range)>
{
let mut violations: Vec<(usize, Range)> = Vec::new();
let mut index: usize = 0;
for &(guest, amount) in proposed_allocations.iter()
{
// we want to get the guest's forbidden range with the greatest
// min value that is less than amount. This would probably best
// be done with a binary search, but for now iterative is fine
for range in guest.forbidden_ranges.clone()
{
if range.min < amount && range.max > amount
{
violations.push((index, range));
}
}
index = index + 1;
}
violations
}
fn public_auction_function
(
guest_list: &Vec<&Guest>,
available_memory: u64
)
-> Vec<u64>
{
auction_with_pinned_allocations(guest_list, available_memory, Vec::new())
}
// returns the list of allocations of the provided memory to the list of
// provided guests which results in the maximal social welfare possible without
// changing the provided pinned allocations
fn auction_with_pinned_allocations
(
guest_list: &Vec<&Guest>,
available_memory: u64,
pinned_allocations: &HashMap<u64, u64> //list of (index, allocation) pairs
)
-> Vec<u64>
{
// so I think the idea is we filter the pinned allocations out in an
// enumerate into a filter into a map that discards the index into a
// collect
// then we put them back into the result of the recursive case with an
// enumerate into a flatmap that returns element +
// contiguous-succesive-elements. So we just need a mutable `index_correction`
// variable that gets incremented every time we insert an element so we
// can keep track of each elements place in the new list so we can map positions in
// the enumerate to positions in the original list
let my_copy =
guest_list
.iter()
.enumerate()
.filter(|(index, &guest)| !pinned_allocations.contains_key(index))
.map(|(index, &guest)| guest.clone())
.collect()
// let mut my_copy = guest_list.clone();
let invalid = invalid_allocations(my_copy.iter().zip(naive_allocation(my_copy, available_memory))
}
// Given the applicable range for an invalid allocation and the amount of available_memory
// this function returns the one to two allocation options available to replace the
// invalid allocation
fn two_options(applicable_range: Range, available_memory: u64) -> Vec<u64>
{
match applicable_range.max.cmp(&available_memory)
{
Ordering::Less | Ordering::Equal =>
vec![applicable_range.min, applicable_range.max],
Ordering::Greater => vec![applicable_range.min],
}
}
//I think I want an immutable list slice that I jut progressively slice more off of
// or present different iterators of maybe?
// the payment rule should be implemented in a different function
// so that we can use this function recursively
// I think what we should really return from this is a list of (index, allocation) pairs
// and take as input a list of inidices of invalid allocations (we can have a helper function
//without that argument that passes an empty list into this function)
// fn auction<'a>(guest_list: &'a mut Vec<&'a Guest>, available_memory: u64) -> Vec<(&'a Guest, u64)>
// {
// if guest_list.is_empty()
// {
// return Vec::new();
// }
//
// // so first we try just try giving as much as
// // possible of the remaining_memory
// // to each guest
// let applicable_range: Range;
// let invalid_guest_number;
// {
// //then we first attempt a naive allocation based solely on the sorting
// //and any upper bounds the guests may have set
// // naive_allocation should maintain the same ordering of geusts as guest_list
// // so indices of proposed_allocation can be used to index into guest_list
// let proposed_allocations = naive_allocation(guest_list, available_memory);
//
// let mut invalid = invalid_allocations(&proposed_allocations);
//
// // yup, commenting out this early return got rid of two of the 3 compile errors
// // and the third is pretty obviously a good thing to error out on because it let me know
// //I was adding two guests for case two instead of just one
// // if invalid.is_empty()
// // {
// // return proposed_allocations;
// // }
//
// // so with this first attempt we want to first check and see if we're
// // assigning someone an amount of memory in one of their forbidden ranges
// // and for each case in which someone was allocated an invalid amount, we
// // need to try two cases.
// // so we just need to try removing the first invalid allocation, which means
// // we can just mutate the guest_list instead of cloning every time
// let (_invalid_guest_number, _applicable_range) = invalid.remove(0);
// invalid_guest_number = _invalid_guest_number;
// applicable_range = _applicable_range;
//
// }
// //so we remove the first invalid allcoation
// let badly_served_guest = guest_list.remove(invalid_guest_number);
//
// // and then we try the two cases with that guest
//
// // So I think the idea is that we try the minimum and maximum of the
// // forbidden range that the invalid value fell into
//
// //case one is no more than the minimum of the forbidden range
// let allocation_amount_one = applicable_range.min;
//
// let mut case_one_proposal = auction(guest_list, available_memory - allocation_amount_one);
//
// case_one_proposal.push((badly_served_guest, allocation_amount_one));
//
// let case_one_welfare = social_welfare(&case_one_proposal);
//
// //case two is at least as much as the maximum of the forbidden range
// let allocation_amount_two = applicable_range.max;
//
// let (case_two_welfare, case_two_proposal) =
// if allocation_amount_two <= available_memory
// {
// let mut inner_case_two_proposal =
// auction(guest_list, available_memory - allocation_amount_two);
//
// inner_case_two_proposal.push((badly_served_guest, allocation_amount_two));
//
// (social_welfare(&inner_case_two_proposal), inner_case_two_proposal)
// }
// else
// {
// (0, Vec::new())
// };
//
//
//
// //return the one with greater welfare, or if equal, the one that allocates less memory
// match case_one_welfare.cmp(&case_two_welfare)
// {
// Ordering::Less => case_two_proposal,
//
// Ordering::Greater => case_one_proposal,
//
// Ordering::Equal => case_one_proposal,
// }
// }
// fn registerGuest(baseMemory: i64)
// {
//
// }
// fn makeBid(mem_unit_price: f64, guest: Guest)
// {
//
// }
fn main()
{
let guest1 =
Guest
{
mem_unit_price: 2,
current_holdings: 1,
forbidden_ranges:
vec!
[
Range{min: 0, max: 3},
Range{min: 4, max: u64::max_value()}
],
base_memory: 10
};
let guest2 =
| {
let mut remaining_memory: u64 = available_memory;
guest_list.iter().map
(
|guest|
{
// if there's no memory left to hand out our job is simple
if remaining_memory == 0
{
0
}
// otherwise get the maximum amount memory this guest
// wants to pay for and give them that
else
{
// if the last forbidden range goes to infinity we want
//to use the minimum of that forbidden range
let upper_bound =
guest | identifier_body |
main.rs | }
total_welfare
}
// assumes that `guest_list` has already been sorted
// just doles out memory in sorted order such that we
// don't give anyone any more memory than they asked for
// it's still possible that we give people too little memory | // which should be checked after getting the result of this
// function
// I think what I really want to return from this is a vector of allocation amounts
// it'll even take up less space than (reference, allocation) pairs and will be much
// less problematic
fn naive_allocation
(
guest_list: &Vec<&Guest>,
available_memory: u64
)
-> Vec<(u64)>
{
let mut remaining_memory: u64 = available_memory;
guest_list.iter().map
(
|guest|
{
// if there's no memory left to hand out our job is simple
if remaining_memory == 0
{
0
}
// otherwise get the maximum amount memory this guest
// wants to pay for and give them that
else
{
// if the last forbidden range goes to infinity we want
//to use the minimum of that forbidden range
let upper_bound =
guest
.forbidden_ranges
.last()
.filter(|range| {range.max == u64::max_value()})
.map(|range| {range.min})
.unwrap_or(u64::max_value());
let mem_to_alloc = min(remaining_memory, upper_bound);
remaining_memory -= mem_to_alloc;
mem_to_alloc
}
}
).collect()
}
// this guy already returns indices instead of references
fn invalid_allocations(proposed_allocations: &Vec<(&Guest, u64)>) -> Vec<(usize, Range)>
{
let mut violations: Vec<(usize, Range)> = Vec::new();
let mut index: usize = 0;
for &(guest, amount) in proposed_allocations.iter()
{
// we want to get the guest's forbidden range with the greatest
// min value that is less than amount. This would probably best
// be done with a binary search, but for now iterative is fine
for range in guest.forbidden_ranges.clone()
{
if range.min < amount && range.max > amount
{
violations.push((index, range));
}
}
index = index + 1;
}
violations
}
fn public_auction_function
(
guest_list: &Vec<&Guest>,
available_memory: u64
)
-> Vec<u64>
{
auction_with_pinned_allocations(guest_list, available_memory, Vec::new())
}
// returns the list of allocations of the provided memory to the list of
// provided guests which results in the maximal social welfare possible without
// changing the provided pinned allocations
fn auction_with_pinned_allocations
(
guest_list: &Vec<&Guest>,
available_memory: u64,
pinned_allocations: &HashMap<u64, u64> //list of (index, allocation) pairs
)
-> Vec<u64>
{
// so I think the idea is we filter the pinned allocations out in an
// enumerate into a filter into a map that discards the index into a
// collect
// then we put them back into the result of the recursive case with an
// enumerate into a flatmap that returns element +
// contiguous-succesive-elements. So we just need a mutable `index_correction`
// variable that gets incremented every time we insert an element so we
// can keep track of each elements place in the new list so we can map positions in
// the enumerate to positions in the original list
let my_copy =
guest_list
.iter()
.enumerate()
.filter(|(index, &guest)| !pinned_allocations.contains_key(index))
.map(|(index, &guest)| guest.clone())
.collect()
// let mut my_copy = guest_list.clone();
let invalid = invalid_allocations(my_copy.iter().zip(naive_allocation(my_copy, available_memory))
}
// Given the applicable range for an invalid allocation and the amount of available_memory
// this function returns the one to two allocation options available to replace the
// invalid allocation
fn two_options(applicable_range: Range, available_memory: u64) -> Vec<u64>
{
match applicable_range.max.cmp(&available_memory)
{
Ordering::Less | Ordering::Equal =>
vec![applicable_range.min, applicable_range.max],
Ordering::Greater => vec![applicable_range.min],
}
}
//I think I want an immutable list slice that I jut progressively slice more off of
// or present different iterators of maybe?
// the payment rule should be implemented in a different function
// so that we can use this function recursively
// I think what we should really return from this is a list of (index, allocation) pairs
// and take as input a list of inidices of invalid allocations (we can have a helper function
//without that argument that passes an empty list into this function)
// fn auction<'a>(guest_list: &'a mut Vec<&'a Guest>, available_memory: u64) -> Vec<(&'a Guest, u64)>
// {
// if guest_list.is_empty()
// {
// return Vec::new();
// }
//
// // so first we try just try giving as much as
// // possible of the remaining_memory
// // to each guest
// let applicable_range: Range;
// let invalid_guest_number;
// {
// //then we first attempt a naive allocation based solely on the sorting
// //and any upper bounds the guests may have set
// // naive_allocation should maintain the same ordering of geusts as guest_list
// // so indices of proposed_allocation can be used to index into guest_list
// let proposed_allocations = naive_allocation(guest_list, available_memory);
//
// let mut invalid = invalid_allocations(&proposed_allocations);
//
// // yup, commenting out this early return got rid of two of the 3 compile errors
// // and the third is pretty obviously a good thing to error out on because it let me know
// //I was adding two guests for case two instead of just one
// // if invalid.is_empty()
// // {
// // return proposed_allocations;
// // }
//
// // so with this first attempt we want to first check and see if we're
// // assigning someone an amount of memory in one of their forbidden ranges
// // and for each case in which someone was allocated an invalid amount, we
// // need to try two cases.
// // so we just need to try removing the first invalid allocation, which means
// // we can just mutate the guest_list instead of cloning every time
// let (_invalid_guest_number, _applicable_range) = invalid.remove(0);
// invalid_guest_number = _invalid_guest_number;
// applicable_range = _applicable_range;
//
// }
// //so we remove the first invalid allcoation
// let badly_served_guest = guest_list.remove(invalid_guest_number);
//
// // and then we try the two cases with that guest
//
// // So I think the idea is that we try the minimum and maximum of the
// // forbidden range that the invalid value fell into
//
// //case one is no more than the minimum of the forbidden range
// let allocation_amount_one = applicable_range.min;
//
// let mut case_one_proposal = auction(guest_list, available_memory - allocation_amount_one);
//
// case_one_proposal.push((badly_served_guest, allocation_amount_one));
//
// let case_one_welfare = social_welfare(&case_one_proposal);
//
// //case two is at least as much as the maximum of the forbidden range
// let allocation_amount_two = applicable_range.max;
//
// let (case_two_welfare, case_two_proposal) =
// if allocation_amount_two <= available_memory
// {
// let mut inner_case_two_proposal =
// auction(guest_list, available_memory - allocation_amount_two);
//
// inner_case_two_proposal.push((badly_served_guest, allocation_amount_two));
//
// (social_welfare(&inner_case_two_proposal), inner_case_two_proposal)
// }
// else
// {
// (0, Vec::new())
// };
//
//
//
// //return the one with greater welfare, or if equal, the one that allocates less memory
// match case_one_welfare.cmp(&case_two_welfare)
// {
// Ordering::Less => case_two_proposal,
//
// Ordering::Greater => case_one_proposal,
//
// Ordering::Equal => case_one_proposal,
// }
// }
// fn registerGuest(baseMemory: i64)
// {
//
// }
// fn makeBid(mem_unit_price: f64, guest: Guest)
// {
//
// }
fn main()
{
let guest1 =
Guest
{
mem_unit_price: 2,
current_holdings: 1,
forbidden_ranges:
vec!
[
Range{min: 0, max: 3},
Range{min: 4, max: u64::max_value()}
],
base_memory: 10
};
let guest2 =
| random_line_split |
|
map.rs | .iter().map(|a| BoundLine(a, self)).collect();
bound.into_iter()
// return self.lines.iter().map(|a| BoundLine(a, self));
}
pub fn iter_sectors(&self) -> std::slice::Iter<Sector> {
self.sectors.iter()
}
pub fn iter_things(&self) -> std::slice::Iter<Thing> {
self.things.iter()
}
pub fn vertex(&self, handle: Handle<Vertex>) -> &Vertex {
&self.vertices[handle.0]
}
pub fn side(&self, handle: Handle<Side>) -> &Side {
&self.sides[handle.0]
}
pub fn sector(&self, handle: Handle<Sector>) -> &Sector {
&self.sectors[handle.0]
}
pub fn bbox(&self) -> Rect {
// TODO ah heck, should include Things too
let points: Vec<_> = self.vertices.iter().map(|v| Point::new(v.x, v.y)).collect();
Rect::from_points(points.iter())
}
pub fn find_player_start(&self) -> Option<Point> {
for thing in &self.things {
if thing.doomednum() == 1 {
return Some(thing.point());
}
}
None
}
pub fn sector_to_polygons(&self, s: usize) -> Vec<Vec<Point>> {
struct Edge<'a> {
line: &'a Line,
side: &'a Side,
facing: Facing,
v0: &'a Vertex,
v1: &'a Vertex,
done: bool,
}
// This is just to convince HashMap to hash on the actual reference, not the underlying
// BareVertex value
struct VertexRef<'a>(&'a Vertex);
impl<'a> PartialEq for VertexRef<'a> {
fn eq(&self, other: &VertexRef) -> bool {
(self.0 as *const _) == (other.0 as *const _)
}
}
impl<'a> Eq for VertexRef<'a> {}
impl<'a> std::hash::Hash for VertexRef<'a> {
fn hash<H: std::hash::Hasher>(&self, state: &mut H) {
(self.0 as *const Vertex).hash(state)
}
}
let mut edges = vec![];
let mut vertices_to_edges = HashMap::new();
// TODO linear scan -- would make more sense to turn the entire map into polygons in one go
for line in &self.lines {
let (frontid, backid) = line.side_indices();
// FIXME need to handle self-referencing sectors, but also
if let Some(front) = line.front.map(|h| &self.sides[h.0]) {
if let Some(back) = line.back.map(|h| &self.sides[h.0]) {
if front.sector == back.sector {
continue;
}
}
}
// TODO seems like a good case for a custom iterator
for &(facing, sideid) in [(Facing::Front, frontid), (Facing::Back, backid)].iter() {
if sideid.is_none() {
continue;
}
// TODO this and the vertices lookups might be bogus and crash...
let side = &self.sides[sideid.unwrap().0];
if side.sector.0 == s {
let v0 = &self.vertices[line.start.0];
let v1 = &self.vertices[line.end.0];
let edge = Edge{
line,
side,
facing,
// TODO should these be swapped depending on the line facing?
v0,
v1,
done: false,
};
edges.push(edge);
vertices_to_edges.entry(VertexRef(v0)).or_insert_with(|| Vec::new()).push(edges.len() - 1);
vertices_to_edges.entry(VertexRef(v1)).or_insert_with(|| Vec::new()).push(edges.len() - 1);
}
}
}
// Trace sectors by starting at the first side's first vertex and attempting to walk from
// there
let mut outlines = Vec::new();
let mut seen_vertices = HashMap::new();
while edges.len() > 0 {
let mut next_vertices = vec![];
for edge in edges.iter() {
// TODO having done-ness for both edges and vertices seems weird, idk
if !seen_vertices.contains_key(&VertexRef(edge.v0)) {
next_vertices.push(edge.v0);
break;
}
if !seen_vertices.contains_key(&VertexRef(edge.v1)) {
next_vertices.push(edge.v1);
break;
}
}
if next_vertices.is_empty() {
break;
}
let mut outline = Vec::new();
while next_vertices.len() > 0 {
let vertices = next_vertices;
next_vertices = Vec::new();
for vertex in vertices.iter() {
if seen_vertices.contains_key(&VertexRef(vertex)) {
continue;
}
seen_vertices.insert(VertexRef(vertex), true);
outline.push(Point::new(vertex.x, vertex.y));
// TODO so, problems occur here if:
// - a vertex has more than two edges
// - special case: double-sided edges are OK! but we have to eliminate
// those, WITHOUT ruining entirely self-referencing sectors
// - a vertex has one edge
for e in vertices_to_edges.get(&VertexRef(vertex)).unwrap().iter() {
let edge = &mut edges[*e];
if edge.done {
// TODO actually this seems weird? why would this happen.
continue;
}
edge.done = true;
if !seen_vertices.contains_key(&VertexRef(edge.v0)) {
next_vertices.push(edge.v0);
}
else if !seen_vertices.contains_key(&VertexRef(edge.v1)) {
next_vertices.push(edge.v1);
}
// Only add EXACTLY ONE vertex at a time for now -- so, assuming simple
// polygons! Figure out the rest, uh, later.
break;
}
}
}
if outline.len() > 0 {
outlines.push(outline);
}
}
outlines
}
}
#[derive(Copy, Clone, Debug)]
pub enum Facing {
Front,
Back,
}
pub struct Handle<T>(usize, PhantomData<*const T>);
// These traits are implemented by hand because #derive'd impls only apply when T implements the
// same trait, but we don't actually own a T, so that bound is unnecessary.
impl<T> Clone for Handle<T> {
fn clone(&self) -> Self {
Handle(self.0, PhantomData)
}
}
impl<T> Copy for Handle<T> {}
impl<T> PartialEq for Handle<T> {
fn eq(&self, other: &Self) -> bool {
self.0 == other.0
}
}
impl<T> Eq for Handle<T> {}
impl<T> std::hash::Hash for Handle<T> {
fn hash<H: std::hash::Hasher>(&self, state: &mut H) {
self.0.hash(state)
}
}
impl<T> From<usize> for Handle<T> {
fn from(index: usize) -> Self {
Handle(index, PhantomData)
}
}
trait MapComponent {}
pub struct Thing {
point: Point,
doomednum: u32,
}
impl Thing {
pub fn point(&self) -> Point {
self.point
}
pub fn doomednum(&self) -> u32 {
self.doomednum
}
}
pub struct Line {
start: Handle<Vertex>,
end: Handle<Vertex>,
flags: u32,
special: usize,
//sector_tag -- oops, different in zdoom...
front: Option<Handle<Side>>,
back: Option<Handle<Side>>,
}
impl Line {
pub fn vertex_indices(&self) -> (Handle<Vertex>, Handle<Vertex>) {
(self.start, self.end)
}
pub fn side_indices(&self) -> (Option<Handle<Side>>, Option<Handle<Side>>) {
(self.front, self.back)
}
pub fn has_special(&self) -> bool {
self.special != 0
}
pub fn blocks_player(&self) -> bool {
self.flags & 1 != 0
}
pub fn is_one_sided(&self) -> bool {
self.front.is_some() != self.back.is_some()
}
pub fn is_two_sided(&self) -> bool {
self.front.is_some() && self.back.is_some()
}
}
// A Line that knows what map it came from, so it can look up its actual sides and vertices
#[derive(Clone, Copy)]
pub struct BoundLine<'a>(&'a Line, &'a Map);
impl<'a> BoundLine<'a> {
pub fn start(&self) -> &Vertex {
self.1.vertex(self.0.start)
}
pub fn end(&self) -> &Vertex {
self.1.vertex(self.0.end)
}
pub fn front(&self) -> Option<&Side> {
self.0.front.map(|s| self.1.side(s))
}
pub fn back(&self) -> Option<&Side> {
self.0.back.map(|s| self.1.side(s))
}
// TODO these are all delegates, eugh
pub fn | vertex_indices | identifier_name |
|
map.rs | -- this should use a method. so should
// new side w/ sector
if bare_line.front_sidedef != -1 {
line.front = Some((bare_line.front_sidedef as usize).into());
}
if bare_line.back_sidedef != -1 {
line.back = Some((bare_line.back_sidedef as usize).into());
}
}
for bare_thing in bare_map.things.iter() {
map.things.push(Thing{
point: Point::new(bare_thing.x as Coord, bare_thing.y as Coord),
doomednum: bare_thing.doomednum as u32,
});
}
map
}
fn side_mut(&mut self, handle: Handle<Side>) -> &mut Side {
&mut self.sides[handle.0]
}
fn line_mut(&mut self, handle: Handle<Line>) -> &mut Line {
&mut self.lines[handle.0]
}
fn add_sector(&mut self) -> Handle<Sector> {
self.sectors.push(Sector{ special: 0, tag: 0, floor_height: 0, ceiling_height: 0 });
(self.sectors.len() - 1).into()
}
fn add_side(&mut self, sector: Handle<Sector>) -> Handle<Side> {
self.sides.push(Side{
id: 0,
lower_texture: "".into(),
middle_texture: "".into(),
upper_texture: "".into(),
sector: sector,
});
(self.sides.len() - 1).into()
}
fn add_vertex(&mut self, x: f64, y: f64) {
self.vertices.push(Vertex{ x, y });
//self.vertices.push(vertex);
//return vertex;
}
fn add_line(&mut self, start: Handle<Vertex>, end: Handle<Vertex>) -> Handle<Line> {
self.lines.push(Line{
start,
end,
flags: 0,
special: 0,
front: None,
back: None,
});
(self.lines.len() - 1).into()
}
pub fn iter_lines(&self) -> <Vec<BoundLine> as IntoIterator>::IntoIter {
let bound: Vec<_> = self.lines.iter().map(|a| BoundLine(a, self)).collect();
bound.into_iter()
// return self.lines.iter().map(|a| BoundLine(a, self));
}
pub fn iter_sectors(&self) -> std::slice::Iter<Sector> {
self.sectors.iter()
}
pub fn iter_things(&self) -> std::slice::Iter<Thing> {
self.things.iter()
}
pub fn vertex(&self, handle: Handle<Vertex>) -> &Vertex {
&self.vertices[handle.0]
}
pub fn side(&self, handle: Handle<Side>) -> &Side {
&self.sides[handle.0]
}
pub fn sector(&self, handle: Handle<Sector>) -> &Sector {
&self.sectors[handle.0]
}
pub fn bbox(&self) -> Rect {
// TODO ah heck, should include Things too
let points: Vec<_> = self.vertices.iter().map(|v| Point::new(v.x, v.y)).collect();
Rect::from_points(points.iter())
}
pub fn find_player_start(&self) -> Option<Point> {
for thing in &self.things {
if thing.doomednum() == 1 {
return Some(thing.point());
}
}
None
}
pub fn sector_to_polygons(&self, s: usize) -> Vec<Vec<Point>> {
struct Edge<'a> {
line: &'a Line,
side: &'a Side,
facing: Facing,
v0: &'a Vertex,
v1: &'a Vertex,
done: bool,
}
// This is just to convince HashMap to hash on the actual reference, not the underlying
// BareVertex value
struct VertexRef<'a>(&'a Vertex);
impl<'a> PartialEq for VertexRef<'a> {
fn eq(&self, other: &VertexRef) -> bool {
(self.0 as *const _) == (other.0 as *const _)
}
}
impl<'a> Eq for VertexRef<'a> {}
impl<'a> std::hash::Hash for VertexRef<'a> {
fn hash<H: std::hash::Hasher>(&self, state: &mut H) {
(self.0 as *const Vertex).hash(state)
}
}
let mut edges = vec![];
let mut vertices_to_edges = HashMap::new();
// TODO linear scan -- would make more sense to turn the entire map into polygons in one go
for line in &self.lines {
let (frontid, backid) = line.side_indices();
// FIXME need to handle self-referencing sectors, but also
if let Some(front) = line.front.map(|h| &self.sides[h.0]) {
if let Some(back) = line.back.map(|h| &self.sides[h.0]) {
if front.sector == back.sector {
continue;
}
}
}
// TODO seems like a good case for a custom iterator
for &(facing, sideid) in [(Facing::Front, frontid), (Facing::Back, backid)].iter() {
if sideid.is_none() {
continue;
}
// TODO this and the vertices lookups might be bogus and crash...
let side = &self.sides[sideid.unwrap().0];
if side.sector.0 == s {
let v0 = &self.vertices[line.start.0];
let v1 = &self.vertices[line.end.0];
let edge = Edge{
line,
side,
facing,
// TODO should these be swapped depending on the line facing?
v0,
v1,
done: false,
};
edges.push(edge);
vertices_to_edges.entry(VertexRef(v0)).or_insert_with(|| Vec::new()).push(edges.len() - 1);
vertices_to_edges.entry(VertexRef(v1)).or_insert_with(|| Vec::new()).push(edges.len() - 1);
}
}
}
// Trace sectors by starting at the first side's first vertex and attempting to walk from
// there
let mut outlines = Vec::new();
let mut seen_vertices = HashMap::new();
while edges.len() > 0 {
let mut next_vertices = vec![];
for edge in edges.iter() {
// TODO having done-ness for both edges and vertices seems weird, idk
if !seen_vertices.contains_key(&VertexRef(edge.v0)) {
next_vertices.push(edge.v0);
break;
}
if !seen_vertices.contains_key(&VertexRef(edge.v1)) {
next_vertices.push(edge.v1);
break;
}
}
if next_vertices.is_empty() {
break;
}
let mut outline = Vec::new();
while next_vertices.len() > 0 {
let vertices = next_vertices;
next_vertices = Vec::new();
for vertex in vertices.iter() {
if seen_vertices.contains_key(&VertexRef(vertex)) {
continue;
}
seen_vertices.insert(VertexRef(vertex), true);
outline.push(Point::new(vertex.x, vertex.y));
// TODO so, problems occur here if:
// - a vertex has more than two edges
// - special case: double-sided edges are OK! but we have to eliminate
// those, WITHOUT ruining entirely self-referencing sectors
// - a vertex has one edge
for e in vertices_to_edges.get(&VertexRef(vertex)).unwrap().iter() {
let edge = &mut edges[*e];
if edge.done {
// TODO actually this seems weird? why would this happen.
continue;
}
edge.done = true;
if !seen_vertices.contains_key(&VertexRef(edge.v0)) {
next_vertices.push(edge.v0);
}
else if !seen_vertices.contains_key(&VertexRef(edge.v1)) {
next_vertices.push(edge.v1);
}
// Only add EXACTLY ONE vertex at a time for now -- so, assuming simple
// polygons! Figure out the rest, uh, later.
break;
}
}
}
if outline.len() > 0 {
outlines.push(outline);
}
}
outlines
}
}
#[derive(Copy, Clone, Debug)]
pub enum Facing {
Front,
Back,
}
pub struct Handle<T>(usize, PhantomData<*const T>);
// These traits are implemented by hand because #derive'd impls only apply when T implements the
// same trait, but we don't actually own a T, so that bound is unnecessary.
impl<T> Clone for Handle<T> {
fn clone(&self) -> Self {
Handle(self.0, PhantomData)
}
}
impl<T> Copy for Handle<T> {}
impl<T> PartialEq for Handle<T> {
fn eq(&self, other: &Self) -> bool {
self.0 == other.0
}
}
impl<T> Eq for Handle<T> {}
impl<T> std::hash::Hash for Handle<T> {
fn hash<H: std::hash::Hasher>(&self, state: &mut H) {
self.0.hash(state) | random_line_split |
||
map.rs | .iter() {
map.add_vertex(bare_vertex.x as f64, bare_vertex.y as f64);
}
for bare_side in bare_map.sides.iter() {
let handle = map.add_side((bare_side.sector as usize).into());
let side = map.side_mut(handle);
side.lower_texture = bare_side.lower_texture.into();
side.middle_texture = bare_side.middle_texture.into();
side.upper_texture = bare_side.upper_texture.into();
}
for bare_line in bare_map.lines.iter() {
let handle = map.add_line((bare_line.v0 as usize).into(), (bare_line.v1 as usize).into());
let line = map.line_mut(handle);
line.flags = bare_line.flags as u32;
// FIXME and here's where we start to go awry -- this should use a method. so should
// new side w/ sector
if bare_line.front_sidedef != -1 {
line.front = Some((bare_line.front_sidedef as usize).into());
}
if bare_line.back_sidedef != -1 {
line.back = Some((bare_line.back_sidedef as usize).into());
}
}
for bare_thing in bare_map.things.iter() {
map.things.push(Thing{
point: Point::new(bare_thing.x as Coord, bare_thing.y as Coord),
doomednum: bare_thing.doomednum as u32,
});
}
map
}
fn side_mut(&mut self, handle: Handle<Side>) -> &mut Side {
&mut self.sides[handle.0]
}
fn line_mut(&mut self, handle: Handle<Line>) -> &mut Line {
&mut self.lines[handle.0]
}
fn add_sector(&mut self) -> Handle<Sector> {
self.sectors.push(Sector{ special: 0, tag: 0, floor_height: 0, ceiling_height: 0 });
(self.sectors.len() - 1).into()
}
fn add_side(&mut self, sector: Handle<Sector>) -> Handle<Side> {
self.sides.push(Side{
id: 0,
lower_texture: "".into(),
middle_texture: "".into(),
upper_texture: "".into(),
sector: sector,
});
(self.sides.len() - 1).into()
}
fn add_vertex(&mut self, x: f64, y: f64) {
self.vertices.push(Vertex{ x, y });
//self.vertices.push(vertex);
//return vertex;
}
fn add_line(&mut self, start: Handle<Vertex>, end: Handle<Vertex>) -> Handle<Line> {
self.lines.push(Line{
start,
end,
flags: 0,
special: 0,
front: None,
back: None,
});
(self.lines.len() - 1).into()
}
pub fn iter_lines(&self) -> <Vec<BoundLine> as IntoIterator>::IntoIter {
let bound: Vec<_> = self.lines.iter().map(|a| BoundLine(a, self)).collect();
bound.into_iter()
// return self.lines.iter().map(|a| BoundLine(a, self));
}
pub fn iter_sectors(&self) -> std::slice::Iter<Sector> |
pub fn iter_things(&self) -> std::slice::Iter<Thing> {
self.things.iter()
}
pub fn vertex(&self, handle: Handle<Vertex>) -> &Vertex {
&self.vertices[handle.0]
}
pub fn side(&self, handle: Handle<Side>) -> &Side {
&self.sides[handle.0]
}
pub fn sector(&self, handle: Handle<Sector>) -> &Sector {
&self.sectors[handle.0]
}
pub fn bbox(&self) -> Rect {
// TODO ah heck, should include Things too
let points: Vec<_> = self.vertices.iter().map(|v| Point::new(v.x, v.y)).collect();
Rect::from_points(points.iter())
}
pub fn find_player_start(&self) -> Option<Point> {
for thing in &self.things {
if thing.doomednum() == 1 {
return Some(thing.point());
}
}
None
}
pub fn sector_to_polygons(&self, s: usize) -> Vec<Vec<Point>> {
struct Edge<'a> {
line: &'a Line,
side: &'a Side,
facing: Facing,
v0: &'a Vertex,
v1: &'a Vertex,
done: bool,
}
// This is just to convince HashMap to hash on the actual reference, not the underlying
// BareVertex value
struct VertexRef<'a>(&'a Vertex);
impl<'a> PartialEq for VertexRef<'a> {
fn eq(&self, other: &VertexRef) -> bool {
(self.0 as *const _) == (other.0 as *const _)
}
}
impl<'a> Eq for VertexRef<'a> {}
impl<'a> std::hash::Hash for VertexRef<'a> {
fn hash<H: std::hash::Hasher>(&self, state: &mut H) {
(self.0 as *const Vertex).hash(state)
}
}
let mut edges = vec![];
let mut vertices_to_edges = HashMap::new();
// TODO linear scan -- would make more sense to turn the entire map into polygons in one go
for line in &self.lines {
let (frontid, backid) = line.side_indices();
// FIXME need to handle self-referencing sectors, but also
if let Some(front) = line.front.map(|h| &self.sides[h.0]) {
if let Some(back) = line.back.map(|h| &self.sides[h.0]) {
if front.sector == back.sector {
continue;
}
}
}
// TODO seems like a good case for a custom iterator
for &(facing, sideid) in [(Facing::Front, frontid), (Facing::Back, backid)].iter() {
if sideid.is_none() {
continue;
}
// TODO this and the vertices lookups might be bogus and crash...
let side = &self.sides[sideid.unwrap().0];
if side.sector.0 == s {
let v0 = &self.vertices[line.start.0];
let v1 = &self.vertices[line.end.0];
let edge = Edge{
line,
side,
facing,
// TODO should these be swapped depending on the line facing?
v0,
v1,
done: false,
};
edges.push(edge);
vertices_to_edges.entry(VertexRef(v0)).or_insert_with(|| Vec::new()).push(edges.len() - 1);
vertices_to_edges.entry(VertexRef(v1)).or_insert_with(|| Vec::new()).push(edges.len() - 1);
}
}
}
// Trace sectors by starting at the first side's first vertex and attempting to walk from
// there
let mut outlines = Vec::new();
let mut seen_vertices = HashMap::new();
while edges.len() > 0 {
let mut next_vertices = vec![];
for edge in edges.iter() {
// TODO having done-ness for both edges and vertices seems weird, idk
if !seen_vertices.contains_key(&VertexRef(edge.v0)) {
next_vertices.push(edge.v0);
break;
}
if !seen_vertices.contains_key(&VertexRef(edge.v1)) {
next_vertices.push(edge.v1);
break;
}
}
if next_vertices.is_empty() {
break;
}
let mut outline = Vec::new();
while next_vertices.len() > 0 {
let vertices = next_vertices;
next_vertices = Vec::new();
for vertex in vertices.iter() {
if seen_vertices.contains_key(&VertexRef(vertex)) {
continue;
}
seen_vertices.insert(VertexRef(vertex), true);
outline.push(Point::new(vertex.x, vertex.y));
// TODO so, problems occur here if:
// - a vertex has more than two edges
// - special case: double-sided edges are OK! but we have to eliminate
// those, WITHOUT ruining entirely self-referencing sectors
// - a vertex has one edge
for e in vertices_to_edges.get(&VertexRef(vertex)).unwrap().iter() {
let edge = &mut edges[*e];
if edge.done {
// TODO actually this seems weird? why would this happen.
continue;
}
edge.done = true;
if !seen_vertices.contains_key(&VertexRef(edge.v0)) {
next_vertices.push(edge.v0);
}
else if !seen_vertices.contains_key(&VertexRef(edge.v1)) {
next_vertices.push(edge.v1);
}
// Only add EXACTLY ONE vertex at a time for now -- so, assuming simple
// polygons! Figure out the rest, uh, later.
break;
}
}
}
if outline.len() > 0 {
outlines.push(outline);
}
}
outlines
}
}
#[derive(Copy, Clone, Debug)]
pub enum Facing {
Front,
Back,
}
pub struct Handle<T | {
self.sectors.iter()
} | identifier_body |
models.py | .Model):
# Model representing a movie genre (e.g. Science Fiction, Non Fiction).
name = models.CharField(
max_length=200,
help_text="Enter a movie genre (e.g. Science Fiction, French Poetry etc.)"
)
def __str__(self):
# String for representing the Model object (in Admin site etc.)
return self.name
class Language(models.Model):
# Model representing a Language (e.g. English, French, Japanese, etc.)
name = models.CharField(max_length=200,
help_text="Enter the movie's natural language (e.g. English, French, Japanese etc.)")
def __str__(self):
# String for representing the Model object (in Admin site etc.)
return self.name
class Movie(models.Model):
# Model representing a movie (but not a specific copy of a movie).
title = models.CharField(max_length=200, null=True, blank=True, help_text='This field will be overwritten if given a valid IMDB id and left blank.')
imdb_link = models.CharField('IMDB Link', max_length=100, blank=True, null=True, help_text='For example, here is <a target="_blank" '
'href="https://www.imdb.com/title/tt3322364/">Concussion\'s link</a>.')
# Foreign Key used because movie can only have one director, but directors can have multiple movies
# Director as a string rather than object because it hasn't been declared yet in file.
director = models.ForeignKey('Director', on_delete=models.SET_NULL, null=True, blank=True, help_text='This field will be overwritten \
if given a valid IMDB id and left blank.')
language = models.ForeignKey('Language', on_delete=models.SET_NULL, null=True, blank=True)
summary = models.TextField(max_length=5000, null=True, blank=True, help_text="Enter a brief description of the movie. This field will \
be overwritten if given a valid IMDB id and left blank.")
# Genre class has already been defined so we can specify the object above.
genre = models.ForeignKey('Genre', on_delete=models.SET_NULL, null=True, blank=True, help_text='This field will be overwritten if given \
a valid IMDB id and left blank.')
tags = TaggableManager(blank=True)
year = models.CharField(max_length=200, null=True, blank=True, help_text='This field will be overwritten if given a valid IMDB id and left blank.')
thumbnail = models.CharField('Thumbnail', max_length=500, blank=True, null=True, help_text='This field will be overwritten \
if given a valid IMDB id and left blank.')
file = S3DirectField(dest='videos', blank=True, null=True)
ads = models.CharField('Google VAMP Ads Link', max_length=1000, blank=True, null=True, help_text="""For example, here is a <a target="_blank"
href="https://pubads.g.doubleclick.net/gampad/ads?sz=640x480&iu=/124319096/external/ad_rule_samples&ciu_szs=300x250&ad_rule=1&impl=s&gdfp_req=1&env=vp&output=vmap&unviewed_position_start=1&cust_params=deployment%3Ddevsite%26sample_ar%3Dpremidpost&cmsid=496&vid=short_onecue&correlator=">
Google VAMP example link</a>.""")
duration = models.CharField(max_length=200)
fps = models.CharField(max_length=200)
dimensions = models.CharField(max_length=200)
max_num_find_articles = models.IntegerField('Max number of research articles', default=5, validators=[MinValueValidator(0), MaxValueValidator(100)], help_text="Default number is 5.")
found_articles = models.TextField('Found Research Articles', max_length=5000, null=True, blank=True, help_text="HTML list output of found research \
articles on Google Scholar. Clear the text to find new articles.")
ratings = GenericRelation(Rating, related_query_name='movie-rating')
class Meta:
ordering = ['title', 'director']
def display_genre(self):
# Creates a string for the Genre. This is required to display genre in Admin.
return ', '.join([genre.name for genre in self.genre.all()[:3]])
display_genre.short_description = 'Genre'
def get_absolute_url(self):
# Returns the url to access a particular movie instance.
return reverse('movie-detail', args=[str(self.id)])
def __str__(self):
# String for representing the Model object.
return self.title if self.title else ''
def get_movie_url(self):
return (self.file).replace(" ", "+")
def get_video_stats(self):
#filename = str(settings.BASE_DIR) + self.file.url
clip = VideoFileClip(self.get_movie_url())
duration = str(datetime.timedelta(seconds=round(clip.duration)))
fps = clip.fps
width, height = clip.size
return [duration, fps, (width, height)]
def get_imdb_stats(self):
ia = imdb.IMDb()
reg = re.compile(r'^.*(ch|co|ev|nm|tt)(\d{7}\d*)\/?$')
id_found = reg.match(self.imdb_link)
if id_found:
movie = ia.get_movie(id_found.group(2))
return [movie['year'], movie['directors'][0], movie['genres'][0], movie['title'], movie.get('plot')[0], movie['cover url']]
else:
raise Exception(f"No imdb match found for imdb link: {self.imdb_link}")
def get_research_articles(self, max_num):
# Search string for Google Scholar to look for.
# e.g. "{self.title} {self.director.name}" would equate to "Concussion Peter Landesman" for the movie Concussion.
search_str = f'{self.title} {self.director.name}'
output = f""
try:
pg = ProxyGenerator()
ip = os.environ['PROXY_IP']
pg.SingleProxy(http = ip, https = ip)
o = scholarly.use_proxy(pg)
search_query = scholarly.search_pubs(search_str)
for i in range(0, max_num):
curr = next(search_query)
# For debugging purposes, this is how you pretty print the search query's contents.
#scholarly.pprint(curr)
# Grab the title of the article.
title = curr['bib']['title']
# Begin our formatted html output for each found research article.
output += f"""
<li>
"""
# See if a publication url (i.e. curr['pub_url']) exists. If so, add an external link to it.
if 'pub_url' in curr:
output += f"""
<a target='_blank' href=\"{curr['pub_url']}\">{title}</a>
"""
else:
output += f"""
{title}
"""
output += f"""
<br>
"""
# Writes the abstract (i.e.curr['bib']['abstract']) if it exists.
if 'bib' in curr and 'abstract' in curr['bib']:
output += f"""
<p>{curr['bib']['abstract']}</p>
"""
output += f"""
</li>
"""
except Exception as e:
pass
# Useful for seeing errors in your terminal. Replace pass with the print statement below.
#print(sys.stderr, e)
return output
def save(self, *args, **kwargs):
|
# Checks if a director name already exists. If not, create and assign to the movie.
director = None
try:
directors = Director.objects.all()
for d in directors:
if (str(d) == str(imdb_stats[1])):
director = d
break
orig.director = director if director is not None else Director.objects.create(name=imdb_stats[1])
except:
orig.director = Director.objects.create(name=imdb_stats[1])
# Checks if a genre name already exists. If not, create and assign to the movie.
genre = None
try:
genres = Genre.objects.all()
for g in genres:
if (str(g) == str(imdb_stats[2])):
genre = g
break
genre = genre if genre is not None else Genre.objects.create(name=imdb_stats[2])
except:
genre = Genre.objects.create(name=imdb_stats[2])
| super(Movie, self).save(*args, **kwargs)
# Uses a custom save to end date any subCases
orig = Movie.objects.get(id=self.id)
fields_to_update = []
try:
specs = self.get_video_stats()
orig.duration = specs[0]
orig.fps = specs[1]
orig.dimensions = specs[2]
fields_to_update.extend(['duration', 'fps', 'dimensions'])
except Exception as e:
pass
try:
imdb_stats = self.get_imdb_stats()
orig.title = imdb_stats[3]
orig.year = imdb_stats[0]
orig.thumbnail = imdb_stats[5] | identifier_body |
models.py | .Model):
# Model representing a movie genre (e.g. Science Fiction, Non Fiction).
name = models.CharField(
max_length=200,
help_text="Enter a movie genre (e.g. Science Fiction, French Poetry etc.)"
)
def __str__(self):
# String for representing the Model object (in Admin site etc.)
return self.name
class Language(models.Model):
# Model representing a Language (e.g. English, French, Japanese, etc.)
name = models.CharField(max_length=200,
help_text="Enter the movie's natural language (e.g. English, French, Japanese etc.)")
def __str__(self):
# String for representing the Model object (in Admin site etc.)
return self.name
class Movie(models.Model):
# Model representing a movie (but not a specific copy of a movie).
title = models.CharField(max_length=200, null=True, blank=True, help_text='This field will be overwritten if given a valid IMDB id and left blank.')
imdb_link = models.CharField('IMDB Link', max_length=100, blank=True, null=True, help_text='For example, here is <a target="_blank" '
'href="https://www.imdb.com/title/tt3322364/">Concussion\'s link</a>.')
# Foreign Key used because movie can only have one director, but directors can have multiple movies
# Director as a string rather than object because it hasn't been declared yet in file.
director = models.ForeignKey('Director', on_delete=models.SET_NULL, null=True, blank=True, help_text='This field will be overwritten \
if given a valid IMDB id and left blank.')
language = models.ForeignKey('Language', on_delete=models.SET_NULL, null=True, blank=True)
summary = models.TextField(max_length=5000, null=True, blank=True, help_text="Enter a brief description of the movie. This field will \
be overwritten if given a valid IMDB id and left blank.")
# Genre class has already been defined so we can specify the object above.
genre = models.ForeignKey('Genre', on_delete=models.SET_NULL, null=True, blank=True, help_text='This field will be overwritten if given \
a valid IMDB id and left blank.')
tags = TaggableManager(blank=True)
year = models.CharField(max_length=200, null=True, blank=True, help_text='This field will be overwritten if given a valid IMDB id and left blank.')
thumbnail = models.CharField('Thumbnail', max_length=500, blank=True, null=True, help_text='This field will be overwritten \
if given a valid IMDB id and left blank.')
file = S3DirectField(dest='videos', blank=True, null=True)
ads = models.CharField('Google VAMP Ads Link', max_length=1000, blank=True, null=True, help_text="""For example, here is a <a target="_blank"
href="https://pubads.g.doubleclick.net/gampad/ads?sz=640x480&iu=/124319096/external/ad_rule_samples&ciu_szs=300x250&ad_rule=1&impl=s&gdfp_req=1&env=vp&output=vmap&unviewed_position_start=1&cust_params=deployment%3Ddevsite%26sample_ar%3Dpremidpost&cmsid=496&vid=short_onecue&correlator=">
Google VAMP example link</a>.""")
duration = models.CharField(max_length=200) | max_num_find_articles = models.IntegerField('Max number of research articles', default=5, validators=[MinValueValidator(0), MaxValueValidator(100)], help_text="Default number is 5.")
found_articles = models.TextField('Found Research Articles', max_length=5000, null=True, blank=True, help_text="HTML list output of found research \
articles on Google Scholar. Clear the text to find new articles.")
ratings = GenericRelation(Rating, related_query_name='movie-rating')
class Meta:
ordering = ['title', 'director']
def display_genre(self):
# Creates a string for the Genre. This is required to display genre in Admin.
return ', '.join([genre.name for genre in self.genre.all()[:3]])
display_genre.short_description = 'Genre'
def get_absolute_url(self):
# Returns the url to access a particular movie instance.
return reverse('movie-detail', args=[str(self.id)])
def __str__(self):
# String for representing the Model object.
return self.title if self.title else ''
def get_movie_url(self):
return (self.file).replace(" ", "+")
def get_video_stats(self):
#filename = str(settings.BASE_DIR) + self.file.url
clip = VideoFileClip(self.get_movie_url())
duration = str(datetime.timedelta(seconds=round(clip.duration)))
fps = clip.fps
width, height = clip.size
return [duration, fps, (width, height)]
def get_imdb_stats(self):
ia = imdb.IMDb()
reg = re.compile(r'^.*(ch|co|ev|nm|tt)(\d{7}\d*)\/?$')
id_found = reg.match(self.imdb_link)
if id_found:
movie = ia.get_movie(id_found.group(2))
return [movie['year'], movie['directors'][0], movie['genres'][0], movie['title'], movie.get('plot')[0], movie['cover url']]
else:
raise Exception(f"No imdb match found for imdb link: {self.imdb_link}")
def get_research_articles(self, max_num):
# Search string for Google Scholar to look for.
# e.g. "{self.title} {self.director.name}" would equate to "Concussion Peter Landesman" for the movie Concussion.
search_str = f'{self.title} {self.director.name}'
output = f""
try:
pg = ProxyGenerator()
ip = os.environ['PROXY_IP']
pg.SingleProxy(http = ip, https = ip)
o = scholarly.use_proxy(pg)
search_query = scholarly.search_pubs(search_str)
for i in range(0, max_num):
curr = next(search_query)
# For debugging purposes, this is how you pretty print the search query's contents.
#scholarly.pprint(curr)
# Grab the title of the article.
title = curr['bib']['title']
# Begin our formatted html output for each found research article.
output += f"""
<li>
"""
# See if a publication url (i.e. curr['pub_url']) exists. If so, add an external link to it.
if 'pub_url' in curr:
output += f"""
<a target='_blank' href=\"{curr['pub_url']}\">{title}</a>
"""
else:
output += f"""
{title}
"""
output += f"""
<br>
"""
# Writes the abstract (i.e.curr['bib']['abstract']) if it exists.
if 'bib' in curr and 'abstract' in curr['bib']:
output += f"""
<p>{curr['bib']['abstract']}</p>
"""
output += f"""
</li>
"""
except Exception as e:
pass
# Useful for seeing errors in your terminal. Replace pass with the print statement below.
#print(sys.stderr, e)
return output
def save(self, *args, **kwargs):
super(Movie, self).save(*args, **kwargs)
# Uses a custom save to end date any subCases
orig = Movie.objects.get(id=self.id)
fields_to_update = []
try:
specs = self.get_video_stats()
orig.duration = specs[0]
orig.fps = specs[1]
orig.dimensions = specs[2]
fields_to_update.extend(['duration', 'fps', 'dimensions'])
except Exception as e:
pass
try:
imdb_stats = self.get_imdb_stats()
orig.title = imdb_stats[3]
orig.year = imdb_stats[0]
orig.thumbnail = imdb_stats[5]
# Checks if a director name already exists. If not, create and assign to the movie.
director = None
try:
directors = Director.objects.all()
for d in directors:
if (str(d) == str(imdb_stats[1])):
director = d
break
orig.director = director if director is not None else Director.objects.create(name=imdb_stats[1])
except:
orig.director = Director.objects.create(name=imdb_stats[1])
# Checks if a genre name already exists. If not, create and assign to the movie.
genre = None
try:
genres = Genre.objects.all()
for g in genres:
if (str(g) == str(imdb_stats[2])):
genre = g
break
genre = genre if genre is not None else Genre.objects.create(name=imdb_stats[2])
except:
genre = Genre.objects.create(name=imdb_stats[2])
orig | fps = models.CharField(max_length=200)
dimensions = models.CharField(max_length=200)
| random_line_split |
models.py | .Model):
# Model representing a movie genre (e.g. Science Fiction, Non Fiction).
name = models.CharField(
max_length=200,
help_text="Enter a movie genre (e.g. Science Fiction, French Poetry etc.)"
)
def __str__(self):
# String for representing the Model object (in Admin site etc.)
return self.name
class Language(models.Model):
# Model representing a Language (e.g. English, French, Japanese, etc.)
name = models.CharField(max_length=200,
help_text="Enter the movie's natural language (e.g. English, French, Japanese etc.)")
def __str__(self):
# String for representing the Model object (in Admin site etc.)
return self.name
class Movie(models.Model):
# Model representing a movie (but not a specific copy of a movie).
title = models.CharField(max_length=200, null=True, blank=True, help_text='This field will be overwritten if given a valid IMDB id and left blank.')
imdb_link = models.CharField('IMDB Link', max_length=100, blank=True, null=True, help_text='For example, here is <a target="_blank" '
'href="https://www.imdb.com/title/tt3322364/">Concussion\'s link</a>.')
# Foreign Key used because movie can only have one director, but directors can have multiple movies
# Director as a string rather than object because it hasn't been declared yet in file.
director = models.ForeignKey('Director', on_delete=models.SET_NULL, null=True, blank=True, help_text='This field will be overwritten \
if given a valid IMDB id and left blank.')
language = models.ForeignKey('Language', on_delete=models.SET_NULL, null=True, blank=True)
summary = models.TextField(max_length=5000, null=True, blank=True, help_text="Enter a brief description of the movie. This field will \
be overwritten if given a valid IMDB id and left blank.")
# Genre class has already been defined so we can specify the object above.
genre = models.ForeignKey('Genre', on_delete=models.SET_NULL, null=True, blank=True, help_text='This field will be overwritten if given \
a valid IMDB id and left blank.')
tags = TaggableManager(blank=True)
year = models.CharField(max_length=200, null=True, blank=True, help_text='This field will be overwritten if given a valid IMDB id and left blank.')
thumbnail = models.CharField('Thumbnail', max_length=500, blank=True, null=True, help_text='This field will be overwritten \
if given a valid IMDB id and left blank.')
file = S3DirectField(dest='videos', blank=True, null=True)
ads = models.CharField('Google VAMP Ads Link', max_length=1000, blank=True, null=True, help_text="""For example, here is a <a target="_blank"
href="https://pubads.g.doubleclick.net/gampad/ads?sz=640x480&iu=/124319096/external/ad_rule_samples&ciu_szs=300x250&ad_rule=1&impl=s&gdfp_req=1&env=vp&output=vmap&unviewed_position_start=1&cust_params=deployment%3Ddevsite%26sample_ar%3Dpremidpost&cmsid=496&vid=short_onecue&correlator=">
Google VAMP example link</a>.""")
duration = models.CharField(max_length=200)
fps = models.CharField(max_length=200)
dimensions = models.CharField(max_length=200)
max_num_find_articles = models.IntegerField('Max number of research articles', default=5, validators=[MinValueValidator(0), MaxValueValidator(100)], help_text="Default number is 5.")
found_articles = models.TextField('Found Research Articles', max_length=5000, null=True, blank=True, help_text="HTML list output of found research \
articles on Google Scholar. Clear the text to find new articles.")
ratings = GenericRelation(Rating, related_query_name='movie-rating')
class Meta:
ordering = ['title', 'director']
def display_genre(self):
# Creates a string for the Genre. This is required to display genre in Admin.
return ', '.join([genre.name for genre in self.genre.all()[:3]])
display_genre.short_description = 'Genre'
def get_absolute_url(self):
# Returns the url to access a particular movie instance.
return reverse('movie-detail', args=[str(self.id)])
def __str__(self):
# String for representing the Model object.
return self.title if self.title else ''
def get_movie_url(self):
return (self.file).replace(" ", "+")
def get_video_stats(self):
#filename = str(settings.BASE_DIR) + self.file.url
clip = VideoFileClip(self.get_movie_url())
duration = str(datetime.timedelta(seconds=round(clip.duration)))
fps = clip.fps
width, height = clip.size
return [duration, fps, (width, height)]
def get_imdb_stats(self):
ia = imdb.IMDb()
reg = re.compile(r'^.*(ch|co|ev|nm|tt)(\d{7}\d*)\/?$')
id_found = reg.match(self.imdb_link)
if id_found:
movie = ia.get_movie(id_found.group(2))
return [movie['year'], movie['directors'][0], movie['genres'][0], movie['title'], movie.get('plot')[0], movie['cover url']]
else:
raise Exception(f"No imdb match found for imdb link: {self.imdb_link}")
def get_research_articles(self, max_num):
# Search string for Google Scholar to look for.
# e.g. "{self.title} {self.director.name}" would equate to "Concussion Peter Landesman" for the movie Concussion.
search_str = f'{self.title} {self.director.name}'
output = f""
try:
pg = ProxyGenerator()
ip = os.environ['PROXY_IP']
pg.SingleProxy(http = ip, https = ip)
o = scholarly.use_proxy(pg)
search_query = scholarly.search_pubs(search_str)
for i in range(0, max_num):
curr = next(search_query)
# For debugging purposes, this is how you pretty print the search query's contents.
#scholarly.pprint(curr)
# Grab the title of the article.
title = curr['bib']['title']
# Begin our formatted html output for each found research article.
output += f"""
<li>
"""
# See if a publication url (i.e. curr['pub_url']) exists. If so, add an external link to it.
if 'pub_url' in curr:
output += f"""
<a target='_blank' href=\"{curr['pub_url']}\">{title}</a>
"""
else:
output += f"""
{title}
"""
output += f"""
<br>
"""
# Writes the abstract (i.e.curr['bib']['abstract']) if it exists.
if 'bib' in curr and 'abstract' in curr['bib']:
output += f"""
<p>{curr['bib']['abstract']}</p>
"""
output += f"""
</li>
"""
except Exception as e:
pass
# Useful for seeing errors in your terminal. Replace pass with the print statement below.
#print(sys.stderr, e)
return output
def | (self, *args, **kwargs):
super(Movie, self).save(*args, **kwargs)
# Uses a custom save to end date any subCases
orig = Movie.objects.get(id=self.id)
fields_to_update = []
try:
specs = self.get_video_stats()
orig.duration = specs[0]
orig.fps = specs[1]
orig.dimensions = specs[2]
fields_to_update.extend(['duration', 'fps', 'dimensions'])
except Exception as e:
pass
try:
imdb_stats = self.get_imdb_stats()
orig.title = imdb_stats[3]
orig.year = imdb_stats[0]
orig.thumbnail = imdb_stats[5]
# Checks if a director name already exists. If not, create and assign to the movie.
director = None
try:
directors = Director.objects.all()
for d in directors:
if (str(d) == str(imdb_stats[1])):
director = d
break
orig.director = director if director is not None else Director.objects.create(name=imdb_stats[1])
except:
orig.director = Director.objects.create(name=imdb_stats[1])
# Checks if a genre name already exists. If not, create and assign to the movie.
genre = None
try:
genres = Genre.objects.all()
for g in genres:
if (str(g) == str(imdb_stats[2])):
genre = g
break
genre = genre if genre is not None else Genre.objects.create(name=imdb_stats[2])
except:
genre = Genre.objects.create(name=imdb_stats[2])
| save | identifier_name |
models.py | movie (but not a specific copy of a movie).
title = models.CharField(max_length=200, null=True, blank=True, help_text='This field will be overwritten if given a valid IMDB id and left blank.')
imdb_link = models.CharField('IMDB Link', max_length=100, blank=True, null=True, help_text='For example, here is <a target="_blank" '
'href="https://www.imdb.com/title/tt3322364/">Concussion\'s link</a>.')
# Foreign Key used because movie can only have one director, but directors can have multiple movies
# Director as a string rather than object because it hasn't been declared yet in file.
director = models.ForeignKey('Director', on_delete=models.SET_NULL, null=True, blank=True, help_text='This field will be overwritten \
if given a valid IMDB id and left blank.')
language = models.ForeignKey('Language', on_delete=models.SET_NULL, null=True, blank=True)
summary = models.TextField(max_length=5000, null=True, blank=True, help_text="Enter a brief description of the movie. This field will \
be overwritten if given a valid IMDB id and left blank.")
# Genre class has already been defined so we can specify the object above.
genre = models.ForeignKey('Genre', on_delete=models.SET_NULL, null=True, blank=True, help_text='This field will be overwritten if given \
a valid IMDB id and left blank.')
tags = TaggableManager(blank=True)
year = models.CharField(max_length=200, null=True, blank=True, help_text='This field will be overwritten if given a valid IMDB id and left blank.')
thumbnail = models.CharField('Thumbnail', max_length=500, blank=True, null=True, help_text='This field will be overwritten \
if given a valid IMDB id and left blank.')
file = S3DirectField(dest='videos', blank=True, null=True)
ads = models.CharField('Google VAMP Ads Link', max_length=1000, blank=True, null=True, help_text="""For example, here is a <a target="_blank"
href="https://pubads.g.doubleclick.net/gampad/ads?sz=640x480&iu=/124319096/external/ad_rule_samples&ciu_szs=300x250&ad_rule=1&impl=s&gdfp_req=1&env=vp&output=vmap&unviewed_position_start=1&cust_params=deployment%3Ddevsite%26sample_ar%3Dpremidpost&cmsid=496&vid=short_onecue&correlator=">
Google VAMP example link</a>.""")
duration = models.CharField(max_length=200)
fps = models.CharField(max_length=200)
dimensions = models.CharField(max_length=200)
max_num_find_articles = models.IntegerField('Max number of research articles', default=5, validators=[MinValueValidator(0), MaxValueValidator(100)], help_text="Default number is 5.")
found_articles = models.TextField('Found Research Articles', max_length=5000, null=True, blank=True, help_text="HTML list output of found research \
articles on Google Scholar. Clear the text to find new articles.")
ratings = GenericRelation(Rating, related_query_name='movie-rating')
class Meta:
ordering = ['title', 'director']
def display_genre(self):
# Creates a string for the Genre. This is required to display genre in Admin.
return ', '.join([genre.name for genre in self.genre.all()[:3]])
display_genre.short_description = 'Genre'
def get_absolute_url(self):
# Returns the url to access a particular movie instance.
return reverse('movie-detail', args=[str(self.id)])
def __str__(self):
# String for representing the Model object.
return self.title if self.title else ''
def get_movie_url(self):
return (self.file).replace(" ", "+")
def get_video_stats(self):
#filename = str(settings.BASE_DIR) + self.file.url
clip = VideoFileClip(self.get_movie_url())
duration = str(datetime.timedelta(seconds=round(clip.duration)))
fps = clip.fps
width, height = clip.size
return [duration, fps, (width, height)]
def get_imdb_stats(self):
ia = imdb.IMDb()
reg = re.compile(r'^.*(ch|co|ev|nm|tt)(\d{7}\d*)\/?$')
id_found = reg.match(self.imdb_link)
if id_found:
movie = ia.get_movie(id_found.group(2))
return [movie['year'], movie['directors'][0], movie['genres'][0], movie['title'], movie.get('plot')[0], movie['cover url']]
else:
raise Exception(f"No imdb match found for imdb link: {self.imdb_link}")
def get_research_articles(self, max_num):
# Search string for Google Scholar to look for.
# e.g. "{self.title} {self.director.name}" would equate to "Concussion Peter Landesman" for the movie Concussion.
search_str = f'{self.title} {self.director.name}'
output = f""
try:
pg = ProxyGenerator()
ip = os.environ['PROXY_IP']
pg.SingleProxy(http = ip, https = ip)
o = scholarly.use_proxy(pg)
search_query = scholarly.search_pubs(search_str)
for i in range(0, max_num):
curr = next(search_query)
# For debugging purposes, this is how you pretty print the search query's contents.
#scholarly.pprint(curr)
# Grab the title of the article.
title = curr['bib']['title']
# Begin our formatted html output for each found research article.
output += f"""
<li>
"""
# See if a publication url (i.e. curr['pub_url']) exists. If so, add an external link to it.
if 'pub_url' in curr:
output += f"""
<a target='_blank' href=\"{curr['pub_url']}\">{title}</a>
"""
else:
output += f"""
{title}
"""
output += f"""
<br>
"""
# Writes the abstract (i.e.curr['bib']['abstract']) if it exists.
if 'bib' in curr and 'abstract' in curr['bib']:
output += f"""
<p>{curr['bib']['abstract']}</p>
"""
output += f"""
</li>
"""
except Exception as e:
pass
# Useful for seeing errors in your terminal. Replace pass with the print statement below.
#print(sys.stderr, e)
return output
def save(self, *args, **kwargs):
super(Movie, self).save(*args, **kwargs)
# Uses a custom save to end date any subCases
orig = Movie.objects.get(id=self.id)
fields_to_update = []
try:
specs = self.get_video_stats()
orig.duration = specs[0]
orig.fps = specs[1]
orig.dimensions = specs[2]
fields_to_update.extend(['duration', 'fps', 'dimensions'])
except Exception as e:
pass
try:
imdb_stats = self.get_imdb_stats()
orig.title = imdb_stats[3]
orig.year = imdb_stats[0]
orig.thumbnail = imdb_stats[5]
# Checks if a director name already exists. If not, create and assign to the movie.
director = None
try:
directors = Director.objects.all()
for d in directors:
if (str(d) == str(imdb_stats[1])):
director = d
break
orig.director = director if director is not None else Director.objects.create(name=imdb_stats[1])
except:
orig.director = Director.objects.create(name=imdb_stats[1])
# Checks if a genre name already exists. If not, create and assign to the movie.
genre = None
try:
genres = Genre.objects.all()
for g in genres:
if (str(g) == str(imdb_stats[2])):
genre = g
break
genre = genre if genre is not None else Genre.objects.create(name=imdb_stats[2])
except:
genre = Genre.objects.create(name=imdb_stats[2])
orig.genre = genre
# Updates values only if their fields are left blank by the admin.
if not self.year:
fields_to_update.append('year')
if not self.genre:
fields_to_update.append('genre')
if not self.title:
fields_to_update.append('title')
if not self.director:
fields_to_update.append('director')
if not self.summary:
orig.summary = imdb_stats[4]
fields_to_update.append('summary')
if not self.thumbnail:
fields_to_update.append('thumbnail')
except Exception as e:
pass
# Searches for research articles by using a single proxy for a Google Scholar search query.
if not self.found_articles:
| orig.found_articles = orig.get_research_articles(self.max_num_find_articles)
fields_to_update.append('found_articles') | conditional_block |
|
main.py | (E-V)
ψ = -1/2ψ'' / (E-1/abs(r))
or, reversed:
ψ'' = -2(E - 1/abs(r))ψ
"""
# todo: Center it up? This approach lags.
# ψ_pp = np.diff(np.diff(ψ))
dx = (x[-1] - x[0]) / x.size
ψ_pp = np.diff(np.diff(ψ)) / dx
ψ_pp = np.append(ψ_pp, np.array([0, 0])) # make the lengths match
ψ_pp_ideal = -2 * (E - 1/np.abs(x)) * ψ
# plt.plot(x, ψ)
# plt.plot(x, ψ_pp)
# plt.xlim(0, 10)
# plt.show()
# For now, assume assume a single protein in the nucleus, at x=0.
ψ_ideal = -1/2 * ψ_pp / (E - 1/np.abs(x))
# plt.plot(x, ψ_ideal)
# plt.plot(x, ψ)
# plt.xlim(0, 10)
# plt.show()
plt.plot(x, ψ)
# plt.plot(x, ψ_pp_ideal)
plt.xlim(0, 10)
plt.show()
# result = (ψ - ψ_ideal) / ψ_ideal
result = (ψ_pp - ψ_pp_ideal) / ψ_pp_ideal
# plt.plot(x, result)
# plt.xlim(0, 10)
# plt.show()
return result
# def check_wf(ψ: Callable[(float, float), ]):
def check_wf_2d(ψ: ndarray):
"""Given a wave function as a set of discrete points, (Or a fn?) determine how much
it close it is to the schrodinger equation by analyzing the derivatives."""
pass
def run_check():
n = 1
E = -2 / (n + 1) ** 2
x, ψ = h_static(E)
print(check_wf_1d(x, ψ, E))
def calc_energy(n: int) -> float:
"""Numerically calculate the energy of a wave function generated
by `h_static`. For n=1, we'd like to see if we can back-calculate E = -1/2."""
E = -2 / (n + 1) ** 2
x, ψ = h_static(E)
# Calculate potential between the e- field and the nucleus point by integrating.
# todo: Let's do this manually first, then try to apply a scipy.integrate approach.
dx = 1
result = 0
ψ2 = np.conj(ψ) * ψ
sample_pts = np.arange(x[0], x[-1], dx)
for pt in sample_pts:
k = 1
Q = 1
V = k * Q / x
q = 1
E = V * q * np.interp([pt], x, ψ2)[0]
result += E / dx
return result
def h2_potential(dist: float) -> float:
"""Given a distance, calculate the potential energy between
2 n=1 S orbital hydrogen atoms"""
pass
def h2_potential(x: float) -> float:
"""Calcualte the electric potential between 2 hydrogen atoms"""
# Start with the perspectic of one atom. Calculate the interaction between
# its nucleus and the other atom's nucleus, and electron.
# Our convention will be attraction is positive potential.
n = 1
E = -2 / (n + 1) ** 2
H = Hydrogen3d([0, 1])
nuc_nuc_V = consts.k * consts.e**2 / x
dx = 2
dv = dx**3
nuc_elec_V = 0
# We'll say the molecules are at the same z and y coordinates,
# but separated on the x axis by input argument `x`.
# Sample point coordinates are centered on the non-POV atom.
# todo: You will gain much precision by taking sample areas
# todo closer together near the nucleus
# `sample_range` applies to all 3 dimensions.
sample_range = np.arange(-10, 10, dx)
sample_pts = []
for j in range(sample_range.size):
for k_ in range(sample_range.size):
for l in range(sample_range.size):
sample_pts.append(Pt(j, k_, l))
# sample_pts = sample_range # 1d
print("num samples: ", len(sample_pts))
for pt in sample_pts:
# We need to integrate over volume, eg by splitting up into
# small cubes.
# Offset the x value by the distance between nuclei.
r = sqrt((pt.x + x)**2 + pt.y**2 + pt.z**2)
ψ_local = H.value(r, 0, 0)
# Divide by the number of sample points: The total answer
# ψ^2 adds up to 1, so this weights each segment evenly.
elec_val = np.conj(ψ_local) * ψ_local
# 2 for both interactions
nuc_elec_V -= 2 * consts.k * consts.e * elec_val / pt.x * dv
elec_elec_V = 0
e_e_factor = len(sample_pts)**2
# todo: You have a problem: WFs past the nuclei aren't attracting/repelling
# todo in the correct direction!
for pt0 in sample_pts:
pass
r0 = sqrt(pt0.x ** 2 + pt0.y ** 2 + pt0.z ** 2)
# We're dealing with S orbitals for now; no need to pass anything beyond
# radius to the `value` method.
ψ_local0 = H.value(r0, 0, 0)
for pt1 in sample_pts:
# todo: We only need to calculate wfs for each pt once!
# todo: Current approach could slow it down
r1 = sqrt(pt1.x**2 + pt1.y**2 + pt1.z**2)
ψ_local1 = H.value(r1, 0, 0)
# These are localized for each pt.
dist = sqrt((pt1.x - pt0.x)**2 + (pt1.y - pt0.y)**2 + (pt1.z - pt0.z)**2)
elec_elec_V += consts.k * consts.e * ψ_local0 * ψ_local1 * dv
print(f"NN: {nuc_nuc_V}, NE: {nuc_elec_V}, EE: {elec_elec_V} Net: {nuc_nuc_V + nuc_elec_V + elec_elec_V}")
# potential etc from both elecs adn the other proton.
def h2_force_pov(x: float) -> float:
"""Calcualte the electric potential between 2 hydrogen atoms. In this
function, we view things from the perspective of the proton of one
of the atoms, and calculate everything else relative to it."""
# Start with the perspectic of one atom. Calculate the interaction between
# its nucleus and the other atom's nucleus, and electron.
# Our convention will be that towards our POV nucleus is positive;
# repulusion from it is negative.
H = Hydrogen3d([0, 1])
# Calculate proton-proton interaction.
nuc_nuc_V = consts.k * consts.e / x
nuc_nuc_F = Vec(nuc_nuc_V * consts.e / x, 0., 0.)
dx = 0.18
dv = dx**3
# We'll say the molecules are at the same z and y coordinates,
# but separated on the x axis by input argument `x`.
# Sample point coordinates are centered on the non-POV atom.
# todo: You will gain much precision by taking sample areas
# todo closer together near the nucleus
# `sample_range` applies to all 3 dimensions.
sample_range = np.arange(-12.1, 12.1, dx) # Don't let 0 be a pt
# todo: Manually entering the pts we want
# Becaus we're dealing with a 3rd power, we need to keep the sample pts minimal. The ones
# near the center should be more finely spaced. (Or perhaps tune the spacing dynamically
# based on changing slopes?)
# sample_range = np.array([-20, -15, -10, -9, -8.5, -8, -7.5, -7, -6.5, -6, -5.5, -5, -])
sample_pts = []
for j in sample_range:
for k_ in sample_range:
for l in sample_range:
sample_pts.append(Vec(j, k_, l))
print("num samples: ", len(sample_pts | ))
print("Sample range: ", sample_range)
# Calculate nucleus-e | conditional_block |
|
main.py |
from numpy import exp, ndarray, sqrt
import numpy as np
import matplotlib.pyplot as plt
import matplotlib
from consts import *
import consts
τ = 2 * np.pi
i = complex(0, 1)
# 2020-11-15
"""
One of your goals is to figure out if you can use hydrogen (1d?) WFs as basis functions to create
arbitrary solns to the Schrodinger equation, thereby making chemistry simulation and modeling
much more computationally efficient.
It appears that for Hydrogen atoms, you can use linear combinations of 1D WFs as basis functions
in 2 adn 3d by choosing the right coefficients, and the right modifier fn (sin, cos etc) across
θ and φ to apply to 2 and 3d situations.
You need to verify that this is correct adn quantify. A challenge is finding accurate 2D orbitals
to compare your results to, and in visualizing and/or quantifying your 3D results to compare
to real results in 3d.
In parallel to verifying this, assume it's right, and try to model a 2-nucleus system. For
example, a H2 molecule. Attempt, in 1D, to find a combination of H atomic orbitals (perhaps
offset in x) that create the H2 molecular orbitals. These orbitals you're attempting to
match can be taken from real data, or by integrating. (May need to break up integration
into three areas, to avoid singularities at each nucleus).
"""
@dataclass
class Hydrogen3d:
"""A Hydrogen 3d superposition"""
# todo: Or is it n = 1, 3, 5...
# coeffs: List[complex] # Positive coefficients: n = 0, 1, 2, 3...
n: int
l: int
m: int
x: np.ndarray
components: List[np.ndarray]
def __init__(self, coeffs: List[complex]):
self.coeffs = coeffs
self.components = []
# n = 1 # todo: Only odd 1d coeffs for now.
n = 0 # todo: Only odd 1d coeffs for now.
for c in self.coeffs:
E = -2 / (n + 1) ** 2
x, ψ = h_static_3d(E)
# if n == 1:
if n == 0:
self.x = x
self.components.append(c * ψ)
# n += 2
n += 1
def value(self, r: float, θ: float, φ: float) -> complex:
"""Get a single value."""
result = 0
for comp in self.components:
result += np.interp([r], self.x, comp)[0]
return result
def value_comp(self, x: float, j: int) -> complex:
"""Get a single value, from a specific component."""
return np.interp([x], self.x, self.components[j])[0]
def plot(self, range_: Tuple[float, float] = (-20, 20), shift: float = 0., size: int = 10_000, show: bool = True) -> None:
ψ = np.zeros(len(self.x), dtype=np.complex128)
for ψi in self.components:
ψ += ψi
# todo: DRY with other series'
plt.plot(self.x + shift, ψ.real)
# plt.plot(self.x, ψ.imag)
# plt.xlim(range_[0], range_[1])
plt.xlim(0, range_[1])
if show:
plt.show()
@dataclass
class Pt:
x: float
y: float
z: float
@dataclass
class Vec:
x: float
y: float
z: float
def __add__(self, other: 'Vec') -> 'Vec':
return Vec(self.x + other.x, self.y + other.y, self.z + other.z)
def __sub__(self, other: 'Vec') -> 'Vec':
return Vec(self.x - other.x, self.y - other.y, self.z - other.z)
def scalar_mul(self, val: float) -> 'Vec':
return Vec(val * self.x, val * self.y, val * self.z)
def length(self) -> float:
return sqrt(self.x**2 + self.y**2 + self.z**2)
matplotlib.use("Qt5Agg")
# import seaborn as sns
# import plotly
# import plotly.graph_objects as go
# A global state var
V_prev: Callable = lambda sx: 0
# Lookin into matrix mechanics, and Feynman path integral approaches too
# orbitals are characterized (in simple cases) by quantum numbers n, l, and m, corresponding to
# energy, angular momentum, and magnetic (ang momentum vec component)
# spin?
# Do electrons (electrically) interact with themselves?
# Breaking up a numerical problem into a number of solveable analytic ones??? Eg set up
# an arbitrary V as a series of step Vs which have anal solns
# Free variables: 2? Energy, and ψ_p_0(ψ). Eg we can set ψ to what we wish, find the ψ_p that
# works with it (and the E set), then normalize.
PRECISION = 100_000
@dataclass
class Nucleus:
n_prot: float
n_neut: float
sx: float
vx: float
def mass(self):
return self.n_prot * m_p + self.n_neut * m_n
def charge(self):
# Charge from protons only.
return self.n_prot * e
def nuc_pot(nuclei: Iterable[Nucleus], sx: float) -> float:
result = 0
for nuclei in nuclei:
# Coulomb potential
result -= e / abs(nuclei.sx - sx)
return result
def ti_schrod_rhs(
E: float, V: Callable, x: float, y: Tuple[complex, complex]
) -> Tuple[complex, complex]:
"""
d²ψ/dx² = 2m/ħ² * (V(x) - E)ψ
"""
ψ, φ = y
ψ_p = φ
φ_p = 2 * m_e / ħ ** 2 * (V(x) - E) * ψ
return ψ_p, φ_p
def solve(E: float, V: Callable, ψ0: float, ψ_p0: float, x_span: Tuple[float, float]):
"""
Calculate the wave function for electrons in an arbitrary potential, at a single snapshot
in time.
"""
rhs = partial(ti_schrod_rhs, E, V)
return solve_ivp(
rhs, x_span, (ψ0, ψ_p0), t_eval=np.linspace(x_span[0], x_span[1], PRECISION)
)
def h_static(E: float, normalize=True) -> Tuple[ndarray, ndarray]:
ψ0 = 0
ψ_p0 = 0.1
x_span = (-100, 0.0000001)
V_elec = partial(nuc_pot, [Nucleus(1, 0, 0, 0)])
# V_elec = partial(nuc_pot, [Nucleus(1, 0, 0, 0), Nucleus(1, 0, 1, 0)])
# Left and right of the x=0 coulomb singularity. Assume odd solution around x=0.
soln_orig = solve(E, V_elec, ψ0, ψ_p0, x_span)
soln_left = soln_orig.y[0]
soln_right = np.flip(soln_left)
soln = np.concatenate([soln_left, -soln_right])
x = np.concatenate([soln_orig.t, np.flip(-soln_orig.t)])
if normalize:
norm = simps(np.conj(soln) * soln, x=x)
return x, soln / norm ** 0.5
return x, soln
def h_static_3d(E: float, normalize=False) -> Tuple[ndarray, ndarray]:
"""We create the radial part of the 3d version from the "radial density" information."""
# todo: Why don't we get a result if we fail to normalize here?
# Normalize the radial part, not the whole thing; this gives us reasonable values,
# without dealing with the asymptote near the origin.
r, ψ = h_static(E, normalize=True)
ψ = sqrt(ψ**2 / r**2)
# Post-process by flipping between 0s, to make up for info lost
# during square root.
ε = 1e-3 # thresh for hit a 0.
ψ_processed = np.copy(ψ)
in_inversion = False
slope_neg_prev = True
for j in range(ψ.size):
if j == | from functools import partial
from typing import List, Iterable, Callable, Tuple | random_line_split |
|
main.py | (self.x - other.x, self.y - other.y, self.z - other.z)
def scalar_mul(self, val: float) -> 'Vec':
return Vec(val * self.x, val * self.y, val * self.z)
def length(self) -> float:
return sqrt(self.x**2 + self.y**2 + self.z**2)
matplotlib.use("Qt5Agg")
# import seaborn as sns
# import plotly
# import plotly.graph_objects as go
# A global state var
V_prev: Callable = lambda sx: 0
# Lookin into matrix mechanics, and Feynman path integral approaches too
# orbitals are characterized (in simple cases) by quantum numbers n, l, and m, corresponding to
# energy, angular momentum, and magnetic (ang momentum vec component)
# spin?
# Do electrons (electrically) interact with themselves?
# Breaking up a numerical problem into a number of solveable analytic ones??? Eg set up
# an arbitrary V as a series of step Vs which have anal solns
# Free variables: 2? Energy, and ψ_p_0(ψ). Eg we can set ψ to what we wish, find the ψ_p that
# works with it (and the E set), then normalize.
PRECISION = 100_000
@dataclass
class Nucleus:
n_prot: float
n_neut: float
sx: float
vx: float
def mass(self):
return self.n_prot * m_p + self.n_neut * m_n
def charge(self):
# Charge from protons only.
return self.n_prot * e
def nuc_pot(nuclei: Iterable[Nucleus], sx: float) -> float:
result = 0
for nuclei in nuclei:
# Coulomb potential
result -= e / abs(nuclei.sx - sx)
return result
def ti_schrod_rhs(
| : Callable, x: float, y: Tuple[complex, complex]
) -> Tuple[complex, complex]:
"""
d²ψ/dx² = 2m/ħ² * (V(x) - E)ψ
"""
ψ, φ = y
ψ_p = φ
φ_p = 2 * m_e / ħ ** 2 * (V(x) - E) * ψ
return ψ_p, φ_p
def solve(E: float, V: Callable, ψ0: float, ψ_p0: float, x_span: Tuple[float, float]):
"""
Calculate the wave function for electrons in an arbitrary potential, at a single snapshot
in time.
"""
rhs = partial(ti_schrod_rhs, E, V)
return solve_ivp(
rhs, x_span, (ψ0, ψ_p0), t_eval=np.linspace(x_span[0], x_span[1], PRECISION)
)
def h_static(E: float, normalize=True) -> Tuple[ndarray, ndarray]:
ψ0 = 0
ψ_p0 = 0.1
x_span = (-100, 0.0000001)
V_elec = partial(nuc_pot, [Nucleus(1, 0, 0, 0)])
# V_elec = partial(nuc_pot, [Nucleus(1, 0, 0, 0), Nucleus(1, 0, 1, 0)])
# Left and right of the x=0 coulomb singularity. Assume odd solution around x=0.
soln_orig = solve(E, V_elec, ψ0, ψ_p0, x_span)
soln_left = soln_orig.y[0]
soln_right = np.flip(soln_left)
soln = np.concatenate([soln_left, -soln_right])
x = np.concatenate([soln_orig.t, np.flip(-soln_orig.t)])
if normalize:
norm = simps(np.conj(soln) * soln, x=x)
return x, soln / norm ** 0.5
return x, soln
def h_static_3d(E: float, normalize=False) -> Tuple[ndarray, ndarray]:
"""We create the radial part of the 3d version from the "radial density" information."""
# todo: Why don't we get a result if we fail to normalize here?
# Normalize the radial part, not the whole thing; this gives us reasonable values,
# without dealing with the asymptote near the origin.
r, ψ = h_static(E, normalize=True)
ψ = sqrt(ψ**2 / r**2)
# Post-process by flipping between 0s, to make up for info lost
# during square root.
ε = 1e-3 # thresh for hit a 0.
ψ_processed = np.copy(ψ)
in_inversion = False
slope_neg_prev = True
for j in range(ψ.size):
if j == 0: # We use slopes; don't mis-index
ψ_processed[j] = ψ[j]
continue
slope_neg = ψ[j] < ψ[j-1]
# Just started or ended an inversion.
if ψ[j] <= ε and slope_neg != slope_neg_prev:
in_inversion = not in_inversion
if in_inversion:
ψ_processed[j] = -ψ[j]
else:
ψ_processed[j] = ψ[j]
slope_neg_prev = slope_neg
if normalize:
norm = simps(np.conj(ψ_processed) * ψ_processed, x=r)
return r, ψ_processed / norm ** 0.5
return r, ψ_processed
def plot_h_static(n: int = 1):
"""This 1d model represents the radial component of the wave function;
ie all of a 2d shell condensed down 2 dimensions to a point."""
# Negative E implies bound state; positive scattering.
# ψ_p0 should be 0 for continuity across the origin.
# E should be a whittaker energy, ie -1/2, -2/9, -1/8, -.08 etc
# Only odd states (n = 1, 3, 5 etc) correspond to 3d H atom.
E = -2 / (n + 1) ** 2
x, ψ = h_static(E)
ψ = ψ**2
fig, ax = plt.subplots()
ax.plot(x, ψ)
ax.grid(True)
plt.xlim(0, 20)
plt.show()
def plot_h_static_3d(n: int = 1):
"""Like H static, but perhaps this is the right model for 3D."""
# todo: Major DRY
E = -2 / (n + 1) ** 2
x, ψ = h_static_3d(E)
fig, ax = plt.subplots()
ax.plot(x, ψ)
ax.grid(True)
plt.xlim(0, 20)
plt.ylim(-0.02, 0.02)
plt.show()
def check_wf_1d(x: ndarray, ψ: ndarray, E: float) -> ndarray:
"""Given a wave function as a set of discrete points, (Or a fn?) determine how much
it close it is to the schrodinger equation by analyzing the derivatives.
The result is a percent diff.
ψ = -1/2ψ'' / (E-V)
ψ = -1/2ψ'' / (E-1/abs(r))
or, reversed:
ψ'' = -2(E - 1/abs(r))ψ
"""
# todo: Center it up? This approach lags.
# ψ_pp = np.diff(np.diff(ψ))
dx = (x[-1] - x[0]) / x.size
ψ_pp = np.diff(np.diff(ψ)) / dx
ψ_pp = np.append(ψ_pp, np.array([0, 0])) # make the lengths match
ψ_pp_ideal = -2 * (E - 1/np.abs(x)) * ψ
# plt.plot(x, ψ)
# plt.plot(x, ψ_pp)
# plt.xlim(0, 10)
# plt.show()
# For now, assume assume a single protein in the nucleus, at x=0.
ψ_ideal = -1/2 * ψ_pp / (E - 1/np.abs(x))
# plt.plot(x, ψ_ideal)
# plt.plot(x, ψ)
# plt.xlim(0, 10)
# plt.show()
plt.plot(x, ψ)
# plt.plot(x, ψ_pp_ideal)
plt.xlim(0, 10)
plt.show()
# result = (ψ - ψ_ideal) / ψ_ideal
result = (ψ_pp - ψ_pp_ideal) / ψ_pp_ideal
# plt.plot(x, result)
# plt.xlim(0, 10)
# plt.show()
return result
# def check_wf( | E: float, V | identifier_name |
main.py | : float
sx: float
vx: float
def mass(self):
return self.n_prot * m_p + self.n_neut * m_n
def charge(self):
# Charge from protons only.
return self.n_prot * e
def nuc_pot(nuclei: Iterable[Nucleus], sx: float) -> float:
result = 0
for nuclei in nuclei:
# Coulomb potential
result -= e / abs(nuclei.sx - sx)
return result
def ti_schrod_rhs(
E: float, V: Callable, x: float, y: Tuple[complex, complex]
) -> Tuple[complex, complex]:
"""
d²ψ/dx² = 2m/ħ² * (V(x) - E)ψ
"""
ψ, φ = y
ψ_p = φ
φ_p = 2 * m_e / ħ ** 2 * (V(x) - E) * ψ
return ψ_p, φ_p
def solve(E: float, V: Callable, ψ0: float, ψ_p0: float, x_span: Tuple[float, float]):
"""
Calculate the wave function for electrons in an arbitrary potential, at a single snapshot
in time.
"""
rhs = partial(ti_schrod_rhs, E, V)
return solve_ivp(
rhs, x_span, (ψ0, ψ_p0), t_eval=np.linspace(x_span[0], x_span[1], PRECISION)
)
def h_static(E: float, normalize=True) -> Tuple[ndarray, ndarray]:
ψ0 = 0
ψ_p0 = 0.1
x_span = (-100, 0.0000001)
V_elec = partial(nuc_pot, [Nucleus(1, 0, 0, 0)])
# V_elec = partial(nuc_pot, [Nucleus(1, 0, 0, 0), Nucleus(1, 0, 1, 0)])
# Left and right of the x=0 coulomb singularity. Assume odd solution around x=0.
soln_orig = solve(E, V_elec, ψ0, ψ_p0, x_span)
soln_left = soln_orig.y[0]
soln_right = np.flip(soln_left)
soln = np.concatenate([soln_left, -soln_right])
x = np.concatenate([soln_orig.t, np.flip(-soln_orig.t)])
if normalize:
norm = simps(np.conj(soln) * soln, x=x)
return x, soln / norm ** 0.5
return x, soln
def h_static_3d(E: float, normalize=False) -> Tuple[ndarray, ndarray]:
"""We create the radial part of the 3d version from the "radial density" information."""
# todo: Why don't we get a result if we fail to normalize here?
# Normalize the radial part, not the whole thing; this gives us reasonable values,
# without dealing with the asymptote near the origin.
r, ψ = h_static(E, normalize=True)
ψ = sqrt(ψ**2 / r**2)
# Post-process by flipping between 0s, to make up for info lost
# during square root.
ε = 1e-3 # thresh for hit a 0.
ψ_processed = np.copy(ψ)
in_inversion = False
slope_neg_prev = True
for j in range(ψ.size):
if j == 0: # We use slopes; don't mis-index
ψ_processed[j] = ψ[j]
continue
slope_neg = ψ[j] < ψ[j-1]
# Just started or ended an inversion.
if ψ[j] <= ε and slope_neg != slope_neg_prev:
in_inversion = not in_inversion
if in_inversion:
ψ_processed[j] = -ψ[j]
else:
ψ_processed[j] = ψ[j]
slope_neg_prev = slope_neg
if normalize:
norm = simps(np.conj(ψ_processed) * ψ_processed, x=r)
return r, ψ_processed / norm ** 0.5
return r, ψ_processed
def plot_h_static(n: int = 1):
"""This 1d model represents the radial component of the wave function;
ie all of a 2d shell condensed down 2 dimensions to a point."""
# Negative E implies bound state; positive scattering.
# ψ_p0 should be 0 for continuity across the origin.
# E should be a whittaker energy, ie -1/2, -2/9, -1/8, -.08 etc
# Only odd states (n = 1, 3, 5 etc) correspond to 3d H atom.
E = -2 / (n + 1) ** 2
x, ψ = h_static(E)
ψ = ψ**2
fig, ax = plt.subplots()
ax.plot(x, ψ)
ax.grid(True)
plt.xlim(0, 20)
plt.show()
def plot_h_static_3d(n: int = 1):
"""Like H static, but perhaps this is the right model for 3D."""
# todo: Major DRY
E = -2 / (n + 1) ** 2
x, ψ = h_static_3d(E)
fig, ax = plt.subplots()
ax.plot(x, ψ)
ax.grid(True)
plt.xlim(0, 20)
plt.ylim(-0.02, 0.02)
plt.show()
def check_wf_1d(x: ndarray, ψ: ndarray, E: float) -> ndarray:
"""Given a wave function as a set of discrete points, (Or a fn?) determine how much
it close it is to the schrodinger equation by analyzing the derivatives.
The result is a percent diff.
ψ = -1/2ψ'' / (E-V)
ψ = -1/2ψ'' / (E-1/abs(r))
or, reversed:
ψ'' = -2(E - 1/abs(r))ψ
"""
# todo: Center it up? This approach lags.
# ψ_pp = np.diff(np.diff(ψ))
dx = (x[-1] - x[0]) / x.size
ψ_pp = np.diff(np.diff(ψ)) / dx
ψ_pp = np.append(ψ_pp, np.array([0, 0])) # make the lengths match
ψ_pp_ideal = -2 * (E - 1/np.abs(x)) * ψ
# plt.plot(x, ψ)
# plt.plot(x, ψ_pp)
# plt.xlim(0, 10)
# plt.show()
# For now, assume assume a single protein in the nucleus, at x=0.
ψ_ideal = -1/2 * ψ_pp / (E - 1/np.abs(x))
# plt.plot(x, ψ_ideal)
# plt.plot(x, ψ)
# plt.xlim(0, 10)
# plt.show()
plt.plot(x, ψ)
# plt.plot(x, ψ_pp_ideal)
plt.xlim(0, 10)
plt.show()
# result = (ψ - ψ_ideal) / ψ_ideal
result = (ψ_pp - ψ_pp_ideal) / ψ_pp_ideal
# plt.plot(x, result)
# plt.xlim(0, 10)
# plt.show()
return result
# def check_wf(ψ: Callable[(float, float), ]):
def check_wf_2d(ψ: ndarray):
"""Given a wave function as a set of discrete points, (Or a fn?) determine how much
it close it is to the schrodinger equation by analyzing the derivatives."""
pass
def run_check():
n = 1
E = -2 / (n + 1) ** 2
x, ψ = h_static(E)
print(check_wf_1d(x, ψ, E))
def calc_energy(n: int) -> float:
"""Numerically calculate the energy of a wave function generated
by `h_static`. For n=1, we'd lik | e to see if we can back-calculate E = -1/2."""
E = -2 / (n + 1) ** 2
x, ψ = h_static(E)
# Calculate potential between the e- field and the nucleus point by integrating.
# todo: Let's do this manually first, then try to apply a scipy.integrate approach.
dx = 1
result = 0
ψ2 = np.conj(ψ) * ψ
sample_pts = np.arange(x[0], x[-1], dx)
for pt in sample_pts:
k = 1
Q = 1
V = k * Q / x
q = 1 | identifier_body |
|
syncmgr.rs | synced up to height {} with hash {}",
height, hash
)
}
Event::Syncing { current, best } => write!(fmt, "Syncing headers {}/{}", current, best),
Event::BlockConnected { height, header } => {
write!(
fmt,
"Block {} connected at height {}",
header.block_hash(),
height
)
}
Event::BlockDisconnected { height, header } => {
write!(
fmt,
"Block {} disconnected at height {}",
header.block_hash(),
height
)
}
Event::BlockDiscovered(from, hash) => {
write!(fmt, "{}: Discovered new block: {}", from, &hash)
}
Event::StaleTip(last_update) => {
write!(
fmt,
"Potential stale tip detected (last update was {})",
last_update
)
}
}
}
}
/// A `getheaders` request sent to a peer.
#[derive(Clone, Debug, PartialEq, Eq)]
struct GetHeaders {
/// Locators hashes.
locators: Locators,
/// Time at which the request was sent.
sent_at: LocalTime,
/// What to do if this request times out.
on_timeout: OnTimeout,
} | pub fn new(config: Config, rng: fastrand::Rng, upstream: U, clock: C) -> Self {
let peers = AddressBook::new(rng.clone());
let last_tip_update = None;
let last_peer_sample = None;
let last_idle = None;
let inflight = HashMap::with_hasher(rng.into());
Self {
peers,
config,
last_tip_update,
last_peer_sample,
last_idle,
inflight,
upstream,
clock,
}
}
/// Initialize the sync manager. Should only be called once.
pub fn initialize<T: BlockReader>(&mut self, tree: &T) {
// TODO: `tip` should return the height.
let (hash, _) = tree.tip();
let height = tree.height();
self.idle(tree);
self.upstream.event(Event::Synced(hash, height));
}
/// Called periodically.
pub fn idle<T: BlockReader>(&mut self, tree: &T) {
let now = self.clock.local_time();
// Nb. The idle timeout is very long: as long as the block interval.
// This shouldn't be a problem, as the sync manager can make progress without it.
if now - self.last_idle.unwrap_or_default() >= IDLE_TIMEOUT {
if !self.sync(tree) {
self.sample_peers(tree);
}
self.last_idle = Some(now);
self.upstream.set_timer(IDLE_TIMEOUT);
}
}
/// Called when a new peer was negotiated.
pub fn peer_negotiated<T: BlockReader>(
&mut self,
socket: Socket,
height: Height,
services: ServiceFlags,
preferred: bool,
link: Link,
tree: &T,
) {
if link.is_outbound() && !services.has(REQUIRED_SERVICES) {
return;
}
if height > self.best_height().unwrap_or_else(|| tree.height()) {
self.upstream.event(Event::PeerHeightUpdated { height });
}
self.register(socket, height, preferred, link);
self.sync(tree);
}
/// Called when a peer disconnected.
pub fn peer_disconnected(&mut self, id: &PeerId) {
self.unregister(id);
}
/// Called when we received a `getheaders` message from a peer.
pub fn received_getheaders<T: BlockReader>(
&mut self,
addr: &PeerId,
(locator_hashes, stop_hash): Locators,
tree: &T,
) {
let max = self.config.max_message_headers;
if self.is_syncing() || max == 0 {
return;
}
let headers = tree.locate_headers(&locator_hashes, stop_hash, max);
if headers.is_empty() {
return;
}
self.upstream.headers(*addr, headers);
}
/// Import blocks into our block tree.
pub fn import_blocks<T: BlockTree, I: Iterator<Item = BlockHeader>>(
&mut self,
blocks: I,
tree: &mut T,
) -> Result<ImportResult, Error> {
match tree.import_blocks(blocks, &self.clock) {
Ok(ImportResult::TipChanged(header, tip, height, reverted, connected)) => {
let result = ImportResult::TipChanged(
header,
tip,
height,
reverted.clone(),
connected.clone(),
);
for (height, header) in reverted {
self.upstream
.event(Event::BlockDisconnected { height, header });
}
for (height, header) in connected {
self.upstream
.event(Event::BlockConnected { height, header });
}
self.upstream.event(Event::Synced(tip, height));
self.broadcast_tip(&tip, tree);
Ok(result)
}
Ok(result @ ImportResult::TipUnchanged) => Ok(result),
Err(err) => Err(err),
}
}
/// Called when we receive headers from a peer.
pub fn received_headers<T: BlockTree>(
&mut self,
from: &PeerId,
headers: Vec<BlockHeader>,
clock: &impl Clock,
tree: &mut T,
) -> Result<ImportResult, store::Error> {
let request = self.inflight.remove(from);
let headers = if let Some(headers) = NonEmpty::from_vec(headers) {
headers
} else {
return Ok(ImportResult::TipUnchanged);
};
let length = headers.len();
if length > MAX_MESSAGE_HEADERS {
log::debug!("Received more than maximum headers allowed from {}", from);
self.record_misbehavior(from);
self.upstream
.disconnect(*from, DisconnectReason::PeerMisbehaving("too many headers"));
return Ok(ImportResult::TipUnchanged);
}
// When unsolicited, we don't want to process too many headers in case of a DoS.
if length > MAX_UNSOLICITED_HEADERS && request.is_none() {
log::debug!("Received {} unsolicited headers from {}", length, from);
return Ok(ImportResult::TipUnchanged);
}
if let Some(peer) = self.peers.get_mut(from) {
peer.last_active = Some(clock.local_time());
} else {
return Ok(ImportResult::TipUnchanged);
}
log::debug!("[sync] Received {} block header(s) from {}", length, from);
let root = headers.first().block_hash();
let best = headers.last().block_hash();
if tree.contains(&best) {
return Ok(ImportResult::TipUnchanged);
}
match self.import_blocks(headers.into_iter(), tree) {
Ok(ImportResult::TipUnchanged) => {
// Try to find a common ancestor that leads up to the first header in
// the list we received.
let locators = (tree.locator_hashes(tree.height()), root);
let timeout = self.config.request_timeout;
self.request(*from, locators, timeout, OnTimeout::Ignore);
Ok(ImportResult::TipUnchanged)
}
Ok(ImportResult::TipChanged(header, tip, height, reverted, connected)) => {
// Update peer height.
if let Some(peer) = self.peers.get_mut(from) {
if height > peer.height {
peer.tip = tip;
peer.height = height;
}
}
// Keep track of when we last updated our tip. This is useful to check
// whether our tip is stale.
self.last_tip_update = Some(clock.local_time());
// If we received less than the maximum number of headers, we must be in sync.
// Otherwise, ask for the next batch of headers.
if length < MAX_MESSAGE_HEADERS {
// If these headers were unsolicited, we may already be ready/synced.
// Otherwise, we're finally in sync.
self.broadcast_tip(&tip, tree);
self.sync(tree);
} else {
let locators = (vec![tip], BlockHash::all_zeros());
let timeout = self.config.request_timeout;
self.request(*from, locators, timeout, OnTimeout::Disconnect);
}
Ok(ImportResult::TipChanged(
header, tip, height, reverted, connected,
))
}
Err(err) => self
.handle_error(from, err)
.map(|()| ImportResult::TipUnchanged),
}
}
fn request(
&mut self,
addr: PeerId,
locators: Locators,
timeout: LocalDuration,
on_timeout: OnTimeout,
) {
// Don't request more than once from the same peer.
if self.inflight.contains_key(&addr) {
return;
}
if let Some(peer) = self.peers.get_mut(&addr) {
debug_assert!(peer.last_asked.as_ref() != Some(&locators));
peer.last_asked = Some(locators.clone());
let sent_at = self.clock.local_time();
let req = GetHeaders {
locators,
sent_at,
on_timeout,
};
self.inflight.insert(addr, req.clone());
self.upstream.get_headers(addr, req.locators |
impl<U: SetTimer + Disconnect + Wire<Event>, C: Clock> SyncManager<U, C> {
/// Create a new sync manager. | random_line_split |
syncmgr.rs | {
/// A block was added to the main chain.
BlockConnected {
/// Block height.
height: Height,
/// Block header.
header: BlockHeader,
},
/// A block was removed from the main chain.
BlockDisconnected {
/// Block height.
height: Height,
/// Block header.
header: BlockHeader,
},
/// A new block was discovered via a peer.
BlockDiscovered(PeerId, BlockHash),
/// Syncing headers.
Syncing {
/// Current block header height.
current: Height,
/// Best known block header height.
best: Height,
},
/// Synced up to the specified hash and height.
Synced(BlockHash, Height),
/// Potential stale tip detected on the active chain.
StaleTip(LocalTime),
/// Peer misbehaved.
PeerMisbehaved(PeerId),
/// Peer height updated.
PeerHeightUpdated {
/// Best height known.
height: Height,
},
}
impl std::fmt::Display for Event {
fn fmt(&self, fmt: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
match self {
Event::PeerMisbehaved(addr) => {
write!(fmt, "{}: Peer misbehaved", addr)
}
Event::PeerHeightUpdated { height } => {
write!(fmt, "Peer height updated to {}", height)
}
Event::Synced(hash, height) => {
write!(
fmt,
"Headers synced up to height {} with hash {}",
height, hash
)
}
Event::Syncing { current, best } => write!(fmt, "Syncing headers {}/{}", current, best),
Event::BlockConnected { height, header } => {
write!(
fmt,
"Block {} connected at height {}",
header.block_hash(),
height
)
}
Event::BlockDisconnected { height, header } => {
write!(
fmt,
"Block {} disconnected at height {}",
header.block_hash(),
height
)
}
Event::BlockDiscovered(from, hash) => {
write!(fmt, "{}: Discovered new block: {}", from, &hash)
}
Event::StaleTip(last_update) => {
write!(
fmt,
"Potential stale tip detected (last update was {})",
last_update
)
}
}
}
}
/// A `getheaders` request sent to a peer.
#[derive(Clone, Debug, PartialEq, Eq)]
struct GetHeaders {
/// Locators hashes.
locators: Locators,
/// Time at which the request was sent.
sent_at: LocalTime,
/// What to do if this request times out.
on_timeout: OnTimeout,
}
impl<U: SetTimer + Disconnect + Wire<Event>, C: Clock> SyncManager<U, C> {
/// Create a new sync manager.
pub fn new(config: Config, rng: fastrand::Rng, upstream: U, clock: C) -> Self {
let peers = AddressBook::new(rng.clone());
let last_tip_update = None;
let last_peer_sample = None;
let last_idle = None;
let inflight = HashMap::with_hasher(rng.into());
Self {
peers,
config,
last_tip_update,
last_peer_sample,
last_idle,
inflight,
upstream,
clock,
}
}
/// Initialize the sync manager. Should only be called once.
pub fn initialize<T: BlockReader>(&mut self, tree: &T) {
// TODO: `tip` should return the height.
let (hash, _) = tree.tip();
let height = tree.height();
self.idle(tree);
self.upstream.event(Event::Synced(hash, height));
}
/// Called periodically.
pub fn idle<T: BlockReader>(&mut self, tree: &T) {
let now = self.clock.local_time();
// Nb. The idle timeout is very long: as long as the block interval.
// This shouldn't be a problem, as the sync manager can make progress without it.
if now - self.last_idle.unwrap_or_default() >= IDLE_TIMEOUT {
if !self.sync(tree) {
self.sample_peers(tree);
}
self.last_idle = Some(now);
self.upstream.set_timer(IDLE_TIMEOUT);
}
}
/// Called when a new peer was negotiated.
pub fn peer_negotiated<T: BlockReader>(
&mut self,
socket: Socket,
height: Height,
services: ServiceFlags,
preferred: bool,
link: Link,
tree: &T,
) {
if link.is_outbound() && !services.has(REQUIRED_SERVICES) {
return;
}
if height > self.best_height().unwrap_or_else(|| tree.height()) {
self.upstream.event(Event::PeerHeightUpdated { height });
}
self.register(socket, height, preferred, link);
self.sync(tree);
}
/// Called when a peer disconnected.
pub fn peer_disconnected(&mut self, id: &PeerId) {
self.unregister(id);
}
/// Called when we received a `getheaders` message from a peer.
pub fn received_getheaders<T: BlockReader>(
&mut self,
addr: &PeerId,
(locator_hashes, stop_hash): Locators,
tree: &T,
) {
let max = self.config.max_message_headers;
if self.is_syncing() || max == 0 {
return;
}
let headers = tree.locate_headers(&locator_hashes, stop_hash, max);
if headers.is_empty() {
return;
}
self.upstream.headers(*addr, headers);
}
/// Import blocks into our block tree.
pub fn import_blocks<T: BlockTree, I: Iterator<Item = BlockHeader>>(
&mut self,
blocks: I,
tree: &mut T,
) -> Result<ImportResult, Error> {
match tree.import_blocks(blocks, &self.clock) {
Ok(ImportResult::TipChanged(header, tip, height, reverted, connected)) => {
let result = ImportResult::TipChanged(
header,
tip,
height,
reverted.clone(),
connected.clone(),
);
for (height, header) in reverted {
self.upstream
.event(Event::BlockDisconnected { height, header });
}
for (height, header) in connected {
self.upstream
.event(Event::BlockConnected { height, header });
}
self.upstream.event(Event::Synced(tip, height));
self.broadcast_tip(&tip, tree);
Ok(result)
}
Ok(result @ ImportResult::TipUnchanged) => Ok(result),
Err(err) => Err(err),
}
}
/// Called when we receive headers from a peer.
pub fn received_headers<T: BlockTree>(
&mut self,
from: &PeerId,
headers: Vec<BlockHeader>,
clock: &impl Clock,
tree: &mut T,
) -> Result<ImportResult, store::Error> {
let request = self.inflight.remove(from);
let headers = if let Some(headers) = NonEmpty::from_vec(headers) {
headers
} else {
return Ok(ImportResult::TipUnchanged);
};
let length = headers.len();
if length > MAX_MESSAGE_HEADERS {
log::debug!("Received more than maximum headers allowed from {}", from);
self.record_misbehavior(from);
self.upstream
.disconnect(*from, DisconnectReason::PeerMisbehaving("too many headers"));
return Ok(ImportResult::TipUnchanged);
}
// When unsolicited, we don't want to process too many headers in case of a DoS.
if length > MAX_UNSOLICITED_HEADERS && request.is_none() {
log::debug!("Received {} unsolicited headers from {}", length, from);
return Ok(ImportResult::TipUnchanged);
}
if let Some(peer) = self.peers.get_mut(from) {
peer.last_active = Some(clock.local_time());
} else {
return Ok(ImportResult::TipUnchanged);
}
log::debug!("[sync] Received {} block header(s) from {}", length, from);
let root = headers.first().block_hash();
let best = headers.last().block_hash();
if tree.contains(&best) {
return Ok(ImportResult::TipUnchanged);
}
match self.import_blocks(headers.into_iter(), tree) {
Ok(ImportResult::TipUnchanged) => {
// Try to find a common ancestor that leads up to the first header in
// the list we received.
let locators = (tree.locator_hashes(tree.height()), root);
let timeout = self.config.request_timeout;
self.request(*from, locators, timeout, OnTimeout::Ignore);
Ok(ImportResult::TipUnchanged)
}
Ok(ImportResult::TipChanged(header, tip, height, reverted, connected)) => {
// Update peer height.
if let Some(peer) = self.peers.get_mut(from) {
if height > peer.height {
peer.tip = tip;
peer.height = height;
}
}
// Keep track of when we last updated our tip. This is useful to check
// whether our tip is stale.
self.last_tip_update = Some(clock.local_time());
// If we received less than the maximum number of headers, we must be in sync.
// | Event | identifier_name |
|
syncmgr.rs | up to height {} with hash {}",
height, hash
)
}
Event::Syncing { current, best } => write!(fmt, "Syncing headers {}/{}", current, best),
Event::BlockConnected { height, header } => {
write!(
fmt,
"Block {} connected at height {}",
header.block_hash(),
height
)
}
Event::BlockDisconnected { height, header } => {
write!(
fmt,
"Block {} disconnected at height {}",
header.block_hash(),
height
)
}
Event::BlockDiscovered(from, hash) => {
write!(fmt, "{}: Discovered new block: {}", from, &hash)
}
Event::StaleTip(last_update) => {
write!(
fmt,
"Potential stale tip detected (last update was {})",
last_update
)
}
}
}
}
/// A `getheaders` request sent to a peer.
#[derive(Clone, Debug, PartialEq, Eq)]
struct GetHeaders {
/// Locators hashes.
locators: Locators,
/// Time at which the request was sent.
sent_at: LocalTime,
/// What to do if this request times out.
on_timeout: OnTimeout,
}
impl<U: SetTimer + Disconnect + Wire<Event>, C: Clock> SyncManager<U, C> {
/// Create a new sync manager.
pub fn new(config: Config, rng: fastrand::Rng, upstream: U, clock: C) -> Self {
let peers = AddressBook::new(rng.clone());
let last_tip_update = None;
let last_peer_sample = None;
let last_idle = None;
let inflight = HashMap::with_hasher(rng.into());
Self {
peers,
config,
last_tip_update,
last_peer_sample,
last_idle,
inflight,
upstream,
clock,
}
}
/// Initialize the sync manager. Should only be called once.
pub fn initialize<T: BlockReader>(&mut self, tree: &T) {
// TODO: `tip` should return the height.
let (hash, _) = tree.tip();
let height = tree.height();
self.idle(tree);
self.upstream.event(Event::Synced(hash, height));
}
/// Called periodically.
pub fn idle<T: BlockReader>(&mut self, tree: &T) {
let now = self.clock.local_time();
// Nb. The idle timeout is very long: as long as the block interval.
// This shouldn't be a problem, as the sync manager can make progress without it.
if now - self.last_idle.unwrap_or_default() >= IDLE_TIMEOUT {
if !self.sync(tree) {
self.sample_peers(tree);
}
self.last_idle = Some(now);
self.upstream.set_timer(IDLE_TIMEOUT);
}
}
/// Called when a new peer was negotiated.
pub fn peer_negotiated<T: BlockReader>(
&mut self,
socket: Socket,
height: Height,
services: ServiceFlags,
preferred: bool,
link: Link,
tree: &T,
) |
/// Called when a peer disconnected.
pub fn peer_disconnected(&mut self, id: &PeerId) {
self.unregister(id);
}
/// Called when we received a `getheaders` message from a peer.
pub fn received_getheaders<T: BlockReader>(
&mut self,
addr: &PeerId,
(locator_hashes, stop_hash): Locators,
tree: &T,
) {
let max = self.config.max_message_headers;
if self.is_syncing() || max == 0 {
return;
}
let headers = tree.locate_headers(&locator_hashes, stop_hash, max);
if headers.is_empty() {
return;
}
self.upstream.headers(*addr, headers);
}
/// Import blocks into our block tree.
pub fn import_blocks<T: BlockTree, I: Iterator<Item = BlockHeader>>(
&mut self,
blocks: I,
tree: &mut T,
) -> Result<ImportResult, Error> {
match tree.import_blocks(blocks, &self.clock) {
Ok(ImportResult::TipChanged(header, tip, height, reverted, connected)) => {
let result = ImportResult::TipChanged(
header,
tip,
height,
reverted.clone(),
connected.clone(),
);
for (height, header) in reverted {
self.upstream
.event(Event::BlockDisconnected { height, header });
}
for (height, header) in connected {
self.upstream
.event(Event::BlockConnected { height, header });
}
self.upstream.event(Event::Synced(tip, height));
self.broadcast_tip(&tip, tree);
Ok(result)
}
Ok(result @ ImportResult::TipUnchanged) => Ok(result),
Err(err) => Err(err),
}
}
/// Called when we receive headers from a peer.
pub fn received_headers<T: BlockTree>(
&mut self,
from: &PeerId,
headers: Vec<BlockHeader>,
clock: &impl Clock,
tree: &mut T,
) -> Result<ImportResult, store::Error> {
let request = self.inflight.remove(from);
let headers = if let Some(headers) = NonEmpty::from_vec(headers) {
headers
} else {
return Ok(ImportResult::TipUnchanged);
};
let length = headers.len();
if length > MAX_MESSAGE_HEADERS {
log::debug!("Received more than maximum headers allowed from {}", from);
self.record_misbehavior(from);
self.upstream
.disconnect(*from, DisconnectReason::PeerMisbehaving("too many headers"));
return Ok(ImportResult::TipUnchanged);
}
// When unsolicited, we don't want to process too many headers in case of a DoS.
if length > MAX_UNSOLICITED_HEADERS && request.is_none() {
log::debug!("Received {} unsolicited headers from {}", length, from);
return Ok(ImportResult::TipUnchanged);
}
if let Some(peer) = self.peers.get_mut(from) {
peer.last_active = Some(clock.local_time());
} else {
return Ok(ImportResult::TipUnchanged);
}
log::debug!("[sync] Received {} block header(s) from {}", length, from);
let root = headers.first().block_hash();
let best = headers.last().block_hash();
if tree.contains(&best) {
return Ok(ImportResult::TipUnchanged);
}
match self.import_blocks(headers.into_iter(), tree) {
Ok(ImportResult::TipUnchanged) => {
// Try to find a common ancestor that leads up to the first header in
// the list we received.
let locators = (tree.locator_hashes(tree.height()), root);
let timeout = self.config.request_timeout;
self.request(*from, locators, timeout, OnTimeout::Ignore);
Ok(ImportResult::TipUnchanged)
}
Ok(ImportResult::TipChanged(header, tip, height, reverted, connected)) => {
// Update peer height.
if let Some(peer) = self.peers.get_mut(from) {
if height > peer.height {
peer.tip = tip;
peer.height = height;
}
}
// Keep track of when we last updated our tip. This is useful to check
// whether our tip is stale.
self.last_tip_update = Some(clock.local_time());
// If we received less than the maximum number of headers, we must be in sync.
// Otherwise, ask for the next batch of headers.
if length < MAX_MESSAGE_HEADERS {
// If these headers were unsolicited, we may already be ready/synced.
// Otherwise, we're finally in sync.
self.broadcast_tip(&tip, tree);
self.sync(tree);
} else {
let locators = (vec![tip], BlockHash::all_zeros());
let timeout = self.config.request_timeout;
self.request(*from, locators, timeout, OnTimeout::Disconnect);
}
Ok(ImportResult::TipChanged(
header, tip, height, reverted, connected,
))
}
Err(err) => self
.handle_error(from, err)
.map(|()| ImportResult::TipUnchanged),
}
}
fn request(
&mut self,
addr: PeerId,
locators: Locators,
timeout: LocalDuration,
on_timeout: OnTimeout,
) {
// Don't request more than once from the same peer.
if self.inflight.contains_key(&addr) {
return;
}
if let Some(peer) = self.peers.get_mut(&addr) {
debug_assert!(peer.last_asked.as_ref() != Some(&locators));
peer.last_asked = Some(locators.clone());
let sent_at = self.clock.local_time();
let req = GetHeaders {
locators,
sent_at,
on_timeout,
};
self.inflight.insert(addr, req.clone());
self.upstream.get_headers(addr, req.locators | {
if link.is_outbound() && !services.has(REQUIRED_SERVICES) {
return;
}
if height > self.best_height().unwrap_or_else(|| tree.height()) {
self.upstream.event(Event::PeerHeightUpdated { height });
}
self.register(socket, height, preferred, link);
self.sync(tree);
} | identifier_body |
syncmgr.rs | Reason::PeerMisbehaving("too many headers"));
return Ok(ImportResult::TipUnchanged);
}
// When unsolicited, we don't want to process too many headers in case of a DoS.
if length > MAX_UNSOLICITED_HEADERS && request.is_none() {
log::debug!("Received {} unsolicited headers from {}", length, from);
return Ok(ImportResult::TipUnchanged);
}
if let Some(peer) = self.peers.get_mut(from) {
peer.last_active = Some(clock.local_time());
} else {
return Ok(ImportResult::TipUnchanged);
}
log::debug!("[sync] Received {} block header(s) from {}", length, from);
let root = headers.first().block_hash();
let best = headers.last().block_hash();
if tree.contains(&best) {
return Ok(ImportResult::TipUnchanged);
}
match self.import_blocks(headers.into_iter(), tree) {
Ok(ImportResult::TipUnchanged) => {
// Try to find a common ancestor that leads up to the first header in
// the list we received.
let locators = (tree.locator_hashes(tree.height()), root);
let timeout = self.config.request_timeout;
self.request(*from, locators, timeout, OnTimeout::Ignore);
Ok(ImportResult::TipUnchanged)
}
Ok(ImportResult::TipChanged(header, tip, height, reverted, connected)) => {
// Update peer height.
if let Some(peer) = self.peers.get_mut(from) {
if height > peer.height {
peer.tip = tip;
peer.height = height;
}
}
// Keep track of when we last updated our tip. This is useful to check
// whether our tip is stale.
self.last_tip_update = Some(clock.local_time());
// If we received less than the maximum number of headers, we must be in sync.
// Otherwise, ask for the next batch of headers.
if length < MAX_MESSAGE_HEADERS {
// If these headers were unsolicited, we may already be ready/synced.
// Otherwise, we're finally in sync.
self.broadcast_tip(&tip, tree);
self.sync(tree);
} else {
let locators = (vec![tip], BlockHash::all_zeros());
let timeout = self.config.request_timeout;
self.request(*from, locators, timeout, OnTimeout::Disconnect);
}
Ok(ImportResult::TipChanged(
header, tip, height, reverted, connected,
))
}
Err(err) => self
.handle_error(from, err)
.map(|()| ImportResult::TipUnchanged),
}
}
fn request(
&mut self,
addr: PeerId,
locators: Locators,
timeout: LocalDuration,
on_timeout: OnTimeout,
) {
// Don't request more than once from the same peer.
if self.inflight.contains_key(&addr) {
return;
}
if let Some(peer) = self.peers.get_mut(&addr) {
debug_assert!(peer.last_asked.as_ref() != Some(&locators));
peer.last_asked = Some(locators.clone());
let sent_at = self.clock.local_time();
let req = GetHeaders {
locators,
sent_at,
on_timeout,
};
self.inflight.insert(addr, req.clone());
self.upstream.get_headers(addr, req.locators);
self.upstream.set_timer(timeout);
}
}
/// Called when we received an `inv` message. This will happen if we are out of sync with a
/// peer, and blocks are being announced. Otherwise, we expect to receive a `headers` message.
pub fn received_inv<T: BlockReader>(&mut self, addr: PeerId, inv: Vec<Inventory>, tree: &T) {
// Don't try to fetch headers from `inv` message while syncing. It's not helpful.
if self.is_syncing() {
return;
}
// Ignore and disconnect peers misbehaving.
if inv.len() > MAX_MESSAGE_INVS {
return;
}
let peer = if let Some(peer) = self.peers.get_mut(&addr) {
peer
} else {
return;
};
let mut best_block = None;
for i in &inv {
if let Inventory::Block(hash) = i {
peer.tip = *hash;
// "Headers-first is the primary method of announcement on the network. If a node
// fell back to sending blocks by inv, it's probably for a re-org. The final block
// hash provided should be the highest."
if !tree.is_known(hash) {
self.upstream.event(Event::BlockDiscovered(addr, *hash));
best_block = Some(hash);
}
}
}
if let Some(stop_hash) = best_block {
let locators = (tree.locator_hashes(tree.height()), *stop_hash);
let timeout = self.config.request_timeout;
// Try to find headers leading up to the `inv` entry.
self.request(addr, locators, timeout, OnTimeout::Retry(3));
}
}
/// Called when we received a tick.
pub fn received_wake<T: BlockReader>(&mut self, tree: &T) {
let local_time = self.clock.local_time();
let timeout = self.config.request_timeout;
let timed_out = self
.inflight
.iter()
.filter_map(|(peer, req)| {
if local_time - req.sent_at >= timeout {
Some((*peer, req.on_timeout, req.clone()))
} else {
None
}
})
.collect::<Vec<_>>();
let mut sync = false;
for (peer, on_timeout, req) in timed_out {
self.inflight.remove(&peer);
match on_timeout {
OnTimeout::Ignore => {
// It's likely that the peer just didn't have the requested header.
}
OnTimeout::Retry(0) | OnTimeout::Disconnect => {
self.upstream
.disconnect(peer, DisconnectReason::PeerTimeout("getheaders"));
sync = true;
}
OnTimeout::Retry(n) => {
if let Some((addr, _)) = self.peers.sample_with(|a, p| {
*a != peer && self.is_request_candidate(a, p, &req.locators.0)
}) {
let addr = *addr;
self.request(addr, req.locators, timeout, OnTimeout::Retry(n - 1));
}
}
}
}
// If some of the requests timed out, force a sync, otherwise just idle.
if sync {
self.sync(tree);
} else {
self.idle(tree);
}
}
/// Get the best known height out of all our peers.
pub fn best_height(&self) -> Option<Height> {
self.peers.iter().map(|(_, p)| p.height).max()
}
/// Are we currently syncing?
pub fn is_syncing(&self) -> bool {
!self.inflight.is_empty()
}
///////////////////////////////////////////////////////////////////////////
fn handle_error(&mut self, from: &PeerId, err: Error) -> Result<(), store::Error> {
match err {
// If this is an error with the underlying store, we have to propagate
// this up, because we can't handle it here.
Error::Store(e) => Err(e),
// If we got a bad block from the peer, we can handle it here.
Error::InvalidBlockPoW
| Error::InvalidBlockTarget(_, _)
| Error::InvalidBlockHash(_, _)
| Error::InvalidBlockHeight(_)
| Error::InvalidBlockTime(_, _) => {
log::debug!("{}: Received invalid headers: {}", from, err);
self.record_misbehavior(from);
self.upstream
.disconnect(*from, DisconnectReason::PeerMisbehaving("invalid headers"));
Ok(())
}
// Harmless errors can be ignored.
Error::DuplicateBlock(_) | Error::BlockMissing(_) => Ok(()),
// TODO: This will be removed.
Error::BlockImportAborted(_, _, _) => Ok(()),
// These shouldn't happen here.
// TODO: Perhaps there's a better way to have this error not show up here.
Error::Interrupted | Error::GenesisMismatch => Ok(()),
}
}
fn record_misbehavior(&mut self, peer: &PeerId) {
self.upstream.event(Event::PeerMisbehaved(*peer));
}
/// Check whether our current tip is stale.
///
/// *Nb. This doesn't check whether we've already requested new blocks.*
fn stale_tip<T: BlockReader>(&self, tree: &T) -> Option<LocalTime> {
let now = self.clock.local_time();
if let Some(last_update) = self.last_tip_update {
if last_update
< now - LocalDuration::from_secs(self.config.params.pow_target_spacing * 3)
{
return Some(last_update);
}
}
// If we don't have the time of the last update, it's probably because we
// are fresh, or restarted our node. In that case we check the last block time
// instead.
let (_, tip) = tree.tip();
let time = LocalTime::from_block_time(tip.time);
if time <= now - TIP_STALE_DURATION | {
return Some(time);
} | conditional_block |
|
dtm.py | /django/BasicBrowser/')
import db as db
from tmv_app.models import *
from scoping.models import Doc, Query
from django.db import connection, transaction
cursor = connection.cursor()
def f_gamma2(docs,gamma,docsizes,docUTset,topic_ids):
vl = []
for d in docs:
if gamma[2][d] > 0.001:
dt = (
docUTset[gamma[0][d]],
topic_ids[gamma[1][d]],
gamma[2][d],
gamma[2][d] / docsizes[gamma[0][d]],
run_id
)
vl.append(dt)
return vl
def tokenize(text):
transtable = {ord(c): None for c in string.punctuation + string.digits}
tokens = nltk.word_tokenize(text.translate(transtable))
tokens = [i for i in tokens if len(i) > 2]
return tokens
def add_features(title):
django.db.connections.close_all()
term, created = Term.objects.get_or_create(title=title)
term.run_id.add(run_id)
django.db.connections.close_all()
return term.pk
class snowball_stemmer(object):
def __init__(self):
self.stemmer = SnowballStemmer("english")
def __call__(self, doc):
return [self.stemmer.stem(t) for t in tokenize(doc)]
def proc_docs(docs):
stoplist = set(nltk.corpus.stopwords.words("english"))
stoplist.add('elsevier')
stoplist.add('rights')
stoplist.add('reserved')
stoplist.add('john')
stoplist.add('wiley')
stoplist.add('sons')
stoplist.add('copyright')
abstracts = [re.split("\([C-c]\) [1-2][0-9]{3} Elsevier",x.content)[0] for x in docs.iterator()]
abstracts = [x.split("Published by Elsevier")[0] for x in abstracts]
abstracts = [x.split("Copyright (C)")[0] for x in abstracts]
abstracts = [re.split("\. \(C\) [1-2][0-9]{3} ",x)[0] for x in abstracts]
docsizes = [len(x) for x in abstracts]
ids = [x.UT for x in docs.iterator()]
PYs = [x.PY for x in docs.iterator()]
return [abstracts, docsizes, ids, stoplist, PYs]
def readInfo(p):
d = {}
with open(p) as f:
for line in f:
(key, val) = line.strip().split(' ',1)
try:
d[key] = int(val)
except:
d[key] = val
return(d)
def | (topic_n,info,topic_ids,vocab_ids,ys):
print(topic_n)
django.db.connections.close_all()
p = "%03d" % (topic_n,)
p = "dtm-output/lda-seq/topic-"+p+"-var-e-log-prob.dat"
tlambda = np.fromfile(p, sep=" ").reshape((info['NUM_TERMS'],info['SEQ_LENGTH']))
for t in range(len(tlambda)):
for py in range(len(tlambda[t])):
score = np.exp(tlambda[t][py])
if score > 0.001:
tt = TopicTerm(
topic_id = topic_ids[topic_n],
term_id = vocab_ids[t],
PY = ys[py],
score = score,
run_id=run_id
)
tt.save()
#db.add_topic_term(topic_n+info['first_topic'], t+info['first_word'], py, score)
django.db.connections.close_all()
#########################################################
## Main function
def main():
try:
qid = int(sys.argv[1])
except:
print("please provide a query ID!")
sys.exit()
#sleep(7200)
Ks = [100,150,200,250]
#Ks = [10,20]
for K in Ks:
#K = 80
n_features=20000
global run_id
run_id = db.init(n_features,1)
stat = RunStats.objects.get(pk=run_id)
stat.method='BD'
stat.save()
stat.query=Query.objects.get(pk=qid)
##########################
## create input folder
if (os.path.isdir('dtm-input')):
shutil.rmtree('dtm-input')
os.mkdir('dtm-input')
yrange = list(range(1990,2017))
#yrange = list(range(2010,2012))
#yrange = list(range(1990,1997))
docs = Doc.objects.filter(
query=Query.objects.get(pk=qid),
content__iregex='\w',
relevant=True,
PY__in=yrange
).order_by('PY')
abstracts, docsizes, ids, stoplist, PYs = proc_docs(docs)
#########################
## Get the features now
print("Extracting word features...")
vectorizer = CountVectorizer(max_df=0.95, min_df=10,
max_features=n_features,
ngram_range=(1,1),
tokenizer=snowball_stemmer(),
stop_words=stoplist)
t0 = time()
dtm = vectorizer.fit_transform(abstracts)
print("done in %0.3fs." % (time() - t0))
del abstracts
gc.collect()
# Get the vocab, add it to db
vocab = vectorizer.get_feature_names()
vocab_ids = []
pool = Pool(processes=8)
vocab_ids.append(pool.map(add_features,vocab))
pool.terminate()
del vocab
vocab_ids = vocab_ids[0]
django.db.connections.close_all()
with open('dtm-input/foo-mult.dat','w') as mult:
for d in range(dtm.shape[0]):
words = find(dtm[d])
uwords = len(words[0])
mult.write(str(uwords) + " ")
for w in range(uwords):
index = words[1][w]
count = words[2][w]
mult.write(str(index)+":"+str(count)+" ")
mult.write('\n')
##########################
##put PY stuff in the seq file
ycounts = docs.values('PY').annotate(
count = models.Count('pk')
)
with open('dtm-input/foo-seq.dat','w') as seq:
seq.write(str(len(yrange)))
for y in ycounts:
seq.write('\n')
seq.write(str(y['count']))
##########################
# Run the dtm
subprocess.Popen([
"/home/galm/software/dtm/dtm/main",
"--ntopics={}".format(K),
"--mode=fit",
"--rng_seed=0",
"--initialize_lda=true",
"--corpus_prefix=/home/galm/projects/sustainability/dtm-input/foo",
"--outname=/home/galm/projects/sustainability/dtm-output",
"--top_chain_var=0.005",
"--alpha=0.01",
"--lda_sequence_min_iter=10",
"--lda_sequence_max_iter=20",
"--lda_max_em_iter=20"
]).wait()
##########################
## Upload the dtm results to the db
info = readInfo("dtm-output/lda-seq/info.dat")
topic_ids = db.add_topics(K)
#################################
# TopicTerms
topics = range(info['NUM_TOPICS'])
pool = Pool(processes=8)
pool.map(partial(
dtm_topic,
info=info,
topic_ids=topic_ids,
vocab_ids=vocab_ids,
ys = yrange
),topics)
pool.terminate()
gc.collect()
######################################
# Doctopics
gamma = np.fromfile('dtm-output/lda-seq/gam.dat', dtype=float,sep=" ")
gamma = gamma.reshape((len(gamma)/info['NUM_TOPICS'],info['NUM_TOPICS']))
gamma = find(csr_matrix(gamma))
glength = len(gamma[0])
chunk_size = 100000
ps = 16
parallel_add = True
all_dts = []
make_t = 0
add_t = 0
def insert_many(values_list):
query='''
INSERT INTO "tmv_app_doctopic"
("doc_id", "topic_id", "score", "scaled_score", "run_id")
VALUES (%s,%s,%s,%s,%s)
'''
cursor = connection.cursor()
cursor.executemany(query,values_list)
for i in range(glength//chunk_size+1):
dts = []
values_list = []
f = i*chunk_size
l = (i+1)*chunk_size
if l > glength:
l = glength
docs = range(f,l)
doc_batches = []
for p in range(ps):
doc_batches.append([x for x in docs if x % ps == p])
pool = Pool(processes=ps)
make_t0 = time | dtm_topic | identifier_name |
dtm.py | /django/BasicBrowser/')
import db as db
from tmv_app.models import *
from scoping.models import Doc, Query
from django.db import connection, transaction
cursor = connection.cursor()
def f_gamma2(docs,gamma,docsizes,docUTset,topic_ids):
vl = []
for d in docs:
if gamma[2][d] > 0.001:
|
return vl
def tokenize(text):
transtable = {ord(c): None for c in string.punctuation + string.digits}
tokens = nltk.word_tokenize(text.translate(transtable))
tokens = [i for i in tokens if len(i) > 2]
return tokens
def add_features(title):
django.db.connections.close_all()
term, created = Term.objects.get_or_create(title=title)
term.run_id.add(run_id)
django.db.connections.close_all()
return term.pk
class snowball_stemmer(object):
def __init__(self):
self.stemmer = SnowballStemmer("english")
def __call__(self, doc):
return [self.stemmer.stem(t) for t in tokenize(doc)]
def proc_docs(docs):
stoplist = set(nltk.corpus.stopwords.words("english"))
stoplist.add('elsevier')
stoplist.add('rights')
stoplist.add('reserved')
stoplist.add('john')
stoplist.add('wiley')
stoplist.add('sons')
stoplist.add('copyright')
abstracts = [re.split("\([C-c]\) [1-2][0-9]{3} Elsevier",x.content)[0] for x in docs.iterator()]
abstracts = [x.split("Published by Elsevier")[0] for x in abstracts]
abstracts = [x.split("Copyright (C)")[0] for x in abstracts]
abstracts = [re.split("\. \(C\) [1-2][0-9]{3} ",x)[0] for x in abstracts]
docsizes = [len(x) for x in abstracts]
ids = [x.UT for x in docs.iterator()]
PYs = [x.PY for x in docs.iterator()]
return [abstracts, docsizes, ids, stoplist, PYs]
def readInfo(p):
d = {}
with open(p) as f:
for line in f:
(key, val) = line.strip().split(' ',1)
try:
d[key] = int(val)
except:
d[key] = val
return(d)
def dtm_topic(topic_n,info,topic_ids,vocab_ids,ys):
print(topic_n)
django.db.connections.close_all()
p = "%03d" % (topic_n,)
p = "dtm-output/lda-seq/topic-"+p+"-var-e-log-prob.dat"
tlambda = np.fromfile(p, sep=" ").reshape((info['NUM_TERMS'],info['SEQ_LENGTH']))
for t in range(len(tlambda)):
for py in range(len(tlambda[t])):
score = np.exp(tlambda[t][py])
if score > 0.001:
tt = TopicTerm(
topic_id = topic_ids[topic_n],
term_id = vocab_ids[t],
PY = ys[py],
score = score,
run_id=run_id
)
tt.save()
#db.add_topic_term(topic_n+info['first_topic'], t+info['first_word'], py, score)
django.db.connections.close_all()
#########################################################
## Main function
def main():
try:
qid = int(sys.argv[1])
except:
print("please provide a query ID!")
sys.exit()
#sleep(7200)
Ks = [100,150,200,250]
#Ks = [10,20]
for K in Ks:
#K = 80
n_features=20000
global run_id
run_id = db.init(n_features,1)
stat = RunStats.objects.get(pk=run_id)
stat.method='BD'
stat.save()
stat.query=Query.objects.get(pk=qid)
##########################
## create input folder
if (os.path.isdir('dtm-input')):
shutil.rmtree('dtm-input')
os.mkdir('dtm-input')
yrange = list(range(1990,2017))
#yrange = list(range(2010,2012))
#yrange = list(range(1990,1997))
docs = Doc.objects.filter(
query=Query.objects.get(pk=qid),
content__iregex='\w',
relevant=True,
PY__in=yrange
).order_by('PY')
abstracts, docsizes, ids, stoplist, PYs = proc_docs(docs)
#########################
## Get the features now
print("Extracting word features...")
vectorizer = CountVectorizer(max_df=0.95, min_df=10,
max_features=n_features,
ngram_range=(1,1),
tokenizer=snowball_stemmer(),
stop_words=stoplist)
t0 = time()
dtm = vectorizer.fit_transform(abstracts)
print("done in %0.3fs." % (time() - t0))
del abstracts
gc.collect()
# Get the vocab, add it to db
vocab = vectorizer.get_feature_names()
vocab_ids = []
pool = Pool(processes=8)
vocab_ids.append(pool.map(add_features,vocab))
pool.terminate()
del vocab
vocab_ids = vocab_ids[0]
django.db.connections.close_all()
with open('dtm-input/foo-mult.dat','w') as mult:
for d in range(dtm.shape[0]):
words = find(dtm[d])
uwords = len(words[0])
mult.write(str(uwords) + " ")
for w in range(uwords):
index = words[1][w]
count = words[2][w]
mult.write(str(index)+":"+str(count)+" ")
mult.write('\n')
##########################
##put PY stuff in the seq file
ycounts = docs.values('PY').annotate(
count = models.Count('pk')
)
with open('dtm-input/foo-seq.dat','w') as seq:
seq.write(str(len(yrange)))
for y in ycounts:
seq.write('\n')
seq.write(str(y['count']))
##########################
# Run the dtm
subprocess.Popen([
"/home/galm/software/dtm/dtm/main",
"--ntopics={}".format(K),
"--mode=fit",
"--rng_seed=0",
"--initialize_lda=true",
"--corpus_prefix=/home/galm/projects/sustainability/dtm-input/foo",
"--outname=/home/galm/projects/sustainability/dtm-output",
"--top_chain_var=0.005",
"--alpha=0.01",
"--lda_sequence_min_iter=10",
"--lda_sequence_max_iter=20",
"--lda_max_em_iter=20"
]).wait()
##########################
## Upload the dtm results to the db
info = readInfo("dtm-output/lda-seq/info.dat")
topic_ids = db.add_topics(K)
#################################
# TopicTerms
topics = range(info['NUM_TOPICS'])
pool = Pool(processes=8)
pool.map(partial(
dtm_topic,
info=info,
topic_ids=topic_ids,
vocab_ids=vocab_ids,
ys = yrange
),topics)
pool.terminate()
gc.collect()
######################################
# Doctopics
gamma = np.fromfile('dtm-output/lda-seq/gam.dat', dtype=float,sep=" ")
gamma = gamma.reshape((len(gamma)/info['NUM_TOPICS'],info['NUM_TOPICS']))
gamma = find(csr_matrix(gamma))
glength = len(gamma[0])
chunk_size = 100000
ps = 16
parallel_add = True
all_dts = []
make_t = 0
add_t = 0
def insert_many(values_list):
query='''
INSERT INTO "tmv_app_doctopic"
("doc_id", "topic_id", "score", "scaled_score", "run_id")
VALUES (%s,%s,%s,%s,%s)
'''
cursor = connection.cursor()
cursor.executemany(query,values_list)
for i in range(glength//chunk_size+1):
dts = []
values_list = []
f = i*chunk_size
l = (i+1)*chunk_size
if l > glength:
l = glength
docs = range(f,l)
doc_batches = []
for p in range(ps):
doc_batches.append([x for x in docs if x % ps == p])
pool = Pool(processes=ps)
make_t0 = time | dt = (
docUTset[gamma[0][d]],
topic_ids[gamma[1][d]],
gamma[2][d],
gamma[2][d] / docsizes[gamma[0][d]],
run_id
)
vl.append(dt) | conditional_block |
dtm.py | /Desktop/django/BasicBrowser/')
import db as db
from tmv_app.models import *
from scoping.models import Doc, Query
from django.db import connection, transaction
cursor = connection.cursor()
def f_gamma2(docs,gamma,docsizes,docUTset,topic_ids):
vl = []
for d in docs:
if gamma[2][d] > 0.001:
dt = (
docUTset[gamma[0][d]],
topic_ids[gamma[1][d]],
gamma[2][d],
gamma[2][d] / docsizes[gamma[0][d]],
run_id
)
vl.append(dt)
return vl
def tokenize(text):
transtable = {ord(c): None for c in string.punctuation + string.digits}
tokens = nltk.word_tokenize(text.translate(transtable))
tokens = [i for i in tokens if len(i) > 2]
return tokens
def add_features(title):
django.db.connections.close_all()
term, created = Term.objects.get_or_create(title=title)
term.run_id.add(run_id)
django.db.connections.close_all()
return term.pk
class snowball_stemmer(object):
def __init__(self):
self.stemmer = SnowballStemmer("english")
def __call__(self, doc):
return [self.stemmer.stem(t) for t in tokenize(doc)]
def proc_docs(docs):
stoplist = set(nltk.corpus.stopwords.words("english"))
stoplist.add('elsevier')
stoplist.add('rights')
stoplist.add('reserved')
stoplist.add('john')
stoplist.add('wiley')
stoplist.add('sons')
stoplist.add('copyright')
abstracts = [re.split("\([C-c]\) [1-2][0-9]{3} Elsevier",x.content)[0] for x in docs.iterator()]
abstracts = [x.split("Published by Elsevier")[0] for x in abstracts]
abstracts = [x.split("Copyright (C)")[0] for x in abstracts]
abstracts = [re.split("\. \(C\) [1-2][0-9]{3} ",x)[0] for x in abstracts]
docsizes = [len(x) for x in abstracts]
ids = [x.UT for x in docs.iterator()]
PYs = [x.PY for x in docs.iterator()]
return [abstracts, docsizes, ids, stoplist, PYs]
def readInfo(p):
d = {}
with open(p) as f:
for line in f:
(key, val) = line.strip().split(' ',1)
try:
d[key] = int(val)
except:
d[key] = val
return(d)
def dtm_topic(topic_n,info,topic_ids,vocab_ids,ys):
print(topic_n)
django.db.connections.close_all()
p = "%03d" % (topic_n,)
p = "dtm-output/lda-seq/topic-"+p+"-var-e-log-prob.dat"
tlambda = np.fromfile(p, sep=" ").reshape((info['NUM_TERMS'],info['SEQ_LENGTH']))
for t in range(len(tlambda)):
for py in range(len(tlambda[t])):
score = np.exp(tlambda[t][py])
if score > 0.001:
tt = TopicTerm(
topic_id = topic_ids[topic_n],
term_id = vocab_ids[t],
PY = ys[py],
score = score,
run_id=run_id
)
tt.save()
#db.add_topic_term(topic_n+info['first_topic'], t+info['first_word'], py, score)
django.db.connections.close_all()
#########################################################
## Main function
def main():
try:
qid = int(sys.argv[1])
except:
print("please provide a query ID!")
sys.exit()
#sleep(7200)
Ks = [100,150,200,250]
#Ks = [10,20]
for K in Ks:
#K = 80
n_features=20000
global run_id
run_id = db.init(n_features,1)
stat = RunStats.objects.get(pk=run_id)
stat.method='BD'
stat.save()
stat.query=Query.objects.get(pk=qid)
##########################
## create input folder
if (os.path.isdir('dtm-input')):
shutil.rmtree('dtm-input')
os.mkdir('dtm-input')
yrange = list(range(1990,2017))
#yrange = list(range(2010,2012))
#yrange = list(range(1990,1997))
docs = Doc.objects.filter(
query=Query.objects.get(pk=qid),
content__iregex='\w',
relevant=True,
PY__in=yrange
).order_by('PY')
abstracts, docsizes, ids, stoplist, PYs = proc_docs(docs)
#########################
## Get the features now
print("Extracting word features...")
vectorizer = CountVectorizer(max_df=0.95, min_df=10,
max_features=n_features,
ngram_range=(1,1),
tokenizer=snowball_stemmer(),
stop_words=stoplist)
t0 = time()
dtm = vectorizer.fit_transform(abstracts)
print("done in %0.3fs." % (time() - t0))
del abstracts
gc.collect()
# Get the vocab, add it to db
vocab = vectorizer.get_feature_names()
vocab_ids = []
pool = Pool(processes=8)
vocab_ids.append(pool.map(add_features,vocab))
pool.terminate()
del vocab
vocab_ids = vocab_ids[0]
django.db.connections.close_all()
with open('dtm-input/foo-mult.dat','w') as mult:
for d in range(dtm.shape[0]):
words = find(dtm[d])
uwords = len(words[0])
mult.write(str(uwords) + " ")
for w in range(uwords):
index = words[1][w]
count = words[2][w]
mult.write(str(index)+":"+str(count)+" ")
mult.write('\n')
##########################
##put PY stuff in the seq file
ycounts = docs.values('PY').annotate(
count = models.Count('pk')
)
with open('dtm-input/foo-seq.dat','w') as seq:
seq.write(str(len(yrange)))
for y in ycounts:
seq.write('\n')
seq.write(str(y['count']))
##########################
# Run the dtm
subprocess.Popen([
"/home/galm/software/dtm/dtm/main",
"--ntopics={}".format(K),
"--mode=fit",
"--rng_seed=0",
"--initialize_lda=true",
"--corpus_prefix=/home/galm/projects/sustainability/dtm-input/foo",
"--outname=/home/galm/projects/sustainability/dtm-output",
"--top_chain_var=0.005",
"--alpha=0.01",
"--lda_sequence_min_iter=10",
"--lda_sequence_max_iter=20",
"--lda_max_em_iter=20"
]).wait()
##########################
## Upload the dtm results to the db
info = readInfo("dtm-output/lda-seq/info.dat")
topic_ids = db.add_topics(K)
#################################
# TopicTerms
topics = range(info['NUM_TOPICS'])
pool = Pool(processes=8)
pool.map(partial(
dtm_topic,
info=info,
topic_ids=topic_ids,
vocab_ids=vocab_ids,
ys = yrange
),topics)
pool.terminate()
gc.collect()
######################################
# Doctopics
gamma = np.fromfile('dtm-output/lda-seq/gam.dat', dtype=float,sep=" ")
gamma = gamma.reshape((len(gamma)/info['NUM_TOPICS'],info['NUM_TOPICS']))
gamma = find(csr_matrix(gamma))
glength = len(gamma[0])
chunk_size = 100000 | parallel_add = True
all_dts = []
make_t = 0
add_t = 0
def insert_many(values_list):
query='''
INSERT INTO "tmv_app_doctopic"
("doc_id", "topic_id", "score", "scaled_score", "run_id")
VALUES (%s,%s,%s,%s,%s)
'''
cursor = connection.cursor()
cursor.executemany(query,values_list)
for i in range(glength//chunk_size+1):
dts = []
values_list = []
f = i*chunk_size
l = (i+1)*chunk_size
if l > glength:
l = glength
docs = range(f,l)
doc_batches = []
for p in range(ps):
doc_batches.append([x for x in docs if x % ps == p])
pool = Pool(processes=ps)
make_t0 = time()
| ps = 16 | random_line_split |
dtm.py | /django/BasicBrowser/')
import db as db
from tmv_app.models import *
from scoping.models import Doc, Query
from django.db import connection, transaction
cursor = connection.cursor()
def f_gamma2(docs,gamma,docsizes,docUTset,topic_ids):
vl = []
for d in docs:
if gamma[2][d] > 0.001:
dt = (
docUTset[gamma[0][d]],
topic_ids[gamma[1][d]],
gamma[2][d],
gamma[2][d] / docsizes[gamma[0][d]],
run_id
)
vl.append(dt)
return vl
def tokenize(text):
|
def add_features(title):
django.db.connections.close_all()
term, created = Term.objects.get_or_create(title=title)
term.run_id.add(run_id)
django.db.connections.close_all()
return term.pk
class snowball_stemmer(object):
def __init__(self):
self.stemmer = SnowballStemmer("english")
def __call__(self, doc):
return [self.stemmer.stem(t) for t in tokenize(doc)]
def proc_docs(docs):
stoplist = set(nltk.corpus.stopwords.words("english"))
stoplist.add('elsevier')
stoplist.add('rights')
stoplist.add('reserved')
stoplist.add('john')
stoplist.add('wiley')
stoplist.add('sons')
stoplist.add('copyright')
abstracts = [re.split("\([C-c]\) [1-2][0-9]{3} Elsevier",x.content)[0] for x in docs.iterator()]
abstracts = [x.split("Published by Elsevier")[0] for x in abstracts]
abstracts = [x.split("Copyright (C)")[0] for x in abstracts]
abstracts = [re.split("\. \(C\) [1-2][0-9]{3} ",x)[0] for x in abstracts]
docsizes = [len(x) for x in abstracts]
ids = [x.UT for x in docs.iterator()]
PYs = [x.PY for x in docs.iterator()]
return [abstracts, docsizes, ids, stoplist, PYs]
def readInfo(p):
d = {}
with open(p) as f:
for line in f:
(key, val) = line.strip().split(' ',1)
try:
d[key] = int(val)
except:
d[key] = val
return(d)
def dtm_topic(topic_n,info,topic_ids,vocab_ids,ys):
print(topic_n)
django.db.connections.close_all()
p = "%03d" % (topic_n,)
p = "dtm-output/lda-seq/topic-"+p+"-var-e-log-prob.dat"
tlambda = np.fromfile(p, sep=" ").reshape((info['NUM_TERMS'],info['SEQ_LENGTH']))
for t in range(len(tlambda)):
for py in range(len(tlambda[t])):
score = np.exp(tlambda[t][py])
if score > 0.001:
tt = TopicTerm(
topic_id = topic_ids[topic_n],
term_id = vocab_ids[t],
PY = ys[py],
score = score,
run_id=run_id
)
tt.save()
#db.add_topic_term(topic_n+info['first_topic'], t+info['first_word'], py, score)
django.db.connections.close_all()
#########################################################
## Main function
def main():
try:
qid = int(sys.argv[1])
except:
print("please provide a query ID!")
sys.exit()
#sleep(7200)
Ks = [100,150,200,250]
#Ks = [10,20]
for K in Ks:
#K = 80
n_features=20000
global run_id
run_id = db.init(n_features,1)
stat = RunStats.objects.get(pk=run_id)
stat.method='BD'
stat.save()
stat.query=Query.objects.get(pk=qid)
##########################
## create input folder
if (os.path.isdir('dtm-input')):
shutil.rmtree('dtm-input')
os.mkdir('dtm-input')
yrange = list(range(1990,2017))
#yrange = list(range(2010,2012))
#yrange = list(range(1990,1997))
docs = Doc.objects.filter(
query=Query.objects.get(pk=qid),
content__iregex='\w',
relevant=True,
PY__in=yrange
).order_by('PY')
abstracts, docsizes, ids, stoplist, PYs = proc_docs(docs)
#########################
## Get the features now
print("Extracting word features...")
vectorizer = CountVectorizer(max_df=0.95, min_df=10,
max_features=n_features,
ngram_range=(1,1),
tokenizer=snowball_stemmer(),
stop_words=stoplist)
t0 = time()
dtm = vectorizer.fit_transform(abstracts)
print("done in %0.3fs." % (time() - t0))
del abstracts
gc.collect()
# Get the vocab, add it to db
vocab = vectorizer.get_feature_names()
vocab_ids = []
pool = Pool(processes=8)
vocab_ids.append(pool.map(add_features,vocab))
pool.terminate()
del vocab
vocab_ids = vocab_ids[0]
django.db.connections.close_all()
with open('dtm-input/foo-mult.dat','w') as mult:
for d in range(dtm.shape[0]):
words = find(dtm[d])
uwords = len(words[0])
mult.write(str(uwords) + " ")
for w in range(uwords):
index = words[1][w]
count = words[2][w]
mult.write(str(index)+":"+str(count)+" ")
mult.write('\n')
##########################
##put PY stuff in the seq file
ycounts = docs.values('PY').annotate(
count = models.Count('pk')
)
with open('dtm-input/foo-seq.dat','w') as seq:
seq.write(str(len(yrange)))
for y in ycounts:
seq.write('\n')
seq.write(str(y['count']))
##########################
# Run the dtm
subprocess.Popen([
"/home/galm/software/dtm/dtm/main",
"--ntopics={}".format(K),
"--mode=fit",
"--rng_seed=0",
"--initialize_lda=true",
"--corpus_prefix=/home/galm/projects/sustainability/dtm-input/foo",
"--outname=/home/galm/projects/sustainability/dtm-output",
"--top_chain_var=0.005",
"--alpha=0.01",
"--lda_sequence_min_iter=10",
"--lda_sequence_max_iter=20",
"--lda_max_em_iter=20"
]).wait()
##########################
## Upload the dtm results to the db
info = readInfo("dtm-output/lda-seq/info.dat")
topic_ids = db.add_topics(K)
#################################
# TopicTerms
topics = range(info['NUM_TOPICS'])
pool = Pool(processes=8)
pool.map(partial(
dtm_topic,
info=info,
topic_ids=topic_ids,
vocab_ids=vocab_ids,
ys = yrange
),topics)
pool.terminate()
gc.collect()
######################################
# Doctopics
gamma = np.fromfile('dtm-output/lda-seq/gam.dat', dtype=float,sep=" ")
gamma = gamma.reshape((len(gamma)/info['NUM_TOPICS'],info['NUM_TOPICS']))
gamma = find(csr_matrix(gamma))
glength = len(gamma[0])
chunk_size = 100000
ps = 16
parallel_add = True
all_dts = []
make_t = 0
add_t = 0
def insert_many(values_list):
query='''
INSERT INTO "tmv_app_doctopic"
("doc_id", "topic_id", "score", "scaled_score", "run_id")
VALUES (%s,%s,%s,%s,%s)
'''
cursor = connection.cursor()
cursor.executemany(query,values_list)
for i in range(glength//chunk_size+1):
dts = []
values_list = []
f = i*chunk_size
l = (i+1)*chunk_size
if l > glength:
l = glength
docs = range(f,l)
doc_batches = []
for p in range(ps):
doc_batches.append([x for x in docs if x % ps == p])
pool = Pool(processes=ps)
make_t0 = time()
| transtable = {ord(c): None for c in string.punctuation + string.digits}
tokens = nltk.word_tokenize(text.translate(transtable))
tokens = [i for i in tokens if len(i) > 2]
return tokens | identifier_body |
Section3.py | _regions(X,y,classifier,test_idx = None,resolution = 0.02):
markers = ("s","x","o","^","v")
colors = ("red","blue","lightgreen","gray","cyan")
cmap = ListedColormap(colors[:len(np.unique(y))])
x1_min,x1_max = X[:,0].min() - 1 ,X[:,0].max() + 1
x2_min,x2_max = X[:,1].min() - 1 ,X[:,1].max() + 1
xx1,xx2 = np.meshgrid(np.arange(x1_min,x1_max,resolution),np.arange(x2_min,x2_max,resolution))
Z = classifier.predict(np.array([xx1.ravel(),xx2.ravel()]).T)
Z = Z.reshape(xx1.shape)
plt.contourf(xx1,xx2,Z,alpha = 0.3 ,cmap = cmap)
plt.xlim(xx1.min(),xx1.max())
plt.ylim(xx2.min(),xx2.max())
for idx ,cl in enumerate(np.unique(y)):
plt.scatter(x = X[y == cl,0] , y = X[y == cl,1] ,alpha = 0.8 , c = colors[idx] ,marker = markers[idx] , label = cl, edgecolor = "black" )
if test_idx:
X_test,y_test = X[test_idx,:],y[test_idx]
plt.scatter(X_test[:,0],X_test[:,1],c = "",edgecolor = "black", | np.hstack((y_train,y_test))
plot_decision_regions(X = X_combined_std,y = y_combined,classifier = ppn,test_idx = range(105,150))
plt.xlabel("petal length")
plt.ylabel("petal width")
plt.legend("upper left")
plt.show()
# In[ ]:
import matplotlib.pyplot as plt
import numpy as np
import math
def sigmoid(z):
return 1.0 / (1.0 + np.exp(-z))
z = np.arange(-7,7,0.1)
phi_z = sigmoid(z)
plt.plot(z,phi_z)
plt.axvline(0.0,color = "k") #axvline >> 垂直線
plt.ylim(-0.1,1.1)
plt.xlabel("z")
plt.ylabel("$\phi$") # phiの記号の記述
plt.yticks([0.0,0.5,1.0])
ax = plt.gca() #gca >> graphic of axis
ax.yaxis.grid(True)
plt.tight_layout()
plt.show()
# In[ ]:
#class label が 0 or 1 の場合
def cost_1(z):
return -np.log(sigmoid(z))
def cost_0(z):
return -np.log(1-sigmoid(z))
z = np.arange(-10,10,0.1)
phi_z = sigmoid(z)
c1 = [cost_1(x) for x in z]
plt.plot(phi_z,c1,label = "J(w) y = 1")
c0 = [cost_0(x) for x in z]
plt.plot(phi_z,c0,label = "J(w) y = -1")
plt.ylim(0.0,5.1)
plt.xlim([0,1])
plt.xlabel("$\phi$(z)")
plt.ylabel("J(w)")
plt.legend(loc = "upper left")
plt.tight_layout()
plt.show()
# In[ ]:
class LogisticRegressionGD(object):
def __init__(self,eta = 0.05,n_iter = 100,random_state = 1):
self.eta = eta
self.n_iter = n_iter
self.random_state = random_state
def fit(self,X,y):
rgen = np.random.RandomState(self.random_state)
self.w = rgen.normal(loc = 0.0,scale = 0.01,size =1+ X.shape[1])
self.gosagun = []
for i in range(self.n_iter):
net_input = self.net_input(X)
output = self.activation(net_input)
errors = (y - output)
self.w[1:] += self.eta * X.T.dot(errors)
self.w[0] += self.eta * errors.sum() #.sum() >> 破壊的メソッド
cost = -y.dot(np.log(output)) - (1-y).dot(np.log(1-output))
self.gosagun.append(cost)
return self
def net_input(self,X):
return np.dot(X,self.w[1:]) + self.w[0]
def activation(self,z):
return 1.0 / (1.0 + np.exp(-np.clip(z,-250,250))) #z >250 の時 z -250 に変換する
def predict(self,X):
return np.where(self.activation(self.net_input(X)) >= 0.5,1,-1)
# In[ ]:
X_train_01_subset = X_train[ (y_train == 0) | (y_train == 1) ]
y_train_01_subset = y_train[ (y_train == 0) | (y_train == 1) ]
lrgd = LogisticRegressionGD(eta = 0.05,n_iter = 1000,random_state = 1)
lrgd.fit(X_train_01_subset,y_train_01_subset)
plot_decision_regions(X_train_01_subset,y_train_01_subset,classifier = lrgd)
plt.xlabel("petal length")
plt.ylabel("petal width")
plt.legend(loc = "upper left")
plt.tight_layout()
plt.show()
# In[ ]:
from sklearn.linear_model import LogisticRegression
lr = LogisticRegression(C = 100.0,random_state = 1)
lr.fit(X_train_std,y_train)
plot_decision_regions(X_combined_std,y_combined,classifier = lr,test_idx = range(105,150))
plt.xlabel("petal length")
plt.ylabel("petal width")
plt.legend(loc = "upper left")
plt.tight_layout()
plt.show()
# In[ ]:
lr.predict_proba(X_test_std) #probability of 1st,2nd,3rd
# In[ ]:
lr.predict_proba(X_test_std[:10,:]).argmax(axis = 1)
# In[ ]:
lr.predict(X_test_std[:10,:])
# In[ ]:
weights ,params = [],[]
for c in np.arange(-5,5):
lr = LogisticRegression(C = 10.0 ** c,random_state = 1)
lr.fit(X_train_std,y_train)
weights.append(lr.coef_[1])
params.append(10.0 ** c)
weights = np.array(weights)
plt.plot(params,weights[:,0],label = "petal length")
plt.plot(params,weights[:,1],label = "petal width")
plt.ylabel("weight coefficient")
plt.xlabel("c")
plt.xscale("log")
plt.show()
# In[ ]:
from sklearn.svm import SVC
svm = SVC(kernel = "linear" , C=1.0 ,random_state = 1)
svm.fit(X_train_std,y_train)
plot_decision_regions(X_combined_std,y_combined,classifier = svm,test_idx = range(105,150))
plt.xlabel("petal length")
plt.ylabel("petal width")
plt.legend(loc = "upper left")
plt.tight_layout()
plt.show()
# In[ ]:
from sklearn.linear_model import SGDClassifier
ppn = SGDClassifier(loss = "perceptron")
lr = SGDClassifier(loss = "log")
svm = SGDClassifier(loss = "hinge")
# In[ ]:
import matplotlib.pyplot as plt
import numpy as np
np.random.seed(1)
X_xor = np.random.randn(200,2)#random.randn(nand.arrayの形)
y_xor = np.logical_xor(X_xor[:,0] > 0 ,X_xor[:,1] > 0)#論理和でTrue or False を割り当てる
y_xor = np.where(y_xor,1,-1) #True >> 1 ,False >> 1の割り当て
plt.scatter(X_xor[y_xor == 1,0],X_xor[y_xor ==1,1] ,c = "b", marker = "x",label = "1")
plt.scatter(X_xor[y_xor == -1,0],X_xor[y_xor == -1,1],c = "r",marker = "s",label = "-1")
plt.xlim([-3,3])
plt.ylim([-3,3])
plt.legend(loc = "best")
plt.tight_layout()
plt.show()
# In[ ]:
svm = SVC(kernel = "rbf",random_state = 1 ,gamma = 0.10,C = 10.0)
svm.fit(X_xor,y_xor)
plot_decision_regions(X_xor,y_xor,classifier = svm)
plt.tight_layout()
plt.show()
# In[ ]:
svm = SVC(kernel = "rbf",random_state = 1,gamma = 0.2 ,C = 1.0)
svm.fit(X_train_std,y_train)
plot_decision_regions(X_combined_std,y_combined,classifier = svm,test_idx = | alpha = 1,marker = "o",s = 100, label = "test_set")
# In[ ]:
X_combined_std = np.vstack((X_train_std,X_test_std))
y_combined = | conditional_block |
Section3.py | _regions(X,y,classifier,test_idx = None,resolution = 0.02):
markers = ("s","x","o","^","v")
colors = ("red","blue","lightgreen","gray","cyan")
cmap = ListedColormap(colors[:len(np.unique(y))])
x1_min,x1_max = X[:,0].min() - 1 ,X[:,0].max() + 1
x2_min,x2_max = X[:,1].min() - 1 ,X[:,1].max() + 1
xx1,xx2 = np.meshgrid(np.arange(x1_min,x1_max,resolution),np.arange(x2_min,x2_max,resolution))
Z = classifier.predict(np.array([xx1.ravel(),xx2.ravel()]).T)
Z = Z.reshape(xx1.shape)
plt.contourf(xx1,xx2,Z,alpha = 0.3 ,cmap = cmap)
plt.xlim(xx1.min(),xx1.max())
plt.ylim(xx2.min(),xx2.max())
for idx ,cl in enumerate(np.unique(y)):
plt.scatter(x = X[y == cl,0] , y = X[y == cl,1] ,alpha = 0.8 , c = colors[idx] ,marker = markers[idx] , label = cl, edgecolor = "black" )
if test_idx:
X_test,y_test = X[test_idx,:],y[test_idx]
plt.scatter(X_test[:,0],X_test[:,1],c = "",edgecolor = "black",alpha = 1,marker = "o",s = 100, label = "test_set")
# In[ ]:
X_combined_std = np.vstack((X_train_std,X_test_std))
y_combined = np.hstack((y_train,y_test))
plot_decision_regions(X = X_combined_std,y = y_combined,classifier = ppn,test_idx = range(105,150))
plt.xlabel("petal length")
plt.ylabel("petal width")
plt.legend("upper left")
plt.show()
# In[ ]:
import matplotlib.pyplot as plt
import numpy as np
import math
def sigmoid(z):
return 1.0 / (1.0 + np.exp(-z))
z = np.arange(-7,7,0.1)
phi_z = sigmoid(z)
plt.plot(z,phi_z)
plt.axvline(0.0,color = "k") #axvline >> 垂直線
plt.ylim(-0.1,1.1)
plt.xlabel("z")
plt.ylabel("$\phi$") # phiの記号の記述
plt.yticks([0.0,0.5,1.0])
ax = plt.gca() #gca >> graphic of axis
ax.yaxis.grid(True)
plt.tight_layout()
plt.show()
# In[ ]:
#class label が 0 or 1 の場合
def cost_1(z):
return -np.log(sigmoid(z))
def cost_0(z):
return -np.log(1-sigmoid(z))
z = np.arange(-10,10,0.1)
phi_z = sigmoid(z)
c1 = [cost_1(x) f | ,c1,label = "J(w) y = 1")
c0 = [cost_0(x) for x in z]
plt.plot(phi_z,c0,label = "J(w) y = -1")
plt.ylim(0.0,5.1)
plt.xlim([0,1])
plt.xlabel("$\phi$(z)")
plt.ylabel("J(w)")
plt.legend(loc = "upper left")
plt.tight_layout()
plt.show()
# In[ ]:
class LogisticRegressionGD(object):
def __init__(self,eta = 0.05,n_iter = 100,random_state = 1):
self.eta = eta
self.n_iter = n_iter
self.random_state = random_state
def fit(self,X,y):
rgen = np.random.RandomState(self.random_state)
self.w = rgen.normal(loc = 0.0,scale = 0.01,size =1+ X.shape[1])
self.gosagun = []
for i in range(self.n_iter):
net_input = self.net_input(X)
output = self.activation(net_input)
errors = (y - output)
self.w[1:] += self.eta * X.T.dot(errors)
self.w[0] += self.eta * errors.sum() #.sum() >> 破壊的メソッド
cost = -y.dot(np.log(output)) - (1-y).dot(np.log(1-output))
self.gosagun.append(cost)
return self
def net_input(self,X):
return np.dot(X,self.w[1:]) + self.w[0]
def activation(self,z):
return 1.0 / (1.0 + np.exp(-np.clip(z,-250,250))) #z >250 の時 z -250 に変換する
def predict(self,X):
return np.where(self.activation(self.net_input(X)) >= 0.5,1,-1)
# In[ ]:
X_train_01_subset = X_train[ (y_train == 0) | (y_train == 1) ]
y_train_01_subset = y_train[ (y_train == 0) | (y_train == 1) ]
lrgd = LogisticRegressionGD(eta = 0.05,n_iter = 1000,random_state = 1)
lrgd.fit(X_train_01_subset,y_train_01_subset)
plot_decision_regions(X_train_01_subset,y_train_01_subset,classifier = lrgd)
plt.xlabel("petal length")
plt.ylabel("petal width")
plt.legend(loc = "upper left")
plt.tight_layout()
plt.show()
# In[ ]:
from sklearn.linear_model import LogisticRegression
lr = LogisticRegression(C = 100.0,random_state = 1)
lr.fit(X_train_std,y_train)
plot_decision_regions(X_combined_std,y_combined,classifier = lr,test_idx = range(105,150))
plt.xlabel("petal length")
plt.ylabel("petal width")
plt.legend(loc = "upper left")
plt.tight_layout()
plt.show()
# In[ ]:
lr.predict_proba(X_test_std) #probability of 1st,2nd,3rd
# In[ ]:
lr.predict_proba(X_test_std[:10,:]).argmax(axis = 1)
# In[ ]:
lr.predict(X_test_std[:10,:])
# In[ ]:
weights ,params = [],[]
for c in np.arange(-5,5):
lr = LogisticRegression(C = 10.0 ** c,random_state = 1)
lr.fit(X_train_std,y_train)
weights.append(lr.coef_[1])
params.append(10.0 ** c)
weights = np.array(weights)
plt.plot(params,weights[:,0],label = "petal length")
plt.plot(params,weights[:,1],label = "petal width")
plt.ylabel("weight coefficient")
plt.xlabel("c")
plt.xscale("log")
plt.show()
# In[ ]:
from sklearn.svm import SVC
svm = SVC(kernel = "linear" , C=1.0 ,random_state = 1)
svm.fit(X_train_std,y_train)
plot_decision_regions(X_combined_std,y_combined,classifier = svm,test_idx = range(105,150))
plt.xlabel("petal length")
plt.ylabel("petal width")
plt.legend(loc = "upper left")
plt.tight_layout()
plt.show()
# In[ ]:
from sklearn.linear_model import SGDClassifier
ppn = SGDClassifier(loss = "perceptron")
lr = SGDClassifier(loss = "log")
svm = SGDClassifier(loss = "hinge")
# In[ ]:
import matplotlib.pyplot as plt
import numpy as np
np.random.seed(1)
X_xor = np.random.randn(200,2)#random.randn(nand.arrayの形)
y_xor = np.logical_xor(X_xor[:,0] > 0 ,X_xor[:,1] > 0)#論理和でTrue or False を割り当てる
y_xor = np.where(y_xor,1,-1) #True >> 1 ,False >> 1の割り当て
plt.scatter(X_xor[y_xor == 1,0],X_xor[y_xor ==1,1] ,c = "b", marker = "x",label = "1")
plt.scatter(X_xor[y_xor == -1,0],X_xor[y_xor == -1,1],c = "r",marker = "s",label = "-1")
plt.xlim([-3,3])
plt.ylim([-3,3])
plt.legend(loc = "best")
plt.tight_layout()
plt.show()
# In[ ]:
svm = SVC(kernel = "rbf",random_state = 1 ,gamma = 0.10,C = 10.0)
svm.fit(X_xor,y_xor)
plot_decision_regions(X_xor,y_xor,classifier = svm)
plt.tight_layout()
plt.show()
# In[ ]:
svm = SVC(kernel = "rbf",random_state = 1,gamma = 0.2 ,C = 1.0)
svm.fit(X_train_std,y_train)
plot_decision_regions(X_combined_std,y_combined,classifier = svm,test_idx | or x in z]
plt.plot(phi_z | identifier_body |
Section3.py | (X,y,classifier,test_idx = None,resolution = 0.02):
markers = ("s","x","o","^","v")
colors = ("red","blue","lightgreen","gray","cyan")
cmap = ListedColormap(colors[:len(np.unique(y))])
x1_min,x1_max = X[:,0].min() - 1 ,X[:,0].max() + 1
x2_min,x2_max = X[:,1].min() - 1 ,X[:,1].max() + 1
xx1,xx2 = np.meshgrid(np.arange(x1_min,x1_max,resolution),np.arange(x2_min,x2_max,resolution))
Z = classifier.predict(np.array([xx1.ravel(),xx2.ravel()]).T)
Z = Z.reshape(xx1.shape)
plt.contourf(xx1,xx2,Z,alpha = 0.3 ,cmap = cmap)
plt.xlim(xx1.min(),xx1.max())
plt.ylim(xx2.min(),xx2.max())
for idx ,cl in enumerate(np.unique(y)):
plt.scatter(x = X[y == cl,0] , y = X[y == cl,1] ,alpha = 0.8 , c = colors[idx] ,marker = markers[idx] , label = cl, edgecolor = "black" )
if test_idx:
X_test,y_test = X[test_idx,:],y[test_idx]
plt.scatter(X_test[:,0],X_test[:,1],c = "",edgecolor = "black",alpha = 1,marker = "o",s = 100, label = "test_set")
# In[ ]:
X_combined_std = np.vstack((X_train_std,X_test_std))
y_combined = np.hstack((y_train,y_test))
plot_decision_regions(X = X_combined_std,y = y_combined,classifier = ppn,test_idx = range(105,150))
plt.xlabel("petal length")
plt.ylabel("petal width")
plt.legend("upper left")
plt.show()
# In[ ]:
import matplotlib.pyplot as plt
import numpy as np
import math
def sigmoid(z):
return 1.0 / (1.0 + np.exp(-z))
z = np.arange(-7,7,0.1)
phi_z = sigmoid(z)
plt.plot(z,phi_z)
plt.axvline(0.0,color = "k") #axvline >> 垂直線
plt.ylim(-0.1,1.1)
plt.xlabel("z")
plt.ylabel("$\phi$") # phiの記号の記述
plt.yticks([0.0,0.5,1.0])
ax = plt.gca() #gca >> graphic of axis
ax.yaxis.grid(True)
plt.tight_layout()
plt.show()
# In[ ]:
#class label が 0 or 1 の場合
def cost_1(z):
return -np.log(sigmoid(z))
def cost_0(z):
return -np.log(1-sigmoid(z))
z = np.arange(-10,10,0.1)
phi_z = sigmoid(z)
c1 = [cost_1(x) for x in z]
plt.plot(phi_z,c1,label = "J(w) y = 1")
c0 = [cost_0(x) for x in z]
plt.plot(phi_z,c0,label = "J(w) y = -1")
plt.ylim(0.0,5.1)
plt.xlim([0,1])
plt.xlabel("$\phi$(z)")
plt.ylabel("J(w)")
plt.legend(loc = "upper left")
plt.tight_layout()
plt.show()
# In[ ]:
class LogisticRegressionGD(object):
def __init__(self,eta = 0.05,n_iter = 100,random_state = 1):
self.eta = eta
self.n_iter = n_iter
self.random_state = random_state
def fit(self,X,y):
rgen = np.random.RandomState(self.random_state)
self.w = rgen.normal(loc = 0.0,scale = 0.01,size =1+ X.shape[1])
self.gosagun = []
for i in range(self.n_iter):
net_input = self.net_input(X)
output = self.activation(net_input)
errors = (y - output)
self.w[1:] += self.eta * X.T.dot(errors)
self.w[0] += self.eta * errors.sum() #.sum() >> 破壊的メソッド
cost = -y.dot(np.log(output)) - (1-y).dot(np.log(1-output))
self.gosagun.append(cost)
return self
def net_input(self,X):
return np.dot(X,self.w[1:]) + self.w[0]
def activation(self,z):
return 1.0 / (1.0 + np.exp(-np.clip(z,-250,250))) #z >250 の時 z -250 に変換する
def predict(self,X):
return np.where(self.activation(self.net_input(X)) >= 0.5,1,-1)
# In[ ]:
X_train_01_subset = X_train[ (y_train == 0) | (y_train == 1) ]
y_train_01_subset = y_train[ (y_train == 0) | (y_train == 1) ]
lrgd = LogisticRegressionGD(eta = 0.05,n_iter = 1000,random_state = 1)
lrgd.fit(X_train_01_subset,y_train_01_subset)
plot_decision_regions(X_train_01_subset,y_train_01_subset,classifier = lrgd)
plt.xlabel("petal length")
plt.ylabel("petal width")
plt.legend(loc = "upper left")
plt.tight_layout()
plt.show()
# In[ ]:
from sklearn.linear_model import LogisticRegression
lr = LogisticRegression(C = 100.0,random_state = 1)
lr.fit(X_train_std,y_train)
plot_decision_regions(X_combined_std,y_combined,classifier = lr,test_idx = range(105,150))
plt.xlabel("petal length")
plt.ylabel("petal width")
plt.legend(loc = "upper left")
plt.tight_layout()
plt.show()
# In[ ]:
lr.predict_proba(X_test_std) #probability of 1st,2nd,3rd
# In[ ]:
lr.predict_proba(X_test_std[:10,:]).argmax(axis = 1)
# In[ ]:
lr.predict(X_test_std[:10,:])
# In[ ]:
weights ,params = [],[]
for c in np.arange(-5,5):
lr = LogisticRegression(C = 10.0 ** c,random_state = 1)
lr.fit(X_train_std,y_train)
weights.append(lr.coef_[1])
params.append(10.0 ** c)
weights = np.array(weights)
plt.plot(params,weights[:,0],label = "petal length")
plt.plot(params,weights[:,1],label = "petal width")
plt.ylabel("weight coefficient")
plt.xlabel("c")
plt.xscale("log")
plt.show()
# In[ ]:
from sklearn.svm import SVC
svm = SVC(kernel = "linear" , C=1.0 ,random_state = 1)
svm.fit(X_train_std,y_train)
plot_decision_regions(X_combined_std,y_combined,classifier = svm,test_idx = range(105,150))
plt.xlabel("petal length")
plt.ylabel("petal width")
plt.legend(loc = "upper left")
plt.tight_layout()
plt.show()
# In[ ]:
from sklearn.linear_model import SGDClassifier
ppn = SGDClassifier(loss = "perceptron")
lr = SGDClassifier(loss = "log")
svm = SGDClassifier(loss = "hinge")
# In[ ]:
import matplotlib.pyplot as plt
import numpy as np
np.random.seed(1)
X_xor = np.random.randn(200,2)#random.randn(nand.arrayの形)
y_xor = np.logical_xor(X_xor[:,0] > 0 ,X_xor[:,1] > 0)#論理和でTrue or False を割り当てる
y_xor = np.where(y_xor,1,-1) #True >> 1 ,False >> 1の割り当て
plt.scatter(X_xor[y_xor == 1,0],X_xor[y_xor ==1,1] ,c = "b", marker = "x",label = "1")
plt.scatter(X_xor[y_xor == -1,0],X_xor[y_xor == -1,1],c = "r",marker = "s",label = "-1")
plt.xlim([-3,3])
plt.ylim([-3,3])
plt.legend(loc = "best")
|
# In[ ]:
svm = SVC(kernel = "rbf",random_state = 1 ,gamma = 0.10,C = 10.0)
svm.fit(X_xor,y_xor)
plot_decision_regions(X_xor,y_xor,classifier = svm)
plt.tight_layout()
plt.show()
# In[ ]:
svm = SVC(kernel = "rbf",random_state = 1,gamma = 0.2 ,C = 1.0)
svm.fit(X_train_std,y_train)
plot_decision_regions(X_combined_std,y_combined,classifier = svm,test | plt.tight_layout()
plt.show()
| random_line_split |
Section3.py | _regions(X,y,classifier,test_idx = None,resolution = 0.02):
markers = ("s","x","o","^","v")
colors = ("red","blue","lightgreen","gray","cyan")
cmap = ListedColormap(colors[:len(np.unique(y))])
x1_min,x1_max = X[:,0].min() - 1 ,X[:,0].max() + 1
x2_min,x2_max = X[:,1].min() - 1 ,X[:,1].max() + 1
xx1,xx2 = np.meshgrid(np.arange(x1_min,x1_max,resolution),np.arange(x2_min,x2_max,resolution))
Z = classifier.predict(np.array([xx1.ravel(),xx2.ravel()]).T)
Z = Z.reshape(xx1.shape)
plt.contourf(xx1,xx2,Z,alpha = 0.3 ,cmap = cmap)
plt.xlim(xx1.min(),xx1.max())
plt.ylim(xx2.min(),xx2.max())
for idx ,cl in enumerate(np.unique(y)):
plt.scatter(x = X[y == cl,0] , y = X[y == cl,1] ,alpha = 0.8 , c = colors[idx] ,marker = markers[idx] , label = cl, edgecolor = "black" )
if test_idx:
X_test,y_test = X[test_idx,:],y[test_idx]
plt.scatter(X_test[:,0],X_test[:,1],c = "",edgecolor = "black",alpha = 1,marker = "o",s = 100, label = "test_set")
# In[ ]:
X_combined_std = np.vstack((X_train_std,X_test_std))
y_combined = np.hstack((y_train,y_test))
plot_decision_regions(X = X_combined_std,y = y_combined,classifier = ppn,test_idx = range(105,150))
plt.xlabel("petal length")
plt.ylabel("petal width")
plt.legend("upper left")
plt.show()
# In[ ]:
import matplotlib.pyplot as plt
import numpy as np
import math
def sigmoid(z):
return 1.0 / (1.0 + np.exp(-z))
z = np.arange(-7,7,0.1)
phi_z = sigmoid(z)
plt.plot(z,phi_z)
plt.axvline(0.0,color = "k") #axvline >> 垂直線
plt.ylim(-0.1,1.1)
plt.xlabel("z")
plt.ylabel("$\phi$") # phiの記号の記述
plt.yticks([0.0,0.5,1.0])
ax = plt.gca() #gca >> graphic of axis
ax.yaxis.grid(True)
plt.tight_layout()
plt.show()
# In[ ]:
#class label が 0 or 1 の場合
def cost_1(z):
return -np.log(sigmoid(z))
def cost_0(z):
return -np.log(1-sigmoid(z))
z = np.arange(-10,10,0.1)
phi_z = sigmoid(z)
c1 = [cost_1(x) for x in z]
plt.plot(phi_z,c1,label = "J(w) y = 1")
c0 = [cost_0(x) for x in z]
plt.plot(phi_z,c0,label = "J(w) y = -1")
plt.ylim(0.0,5.1)
plt.xlim([0,1])
plt.xlabel("$\phi$(z)")
plt.ylabel("J(w)")
plt.legend(loc = "upper left")
plt.tight_layout()
plt.show()
# In[ ]:
class LogisticRegressionGD(object):
def __init__(self,eta = 0.05,n_iter = 100,random_state = 1):
self.eta = eta
self.n_iter = n | random_state = random_state
def fit(self,X,y):
rgen = np.random.RandomState(self.random_state)
self.w = rgen.normal(loc = 0.0,scale = 0.01,size =1+ X.shape[1])
self.gosagun = []
for i in range(self.n_iter):
net_input = self.net_input(X)
output = self.activation(net_input)
errors = (y - output)
self.w[1:] += self.eta * X.T.dot(errors)
self.w[0] += self.eta * errors.sum() #.sum() >> 破壊的メソッド
cost = -y.dot(np.log(output)) - (1-y).dot(np.log(1-output))
self.gosagun.append(cost)
return self
def net_input(self,X):
return np.dot(X,self.w[1:]) + self.w[0]
def activation(self,z):
return 1.0 / (1.0 + np.exp(-np.clip(z,-250,250))) #z >250 の時 z -250 に変換する
def predict(self,X):
return np.where(self.activation(self.net_input(X)) >= 0.5,1,-1)
# In[ ]:
X_train_01_subset = X_train[ (y_train == 0) | (y_train == 1) ]
y_train_01_subset = y_train[ (y_train == 0) | (y_train == 1) ]
lrgd = LogisticRegressionGD(eta = 0.05,n_iter = 1000,random_state = 1)
lrgd.fit(X_train_01_subset,y_train_01_subset)
plot_decision_regions(X_train_01_subset,y_train_01_subset,classifier = lrgd)
plt.xlabel("petal length")
plt.ylabel("petal width")
plt.legend(loc = "upper left")
plt.tight_layout()
plt.show()
# In[ ]:
from sklearn.linear_model import LogisticRegression
lr = LogisticRegression(C = 100.0,random_state = 1)
lr.fit(X_train_std,y_train)
plot_decision_regions(X_combined_std,y_combined,classifier = lr,test_idx = range(105,150))
plt.xlabel("petal length")
plt.ylabel("petal width")
plt.legend(loc = "upper left")
plt.tight_layout()
plt.show()
# In[ ]:
lr.predict_proba(X_test_std) #probability of 1st,2nd,3rd
# In[ ]:
lr.predict_proba(X_test_std[:10,:]).argmax(axis = 1)
# In[ ]:
lr.predict(X_test_std[:10,:])
# In[ ]:
weights ,params = [],[]
for c in np.arange(-5,5):
lr = LogisticRegression(C = 10.0 ** c,random_state = 1)
lr.fit(X_train_std,y_train)
weights.append(lr.coef_[1])
params.append(10.0 ** c)
weights = np.array(weights)
plt.plot(params,weights[:,0],label = "petal length")
plt.plot(params,weights[:,1],label = "petal width")
plt.ylabel("weight coefficient")
plt.xlabel("c")
plt.xscale("log")
plt.show()
# In[ ]:
from sklearn.svm import SVC
svm = SVC(kernel = "linear" , C=1.0 ,random_state = 1)
svm.fit(X_train_std,y_train)
plot_decision_regions(X_combined_std,y_combined,classifier = svm,test_idx = range(105,150))
plt.xlabel("petal length")
plt.ylabel("petal width")
plt.legend(loc = "upper left")
plt.tight_layout()
plt.show()
# In[ ]:
from sklearn.linear_model import SGDClassifier
ppn = SGDClassifier(loss = "perceptron")
lr = SGDClassifier(loss = "log")
svm = SGDClassifier(loss = "hinge")
# In[ ]:
import matplotlib.pyplot as plt
import numpy as np
np.random.seed(1)
X_xor = np.random.randn(200,2)#random.randn(nand.arrayの形)
y_xor = np.logical_xor(X_xor[:,0] > 0 ,X_xor[:,1] > 0)#論理和でTrue or False を割り当てる
y_xor = np.where(y_xor,1,-1) #True >> 1 ,False >> 1の割り当て
plt.scatter(X_xor[y_xor == 1,0],X_xor[y_xor ==1,1] ,c = "b", marker = "x",label = "1")
plt.scatter(X_xor[y_xor == -1,0],X_xor[y_xor == -1,1],c = "r",marker = "s",label = "-1")
plt.xlim([-3,3])
plt.ylim([-3,3])
plt.legend(loc = "best")
plt.tight_layout()
plt.show()
# In[ ]:
svm = SVC(kernel = "rbf",random_state = 1 ,gamma = 0.10,C = 10.0)
svm.fit(X_xor,y_xor)
plot_decision_regions(X_xor,y_xor,classifier = svm)
plt.tight_layout()
plt.show()
# In[ ]:
svm = SVC(kernel = "rbf",random_state = 1,gamma = 0.2 ,C = 1.0)
svm.fit(X_train_std,y_train)
plot_decision_regions(X_combined_std,y_combined,classifier = svm,test_idx | _iter
self. | identifier_name |
pipeline_cutouts.py | cutout.wcs.wcs.crpix[1])
hdu.header.set('CDELT1', cutout.wcs.wcs.cdelt[0])
hdu.header.set('CDELT2', cutout.wcs.wcs.cdelt[1])
hdu.header.set('NAXIS1', cutout.wcs.pixel_shape[0])
hdu.header.set('NAXIS2', cutout.wcs.pixel_shape[1])
return hdu
# ------ ------ ------ ------ ------ ------ ------ ------ ------ ------
def do_primarybeam_correction(pbname, imagename):
print(' Preparing to apply the primary beam correction to {0}'.format(imagename))
hdu = fits.open(imagename)[0]
pb = fits.open(pbname)[0]
wcs = WCS(pb.header)
# cutout pb field of view to match image field of view
x_size = hdu.header['NAXIS1']
x_pixel_deg = hdu.header['CDELT2'] # CDELT1 is negative, so take positive one
size = (x_size*x_pixel_deg*u.degree, x_size*x_pixel_deg*u.degree) # angular size of cutout, using astropy coord. approx 32768*0.6 arcseconds.
position = SkyCoord(pb.header['CRVAL1']*u.degree, pb.header['CRVAL2']*u.degree) # RA and DEC of beam PB pointing
print(' Cutting out image FOV from primary beam image...')
cutout = Cutout2D(pb.data[0,0,:,:], position=position, size=size, mode='trim', wcs=wcs.celestial, copy=True)
# Update the FITS header with the cutout WCS by hand using my own function
# don't use cutout.wcs.to_header() because it doesn't account for the freq and stokes axes. is only compatible with 2D fits images.
#pb.header.update(cutout.wcs.to_header()) #
pb = update_header_from_cutout2D(pb, cutout)
# write updated fits file to disk
pb.writeto(pbname[:-5]+'_cutout.fits', overwrite=True) # Write the cutout to a new FITS file
# regrid PB image cutout to match pixel scale of the image FOV
print(' Regridding image...')
# get header of image to match PB to
montage.mGetHdr(imagename, 'hdu_tmp.hdr')
# regrid pb image (270 pixels) to size of ref image (32k pixels)
montage.reproject(in_images=pbname[:-5]+'_cutout.fits', out_images=pbname[:-5]+'_cutout_regrid.fits', header='hdu_tmp.hdr', exact_size=True)
os.remove('hdu_tmp.hdr') # get rid of header text file saved to disk
# update montage output to float32
pb = fits.open(pbname[:-5]+'_cutout_regrid.fits', mode='update')
newdata = np.zeros((1,1,pb[0].data.shape[0], pb[0].data.shape[1]), dtype=np.float32)
newdata[0,0,:,:] = pb[0].data
pb[0].data = newdata # naxis will automatically update to 4 in the header
# fix nans introduced in primary beam by montage at edges and write to new file
print(' A small buffer of NaNs is introduced around the image by Montage when regridding to match the size, \n these have been set to the value of their nearest neighbours to maintain the same image dimensions')
mask = np.isnan(pb[0].data)
pb[0].data[mask] = np.interp(np.flatnonzero(mask), np.flatnonzero(~mask), pb[0].data[~mask])
pb.flush()
pb.close()
# apply primary beam correction
pb = fits.open(pbname[:-5]+'_cutout_regrid.fits')[0]
hdu.data = hdu.data / pb.data
hdu.writeto(imagename[:-5]+'_PBCOR.fits', overwrite=True)
print(' Primary beam correction applied to {0}'.format(imagename[:-5]+'_PBCOR.fits') )
# ------ ------ ------ ------ ------ ------ ------ ------ ------ ------
def do_image_chopping(input_image, split_into):
hdu = fits.open(input_image)[0]
wcs = WCS(hdu.header)
# currently hard coded to only accept square images
im_width = hdu.header['NAXIS1'] # get image width
print(' Input fits image dimensions: {0}'.format(im_width))
print(' Cutting into {0} images of dimensions {1}'.format(split_into**2, im_width/split_into))
# get centre positions for each new fits image. assuming x=y. divide image width by split_into*2
positions = np.array(range(1,(split_into*2),2))*(im_width/(split_into*2))
# round to integer as in pixel coordinates. this approximation shouldn't matter since we include a buffer later
positions = positions.astype(int) # keep as original
positions_x = positions # make copy to append to in loop
positions_y = positions # make copy to append to in loop
# Make a 2D array of all centre positions. length = split_into**2.
for i in range(split_into-1):
# stack x coords repeating split_into times.
positions_x = np.hstack(( positions_x, positions )) # e.g. [ x1, x2, x3, x4, x1, x2, x3, x4, repeat split_into times]
# stack y coords, but np.roll shifts array indices by 1 to get different combinations
positions_y = np.hstack(( positions_y, np.roll(positions,i+1) )) # e.g. [ (y1, y2, y3, y4), (y2, y3, y4, y1), (y3, y4, y1, y2), ... ]
# create 2D array with coordinates: [ [x1,y1], [x2,y2], [x3,y3]... ]
position_coords_inpixels = np.array([positions_x,positions_y]).T
# create buffer of 5% so images overlap. This can be small... only needs to account for image edge cutting through
size = (im_width/split_into) * 1.05 # e.g. 4000 pixel image becomes 4200. sifting to remove duplicates later
# size array needs to be same shape as position_coords_inpixels
size_inpixels = np.array([[size,size]]*(split_into**2)).astype(int)
# loop over images to be cut out
plt.figure() # plot original image and overlay cutout boundaries at the end.
data[data<1e-7]=1e-7 # min pixel brightness to display
data[data>1e-5]=1e-5 # max pixel brightness to display
plt.imshow(hdu.data[0,0,:,:], origin='lower')
colourlist=iter(cm.rainbow(np.linspace(0,1,split_into**2))) # each cutout a different colour
for i in range(split_into**2):
pri | # get cutout file names, must be in same order so they are matched correctly
images_560 = sorted(glob.glob('560*_cutout.fits'))
images_1400 = sorted(glob.glob('1400*_cutout.fits'))
# loop over image cutouts to make cube for each of them
for file560, file1400, i in zip(images_560, images_1400, range(len(images_560))):
print(' Making cube {0} of {1}'.format(i, len(images_560)-1))
hdu560 = fits.open(file560)[0]
hdu1400 = fits.open(file1400)[0]
# make cube from the input files along freq axis
cube = np.zeros((2,hdu560.data.shape[0],hdu560.data.shape[1]))
cube[0,:,:] | nt(' Cutting out image {0} of {1}'.format(i+1, split_into**2))
cutout = Cutout2D(hdu.data[0,0,:,:], position=tuple(position_coords_inpixels[i], size=tuple(size_inpixels[i]), mode='trim', wcs=wcs.celestial, copy=True)
cutout.plot_on_original(color=next(colourlist))
# Update the FITS header with the cutout WCS by hand using my own function
hdu = update_header_from_cutout2D(hdu, cutout)
hdu.writeto(input_image[:-5]+'_'+str(i)+'_cutout.fits', overwrite=True) # Write the cutout to a new FITS file
print(' Saving cutout arrangement as {0}'.format(input_image+'_cutouts.png'))
plt.savefig(input_image+'_cutout_annotation.png')
# ------ ------ ------ ------ ------ ------ ------ ------ ------ ------
# make image cube for pybdsf spectral index mode, looping over all cutouts
def make_image_cubes_for_cutouts():
| conditional_block |
pipeline_cutouts.py | j, name ):
with open(name + '.pkl', 'wb') as f:
pickle.dump(obj, f, pickle.HIGHEST_PROTOCOL)
def load_obj(name ):
with open(name + '.pkl', 'rb') as f:
return pickle.load(f)
# ------ ------ ------ ------ ------ ------ ------ ------ ------ ------
def update_header_from_cutout2D(hdu, cutout):
# update data
newdata = np.zeros((1,1,cutout.data.shape[0], cutout.data.shape[1]), dtype=np.float32)
newdata[0,0,:,:] = cutout.data
hdu.data = newdata
# update header cards returned from cutout2D wcs:
hdu.header.set('CRVAL1', cutout.wcs.wcs.crval[0])
hdu.header.set('CRVAL2', cutout.wcs.wcs.crval[1])
hdu.header.set('CRPIX1', cutout.wcs.wcs.crpix[0])
hdu.header.set('CRPIX2', cutout.wcs.wcs.crpix[1])
hdu.header.set('CDELT1', cutout.wcs.wcs.cdelt[0])
hdu.header.set('CDELT2', cutout.wcs.wcs.cdelt[1])
hdu.header.set('NAXIS1', cutout.wcs.pixel_shape[0])
hdu.header.set('NAXIS2', cutout.wcs.pixel_shape[1])
return hdu
# ------ ------ ------ ------ ------ ------ ------ ------ ------ ------
def do_primarybeam_correction(pbname, imagename):
print(' Preparing to apply the primary beam correction to {0}'.format(imagename))
hdu = fits.open(imagename)[0]
pb = fits.open(pbname)[0]
wcs = WCS(pb.header)
# cutout pb field of view to match image field of view
x_size = hdu.header['NAXIS1']
x_pixel_deg = hdu.header['CDELT2'] # CDELT1 is negative, so take positive one
size = (x_size*x_pixel_deg*u.degree, x_size*x_pixel_deg*u.degree) # angular size of cutout, using astropy coord. approx 32768*0.6 arcseconds.
position = SkyCoord(pb.header['CRVAL1']*u.degree, pb.header['CRVAL2']*u.degree) # RA and DEC of beam PB pointing
print(' Cutting out image FOV from primary beam image...')
cutout = Cutout2D(pb.data[0,0,:,:], position=position, size=size, mode='trim', wcs=wcs.celestial, copy=True)
# Update the FITS header with the cutout WCS by hand using my own function
# don't use cutout.wcs.to_header() because it doesn't account for the freq and stokes axes. is only compatible with 2D fits images.
#pb.header.update(cutout.wcs.to_header()) #
pb = update_header_from_cutout2D(pb, cutout)
# write updated fits file to disk
pb.writeto(pbname[:-5]+'_cutout.fits', overwrite=True) # Write the cutout to a new FITS file
# regrid PB image cutout to match pixel scale of the image FOV
print(' Regridding image...')
# get header of image to match PB to
montage.mGetHdr(imagename, 'hdu_tmp.hdr')
# regrid pb image (270 pixels) to size of ref image (32k pixels)
montage.reproject(in_images=pbname[:-5]+'_cutout.fits', out_images=pbname[:-5]+'_cutout_regrid.fits', header='hdu_tmp.hdr', exact_size=True)
os.remove('hdu_tmp.hdr') # get rid of header text file saved to disk
# update montage output to float32
pb = fits.open(pbname[:-5]+'_cutout_regrid.fits', mode='update')
newdata = np.zeros((1,1,pb[0].data.shape[0], pb[0].data.shape[1]), dtype=np.float32)
newdata[0,0,:,:] = pb[0].data
pb[0].data = newdata # naxis will automatically update to 4 in the header
# fix nans introduced in primary beam by montage at edges and write to new file
print(' A small buffer of NaNs is introduced around the image by Montage when regridding to match the size, \n these have been set to the value of their nearest neighbours to maintain the same image dimensions')
mask = np.isnan(pb[0].data)
pb[0].data[mask] = np.interp(np.flatnonzero(mask), np.flatnonzero(~mask), pb[0].data[~mask])
pb.flush()
pb.close()
# apply primary beam correction
pb = fits.open(pbname[:-5]+'_cutout_regrid.fits')[0]
hdu.data = hdu.data / pb.data
hdu.writeto(imagename[:-5]+'_PBCOR.fits', overwrite=True)
print(' Primary beam correction applied to {0}'.format(imagename[:-5]+'_PBCOR.fits') )
# ------ ------ ------ ------ ------ ------ ------ ------ ------ ------
def do_image_chopping(input_image, split_into):
hdu = fits.open(input_image)[0]
wcs = WCS(hdu.header)
# currently hard coded to only accept square images
im_width = hdu.header['NAXIS1'] # get image width
print(' Input fits image dimensions: {0}'.format(im_width))
print(' Cutting into {0} images of dimensions {1}'.format(split_into**2, im_width/split_into))
# get centre positions for each new fits image. assuming x=y. divide image width by split_into*2
positions = np.array(range(1,(split_into*2),2))*(im_width/(split_into*2))
# round to integer as in pixel coordinates. this approximation shouldn't matter since we include a buffer later
positions = positions.astype(int) # keep as original
positions_x = positions # make copy to append to in loop
positions_y = positions # make copy to append to in loop
# Make a 2D array of all centre positions. length = split_into**2.
for i in range(split_into-1):
# stack x coords repeating split_into times.
positions_x = np.hstack(( positions_x, positions )) # e.g. [ x1, x2, x3, x4, x1, x2, x3, x4, repeat split_into times]
# stack y coords, but np.roll shifts array indices by 1 to get different combinations
positions_y = np.hstack(( positions_y, np.roll(positions,i+1) )) # e.g. [ (y1, y2, y3, y4), (y2, y3, y4, y1), (y3, y4, y1, y2), ... ]
# create 2D array with coordinates: [ [x1,y1], [x2,y2], [x3,y3]... ]
position_coords_inpixels = np.array([positions_x,positions_y]).T
# create buffer of 5% so images overlap. This can be small... only needs to account for image edge cutting through
size = (im_width/split_into) * 1.05 # e.g. 4000 pixel image becomes 4200. sifting to remove duplicates later
# size array needs to be same shape as position_coords_inpixels
size_inpixels = np.array([[size,size]]*(split_into**2)).astype(int)
# loop over images to be cut out
plt.figure() # plot original image and overlay cutout boundaries at the end.
data[data<1e-7]=1e-7 # min pixel brightness to display
data[data>1e-5]=1e-5 # max pixel brightness to display
plt.imshow(hdu.data[0,0,:,:], origin='lower')
colourlist=iter(cm.rainbow(np.linspace(0,1,split_into**2))) # each cutout a different colour
for i in range(split_into**2):
print(' Cutting out image {0} of {1}'.format(i+1, split_into**2))
cutout = Cutout2D(hdu.data[0,0,:,:], position=tuple(position_coords_inpixels[i], size=tuple(size_inpixels[i]), mode='trim', wcs=wcs.celestial, copy=True)
cutout.plot_on_original(color=next(colourlist))
# Update the FITS header with the cutout WCS by hand using my own function
hdu = update_header_from_cutout2D(hdu, cutout)
hdu.writeto(input_image[:-5]+'_'+str(i)+'_cutout.fits', overwrite=True) # Write the cutout to a new FITS file
print(' Saving cutout arrangement as {0}'.format(input_image+'_cutouts.png'))
plt.savefig(input_image+'_cutout_annotation.png')
# ------ ------ ------ ------ ------ ------ ------ ------ ------ ------
# make image cube for pybdsf spectral index mode, | e_obj(ob | identifier_name |
|
pipeline_cutouts.py | ', cutout.wcs.wcs.crpix[1])
hdu.header.set('CDELT1', cutout.wcs.wcs.cdelt[0])
hdu.header.set('CDELT2', cutout.wcs.wcs.cdelt[1])
hdu.header.set('NAXIS1', cutout.wcs.pixel_shape[0])
hdu.header.set('NAXIS2', cutout.wcs.pixel_shape[1])
return hdu
# ------ ------ ------ ------ ------ ------ ------ ------ ------ ------
def do_primarybeam_correction(pbname, imagename):
print(' Preparing to apply the primary beam correction to {0}'.format(imagename))
hdu = fits.open(imagename)[0]
pb = fits.open(pbname)[0]
wcs = WCS(pb.header)
# cutout pb field of view to match image field of view
x_size = hdu.header['NAXIS1']
x_pixel_deg = hdu.header['CDELT2'] # CDELT1 is negative, so take positive one
size = (x_size*x_pixel_deg*u.degree, x_size*x_pixel_deg*u.degree) # angular size of cutout, using astropy coord. approx 32768*0.6 arcseconds.
position = SkyCoord(pb.header['CRVAL1']*u.degree, pb.header['CRVAL2']*u.degree) # RA and DEC of beam PB pointing
print(' Cutting out image FOV from primary beam image...')
cutout = Cutout2D(pb.data[0,0,:,:], position=position, size=size, mode='trim', wcs=wcs.celestial, copy=True)
# Update the FITS header with the cutout WCS by hand using my own function
# don't use cutout.wcs.to_header() because it doesn't account for the freq and stokes axes. is only compatible with 2D fits images.
#pb.header.update(cutout.wcs.to_header()) #
pb = update_header_from_cutout2D(pb, cutout)
# write updated fits file to disk
pb.writeto(pbname[:-5]+'_cutout.fits', overwrite=True) # Write the cutout to a new FITS file
# regrid PB image cutout to match pixel scale of the image FOV
print(' Regridding image...')
# get header of image to match PB to
montage.mGetHdr(imagename, 'hdu_tmp.hdr')
# regrid pb image (270 pixels) to size of ref image (32k pixels)
montage.reproject(in_images=pbname[:-5]+'_cutout.fits', out_images=pbname[:-5]+'_cutout_regrid.fits', header='hdu_tmp.hdr', exact_size=True)
os.remove('hdu_tmp.hdr') # get rid of header text file saved to disk
# update montage output to float32
pb = fits.open(pbname[:-5]+'_cutout_regrid.fits', mode='update')
newdata = np.zeros((1,1,pb[0].data.shape[0], pb[0].data.shape[1]), dtype=np.float32)
newdata[0,0,:,:] = pb[0].data
pb[0].data = newdata # naxis will automatically update to 4 in the header
# fix nans introduced in primary beam by montage at edges and write to new file
print(' A small buffer of NaNs is introduced around the image by Montage when regridding to match the size, \n these have been set to the value of their nearest neighbours to maintain the same image dimensions')
mask = np.isnan(pb[0].data)
pb[0].data[mask] = np.interp(np.flatnonzero(mask), np.flatnonzero(~mask), pb[0].data[~mask])
pb.flush()
pb.close()
# apply primary beam correction
pb = fits.open(pbname[:-5]+'_cutout_regrid.fits')[0]
hdu.data = hdu.data / pb.data
hdu.writeto(imagename[:-5]+'_PBCOR.fits', overwrite=True)
print(' Primary beam correction applied to {0}'.format(imagename[:-5]+'_PBCOR.fits') )
# ------ ------ ------ ------ ------ ------ ------ ------ ------ ------
def do_image_chopping(input_image, split_into):
hdu |
# create 2D array with coordinates: [ [x1,y1], [x2,y2], [x3,y3]... ]
position_coords_inpixels = np.array([positions_x,positions_y]).T
# create buffer of 5% so images overlap. This can be small... only needs to account for image edge cutting through
size = (im_width/split_into) * 1.05 # e.g. 4000 pixel image becomes 4200. sifting to remove duplicates later
# size array needs to be same shape as position_coords_inpixels
size_inpixels = np.array([[size,size]]*(split_into**2)).astype(int)
# loop over images to be cut out
plt.figure() # plot original image and overlay cutout boundaries at the end.
data[data<1e-7]=1e-7 # min pixel brightness to display
data[data>1e-5]=1e-5 # max pixel brightness to display
plt.imshow(hdu.data[0,0,:,:], origin='lower')
colourlist=iter(cm.rainbow(np.linspace(0,1,split_into**2))) # each cutout a different colour
for i in range(split_into**2):
print(' Cutting out image {0} of {1}'.format(i+1, split_into**2))
cutout = Cutout2D(hdu.data[0,0,:,:], position=tuple(position_coords_inpixels[i], size=tuple(size_inpixels[i]), mode='trim', wcs=wcs.celestial, copy=True)
cutout.plot_on_original(color=next(colourlist))
# Update the FITS header with the cutout WCS by hand using my own function
hdu = update_header_from_cutout2D(hdu, cutout)
hdu.writeto(input_image[:-5]+'_'+str(i)+'_cutout.fits', overwrite=True) # Write the cutout to a new FITS file
print(' Saving cutout arrangement as {0}'.format(input_image+'_cutouts.png'))
plt.savefig(input_image+'_cutout_annotation.png')
# ------ ------ ------ ------ ------ ------ ------ ------ ------ ------
# make image cube for pybdsf spectral index mode, looping over all cutouts
def make_image_cubes_for_cutouts():
# get cutout file names, must be in same order so they are matched correctly
images_560 = sorted(glob.glob('560*_cutout.fits'))
images_1400 = sorted(glob.glob('1400*_cutout.fits'))
# loop over image cutouts to make cube for each of them
for file560, file1400, i in zip(images_560, images_1400, range(len(images_560))):
print(' Making cube {0} of {1}'.format(i, len(images_560)-1))
hdu560 = fits.open(file560)[0]
hdu1400 = fits.open(file1400)[0]
# make cube from the input files along freq axis
cube = np.zeros((2,hdu560.data.shape[0],hdu560.data.shape[1]))
cube[0,:,:] = | = fits.open(input_image)[0]
wcs = WCS(hdu.header)
# currently hard coded to only accept square images
im_width = hdu.header['NAXIS1'] # get image width
print(' Input fits image dimensions: {0}'.format(im_width))
print(' Cutting into {0} images of dimensions {1}'.format(split_into**2, im_width/split_into))
# get centre positions for each new fits image. assuming x=y. divide image width by split_into*2
positions = np.array(range(1,(split_into*2),2))*(im_width/(split_into*2))
# round to integer as in pixel coordinates. this approximation shouldn't matter since we include a buffer later
positions = positions.astype(int) # keep as original
positions_x = positions # make copy to append to in loop
positions_y = positions # make copy to append to in loop
# Make a 2D array of all centre positions. length = split_into**2.
for i in range(split_into-1):
# stack x coords repeating split_into times.
positions_x = np.hstack(( positions_x, positions )) # e.g. [ x1, x2, x3, x4, x1, x2, x3, x4, repeat split_into times]
# stack y coords, but np.roll shifts array indices by 1 to get different combinations
positions_y = np.hstack(( positions_y, np.roll(positions,i+1) )) # e.g. [ (y1, y2, y3, y4), (y2, y3, y4, y1), (y3, y4, y1, y2), ... ] | identifier_body |
pipeline_cutouts.py | ', cutout.wcs.wcs.crpix[1])
hdu.header.set('CDELT1', cutout.wcs.wcs.cdelt[0])
hdu.header.set('CDELT2', cutout.wcs.wcs.cdelt[1])
hdu.header.set('NAXIS1', cutout.wcs.pixel_shape[0])
hdu.header.set('NAXIS2', cutout.wcs.pixel_shape[1])
return hdu
# ------ ------ ------ ------ ------ ------ ------ ------ ------ ------
def do_primarybeam_correction(pbname, imagename):
print(' Preparing to apply the primary beam correction to {0}'.format(imagename))
hdu = fits.open(imagename)[0]
pb = fits.open(pbname)[0]
wcs = WCS(pb.header)
# cutout pb field of view to match image field of view
x_size = hdu.header['NAXIS1']
x_pixel_deg = hdu.header['CDELT2'] # CDELT1 is negative, so take positive one
size = (x_size*x_pixel_deg*u.degree, x_size*x_pixel_deg*u.degree) # angular size of cutout, using astropy coord. approx 32768*0.6 arcseconds.
position = SkyCoord(pb.header['CRVAL1']*u.degree, pb.header['CRVAL2']*u.degree) # RA and DEC of beam PB pointing
print(' Cutting out image FOV from primary beam image...')
cutout = Cutout2D(pb.data[0,0,:,:], position=position, size=size, mode='trim', wcs=wcs.celestial, copy=True)
# Update the FITS header with the cutout WCS by hand using my own function
# don't use cutout.wcs.to_header() because it doesn't account for the freq and stokes axes. is only compatible with 2D fits images.
#pb.header.update(cutout.wcs.to_header()) #
pb = update_header_from_cutout2D(pb, cutout)
# write updated fits file to disk
pb.writeto(pbname[:-5]+'_cutout.fits', overwrite=True) # Write the cutout to a new FITS file
# regrid PB image cutout to match pixel scale of the image FOV
print(' Regridding image...')
# get header of image to match PB to
montage.mGetHdr(imagename, 'hdu_tmp.hdr')
# regrid pb image (270 pixels) to size of ref image (32k pixels)
montage.reproject(in_images=pbname[:-5]+'_cutout.fits', out_images=pbname[:-5]+'_cutout_regrid.fits', header='hdu_tmp.hdr', exact_size=True)
os.remove('hdu_tmp.hdr') # get rid of header text file saved to disk
# update montage output to float32
pb = fits.open(pbname[:-5]+'_cutout_regrid.fits', mode='update')
newdata = np.zeros((1,1,pb[0].data.shape[0], pb[0].data.shape[1]), dtype=np.float32)
newdata[0,0,:,:] = pb[0].data
pb[0].data = newdata # naxis will automatically update to 4 in the header
# fix nans introduced in primary beam by montage at edges and write to new file
print(' A small buffer of NaNs is introduced around the image by Montage when regridding to match the size, \n these have been set to the value of their nearest neighbours to maintain the same image dimensions')
mask = np.isnan(pb[0].data)
pb[0].data[mask] = np.interp(np.flatnonzero(mask), np.flatnonzero(~mask), pb[0].data[~mask])
pb.flush()
pb.close()
# apply primary beam correction
pb = fits.open(pbname[:-5]+'_cutout_regrid.fits')[0]
hdu.data = hdu.data / pb.data
hdu.writeto(imagename[:-5]+'_PBCOR.fits', overwrite=True)
print(' Primary beam correction applied to {0}'.format(imagename[:-5]+'_PBCOR.fits') ) |
def do_image_chopping(input_image, split_into):
hdu = fits.open(input_image)[0]
wcs = WCS(hdu.header)
# currently hard coded to only accept square images
im_width = hdu.header['NAXIS1'] # get image width
print(' Input fits image dimensions: {0}'.format(im_width))
print(' Cutting into {0} images of dimensions {1}'.format(split_into**2, im_width/split_into))
# get centre positions for each new fits image. assuming x=y. divide image width by split_into*2
positions = np.array(range(1,(split_into*2),2))*(im_width/(split_into*2))
# round to integer as in pixel coordinates. this approximation shouldn't matter since we include a buffer later
positions = positions.astype(int) # keep as original
positions_x = positions # make copy to append to in loop
positions_y = positions # make copy to append to in loop
# Make a 2D array of all centre positions. length = split_into**2.
for i in range(split_into-1):
# stack x coords repeating split_into times.
positions_x = np.hstack(( positions_x, positions )) # e.g. [ x1, x2, x3, x4, x1, x2, x3, x4, repeat split_into times]
# stack y coords, but np.roll shifts array indices by 1 to get different combinations
positions_y = np.hstack(( positions_y, np.roll(positions,i+1) )) # e.g. [ (y1, y2, y3, y4), (y2, y3, y4, y1), (y3, y4, y1, y2), ... ]
# create 2D array with coordinates: [ [x1,y1], [x2,y2], [x3,y3]... ]
position_coords_inpixels = np.array([positions_x,positions_y]).T
# create buffer of 5% so images overlap. This can be small... only needs to account for image edge cutting through
size = (im_width/split_into) * 1.05 # e.g. 4000 pixel image becomes 4200. sifting to remove duplicates later
# size array needs to be same shape as position_coords_inpixels
size_inpixels = np.array([[size,size]]*(split_into**2)).astype(int)
# loop over images to be cut out
plt.figure() # plot original image and overlay cutout boundaries at the end.
data[data<1e-7]=1e-7 # min pixel brightness to display
data[data>1e-5]=1e-5 # max pixel brightness to display
plt.imshow(hdu.data[0,0,:,:], origin='lower')
colourlist=iter(cm.rainbow(np.linspace(0,1,split_into**2))) # each cutout a different colour
for i in range(split_into**2):
print(' Cutting out image {0} of {1}'.format(i+1, split_into**2))
cutout = Cutout2D(hdu.data[0,0,:,:], position=tuple(position_coords_inpixels[i], size=tuple(size_inpixels[i]), mode='trim', wcs=wcs.celestial, copy=True)
cutout.plot_on_original(color=next(colourlist))
# Update the FITS header with the cutout WCS by hand using my own function
hdu = update_header_from_cutout2D(hdu, cutout)
hdu.writeto(input_image[:-5]+'_'+str(i)+'_cutout.fits', overwrite=True) # Write the cutout to a new FITS file
print(' Saving cutout arrangement as {0}'.format(input_image+'_cutouts.png'))
plt.savefig(input_image+'_cutout_annotation.png')
# ------ ------ ------ ------ ------ ------ ------ ------ ------ ------
# make image cube for pybdsf spectral index mode, looping over all cutouts
def make_image_cubes_for_cutouts():
# get cutout file names, must be in same order so they are matched correctly
images_560 = sorted(glob.glob('560*_cutout.fits'))
images_1400 = sorted(glob.glob('1400*_cutout.fits'))
# loop over image cutouts to make cube for each of them
for file560, file1400, i in zip(images_560, images_1400, range(len(images_560))):
print(' Making cube {0} of {1}'.format(i, len(images_560)-1))
hdu560 = fits.open(file560)[0]
hdu1400 = fits.open(file1400)[0]
# make cube from the input files along freq axis
cube = np.zeros((2,hdu560.data.shape[0],hdu560.data.shape[1]))
cube[0,:,:] |
# ------ ------ ------ ------ ------ ------ ------ ------ ------ ------ | random_line_split |
web.go | `,`description`,`last_update`,`user_id`,`group`,`unread`,`active` FROM home_view WHERE user_id=%d", userId)
if err != nil {
respondError( w, err.Error() )
return
}
feeds := make([]FeedViewItem, len(rows))
for id, row := range rows {
feeds[id] = FeedViewItem{row.Int(0), row.Str(1), row.Str(2), row.Str(3), row.Int64(4), row.Str(6), row.Int(7), row.Int(8)}
}
b, err := json.Marshal(HomeView{feeds})
if err != nil {
respondError( w, err.Error() )
return
}
fmt.Fprint(w, string(b))
//c.Format = goweb.JSON_FORMAT
//c.RespondWithData(HomeView{feeds})
}
func getFeedsQueryFromForm(r *http.Request) (string, error) {
group := r.FormValue("group")
feed := r.FormValue("feed")
starred := r.FormValue("starred")
if group != "" {
return fmt.Sprintf("`group`='%s'", gConn.Escape(group)), nil
} else if feed != "" {
feedId, err := strconv.ParseInt(feed, 10, 64)
if err != nil {
return "", err
}
return fmt.Sprintf("`feedid`=%d", feedId), nil
} else if starred != "" {
return fmt.Sprintf("`label`='star'"), nil
}
return "", nil
}
func serveFeedItems(w http.ResponseWriter, r *http.Request) {
var userId uint32 = 1
searchQuery := fmt.Sprintf("userid=%d", userId)
start, err := strconv.ParseInt(r.FormValue("start"), 10, 64)
if err != nil {
start = 0
}
extraSearch, err := getFeedsQueryFromForm(r)
if err != nil {
respondError( w, err.Error() )
return
}
if extraSearch != "" {
searchQuery = fmt.Sprintf("%s AND %s", searchQuery, extraSearch )
}
conn := GetConnection().Clone()
if err := conn.Connect(); err != nil {
respondError( w, err.Error() )
return
}
defer conn.Close()
rows, _, err := conn.Query("SELECT `id`,`title`,`published`,`updated`,`link`,`author`,`feedtitle`,`feedid`,`is_read`,`label` FROM `entrylist` WHERE %s ORDER BY `updated` DESC LIMIT %d,100", searchQuery, start)
if err != nil {
respondError( w, err.Error() )
return
}
feeds := make([]ChannelViewItem, len(rows))
for id, row := range rows {
feeds[id] = ChannelViewItem{row.Int(0), row.Str(1), row.Str(2), row.Int64(3), row.Str(4), row.Str(5), row.Str(6), row.Int(7), row.Int(8), row.Str(9) }
}
b, err := json.Marshal(ChannelView{feeds})
if err != nil {
respondError( w, err.Error() )
return
}
fmt.Fprint(w, string(b))
//c.Format = goweb.JSON_FORMAT
//c.RespondWithData(ChannelView{feeds})
}
type FeedEntryModel struct {
Content string
}
func serveGetItem(w http.ResponseWriter, r *http.Request) | if err != nil {
respondError( w, err.Error() )
return
}
if len(rows) == 0 {
fmt.Fprint(w, "{\"error\": \"Could not find entry\"}")
return
}
row := rows[0]
b, err := json.Marshal(FeedEntryModel{row.Str(0)})
if err != nil {
respondError( w, err.Error() )
return
}
//TODO:The update and replace statments should be in a transaction
//TODO: Only insert into user_feed_readitems when the items is newer than the user_feeds.newest_read
_, _, err = conn.QueryFirst("REPLACE INTO `user_feed_readitems`(user_id,entry_id) VALUES (%d,%d)", userId, id)
if err != nil {
respondError( w, err.Error() )
return
}
/*
_, _, err = GetConnection().QueryFirst("UPDATE user_feed SET unread_items=GREATEST(unread_items-1,0) WHERE user_id=%d AND feed_id=%d", userId, feedId)
if err != nil {
panic(err)
}
*/
//TODO: Extremelly inneficient; Make better method
_, _, err = conn.QueryFirst("CALL update_unread()")
if err != nil {
respondError( w, err.Error() )
return
}
fmt.Fprint(w, string(b))
}
func serveUpdateItemLabels(w http.ResponseWriter, r *http.Request) {
if r.Method != "POST" {
w.WriteHeader(400)
fmt.Fprint(w, "Not a post request.")
return
}
conn := GetConnection().Clone()
if err := conn.Connect(); err != nil {
respondError( w, err.Error() )
return
}
defer conn.Close()
entryId, err := strconv.ParseInt(r.FormValue("id"), 10, 64)
labels := conn.Escape( r.FormValue("labels") )
userId := 1
if err != nil {
respondError( w, err.Error() )
return
}
getQuery := func() string {
if labels == "" {
return fmt.Sprintf("DELETE FROM `user_entry_label` WHERE `user_id`=%d AND `feed_entry_id`=%d", userId, entryId )
}
return fmt.Sprintf("INSERT IGNORE INTO `user_entry_label`(`user_id`,`feed_entry_id`,`label`) VALUES (%d,%d,'%s')", userId, entryId, labels )
}
_, _, err = conn.Query(getQuery())
if err != nil {
respondError( w, err.Error() )
return
}
fmt.Fprint(w, "{\"success\":1}")
}
func updatePriorities(transaction mysql.Transaction, userId int, newPriorities map[string]GroupInfo) error {
for feedIdStr, priority := range newPriorities {
feedId, err := strconv.ParseInt(feedIdStr, 10, 64)
if err != nil {
return err
}
_, _, err = transaction.Query("UPDATE `user_feed` SET `priority`=%d, `group`='%s' WHERE feed_id=%d AND user_id=%d",
priority.Priority,
gConn.Escape(priority.Group),
feedId,
userId)
if err != nil {
return err
}
}
return nil
}
type GroupInfo struct {
Priority int
Group string
}
func serveUpdateOrder(w http.ResponseWriter, r *http.Request) {
if r.Method != "POST" {
w.WriteHeader(400)
fmt.Fprint(w, "Not a post request.")
return
}
userId := 1
decoder := json.NewDecoder(r.Body)
var newPriorities map[string]GroupInfo = make(map[string]GroupInfo)
err := decoder.Decode(&newPriorities)
if err != nil {
respondError( w, err.Error() )
return
}
conn := GetConnection().Clone()
if err := conn.Connect(); err != nil {
respondError( w, err.Error() )
return
}
defer conn.Close()
transaction, err := conn.Begin()
if err != nil {
respondError( w, err.Error() )
return
}
err = updatePriorities(transaction, userId, newPriorities)
if err != nil {
transaction.Rollback()
respondError( w, err.Error() )
return
}
transaction.Commit()
}
func serveMarkRead(w http.ResponseWriter, r *http.Request) {
if r.Method != "POST" {
w.WriteHeader(400)
fmt.Fprint(w, "Not a post request.")
return
}
var userId uint32 = 1
decoder := json.NewDecoder(r.Body)
var readFeedsIds []int
err := decoder.Decode(&readFeedsIds)
if err != nil {
respondError( w, err.Error() )
return
}
if len(readFeedsIds) == 0 {
respondError( w, "The array of feeds to mark as read is empty." )
return
}
readFeeds := []string {}
for _, id := range(readFeedsIds) {
readFeeds = append(readFeeds, strconv.Itoa( id ) )
}
conn := GetConnection().Clone()
| {
id, err := strconv.ParseInt(r.FormValue("id"), 10, 64)
userId := 1
if err != nil {
respondError( w, err.Error() )
return
}
conn := GetConnection().Clone()
if err := conn.Connect(); err != nil {
respondError( w, err.Error() )
return
}
defer conn.Close()
rows, _, err := conn.Query("SELECT `content` FROM `feed_entry` WHERE id=%d", id)
| identifier_body |
web.go | `,`description`,`last_update`,`user_id`,`group`,`unread`,`active` FROM home_view WHERE user_id=%d", userId)
if err != nil {
respondError( w, err.Error() )
return
}
feeds := make([]FeedViewItem, len(rows))
for id, row := range rows {
feeds[id] = FeedViewItem{row.Int(0), row.Str(1), row.Str(2), row.Str(3), row.Int64(4), row.Str(6), row.Int(7), row.Int(8)}
}
b, err := json.Marshal(HomeView{feeds})
if err != nil {
respondError( w, err.Error() )
return
}
fmt.Fprint(w, string(b))
//c.Format = goweb.JSON_FORMAT
//c.RespondWithData(HomeView{feeds})
}
func getFeedsQueryFromForm(r *http.Request) (string, error) {
group := r.FormValue("group")
feed := r.FormValue("feed")
starred := r.FormValue("starred")
if group != "" {
return fmt.Sprintf("`group`='%s'", gConn.Escape(group)), nil
} else if feed != "" {
feedId, err := strconv.ParseInt(feed, 10, 64)
if err != nil {
return "", err
}
return fmt.Sprintf("`feedid`=%d", feedId), nil
} else if starred != "" {
return fmt.Sprintf("`label`='star'"), nil
}
return "", nil
}
func serveFeedItems(w http.ResponseWriter, r *http.Request) {
var userId uint32 = 1
searchQuery := fmt.Sprintf("userid=%d", userId)
start, err := strconv.ParseInt(r.FormValue("start"), 10, 64)
if err != nil {
start = 0
}
extraSearch, err := getFeedsQueryFromForm(r)
if err != nil {
respondError( w, err.Error() )
return
}
if extraSearch != "" {
searchQuery = fmt.Sprintf("%s AND %s", searchQuery, extraSearch )
}
conn := GetConnection().Clone()
if err := conn.Connect(); err != nil {
respondError( w, err.Error() )
return
}
defer conn.Close()
rows, _, err := conn.Query("SELECT `id`,`title`,`published`,`updated`,`link`,`author`,`feedtitle`,`feedid`,`is_read`,`label` FROM `entrylist` WHERE %s ORDER BY `updated` DESC LIMIT %d,100", searchQuery, start)
if err != nil {
respondError( w, err.Error() )
return
}
feeds := make([]ChannelViewItem, len(rows))
for id, row := range rows {
feeds[id] = ChannelViewItem{row.Int(0), row.Str(1), row.Str(2), row.Int64(3), row.Str(4), row.Str(5), row.Str(6), row.Int(7), row.Int(8), row.Str(9) }
}
b, err := json.Marshal(ChannelView{feeds})
if err != nil {
respondError( w, err.Error() )
return
}
fmt.Fprint(w, string(b))
//c.Format = goweb.JSON_FORMAT
//c.RespondWithData(ChannelView{feeds})
}
type FeedEntryModel struct {
Content string
}
func serveGetItem(w http.ResponseWriter, r *http.Request) {
id, err := strconv.ParseInt(r.FormValue("id"), 10, 64)
userId := 1
if err != nil {
respondError( w, err.Error() )
return
}
conn := GetConnection().Clone()
if err := conn.Connect(); err != nil {
respondError( w, err.Error() )
return
}
defer conn.Close()
rows, _, err := conn.Query("SELECT `content` FROM `feed_entry` WHERE id=%d", id)
if err != nil {
respondError( w, err.Error() )
return
}
if len(rows) == 0 {
fmt.Fprint(w, "{\"error\": \"Could not find entry\"}")
return
}
row := rows[0]
b, err := json.Marshal(FeedEntryModel{row.Str(0)})
if err != nil {
respondError( w, err.Error() )
return
}
//TODO:The update and replace statments should be in a transaction
//TODO: Only insert into user_feed_readitems when the items is newer than the user_feeds.newest_read
_, _, err = conn.QueryFirst("REPLACE INTO `user_feed_readitems`(user_id,entry_id) VALUES (%d,%d)", userId, id)
if err != nil |
/*
_, _, err = GetConnection().QueryFirst("UPDATE user_feed SET unread_items=GREATEST(unread_items-1,0) WHERE user_id=%d AND feed_id=%d", userId, feedId)
if err != nil {
panic(err)
}
*/
//TODO: Extremelly inneficient; Make better method
_, _, err = conn.QueryFirst("CALL update_unread()")
if err != nil {
respondError( w, err.Error() )
return
}
fmt.Fprint(w, string(b))
}
func serveUpdateItemLabels(w http.ResponseWriter, r *http.Request) {
if r.Method != "POST" {
w.WriteHeader(400)
fmt.Fprint(w, "Not a post request.")
return
}
conn := GetConnection().Clone()
if err := conn.Connect(); err != nil {
respondError( w, err.Error() )
return
}
defer conn.Close()
entryId, err := strconv.ParseInt(r.FormValue("id"), 10, 64)
labels := conn.Escape( r.FormValue("labels") )
userId := 1
if err != nil {
respondError( w, err.Error() )
return
}
getQuery := func() string {
if labels == "" {
return fmt.Sprintf("DELETE FROM `user_entry_label` WHERE `user_id`=%d AND `feed_entry_id`=%d", userId, entryId )
}
return fmt.Sprintf("INSERT IGNORE INTO `user_entry_label`(`user_id`,`feed_entry_id`,`label`) VALUES (%d,%d,'%s')", userId, entryId, labels )
}
_, _, err = conn.Query(getQuery())
if err != nil {
respondError( w, err.Error() )
return
}
fmt.Fprint(w, "{\"success\":1}")
}
func updatePriorities(transaction mysql.Transaction, userId int, newPriorities map[string]GroupInfo) error {
for feedIdStr, priority := range newPriorities {
feedId, err := strconv.ParseInt(feedIdStr, 10, 64)
if err != nil {
return err
}
_, _, err = transaction.Query("UPDATE `user_feed` SET `priority`=%d, `group`='%s' WHERE feed_id=%d AND user_id=%d",
priority.Priority,
gConn.Escape(priority.Group),
feedId,
userId)
if err != nil {
return err
}
}
return nil
}
type GroupInfo struct {
Priority int
Group string
}
func serveUpdateOrder(w http.ResponseWriter, r *http.Request) {
if r.Method != "POST" {
w.WriteHeader(400)
fmt.Fprint(w, "Not a post request.")
return
}
userId := 1
decoder := json.NewDecoder(r.Body)
var newPriorities map[string]GroupInfo = make(map[string]GroupInfo)
err := decoder.Decode(&newPriorities)
if err != nil {
respondError( w, err.Error() )
return
}
conn := GetConnection().Clone()
if err := conn.Connect(); err != nil {
respondError( w, err.Error() )
return
}
defer conn.Close()
transaction, err := conn.Begin()
if err != nil {
respondError( w, err.Error() )
return
}
err = updatePriorities(transaction, userId, newPriorities)
if err != nil {
transaction.Rollback()
respondError( w, err.Error() )
return
}
transaction.Commit()
}
func serveMarkRead(w http.ResponseWriter, r *http.Request) {
if r.Method != "POST" {
w.WriteHeader(400)
fmt.Fprint(w, "Not a post request.")
return
}
var userId uint32 = 1
decoder := json.NewDecoder(r.Body)
var readFeedsIds []int
err := decoder.Decode(&readFeedsIds)
if err != nil {
respondError( w, err.Error() )
return
}
if len(readFeedsIds) == 0 {
respondError( w, "The array of feeds to mark as read is empty." )
return
}
readFeeds := []string {}
for _, id := range(readFeedsIds) {
readFeeds = append(readFeeds, strconv.Itoa( id ) )
}
conn := GetConnection().Clone | {
respondError( w, err.Error() )
return
} | conditional_block |
web.go | `,`description`,`last_update`,`user_id`,`group`,`unread`,`active` FROM home_view WHERE user_id=%d", userId)
if err != nil {
respondError( w, err.Error() )
return
}
feeds := make([]FeedViewItem, len(rows))
for id, row := range rows {
feeds[id] = FeedViewItem{row.Int(0), row.Str(1), row.Str(2), row.Str(3), row.Int64(4), row.Str(6), row.Int(7), row.Int(8)}
}
b, err := json.Marshal(HomeView{feeds})
if err != nil {
respondError( w, err.Error() )
return
}
fmt.Fprint(w, string(b))
//c.Format = goweb.JSON_FORMAT
//c.RespondWithData(HomeView{feeds})
}
func getFeedsQueryFromForm(r *http.Request) (string, error) {
group := r.FormValue("group")
feed := r.FormValue("feed")
starred := r.FormValue("starred")
if group != "" {
return fmt.Sprintf("`group`='%s'", gConn.Escape(group)), nil
} else if feed != "" {
feedId, err := strconv.ParseInt(feed, 10, 64)
if err != nil {
return "", err
}
return fmt.Sprintf("`feedid`=%d", feedId), nil
} else if starred != "" {
return fmt.Sprintf("`label`='star'"), nil
}
return "", nil
}
func | (w http.ResponseWriter, r *http.Request) {
var userId uint32 = 1
searchQuery := fmt.Sprintf("userid=%d", userId)
start, err := strconv.ParseInt(r.FormValue("start"), 10, 64)
if err != nil {
start = 0
}
extraSearch, err := getFeedsQueryFromForm(r)
if err != nil {
respondError( w, err.Error() )
return
}
if extraSearch != "" {
searchQuery = fmt.Sprintf("%s AND %s", searchQuery, extraSearch )
}
conn := GetConnection().Clone()
if err := conn.Connect(); err != nil {
respondError( w, err.Error() )
return
}
defer conn.Close()
rows, _, err := conn.Query("SELECT `id`,`title`,`published`,`updated`,`link`,`author`,`feedtitle`,`feedid`,`is_read`,`label` FROM `entrylist` WHERE %s ORDER BY `updated` DESC LIMIT %d,100", searchQuery, start)
if err != nil {
respondError( w, err.Error() )
return
}
feeds := make([]ChannelViewItem, len(rows))
for id, row := range rows {
feeds[id] = ChannelViewItem{row.Int(0), row.Str(1), row.Str(2), row.Int64(3), row.Str(4), row.Str(5), row.Str(6), row.Int(7), row.Int(8), row.Str(9) }
}
b, err := json.Marshal(ChannelView{feeds})
if err != nil {
respondError( w, err.Error() )
return
}
fmt.Fprint(w, string(b))
//c.Format = goweb.JSON_FORMAT
//c.RespondWithData(ChannelView{feeds})
}
type FeedEntryModel struct {
Content string
}
func serveGetItem(w http.ResponseWriter, r *http.Request) {
id, err := strconv.ParseInt(r.FormValue("id"), 10, 64)
userId := 1
if err != nil {
respondError( w, err.Error() )
return
}
conn := GetConnection().Clone()
if err := conn.Connect(); err != nil {
respondError( w, err.Error() )
return
}
defer conn.Close()
rows, _, err := conn.Query("SELECT `content` FROM `feed_entry` WHERE id=%d", id)
if err != nil {
respondError( w, err.Error() )
return
}
if len(rows) == 0 {
fmt.Fprint(w, "{\"error\": \"Could not find entry\"}")
return
}
row := rows[0]
b, err := json.Marshal(FeedEntryModel{row.Str(0)})
if err != nil {
respondError( w, err.Error() )
return
}
//TODO:The update and replace statments should be in a transaction
//TODO: Only insert into user_feed_readitems when the items is newer than the user_feeds.newest_read
_, _, err = conn.QueryFirst("REPLACE INTO `user_feed_readitems`(user_id,entry_id) VALUES (%d,%d)", userId, id)
if err != nil {
respondError( w, err.Error() )
return
}
/*
_, _, err = GetConnection().QueryFirst("UPDATE user_feed SET unread_items=GREATEST(unread_items-1,0) WHERE user_id=%d AND feed_id=%d", userId, feedId)
if err != nil {
panic(err)
}
*/
//TODO: Extremelly inneficient; Make better method
_, _, err = conn.QueryFirst("CALL update_unread()")
if err != nil {
respondError( w, err.Error() )
return
}
fmt.Fprint(w, string(b))
}
func serveUpdateItemLabels(w http.ResponseWriter, r *http.Request) {
if r.Method != "POST" {
w.WriteHeader(400)
fmt.Fprint(w, "Not a post request.")
return
}
conn := GetConnection().Clone()
if err := conn.Connect(); err != nil {
respondError( w, err.Error() )
return
}
defer conn.Close()
entryId, err := strconv.ParseInt(r.FormValue("id"), 10, 64)
labels := conn.Escape( r.FormValue("labels") )
userId := 1
if err != nil {
respondError( w, err.Error() )
return
}
getQuery := func() string {
if labels == "" {
return fmt.Sprintf("DELETE FROM `user_entry_label` WHERE `user_id`=%d AND `feed_entry_id`=%d", userId, entryId )
}
return fmt.Sprintf("INSERT IGNORE INTO `user_entry_label`(`user_id`,`feed_entry_id`,`label`) VALUES (%d,%d,'%s')", userId, entryId, labels )
}
_, _, err = conn.Query(getQuery())
if err != nil {
respondError( w, err.Error() )
return
}
fmt.Fprint(w, "{\"success\":1}")
}
func updatePriorities(transaction mysql.Transaction, userId int, newPriorities map[string]GroupInfo) error {
for feedIdStr, priority := range newPriorities {
feedId, err := strconv.ParseInt(feedIdStr, 10, 64)
if err != nil {
return err
}
_, _, err = transaction.Query("UPDATE `user_feed` SET `priority`=%d, `group`='%s' WHERE feed_id=%d AND user_id=%d",
priority.Priority,
gConn.Escape(priority.Group),
feedId,
userId)
if err != nil {
return err
}
}
return nil
}
type GroupInfo struct {
Priority int
Group string
}
func serveUpdateOrder(w http.ResponseWriter, r *http.Request) {
if r.Method != "POST" {
w.WriteHeader(400)
fmt.Fprint(w, "Not a post request.")
return
}
userId := 1
decoder := json.NewDecoder(r.Body)
var newPriorities map[string]GroupInfo = make(map[string]GroupInfo)
err := decoder.Decode(&newPriorities)
if err != nil {
respondError( w, err.Error() )
return
}
conn := GetConnection().Clone()
if err := conn.Connect(); err != nil {
respondError( w, err.Error() )
return
}
defer conn.Close()
transaction, err := conn.Begin()
if err != nil {
respondError( w, err.Error() )
return
}
err = updatePriorities(transaction, userId, newPriorities)
if err != nil {
transaction.Rollback()
respondError( w, err.Error() )
return
}
transaction.Commit()
}
func serveMarkRead(w http.ResponseWriter, r *http.Request) {
if r.Method != "POST" {
w.WriteHeader(400)
fmt.Fprint(w, "Not a post request.")
return
}
var userId uint32 = 1
decoder := json.NewDecoder(r.Body)
var readFeedsIds []int
err := decoder.Decode(&readFeedsIds)
if err != nil {
respondError( w, err.Error() )
return
}
if len(readFeedsIds) == 0 {
respondError( w, "The array of feeds to mark as read is empty." )
return
}
readFeeds := []string {}
for _, id := range(readFeedsIds) {
readFeeds = append(readFeeds, strconv.Itoa( id ) )
}
conn := GetConnection().Clone()
| serveFeedItems | identifier_name |
web.go | link`,`description`,`last_update`,`user_id`,`group`,`unread`,`active` FROM home_view WHERE user_id=%d", userId)
if err != nil {
respondError( w, err.Error() )
return
}
feeds := make([]FeedViewItem, len(rows))
for id, row := range rows {
feeds[id] = FeedViewItem{row.Int(0), row.Str(1), row.Str(2), row.Str(3), row.Int64(4), row.Str(6), row.Int(7), row.Int(8)}
}
b, err := json.Marshal(HomeView{feeds})
if err != nil {
respondError( w, err.Error() )
return
}
fmt.Fprint(w, string(b))
//c.Format = goweb.JSON_FORMAT
//c.RespondWithData(HomeView{feeds})
}
func getFeedsQueryFromForm(r *http.Request) (string, error) {
group := r.FormValue("group")
feed := r.FormValue("feed")
starred := r.FormValue("starred")
if group != "" {
return fmt.Sprintf("`group`='%s'", gConn.Escape(group)), nil
} else if feed != "" {
feedId, err := strconv.ParseInt(feed, 10, 64)
if err != nil {
return "", err
}
return fmt.Sprintf("`feedid`=%d", feedId), nil
} else if starred != "" {
return fmt.Sprintf("`label`='star'"), nil
}
return "", nil
}
func serveFeedItems(w http.ResponseWriter, r *http.Request) {
var userId uint32 = 1
searchQuery := fmt.Sprintf("userid=%d", userId)
start, err := strconv.ParseInt(r.FormValue("start"), 10, 64)
if err != nil {
start = 0
}
extraSearch, err := getFeedsQueryFromForm(r)
if err != nil {
respondError( w, err.Error() )
return
}
if extraSearch != "" {
searchQuery = fmt.Sprintf("%s AND %s", searchQuery, extraSearch )
}
conn := GetConnection().Clone()
if err := conn.Connect(); err != nil {
respondError( w, err.Error() )
return
}
defer conn.Close()
rows, _, err := conn.Query("SELECT `id`,`title`,`published`,`updated`,`link`,`author`,`feedtitle`,`feedid`,`is_read`,`label` FROM `entrylist` WHERE %s ORDER BY `updated` DESC LIMIT %d,100", searchQuery, start)
if err != nil {
respondError( w, err.Error() )
return
}
feeds := make([]ChannelViewItem, len(rows))
for id, row := range rows {
feeds[id] = ChannelViewItem{row.Int(0), row.Str(1), row.Str(2), row.Int64(3), row.Str(4), row.Str(5), row.Str(6), row.Int(7), row.Int(8), row.Str(9) }
}
b, err := json.Marshal(ChannelView{feeds})
if err != nil {
respondError( w, err.Error() )
return
}
fmt.Fprint(w, string(b))
//c.Format = goweb.JSON_FORMAT
//c.RespondWithData(ChannelView{feeds})
}
type FeedEntryModel struct {
Content string
}
func serveGetItem(w http.ResponseWriter, r *http.Request) {
id, err := strconv.ParseInt(r.FormValue("id"), 10, 64)
userId := 1
if err != nil {
respondError( w, err.Error() )
return
}
conn := GetConnection().Clone()
if err := conn.Connect(); err != nil {
respondError( w, err.Error() )
return
}
defer conn.Close()
rows, _, err := conn.Query("SELECT `content` FROM `feed_entry` WHERE id=%d", id)
if err != nil {
respondError( w, err.Error() )
return
}
if len(rows) == 0 {
fmt.Fprint(w, "{\"error\": \"Could not find entry\"}")
return
}
row := rows[0]
b, err := json.Marshal(FeedEntryModel{row.Str(0)})
if err != nil {
respondError( w, err.Error() )
return
}
//TODO:The update and replace statments should be in a transaction
//TODO: Only insert into user_feed_readitems when the items is newer than the user_feeds.newest_read
_, _, err = conn.QueryFirst("REPLACE INTO `user_feed_readitems`(user_id,entry_id) VALUES (%d,%d)", userId, id)
if err != nil {
respondError( w, err.Error() )
return
}
/*
_, _, err = GetConnection().QueryFirst("UPDATE user_feed SET unread_items=GREATEST(unread_items-1,0) WHERE user_id=%d AND feed_id=%d", userId, feedId)
if err != nil {
panic(err)
}
*/ | respondError( w, err.Error() )
return
}
fmt.Fprint(w, string(b))
}
func serveUpdateItemLabels(w http.ResponseWriter, r *http.Request) {
if r.Method != "POST" {
w.WriteHeader(400)
fmt.Fprint(w, "Not a post request.")
return
}
conn := GetConnection().Clone()
if err := conn.Connect(); err != nil {
respondError( w, err.Error() )
return
}
defer conn.Close()
entryId, err := strconv.ParseInt(r.FormValue("id"), 10, 64)
labels := conn.Escape( r.FormValue("labels") )
userId := 1
if err != nil {
respondError( w, err.Error() )
return
}
getQuery := func() string {
if labels == "" {
return fmt.Sprintf("DELETE FROM `user_entry_label` WHERE `user_id`=%d AND `feed_entry_id`=%d", userId, entryId )
}
return fmt.Sprintf("INSERT IGNORE INTO `user_entry_label`(`user_id`,`feed_entry_id`,`label`) VALUES (%d,%d,'%s')", userId, entryId, labels )
}
_, _, err = conn.Query(getQuery())
if err != nil {
respondError( w, err.Error() )
return
}
fmt.Fprint(w, "{\"success\":1}")
}
func updatePriorities(transaction mysql.Transaction, userId int, newPriorities map[string]GroupInfo) error {
for feedIdStr, priority := range newPriorities {
feedId, err := strconv.ParseInt(feedIdStr, 10, 64)
if err != nil {
return err
}
_, _, err = transaction.Query("UPDATE `user_feed` SET `priority`=%d, `group`='%s' WHERE feed_id=%d AND user_id=%d",
priority.Priority,
gConn.Escape(priority.Group),
feedId,
userId)
if err != nil {
return err
}
}
return nil
}
type GroupInfo struct {
Priority int
Group string
}
func serveUpdateOrder(w http.ResponseWriter, r *http.Request) {
if r.Method != "POST" {
w.WriteHeader(400)
fmt.Fprint(w, "Not a post request.")
return
}
userId := 1
decoder := json.NewDecoder(r.Body)
var newPriorities map[string]GroupInfo = make(map[string]GroupInfo)
err := decoder.Decode(&newPriorities)
if err != nil {
respondError( w, err.Error() )
return
}
conn := GetConnection().Clone()
if err := conn.Connect(); err != nil {
respondError( w, err.Error() )
return
}
defer conn.Close()
transaction, err := conn.Begin()
if err != nil {
respondError( w, err.Error() )
return
}
err = updatePriorities(transaction, userId, newPriorities)
if err != nil {
transaction.Rollback()
respondError( w, err.Error() )
return
}
transaction.Commit()
}
func serveMarkRead(w http.ResponseWriter, r *http.Request) {
if r.Method != "POST" {
w.WriteHeader(400)
fmt.Fprint(w, "Not a post request.")
return
}
var userId uint32 = 1
decoder := json.NewDecoder(r.Body)
var readFeedsIds []int
err := decoder.Decode(&readFeedsIds)
if err != nil {
respondError( w, err.Error() )
return
}
if len(readFeedsIds) == 0 {
respondError( w, "The array of feeds to mark as read is empty." )
return
}
readFeeds := []string {}
for _, id := range(readFeedsIds) {
readFeeds = append(readFeeds, strconv.Itoa( id ) )
}
conn := GetConnection().Clone()
| //TODO: Extremelly inneficient; Make better method
_, _, err = conn.QueryFirst("CALL update_unread()")
if err != nil { | random_line_split |
sandbox.ts | component onto it.
*
* To prevent a component for being initialized (for example when you want to initialize it at a later moment)
* just add a `data-skip` attribute to its root element.
*
* @class
* @param {object} config
* @param {Component[]|[Component, object][]} [config.components] Array of components constructor or array with [ComponentConstructor, options]
* @param {HTMLElement|string} [config.root=document.body] Root element of the sandbox. Either a DOM element or a CSS selector
* @param {string} [config.id] ID of the sandbox
* @property {string} $id Sandbox internal id
* @property {HTMLElement} $el Sandbox root DOM element
* @property {Context} $ctx Internal [context](/packages/yuzu-application/api/context). Used to share data across child instances
* @property {object[]} $registry Registered components storage
* @property {Map} $instances Running instances storage
* @returns {Sandbox}
*/
export class Sandbox<S = {}> extends Component<S, ISandboxOptions> {
public static SB_DATA_ATTR = 'data-yuzu-sb';
public defaultOptions(): ISandboxOptions {
return {
components: [],
context: createContext(),
id: '',
root: document.body,
};
}
public $id: string;
public $ctx?: IContext;
public $registry: ISandboxRegistryEntry[] = [];
public $instances = new Map<
string | entrySelectorFn,
Component<any, any>[]
>();
/**
* Creates a sandbox instance.
*
* @constructor
*/
public constructor(options: Partial<ISandboxOptions> = {}) {
super(options);
const { components = [], id } = this.options;
this.$id = id || nextSbUid('_sbx-');
components.forEach((config) => {
if (!Array.isArray(config)) {
if (config.root) {
this.register({ component: config, selector: config.root });
}
if (process.env.NODE_ENV !== 'production') {
!config.root &&
this.$warn(
`Skipping component ${config.displayName ||
config.name} because static "root" selector is missing`,
);
}
} else {
const [component, params = {}] = config;
const selector = component.root || params.selector;
if (selector) {
this.register({ component, selector, ...params });
}
if (process.env.NODE_ENV !== 'production') {
!selector &&
this.$warn(
`Skipping component ${component.displayName ||
component.name} because a static "root" selector is missing and no "selector" param is passed-in`,
);
}
}
});
return this;
}
/**
* ```js
* register(params)
* ```
*
* Registers a new component into the sandbox. The registered components
* will be traversed on `.mount()` initializing every matching component.
*
* @param {object} params Every property other than `component` and `selector` will be used as component option
* @param {Component} params.component Component constructor
* @param {string} params.selector Child component root CSS selector
* @example
* sandbox.register({
* component: Counter,
* selector: '.Counter',
* theme: 'dark' // <-- instance options
* });
*/
public register<C extends Component<any, any>>(params: {
component: IComponentConstructable<C>;
selector: string | entrySelectorFn;
[key: string]: any;
}): void {
invariant(
Component.isComponent(params.component),
'Missing or invalid `component` property',
);
invariant(
typeof params.selector === 'string' ||
typeof params.selector === 'function',
'Missing `selector` property',
);
this.$registry.push(params);
}
/**
* ```js
* start([data])
* ```
*
* **DEPRECATED!** Use `sandbox.mount(root)` instead.
*
* Starts the sandbox with an optional context.
*
* The store will be available inside each component at `this.$context`.
*
* @deprecated
* @param {object} [data] Optional context data object to be injected into the child components.
* @fires Sandbox#beforeStart
* @fires Sandbox#start Events dispatched after all components are initialized
* @returns {Sandbox}
* @example
* sandbox.start();
*
* // with context data
* sandbox.start({ globalTheme: 'dark' });
*/
public start(data = {}): this {
Object.defineProperty(this, '$legacyStart', { value: true });
if (process.env.NODE_ENV !== 'production') {
this.$warn(`Sandbox.start is deprecated. Use the "mount" method instead`);
}
this.mount(this.options.root);
this.setup();
this.$ctx && this.$ctx.update(data);
this.discover();
return this;
}
/**
* ```js
* mount([el], [state])
* ```
*
* Enhances `Component.mount()` by firing the child components discovery logic.
* By default will use `document.body` as mount element.
*
* @param {string|Element} el Component's root element
* @param {object|null} [state={}] Initial state
* @fires Sandbox#beforeStart
* @fires Sandbox#start Events dispatched after all components are initialized
* @returns {Sandbox}
*/
public mount(el: string | Element, state: Partial<S> | null = {}): this {
super.mount(el, state);
this.$el.setAttribute(Sandbox.SB_DATA_ATTR, '');
if (!this.hasOwnProperty('$legacyStart')) {
this.setup();
this.discover();
}
return this;
}
/**
* Setups the sandbox context passed in the options.
*
* @ignore
*/
public setup(): void {
this.$ctx = this.options.context;
this.$ctx.inject(this);
}
/**
* Initializes the sandbox child components.
*
* @ignore
* @returns {Promise}
*/
public discover(): Promise<void> {
invariant(isElement(this.$el), '"this.$el" is not a DOM element');
this.emit('beforeStart');
const sbSelector = `[${Sandbox.SB_DATA_ATTR}]`;
const ret = this.$registry.map(
async ({ component: ComponentConstructor, selector, ...options }) => {
if (this.$instances.has(selector)) {
this.$warn(
`Component ${ComponentConstructor} already initialized on ${selector}`,
);
return;
}
const targets = this.resolveSelector(selector);
let instances: Promise<Component<any, any>>[] | undefined;
if (targets === true) {
instances = [this.createInstance(ComponentConstructor, options)];
} else if (Array.isArray(targets)) {
const { $el } = this;
instances = targets
.filter((el) => {
return (
isElement(el) &&
!el.dataset.skip &&
!el.closest('[data-skip]') &&
el.closest(sbSelector) === $el
);
})
.map((el) => {
return this.createInstance(ComponentConstructor, options, el);
});
}
if (instances) {
this.$instances.set(selector, await Promise.all(instances));
}
return true;
},
);
return Promise.all(ret).then(() => {
this.emit('start');
});
}
/**
* Resolves a configured component selector to a list of DOM nodes or a boolean (for detached components)
*
* @ignore
* @param {string|function} selector Selector string or function.
* @returns {HTMLElement[]|boolean}
*/
public resolveSelector(
selector: string | entrySelectorFn,
): HTMLElement[] | boolean {
let targets = evaluate(selector, this);
if (typeof targets === 'string') {
targets = this.findNodes(targets) as HTMLElement[];
}
return targets;
}
/**
* Creates a component instance.
* Reads inline components from the passed-in root DOM element.
*
* @ignore
* @param {object} options instance options
* @param {HTMLElement} [el] Root element
* @returns {Component}
*/
public createInstance<C extends Component<any, any>>(
ComponentConstructor: IComponentConstructable<C>,
options: Record<string, any>,
el?: HTMLElement,
): Promise<C> {
const inlineOptions = el ? datasetParser(el) : {};
return this.setRef({
id: nextChildUid(this.$id + '-c.'),
...options,
...inlineOptions,
component: ComponentConstructor,
el,
});
}
/**
* ```js
* stop()
* ```
*
* **DEPRECATED!** Use `sandbox.destroy()` instead.
*
* Stops every running component, clears sandbox events and destroys the instance.
*
* @deprecated
* @fires Sandbox#beforeStop
* @fires Sandbox#stop
* @returns {Promise<void>}
* @example
* sandbox.stop();
*/
public async stop(): Promise<void> {
if (process.env.NODE_ENV !== 'production') | {
this.$warn(
`Sandbox.stop is deprecated. Use the "destroy" method instead`,
);
} | conditional_block |
|
sandbox.ts | [],
context: createContext(),
id: '',
root: document.body,
};
}
public $id: string;
public $ctx?: IContext;
public $registry: ISandboxRegistryEntry[] = [];
public $instances = new Map<
string | entrySelectorFn,
Component<any, any>[]
>();
/**
* Creates a sandbox instance.
*
* @constructor
*/
public constructor(options: Partial<ISandboxOptions> = {}) {
super(options);
const { components = [], id } = this.options;
this.$id = id || nextSbUid('_sbx-');
components.forEach((config) => {
if (!Array.isArray(config)) {
if (config.root) {
this.register({ component: config, selector: config.root });
}
if (process.env.NODE_ENV !== 'production') {
!config.root &&
this.$warn(
`Skipping component ${config.displayName ||
config.name} because static "root" selector is missing`,
);
}
} else {
const [component, params = {}] = config;
const selector = component.root || params.selector;
if (selector) {
this.register({ component, selector, ...params });
}
if (process.env.NODE_ENV !== 'production') {
!selector &&
this.$warn(
`Skipping component ${component.displayName ||
component.name} because a static "root" selector is missing and no "selector" param is passed-in`,
);
}
}
});
return this;
}
/**
* ```js
* register(params)
* ```
*
* Registers a new component into the sandbox. The registered components
* will be traversed on `.mount()` initializing every matching component.
*
* @param {object} params Every property other than `component` and `selector` will be used as component option
* @param {Component} params.component Component constructor
* @param {string} params.selector Child component root CSS selector
* @example
* sandbox.register({
* component: Counter,
* selector: '.Counter',
* theme: 'dark' // <-- instance options
* });
*/
public register<C extends Component<any, any>>(params: {
component: IComponentConstructable<C>;
selector: string | entrySelectorFn;
[key: string]: any;
}): void {
invariant(
Component.isComponent(params.component),
'Missing or invalid `component` property',
);
invariant(
typeof params.selector === 'string' ||
typeof params.selector === 'function',
'Missing `selector` property',
);
this.$registry.push(params);
}
/**
* ```js
* start([data])
* ```
*
* **DEPRECATED!** Use `sandbox.mount(root)` instead.
*
* Starts the sandbox with an optional context.
*
* The store will be available inside each component at `this.$context`.
*
* @deprecated
* @param {object} [data] Optional context data object to be injected into the child components.
* @fires Sandbox#beforeStart
* @fires Sandbox#start Events dispatched after all components are initialized
* @returns {Sandbox}
* @example
* sandbox.start();
*
* // with context data
* sandbox.start({ globalTheme: 'dark' });
*/
public start(data = {}): this {
Object.defineProperty(this, '$legacyStart', { value: true });
if (process.env.NODE_ENV !== 'production') {
this.$warn(`Sandbox.start is deprecated. Use the "mount" method instead`);
}
this.mount(this.options.root);
this.setup();
this.$ctx && this.$ctx.update(data);
this.discover();
return this;
}
/**
* ```js
* mount([el], [state])
* ```
*
* Enhances `Component.mount()` by firing the child components discovery logic.
* By default will use `document.body` as mount element.
*
* @param {string|Element} el Component's root element
* @param {object|null} [state={}] Initial state
* @fires Sandbox#beforeStart
* @fires Sandbox#start Events dispatched after all components are initialized
* @returns {Sandbox}
*/
public mount(el: string | Element, state: Partial<S> | null = {}): this {
super.mount(el, state);
this.$el.setAttribute(Sandbox.SB_DATA_ATTR, '');
if (!this.hasOwnProperty('$legacyStart')) {
this.setup();
this.discover();
}
return this;
}
/**
* Setups the sandbox context passed in the options.
*
* @ignore
*/
public setup(): void {
this.$ctx = this.options.context;
this.$ctx.inject(this);
}
/**
* Initializes the sandbox child components.
*
* @ignore
* @returns {Promise}
*/
public discover(): Promise<void> {
invariant(isElement(this.$el), '"this.$el" is not a DOM element');
this.emit('beforeStart');
const sbSelector = `[${Sandbox.SB_DATA_ATTR}]`;
const ret = this.$registry.map(
async ({ component: ComponentConstructor, selector, ...options }) => {
if (this.$instances.has(selector)) {
this.$warn(
`Component ${ComponentConstructor} already initialized on ${selector}`,
);
return;
}
const targets = this.resolveSelector(selector);
let instances: Promise<Component<any, any>>[] | undefined;
if (targets === true) {
instances = [this.createInstance(ComponentConstructor, options)];
} else if (Array.isArray(targets)) {
const { $el } = this;
instances = targets
.filter((el) => {
return (
isElement(el) &&
!el.dataset.skip &&
!el.closest('[data-skip]') &&
el.closest(sbSelector) === $el
);
})
.map((el) => {
return this.createInstance(ComponentConstructor, options, el);
});
}
if (instances) {
this.$instances.set(selector, await Promise.all(instances));
}
return true;
},
);
return Promise.all(ret).then(() => {
this.emit('start');
});
}
/**
* Resolves a configured component selector to a list of DOM nodes or a boolean (for detached components)
*
* @ignore
* @param {string|function} selector Selector string or function.
* @returns {HTMLElement[]|boolean}
*/
public resolveSelector(
selector: string | entrySelectorFn,
): HTMLElement[] | boolean {
let targets = evaluate(selector, this);
if (typeof targets === 'string') {
targets = this.findNodes(targets) as HTMLElement[];
}
return targets;
}
/**
* Creates a component instance.
* Reads inline components from the passed-in root DOM element.
*
* @ignore
* @param {object} options instance options
* @param {HTMLElement} [el] Root element
* @returns {Component}
*/
public createInstance<C extends Component<any, any>>(
ComponentConstructor: IComponentConstructable<C>,
options: Record<string, any>,
el?: HTMLElement,
): Promise<C> {
const inlineOptions = el ? datasetParser(el) : {};
return this.setRef({
id: nextChildUid(this.$id + '-c.'),
...options,
...inlineOptions,
component: ComponentConstructor,
el,
});
}
/**
* ```js
* stop()
* ```
*
* **DEPRECATED!** Use `sandbox.destroy()` instead.
*
* Stops every running component, clears sandbox events and destroys the instance.
*
* @deprecated
* @fires Sandbox#beforeStop
* @fires Sandbox#stop
* @returns {Promise<void>}
* @example
* sandbox.stop();
*/
public async stop(): Promise<void> {
if (process.env.NODE_ENV !== 'production') {
this.$warn(
`Sandbox.stop is deprecated. Use the "destroy" method instead`,
);
}
return this.destroy();
}
/**
* ```js
* destroy()
* ```
*
* Enhances `Component.destroy()`.
* Stops every running component, clears sandbox events and destroys the instance.
*
* @deprecated
* @fires Sandbox#beforeStop
* @fires Sandbox#stop
* @returns {Promise<void>}
* @example
* sandbox.destroy();
*/
public async destroy(): Promise<void> {
this.emit('beforeStop');
await this.beforeDestroy();
this.removeListeners();
try {
if (this.$el) {
this.$el.removeAttribute(Sandbox.SB_DATA_ATTR);
}
await this.destroyRefs();
this.$active = false;
} catch (e) {
this.emit('error', e);
return Promise.reject(e);
}
this.$instances.clear();
this.emit('stop');
this.clear();
return super.destroy();
}
/**
* Removes events and associated store
*
* @ignore
*/
public clear(): void {
this.$ctx = undefined; // release the context
this.off('beforeStart');
this.off('start'); | this.off('error'); | random_line_split |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.