file_name
large_stringlengths 4
140
| prefix
large_stringlengths 0
12.1k
| suffix
large_stringlengths 0
12k
| middle
large_stringlengths 0
7.51k
| fim_type
large_stringclasses 4
values |
---|---|---|---|---|
script.js | send too many ajax requests
*/
function addRowToTable(action, name, instructions) | 'src', 'picture.php?' + $.param({'medicineName': name})
);
newElement.children('td').eq(0).append(pictureElement);
pictureElement.ready(function () {
pictureElement.height(pictureElement.width());
});
} else if (action === "update") {
row = 'table tbody tr[name=' + htmlName + '] td:nth-child(2)';
$(row).html(instructions);
} else if (action === "remove") {
row = 'table tbody tr[name=' + htmlName + ']';
console.log(row);
$(row).remove();
} else {
console.error(action);
}
}
function extractName(medicineName) {
'use strict';
return medicineName.substr(medicineName.indexOf('_') + 1);
}
$(document).ready(function () {
'use strict';
var request, qrcode, width, nameElement, dataElement, text;
//jQuery('#qrcode').qrcode("this plugin is great");
//new QRCode(document.getElementById("qrcode"), "http://jindo.dev.naver.com/collie");
width = $('button#add_bttn').width();
console.log(width);
// turn #qrcode div into QR code
qrcode = new QRCode("qrcode", {
text: " ",
width: 170,
height: 170,
colorDark : "#000000",
colorLight : "#ffffff",
correctLevel : QRCode.CorrectLevel.H
});
nameElement = $("textarea[name='medicineName']");
dataElement = $("textarea[name='medicineData']");
nameElement.on('change keydown paste input', function () {
// whenever name field (nameElement) content changes, set QR code value arccordingly.
console.log("NCHAN");
var name, isEmpty, googleElement;
name = $.trim($(this).val());
isEmpty = true;
//console.log(name);
qrcode.makeCode(name);
//console.log(name);
$('table#medicineTable').find('tr').each(function () {
var rowName = extractName($(this).attr('name'));
//console.log(rowName);
if (rowName == 'google') {
} else if (rowName.indexOf(name) !== -1) {
$(this).css('display', 'table-row');
isEmpty = false;
} else {
$(this).css('display', 'none');
}
});
if (isEmpty == true) {
$("tr[name='google']").css("display", "table-row");
googleElement = $("tr[name='google'] td:nth-child(2)");
googleElement.empty();
googleElement.append("Internal search has no results. Google search ");
googleElement.append($('<b/>', {"id": "google_query"}).text(name));
googleElement.append(' instead?');
//console.log("EMPTY");
} else {
$("tr[name='google']").css("display", "none");
}
});
/*
setTimeout(
function () {nameElement.trigger("input");},
0
);
*/
$('div#searchTable').height($('div#formDiv').height());
//Ajax call to getAll.php to get all medicine entries in database
request = $.ajax({
url: 'getAll.php',
type: 'post',
success: function (data, textStatus, xhr) {
var name, instructions, medicines, htmlName, row, k;
console.log("RAWDATA", data);
medicines = JSON.parse(data);
//add medicine names + instructions to table
for (k = 0; k < medicines.length; k += 1) {
row = medicines[k];
console.log(row);
console.log(row.name, row.instructions);
addRowToTable('insert', row.name, row.instructions);
}
//console.log(medicines);
nameElement.trigger("input")
},
error: function (xhr, desc, err) {
console.log(xhr);
console.log("Details: " + desc + "\nError:" + err);
}
});
$('table#medicineTable').on('click', 'tr', function () {
console.log($(this).attr('name'));
var medicineName, medicineData, pictureElement, query;
if ($(this).hasClass('clicked')) { return; }
if ($(this).attr('id') === 'google_search_tr') {
query = $("b#google_query").text();
window.location.replace("externalSearch.php?q=" + query);
return;
}
medicineName = $(this).attr('name');
medicineData = $(this).children('td').eq(1).text();
medicineName = extractName(medicineName);
$('table#medicineTable *').removeClass('clicked');
$(this).addClass('clicked');
qrcode.makeCode(medicineName);
$("textarea[name='medicineName']").val(medicineName);
$("textarea[name='medicineData']").val(medicineData);
});
// Bind to the submit event of our form
console.log("potato", $("p#removeBttn"));
$("p#removeBttn").click(function (event) {
var form, name, instructions, inputs;
console.log('hiya');
event.preventDefault();
if (request) {request.abort(); }
form = $(this);
inputs = form.find("input, select, button, textarea");
inputs.prop("disabled", true);
console.log(inputs);
// Let's select and cache all the fields
name = $.trim(nameElement.val());
instructions = $.trim(dataElement.val());
if (name === '') {
alert("name is empty!");
return;
}
request = $.ajax({
url: 'remove.php',
type: 'post',
data: {'name': name, 'instructions': instructions},
success: function (data, textStatus, xhr) {
var name, instructions, action, result, htmlName, row;
console.log("RAWDATA", data);
result = JSON.parse(data);
/*
result[0] is action, is "update" if php script updated database
and "insert" if php script inserted new medicine
*/
action = result[0];
name = result[1];
instructions = result[2];
console.log(action);
if (action === 'remove') {
//console.log("REMOVING");
addRowToTable('remove', name, instructions);
}
console.log(result);
inputs.prop("disabled", false);
},
error: function (xhr, desc, err) {
console.log(xhr);
console.log("Details: " + desc + "\nError:" + err);
inputs.prop("disabled", false);
}
});
});
$("form#editForm").submit('click', function (event) {
var form, name, instructions, inputs;
console.log('hiya');
// Prevent default posting of form - put here to work in case of errors
event.preventDefault();
// Abort any pending request
if (request) {request.abort(); }
// setup some local variables
form = $(this);
//name = $("form textarea[name='medicineName']").val();
//instructions = $("form textarea[name='medicineName']").val();
//console.log("NAME", name, instructions);
// Let's select and cache all the fields
name = $.trim(nameElement.val());
instructions = $.trim(dataElement.val());
if (name === '' || instructions === '') {
alert("name or instructions is empty!");
return;
}
inputs = form.find("input, select, button, textarea");
// Serialize the data in the form
//serializedData = form.serialize();
console.log(inputs);
// Let's disable the inputs for the duration of the Ajax request.
// Note: we disable elements AFTER the form data has been serialized.
// Disabled form elements will not be serialized.
inputs.prop("disabled", true);
// Fire off the request to /form.php
//console.log("SERIAL", serializedData, inputs);
request = $.ajax({
url: 'docterEnter.php',
type: 'post',
data: {'name': name, 'instructions': instructions},
success: function (data, textStatus, xhr) {
var name, instructions, action, result, htmlName, row;
console.log("RAWDATA", data);
result = JSON.parse(data);
/*
result[0] is action, value is "update" if php script updated database
and "insert" if php script inserted new medicine into database
| {
'use strict';
var htmlName, newElement, pictureElement, medicineName, row;
htmlName = JSON.stringify("medicine_" + name);
console.log(name, instructions);
if (action === "insert") {
newElement = $("<tr name=" + htmlName + ">").append(
$('<td/>').text(name),
$('<td/>').text(instructions)
);
$('table tbody').append(newElement);
console.log("APPENDED");
pictureElement = $(
"<iframe id='picture' scrolling='no'/>",
{'class': "images"}
);
pictureElement.attr( | identifier_body |
script.js | send too many ajax requests
*/
function | (action, name, instructions) {
'use strict';
var htmlName, newElement, pictureElement, medicineName, row;
htmlName = JSON.stringify("medicine_" + name);
console.log(name, instructions);
if (action === "insert") {
newElement = $("<tr name=" + htmlName + ">").append(
$('<td/>').text(name),
$('<td/>').text(instructions)
);
$('table tbody').append(newElement);
console.log("APPENDED");
pictureElement = $(
"<iframe id='picture' scrolling='no'/>",
{'class': "images"}
);
pictureElement.attr(
'src', 'picture.php?' + $.param({'medicineName': name})
);
newElement.children('td').eq(0).append(pictureElement);
pictureElement.ready(function () {
pictureElement.height(pictureElement.width());
});
} else if (action === "update") {
row = 'table tbody tr[name=' + htmlName + '] td:nth-child(2)';
$(row).html(instructions);
} else if (action === "remove") {
row = 'table tbody tr[name=' + htmlName + ']';
console.log(row);
$(row).remove();
} else {
console.error(action);
}
}
function extractName(medicineName) {
'use strict';
return medicineName.substr(medicineName.indexOf('_') + 1);
}
$(document).ready(function () {
'use strict';
var request, qrcode, width, nameElement, dataElement, text;
//jQuery('#qrcode').qrcode("this plugin is great");
//new QRCode(document.getElementById("qrcode"), "http://jindo.dev.naver.com/collie");
width = $('button#add_bttn').width();
console.log(width);
// turn #qrcode div into QR code
qrcode = new QRCode("qrcode", {
text: " ",
width: 170,
height: 170,
colorDark : "#000000",
colorLight : "#ffffff",
correctLevel : QRCode.CorrectLevel.H
});
nameElement = $("textarea[name='medicineName']");
dataElement = $("textarea[name='medicineData']");
nameElement.on('change keydown paste input', function () {
// whenever name field (nameElement) content changes, set QR code value arccordingly.
console.log("NCHAN");
var name, isEmpty, googleElement;
name = $.trim($(this).val());
isEmpty = true;
//console.log(name);
qrcode.makeCode(name);
//console.log(name);
$('table#medicineTable').find('tr').each(function () {
var rowName = extractName($(this).attr('name'));
//console.log(rowName);
if (rowName == 'google') {
} else if (rowName.indexOf(name) !== -1) {
$(this).css('display', 'table-row');
isEmpty = false;
} else {
$(this).css('display', 'none');
}
});
if (isEmpty == true) {
$("tr[name='google']").css("display", "table-row");
googleElement = $("tr[name='google'] td:nth-child(2)");
googleElement.empty();
googleElement.append("Internal search has no results. Google search ");
googleElement.append($('<b/>', {"id": "google_query"}).text(name));
googleElement.append(' instead?');
//console.log("EMPTY");
} else {
$("tr[name='google']").css("display", "none");
}
});
/*
setTimeout(
function () {nameElement.trigger("input");},
0
);
*/
$('div#searchTable').height($('div#formDiv').height());
//Ajax call to getAll.php to get all medicine entries in database
request = $.ajax({
url: 'getAll.php',
type: 'post',
success: function (data, textStatus, xhr) {
var name, instructions, medicines, htmlName, row, k;
console.log("RAWDATA", data);
medicines = JSON.parse(data);
//add medicine names + instructions to table
for (k = 0; k < medicines.length; k += 1) {
row = medicines[k];
console.log(row);
console.log(row.name, row.instructions);
addRowToTable('insert', row.name, row.instructions);
}
//console.log(medicines);
nameElement.trigger("input")
},
error: function (xhr, desc, err) {
console.log(xhr);
console.log("Details: " + desc + "\nError:" + err);
}
});
$('table#medicineTable').on('click', 'tr', function () {
console.log($(this).attr('name'));
var medicineName, medicineData, pictureElement, query;
if ($(this).hasClass('clicked')) { return; }
if ($(this).attr('id') === 'google_search_tr') {
query = $("b#google_query").text();
window.location.replace("externalSearch.php?q=" + query);
return;
}
medicineName = $(this).attr('name');
medicineData = $(this).children('td').eq(1).text();
medicineName = extractName(medicineName);
$('table#medicineTable *').removeClass('clicked');
$(this).addClass('clicked');
qrcode.makeCode(medicineName);
$("textarea[name='medicineName']").val(medicineName);
$("textarea[name='medicineData']").val(medicineData);
});
// Bind to the submit event of our form
console.log("potato", $("p#removeBttn"));
$("p#removeBttn").click(function (event) {
var form, name, instructions, inputs;
console.log('hiya');
event.preventDefault();
if (request) {request.abort(); }
form = $(this);
inputs = form.find("input, select, button, textarea");
inputs.prop("disabled", true);
console.log(inputs);
// Let's select and cache all the fields
name = $.trim(nameElement.val());
instructions = $.trim(dataElement.val());
if (name === '') {
alert("name is empty!");
return;
}
request = $.ajax({
url: 'remove.php',
type: 'post',
data: {'name': name, 'instructions': instructions},
success: function (data, textStatus, xhr) {
var name, instructions, action, result, htmlName, row;
console.log("RAWDATA", data);
result = JSON.parse(data);
/*
result[0] is action, is "update" if php script updated database
and "insert" if php script inserted new medicine
*/
action = result[0];
name = result[1];
instructions = result[2];
console.log(action);
if (action === 'remove') {
//console.log("REMOVING");
addRowToTable('remove', name, instructions);
}
console.log(result);
inputs.prop("disabled", false);
},
error: function (xhr, desc, err) {
console.log(xhr);
console.log("Details: " + desc + "\nError:" + err);
inputs.prop("disabled", false);
}
});
});
$("form#editForm").submit('click', function (event) {
var form, name, instructions, inputs;
console.log('hiya');
// Prevent default posting of form - put here to work in case of errors
event.preventDefault();
// Abort any pending request
if (request) {request.abort(); }
// setup some local variables
form = $(this);
//name = $("form textarea[name='medicineName']").val();
//instructions = $("form textarea[name='medicineName']").val();
//console.log("NAME", name, instructions);
// Let's select and cache all the fields
name = $.trim(nameElement.val());
instructions = $.trim(dataElement.val());
if (name === '' || instructions === '') {
alert("name or instructions is empty!");
return;
}
inputs = form.find("input, select, button, textarea");
// Serialize the data in the form
//serializedData = form.serialize();
console.log(inputs);
// Let's disable the inputs for the duration of the Ajax request.
// Note: we disable elements AFTER the form data has been serialized.
// Disabled form elements will not be serialized.
inputs.prop("disabled", true);
// Fire off the request to /form.php
//console.log("SERIAL", serializedData, inputs);
request = $.ajax({
url: 'docterEnter.php',
type: 'post',
data: {'name': name, 'instructions': instructions},
success: function (data, textStatus, xhr) {
var name, instructions, action, result, htmlName, row;
console.log("RAWDATA", data);
result = JSON.parse(data);
/*
result[0] is action, value is "update" if php script updated database
and "insert" if php script inserted new medicine into database
| addRowToTable | identifier_name |
script.js | send too many ajax requests
*/
function addRowToTable(action, name, instructions) {
'use strict';
var htmlName, newElement, pictureElement, medicineName, row;
htmlName = JSON.stringify("medicine_" + name);
console.log(name, instructions);
if (action === "insert") {
newElement = $("<tr name=" + htmlName + ">").append(
$('<td/>').text(name),
$('<td/>').text(instructions)
);
$('table tbody').append(newElement);
console.log("APPENDED");
pictureElement = $(
"<iframe id='picture' scrolling='no'/>",
{'class': "images"}
);
pictureElement.attr(
'src', 'picture.php?' + $.param({'medicineName': name})
);
newElement.children('td').eq(0).append(pictureElement);
pictureElement.ready(function () {
pictureElement.height(pictureElement.width());
});
} else if (action === "update") {
row = 'table tbody tr[name=' + htmlName + '] td:nth-child(2)';
$(row).html(instructions);
} else if (action === "remove") {
row = 'table tbody tr[name=' + htmlName + ']';
console.log(row);
$(row).remove();
} else {
console.error(action);
}
}
function extractName(medicineName) {
'use strict';
return medicineName.substr(medicineName.indexOf('_') + 1);
}
$(document).ready(function () {
'use strict';
var request, qrcode, width, nameElement, dataElement, text;
//jQuery('#qrcode').qrcode("this plugin is great");
//new QRCode(document.getElementById("qrcode"), "http://jindo.dev.naver.com/collie");
width = $('button#add_bttn').width();
console.log(width);
// turn #qrcode div into QR code
qrcode = new QRCode("qrcode", {
text: " ",
width: 170,
height: 170,
colorDark : "#000000",
colorLight : "#ffffff",
correctLevel : QRCode.CorrectLevel.H
});
nameElement = $("textarea[name='medicineName']");
dataElement = $("textarea[name='medicineData']");
nameElement.on('change keydown paste input', function () {
// whenever name field (nameElement) content changes, set QR code value arccordingly.
console.log("NCHAN");
var name, isEmpty, googleElement;
name = $.trim($(this).val());
isEmpty = true;
//console.log(name);
qrcode.makeCode(name);
//console.log(name);
$('table#medicineTable').find('tr').each(function () {
var rowName = extractName($(this).attr('name'));
//console.log(rowName);
if (rowName == 'google') {
} else if (rowName.indexOf(name) !== -1) {
$(this).css('display', 'table-row');
isEmpty = false;
} else {
$(this).css('display', 'none');
}
});
if (isEmpty == true) {
$("tr[name='google']").css("display", "table-row");
googleElement = $("tr[name='google'] td:nth-child(2)");
googleElement.empty();
googleElement.append("Internal search has no results. Google search ");
googleElement.append($('<b/>', {"id": "google_query"}).text(name));
googleElement.append(' instead?');
//console.log("EMPTY");
} else |
});
/*
setTimeout(
function () {nameElement.trigger("input");},
0
);
*/
$('div#searchTable').height($('div#formDiv').height());
//Ajax call to getAll.php to get all medicine entries in database
request = $.ajax({
url: 'getAll.php',
type: 'post',
success: function (data, textStatus, xhr) {
var name, instructions, medicines, htmlName, row, k;
console.log("RAWDATA", data);
medicines = JSON.parse(data);
//add medicine names + instructions to table
for (k = 0; k < medicines.length; k += 1) {
row = medicines[k];
console.log(row);
console.log(row.name, row.instructions);
addRowToTable('insert', row.name, row.instructions);
}
//console.log(medicines);
nameElement.trigger("input")
},
error: function (xhr, desc, err) {
console.log(xhr);
console.log("Details: " + desc + "\nError:" + err);
}
});
$('table#medicineTable').on('click', 'tr', function () {
console.log($(this).attr('name'));
var medicineName, medicineData, pictureElement, query;
if ($(this).hasClass('clicked')) { return; }
if ($(this).attr('id') === 'google_search_tr') {
query = $("b#google_query").text();
window.location.replace("externalSearch.php?q=" + query);
return;
}
medicineName = $(this).attr('name');
medicineData = $(this).children('td').eq(1).text();
medicineName = extractName(medicineName);
$('table#medicineTable *').removeClass('clicked');
$(this).addClass('clicked');
qrcode.makeCode(medicineName);
$("textarea[name='medicineName']").val(medicineName);
$("textarea[name='medicineData']").val(medicineData);
});
// Bind to the submit event of our form
console.log("potato", $("p#removeBttn"));
$("p#removeBttn").click(function (event) {
var form, name, instructions, inputs;
console.log('hiya');
event.preventDefault();
if (request) {request.abort(); }
form = $(this);
inputs = form.find("input, select, button, textarea");
inputs.prop("disabled", true);
console.log(inputs);
// Let's select and cache all the fields
name = $.trim(nameElement.val());
instructions = $.trim(dataElement.val());
if (name === '') {
alert("name is empty!");
return;
}
request = $.ajax({
url: 'remove.php',
type: 'post',
data: {'name': name, 'instructions': instructions},
success: function (data, textStatus, xhr) {
var name, instructions, action, result, htmlName, row;
console.log("RAWDATA", data);
result = JSON.parse(data);
/*
result[0] is action, is "update" if php script updated database
and "insert" if php script inserted new medicine
*/
action = result[0];
name = result[1];
instructions = result[2];
console.log(action);
if (action === 'remove') {
//console.log("REMOVING");
addRowToTable('remove', name, instructions);
}
console.log(result);
inputs.prop("disabled", false);
},
error: function (xhr, desc, err) {
console.log(xhr);
console.log("Details: " + desc + "\nError:" + err);
inputs.prop("disabled", false);
}
});
});
$("form#editForm").submit('click', function (event) {
var form, name, instructions, inputs;
console.log('hiya');
// Prevent default posting of form - put here to work in case of errors
event.preventDefault();
// Abort any pending request
if (request) {request.abort(); }
// setup some local variables
form = $(this);
//name = $("form textarea[name='medicineName']").val();
//instructions = $("form textarea[name='medicineName']").val();
//console.log("NAME", name, instructions);
// Let's select and cache all the fields
name = $.trim(nameElement.val());
instructions = $.trim(dataElement.val());
if (name === '' || instructions === '') {
alert("name or instructions is empty!");
return;
}
inputs = form.find("input, select, button, textarea");
// Serialize the data in the form
//serializedData = form.serialize();
console.log(inputs);
// Let's disable the inputs for the duration of the Ajax request.
// Note: we disable elements AFTER the form data has been serialized.
// Disabled form elements will not be serialized.
inputs.prop("disabled", true);
// Fire off the request to /form.php
//console.log("SERIAL", serializedData, inputs);
request = $.ajax({
url: 'docterEnter.php',
type: 'post',
data: {'name': name, 'instructions': instructions},
success: function (data, textStatus, xhr) {
var name, instructions, action, result, htmlName, row;
console.log("RAWDATA", data);
result = JSON.parse(data);
/*
result[0] is action, value is "update" if php script updated database
and "insert" if php script inserted new medicine into database | {
$("tr[name='google']").css("display", "none");
} | conditional_block |
script.js | 't send too many ajax requests
*/
function addRowToTable(action, name, instructions) {
'use strict';
var htmlName, newElement, pictureElement, medicineName, row;
htmlName = JSON.stringify("medicine_" + name);
console.log(name, instructions);
if (action === "insert") {
newElement = $("<tr name=" + htmlName + ">").append(
$('<td/>').text(name),
$('<td/>').text(instructions)
);
$('table tbody').append(newElement);
console.log("APPENDED");
pictureElement = $(
"<iframe id='picture' scrolling='no'/>",
{'class': "images"}
);
pictureElement.attr(
'src', 'picture.php?' + $.param({'medicineName': name})
);
newElement.children('td').eq(0).append(pictureElement);
pictureElement.ready(function () {
pictureElement.height(pictureElement.width());
});
} else if (action === "update") {
row = 'table tbody tr[name=' + htmlName + '] td:nth-child(2)';
$(row).html(instructions);
} else if (action === "remove") {
row = 'table tbody tr[name=' + htmlName + ']';
console.log(row);
$(row).remove();
} else {
console.error(action);
}
}
function extractName(medicineName) {
'use strict';
return medicineName.substr(medicineName.indexOf('_') + 1);
}
$(document).ready(function () {
'use strict';
var request, qrcode, width, nameElement, dataElement, text;
//jQuery('#qrcode').qrcode("this plugin is great");
//new QRCode(document.getElementById("qrcode"), "http://jindo.dev.naver.com/collie");
width = $('button#add_bttn').width();
console.log(width);
// turn #qrcode div into QR code
qrcode = new QRCode("qrcode", {
text: " ",
width: 170,
height: 170,
colorDark : "#000000",
colorLight : "#ffffff",
correctLevel : QRCode.CorrectLevel.H
});
nameElement = $("textarea[name='medicineName']");
dataElement = $("textarea[name='medicineData']");
nameElement.on('change keydown paste input', function () {
// whenever name field (nameElement) content changes, set QR code value arccordingly.
console.log("NCHAN");
var name, isEmpty, googleElement;
name = $.trim($(this).val());
isEmpty = true;
//console.log(name);
qrcode.makeCode(name);
//console.log(name);
$('table#medicineTable').find('tr').each(function () {
var rowName = extractName($(this).attr('name'));
//console.log(rowName);
if (rowName == 'google') {
} else if (rowName.indexOf(name) !== -1) {
$(this).css('display', 'table-row');
isEmpty = false;
} else {
$(this).css('display', 'none');
}
});
if (isEmpty == true) {
$("tr[name='google']").css("display", "table-row");
googleElement = $("tr[name='google'] td:nth-child(2)");
googleElement.empty();
googleElement.append("Internal search has no results. Google search ");
googleElement.append($('<b/>', {"id": "google_query"}).text(name));
googleElement.append(' instead?');
//console.log("EMPTY");
} else {
$("tr[name='google']").css("display", "none");
}
});
/*
setTimeout(
function () {nameElement.trigger("input");},
0
);
*/
$('div#searchTable').height($('div#formDiv').height());
//Ajax call to getAll.php to get all medicine entries in database
request = $.ajax({
url: 'getAll.php',
type: 'post',
success: function (data, textStatus, xhr) {
var name, instructions, medicines, htmlName, row, k;
console.log("RAWDATA", data);
medicines = JSON.parse(data);
//add medicine names + instructions to table
for (k = 0; k < medicines.length; k += 1) {
row = medicines[k];
console.log(row);
console.log(row.name, row.instructions);
addRowToTable('insert', row.name, row.instructions);
}
//console.log(medicines);
nameElement.trigger("input")
},
error: function (xhr, desc, err) {
console.log(xhr);
console.log("Details: " + desc + "\nError:" + err);
}
});
$('table#medicineTable').on('click', 'tr', function () {
console.log($(this).attr('name'));
var medicineName, medicineData, pictureElement, query;
if ($(this).hasClass('clicked')) { return; }
if ($(this).attr('id') === 'google_search_tr') {
query = $("b#google_query").text();
window.location.replace("externalSearch.php?q=" + query);
return;
}
medicineName = $(this).attr('name');
medicineData = $(this).children('td').eq(1).text();
medicineName = extractName(medicineName);
$('table#medicineTable *').removeClass('clicked');
$(this).addClass('clicked');
qrcode.makeCode(medicineName);
$("textarea[name='medicineName']").val(medicineName);
$("textarea[name='medicineData']").val(medicineData);
});
// Bind to the submit event of our form
console.log("potato", $("p#removeBttn"));
$("p#removeBttn").click(function (event) {
var form, name, instructions, inputs;
console.log('hiya');
event.preventDefault();
if (request) {request.abort(); }
form = $(this);
inputs = form.find("input, select, button, textarea");
inputs.prop("disabled", true);
console.log(inputs);
// Let's select and cache all the fields
name = $.trim(nameElement.val());
instructions = $.trim(dataElement.val());
if (name === '') {
alert("name is empty!");
return;
}
request = $.ajax({
url: 'remove.php',
type: 'post',
data: {'name': name, 'instructions': instructions},
success: function (data, textStatus, xhr) {
var name, instructions, action, result, htmlName, row;
console.log("RAWDATA", data);
result = JSON.parse(data);
/*
result[0] is action, is "update" if php script updated database
and "insert" if php script inserted new medicine
*/
action = result[0];
name = result[1];
instructions = result[2];
console.log(action);
if (action === 'remove') {
//console.log("REMOVING");
addRowToTable('remove', name, instructions);
}
console.log(result);
inputs.prop("disabled", false);
},
error: function (xhr, desc, err) {
console.log(xhr);
console.log("Details: " + desc + "\nError:" + err);
inputs.prop("disabled", false);
}
});
});
$("form#editForm").submit('click', function (event) {
var form, name, instructions, inputs;
console.log('hiya');
// Prevent default posting of form - put here to work in case of errors
event.preventDefault();
// Abort any pending request
if (request) {request.abort(); }
// setup some local variables
form = $(this);
//name = $("form textarea[name='medicineName']").val();
//instructions = $("form textarea[name='medicineName']").val();
//console.log("NAME", name, instructions);
// Let's select and cache all the fields
name = $.trim(nameElement.val());
instructions = $.trim(dataElement.val());
if (name === '' || instructions === '') {
alert("name or instructions is empty!");
return;
} | // Serialize the data in the form
//serializedData = form.serialize();
console.log(inputs);
// Let's disable the inputs for the duration of the Ajax request.
// Note: we disable elements AFTER the form data has been serialized.
// Disabled form elements will not be serialized.
inputs.prop("disabled", true);
// Fire off the request to /form.php
//console.log("SERIAL", serializedData, inputs);
request = $.ajax({
url: 'docterEnter.php',
type: 'post',
data: {'name': name, 'instructions': instructions},
success: function (data, textStatus, xhr) {
var name, instructions, action, result, htmlName, row;
console.log("RAWDATA", data);
result = JSON.parse(data);
/*
result[0] is action, value is "update" if php script updated database
and "insert" if php script inserted new medicine into database
|
inputs = form.find("input, select, button, textarea"); | random_line_split |
fourier.py | the paper
"Rethinking Attention with Performers" https://arxiv.org/pdf/2009.14794.pdf
and the traditional random Fourier features that approximate the RBF kernel.
"""
from math import sqrt, log
from typing import Optional
from einops import repeat
import torch
from .base import Kernel
def orthogonal_random_matrix_(
num_rows: int,
num_columns: int,
scaling: float = 0,
device: Optional[torch.device] = None
):
| else:
raise ValueError(f"Invalid scaling {scaling}")
return torch.diag(multiplier) @ final_matrix
def orthogonal_matrix_chunk(
cols: int,
device: torch.device = None
) -> torch.Tensor:
unstructured_block = torch.randn((cols, cols), device=device)
q, _ = torch.qr(unstructured_block.cpu(), some=True)
q = q.to(device)
return q.t()
class RandomFourierFeatures(Kernel):
"""Random Fourier Features for the RBF kernel according to [1].
[1]: "Weighted Sums of Random Kitchen Sinks: Replacing minimization with
randomization in learning" by A. Rahimi and Benjamin Recht.
Arguments
---------
hidden_size: int, The input query dimensions in order to sample
the noise matrix
softmax_temp: float, The temerature for the Gaussian kernel
approximation exp(-t * |x-y|^2)
(default: 1/sqrt(query_dimensions))
orthogonal: bool, When True the random matrix is initialized for
orthogonal random features to reduce the approximation
variance (default: False)
"""
def __init__(
self,
head_size: int,
kernel_size: Optional[int] = None,
softmax_temp: Optional[float] = None,
orthogonal: bool = False
):
super(RandomFourierFeatures, self).__init__(head_size)
assert kernel_size % 2 == 0, "kernel size not divisible by 2"
self.kernel_size = kernel_size
self.orthogonal = orthogonal
self.softmax_temp = (
1/sqrt(head_size) if softmax_temp is None
else softmax_temp
)
# Make a buffer for storing the sampled omega
self.register_buffer(
"omega",
torch.zeros(head_size, self.kernel_size//2)
)
def new_kernel(self):
if self.orthogonal:
orthogonal_random_matrix_(self.omega)
else:
self.omega.normal_()
def forward(
self,
x: torch.Tensor,
**kwargs
) -> torch.Tensor:
x = x * sqrt(self.softmax_temp)
u = x.unsqueeze(-2).matmul(self.omega).squeeze(-2)
phi = torch.cat([torch.cos(u), torch.sin(u)], dim=-1)
return phi * sqrt(2/self.kernel_size)
class SmoothedRandomFourierFeatures(RandomFourierFeatures):
"""Simply add a constant value to the dot product in order to avoid
possible numerical instabilities when the feature map is slightly
negative.
Implements K(x, y) = exp(-|x-y|^2) + s.
Arguments
---------
query_dimensions: int, The input query dimensions in order to sample
the noise matrix
n_dims: int, The size of the feature map (should be divisible by 2)
(default: query_dimensions)
softmax_temp: float, The temerature for the Gaussian kernel
approximation exp(-t * |x-y|^2)
(default: 1/sqrt(query_dimensions))
orthogonal: bool, When True the random matrix is initialized for
orthogonal random features to reduce the approximation
variance (default: False)
smoothing: float, The smoothing parameter to add to the dot product.
"""
def __init__(self, query_dimensions, n_dims=None, softmax_temp=None,
orthogonal=False, smoothing=1.0):
super(SmoothedRandomFourierFeatures, self).__init__(
query_dimensions,
n_dims=query_dimensions-1 if n_dims is None else n_dims-1,
softmax_temp=softmax_temp,
orthogonal=orthogonal,
)
self.smoothing = smoothing
def forward(self, x):
y = super().forward(x)
smoothing = torch.full(
y.shape[:-1] + (1,),
self.smoothing,
dtype=y.dtype,
device=y.device
)
return torch.cat([y, smoothing], dim=-1)
class SoftmaxKernel(Kernel):
"""Positive orthogonal random features that approximate the softmax kernel.
Basically implementation of Lemma 1 from "Rethinking Attention with
Performers".
Arguments
---------
head_size: int, The input query dimensions in order to sample
the noise matrix
kernel_size: int, The size of the feature map (should be divisible by 2)
(default: query_dimensions)
softmax_temp: float, The temerature for the softmax approximation
(default: 1/sqrt(query_dimensions))
orthogonal: bool, If set to true then the random matrix should be
orthogonal which results in lower approximation variance
(default: True)
stabilize: bool, If set to True subtract the max norm from the
exponentials to make sure that there are no infinities. It
is equivalent to a robust implementation of softmax where
the max is subtracted before the exponentiation.
(default: False)
"""
def __init__(
self,
head_size: int,
kernel_size: Optional[int] = None,
ortho_scaling: Optional[float] = 0,
causal: bool = False,
orthogonal: bool = True,
eps: float = 1e-6
):
super(SoftmaxKernel, self).__init__(head_size)
kernel_size = int(self.head_size * log(self.head_size)) if kernel_size is None else kernel_size
self.kernel_size = kernel_size
self.ortho_scaling = ortho_scaling
self.causal = causal
self.orthogonal = orthogonal
self.eps = eps
self.register_buffer(
"omegas",
self.new_kernel()
)
if self.causal:
raise NotImplementedError("linear causal attention not yet implemented")
def new_kernel(
self,
device: Optional[torch.device] = "cpu"
):
return orthogonal_random_matrix_(
self.kernel_size,
self.head_size,
scaling=self.ortho_scaling,
device=device
)
def forward(
self,
x: torch.Tensor,
is_query: bool,
normalize_data: bool = True,
) -> torch.Tensor:
b, h, *_ = x.shape
if normalize_data:
x_norm = 1. / (x.shape[-1] ** 0.25)
else:
x_norm = 1.
ratio = 1. / (self.omegas.shape[0] ** 0.5)
projection_matrix = repeat(self.omegas, 'j d -> b h j d', b=b, h=h)
data_dash = torch.einsum('...id,...jd->...ij', (x_norm * x), projection_matrix)
diag_x = torch.sum(x ** 2, dim=-1)
diag_x = ((diag_x / 2.0) * (x_norm ** 2)).unsqueeze(dim=-1)
if is_query:
data_dash = ratio * (
torch.exp(data_dash - diag_x - torch.max(data_dash, dim=-1, keepdim=True).values) + self.eps
)
else:
data_dash = ratio * (
torch.exp(data_dash - diag_x - torch.max(data_dash)) + self.eps
)
return data_dash
class GeneralizedRandomFeatures(RandomFourierFeatures):
"""Implements the generalized random Fourier features from Performers.
It computes φ(χ) = [f(ω_1 χ), f(ω_2 χ), ..., f(ω_n χ)] where f(.) is the
passed in `kernel_fn`.
Arguments
---------
query_dimensions: int, The input query dimensions in order to sample
the noise matrix
n_dims: int, The size of the feature map (default: query_dimensions)
softmax_temp: float, A normalizer for the dot products that is
multiplied to the input features before the feature map
application (default: 1.0)
orthogonal: bool, If set to true then the random matrix should be
orthogonal which results in lower approximation variance
(default: True)
kernel_fn: callable, defines the f used for the feature map.
(default: relu)
"""
def __init__(self, query_dimensions, n_dims=None, softmax_temp= | num_full_blocks = int(num_rows / num_columns)
block_list = []
for _ in range(num_full_blocks):
q = orthogonal_matrix_chunk(num_columns, device)
block_list.append(q)
remaining_rows = num_rows - (num_full_blocks * num_columns)
if remaining_rows > 0:
q = orthogonal_matrix_chunk(num_columns, device)
block_list.append(q[:remaining_rows])
final_matrix = torch.cat(block_list)
if scaling == 0:
multiplier = torch.randn((num_rows, num_columns), device=device)\
.norm(dim=1)
elif scaling == 1:
multiplier = sqrt((float(num_columns))) * torch.ones((num_rows,), device=device) | identifier_body |
fourier.py | paper
"Rethinking Attention with Performers" https://arxiv.org/pdf/2009.14794.pdf
and the traditional random Fourier features that approximate the RBF kernel.
"""
from math import sqrt, log
from typing import Optional
from einops import repeat
import torch
from .base import Kernel
def orthogonal_random_matrix_(
num_rows: int,
num_columns: int,
scaling: float = 0,
device: Optional[torch.device] = None
):
num_full_blocks = int(num_rows / num_columns)
block_list = []
for _ in range(num_full_blocks):
q = orthogonal_matrix_chunk(num_columns, device)
block_list.append(q)
remaining_rows = num_rows - (num_full_blocks * num_columns)
if remaining_rows > 0:
q = orthogonal_matrix_chunk(num_columns, device)
block_list.append(q[:remaining_rows])
final_matrix = torch.cat(block_list)
if scaling == 0:
multiplier = torch.randn((num_rows, num_columns), device=device)\
.norm(dim=1)
elif scaling == 1:
|
else:
raise ValueError(f"Invalid scaling {scaling}")
return torch.diag(multiplier) @ final_matrix
def orthogonal_matrix_chunk(
cols: int,
device: torch.device = None
) -> torch.Tensor:
unstructured_block = torch.randn((cols, cols), device=device)
q, _ = torch.qr(unstructured_block.cpu(), some=True)
q = q.to(device)
return q.t()
class RandomFourierFeatures(Kernel):
"""Random Fourier Features for the RBF kernel according to [1].
[1]: "Weighted Sums of Random Kitchen Sinks: Replacing minimization with
randomization in learning" by A. Rahimi and Benjamin Recht.
Arguments
---------
hidden_size: int, The input query dimensions in order to sample
the noise matrix
softmax_temp: float, The temerature for the Gaussian kernel
approximation exp(-t * |x-y|^2)
(default: 1/sqrt(query_dimensions))
orthogonal: bool, When True the random matrix is initialized for
orthogonal random features to reduce the approximation
variance (default: False)
"""
def __init__(
self,
head_size: int,
kernel_size: Optional[int] = None,
softmax_temp: Optional[float] = None,
orthogonal: bool = False
):
super(RandomFourierFeatures, self).__init__(head_size)
assert kernel_size % 2 == 0, "kernel size not divisible by 2"
self.kernel_size = kernel_size
self.orthogonal = orthogonal
self.softmax_temp = (
1/sqrt(head_size) if softmax_temp is None
else softmax_temp
)
# Make a buffer for storing the sampled omega
self.register_buffer(
"omega",
torch.zeros(head_size, self.kernel_size//2)
)
def new_kernel(self):
if self.orthogonal:
orthogonal_random_matrix_(self.omega)
else:
self.omega.normal_()
def forward(
self,
x: torch.Tensor,
**kwargs
) -> torch.Tensor:
x = x * sqrt(self.softmax_temp)
u = x.unsqueeze(-2).matmul(self.omega).squeeze(-2)
phi = torch.cat([torch.cos(u), torch.sin(u)], dim=-1)
return phi * sqrt(2/self.kernel_size)
class SmoothedRandomFourierFeatures(RandomFourierFeatures):
"""Simply add a constant value to the dot product in order to avoid
possible numerical instabilities when the feature map is slightly
negative.
Implements K(x, y) = exp(-|x-y|^2) + s.
Arguments
---------
query_dimensions: int, The input query dimensions in order to sample
the noise matrix
n_dims: int, The size of the feature map (should be divisible by 2)
(default: query_dimensions)
softmax_temp: float, The temerature for the Gaussian kernel
approximation exp(-t * |x-y|^2)
(default: 1/sqrt(query_dimensions))
orthogonal: bool, When True the random matrix is initialized for
orthogonal random features to reduce the approximation
variance (default: False)
smoothing: float, The smoothing parameter to add to the dot product.
"""
def __init__(self, query_dimensions, n_dims=None, softmax_temp=None,
orthogonal=False, smoothing=1.0):
super(SmoothedRandomFourierFeatures, self).__init__(
query_dimensions,
n_dims=query_dimensions-1 if n_dims is None else n_dims-1,
softmax_temp=softmax_temp,
orthogonal=orthogonal,
)
self.smoothing = smoothing
def forward(self, x):
y = super().forward(x)
smoothing = torch.full(
y.shape[:-1] + (1,),
self.smoothing,
dtype=y.dtype,
device=y.device
)
return torch.cat([y, smoothing], dim=-1)
class SoftmaxKernel(Kernel):
"""Positive orthogonal random features that approximate the softmax kernel.
Basically implementation of Lemma 1 from "Rethinking Attention with
Performers".
Arguments
---------
head_size: int, The input query dimensions in order to sample
the noise matrix
kernel_size: int, The size of the feature map (should be divisible by 2)
(default: query_dimensions)
softmax_temp: float, The temerature for the softmax approximation
(default: 1/sqrt(query_dimensions))
orthogonal: bool, If set to true then the random matrix should be
orthogonal which results in lower approximation variance
(default: True)
stabilize: bool, If set to True subtract the max norm from the
exponentials to make sure that there are no infinities. It
is equivalent to a robust implementation of softmax where
the max is subtracted before the exponentiation.
(default: False)
"""
def __init__(
self,
head_size: int,
kernel_size: Optional[int] = None,
ortho_scaling: Optional[float] = 0,
causal: bool = False,
orthogonal: bool = True,
eps: float = 1e-6
):
super(SoftmaxKernel, self).__init__(head_size)
kernel_size = int(self.head_size * log(self.head_size)) if kernel_size is None else kernel_size
self.kernel_size = kernel_size
self.ortho_scaling = ortho_scaling
self.causal = causal
self.orthogonal = orthogonal
self.eps = eps
self.register_buffer(
"omegas",
self.new_kernel()
)
if self.causal:
raise NotImplementedError("linear causal attention not yet implemented")
def new_kernel(
self,
device: Optional[torch.device] = "cpu"
):
return orthogonal_random_matrix_(
self.kernel_size,
self.head_size,
scaling=self.ortho_scaling,
device=device
)
def forward(
self,
x: torch.Tensor,
is_query: bool,
normalize_data: bool = True,
) -> torch.Tensor:
b, h, *_ = x.shape
if normalize_data:
x_norm = 1. / (x.shape[-1] ** 0.25)
else:
x_norm = 1.
ratio = 1. / (self.omegas.shape[0] ** 0.5)
projection_matrix = repeat(self.omegas, 'j d -> b h j d', b=b, h=h)
data_dash = torch.einsum('...id,...jd->...ij', (x_norm * x), projection_matrix)
diag_x = torch.sum(x ** 2, dim=-1)
diag_x = ((diag_x / 2.0) * (x_norm ** 2)).unsqueeze(dim=-1)
if is_query:
data_dash = ratio * (
torch.exp(data_dash - diag_x - torch.max(data_dash, dim=-1, keepdim=True).values) + self.eps
)
else:
data_dash = ratio * (
torch.exp(data_dash - diag_x - torch.max(data_dash)) + self.eps
)
return data_dash
class GeneralizedRandomFeatures(RandomFourierFeatures):
"""Implements the generalized random Fourier features from Performers.
It computes φ(χ) = [f(ω_1 χ), f(ω_2 χ), ..., f(ω_n χ)] where f(.) is the
passed in `kernel_fn`.
Arguments
---------
query_dimensions: int, The input query dimensions in order to sample
the noise matrix
n_dims: int, The size of the feature map (default: query_dimensions)
softmax_temp: float, A normalizer for the dot products that is
multiplied to the input features before the feature map
application (default: 1.0)
orthogonal: bool, If set to true then the random matrix should be
orthogonal which results in lower approximation variance
(default: True)
kernel_fn: callable, defines the f used for the feature map.
(default: relu)
"""
def __init__(self, query_dimensions, n_dims=None, softmax_temp | multiplier = sqrt((float(num_columns))) * torch.ones((num_rows,), device=device) | conditional_block |
fourier.py | the paper
"Rethinking Attention with Performers" https://arxiv.org/pdf/2009.14794.pdf
and the traditional random Fourier features that approximate the RBF kernel.
"""
from math import sqrt, log
from typing import Optional
from einops import repeat
import torch
from .base import Kernel
def orthogonal_random_matrix_(
num_rows: int,
num_columns: int,
scaling: float = 0,
device: Optional[torch.device] = None
):
num_full_blocks = int(num_rows / num_columns)
block_list = []
for _ in range(num_full_blocks):
q = orthogonal_matrix_chunk(num_columns, device)
block_list.append(q)
remaining_rows = num_rows - (num_full_blocks * num_columns)
if remaining_rows > 0:
q = orthogonal_matrix_chunk(num_columns, device)
block_list.append(q[:remaining_rows])
final_matrix = torch.cat(block_list)
if scaling == 0:
multiplier = torch.randn((num_rows, num_columns), device=device)\
.norm(dim=1)
elif scaling == 1:
multiplier = sqrt((float(num_columns))) * torch.ones((num_rows,), device=device)
else:
raise ValueError(f"Invalid scaling {scaling}")
return torch.diag(multiplier) @ final_matrix
def orthogonal_matrix_chunk(
cols: int,
device: torch.device = None
) -> torch.Tensor:
unstructured_block = torch.randn((cols, cols), device=device)
q, _ = torch.qr(unstructured_block.cpu(), some=True)
q = q.to(device)
return q.t()
class RandomFourierFeatures(Kernel):
"""Random Fourier Features for the RBF kernel according to [1].
[1]: "Weighted Sums of Random Kitchen Sinks: Replacing minimization with
randomization in learning" by A. Rahimi and Benjamin Recht.
Arguments
---------
hidden_size: int, The input query dimensions in order to sample
the noise matrix
softmax_temp: float, The temerature for the Gaussian kernel
approximation exp(-t * |x-y|^2)
(default: 1/sqrt(query_dimensions))
orthogonal: bool, When True the random matrix is initialized for
orthogonal random features to reduce the approximation
variance (default: False)
"""
def __init__(
self,
head_size: int,
kernel_size: Optional[int] = None,
softmax_temp: Optional[float] = None,
orthogonal: bool = False
):
super(RandomFourierFeatures, self).__init__(head_size)
assert kernel_size % 2 == 0, "kernel size not divisible by 2"
self.kernel_size = kernel_size
self.orthogonal = orthogonal
self.softmax_temp = (
1/sqrt(head_size) if softmax_temp is None
else softmax_temp
)
# Make a buffer for storing the sampled omega
self.register_buffer(
"omega",
torch.zeros(head_size, self.kernel_size//2)
)
def new_kernel(self):
if self.orthogonal:
orthogonal_random_matrix_(self.omega)
else:
self.omega.normal_()
def forward(
self,
x: torch.Tensor,
**kwargs
) -> torch.Tensor:
x = x * sqrt(self.softmax_temp)
u = x.unsqueeze(-2).matmul(self.omega).squeeze(-2)
phi = torch.cat([torch.cos(u), torch.sin(u)], dim=-1)
return phi * sqrt(2/self.kernel_size)
class SmoothedRandomFourierFeatures(RandomFourierFeatures):
"""Simply add a constant value to the dot product in order to avoid
possible numerical instabilities when the feature map is slightly
negative.
Implements K(x, y) = exp(-|x-y|^2) + s.
Arguments
---------
query_dimensions: int, The input query dimensions in order to sample
the noise matrix
n_dims: int, The size of the feature map (should be divisible by 2)
(default: query_dimensions)
softmax_temp: float, The temerature for the Gaussian kernel
approximation exp(-t * |x-y|^2)
(default: 1/sqrt(query_dimensions))
orthogonal: bool, When True the random matrix is initialized for
orthogonal random features to reduce the approximation
variance (default: False)
smoothing: float, The smoothing parameter to add to the dot product.
"""
def __init__(self, query_dimensions, n_dims=None, softmax_temp=None,
orthogonal=False, smoothing=1.0):
super(SmoothedRandomFourierFeatures, self).__init__(
query_dimensions,
n_dims=query_dimensions-1 if n_dims is None else n_dims-1,
softmax_temp=softmax_temp,
orthogonal=orthogonal,
)
self.smoothing = smoothing
def forward(self, x):
y = super().forward(x)
smoothing = torch.full(
y.shape[:-1] + (1,),
self.smoothing,
dtype=y.dtype,
device=y.device
)
return torch.cat([y, smoothing], dim=-1)
class SoftmaxKernel(Kernel):
"""Positive orthogonal random features that approximate the softmax kernel.
Basically implementation of Lemma 1 from "Rethinking Attention with
Performers".
Arguments
---------
head_size: int, The input query dimensions in order to sample
the noise matrix
kernel_size: int, The size of the feature map (should be divisible by 2)
(default: query_dimensions)
softmax_temp: float, The temerature for the softmax approximation
(default: 1/sqrt(query_dimensions))
orthogonal: bool, If set to true then the random matrix should be
orthogonal which results in lower approximation variance
(default: True)
stabilize: bool, If set to True subtract the max norm from the
exponentials to make sure that there are no infinities. It
is equivalent to a robust implementation of softmax where
the max is subtracted before the exponentiation.
(default: False)
"""
def __init__(
self,
head_size: int,
kernel_size: Optional[int] = None,
ortho_scaling: Optional[float] = 0,
causal: bool = False,
orthogonal: bool = True,
eps: float = 1e-6
):
super(SoftmaxKernel, self).__init__(head_size)
kernel_size = int(self.head_size * log(self.head_size)) if kernel_size is None else kernel_size
self.kernel_size = kernel_size
self.ortho_scaling = ortho_scaling
self.causal = causal
self.orthogonal = orthogonal
self.eps = eps
self.register_buffer(
"omegas",
self.new_kernel()
)
if self.causal:
raise NotImplementedError("linear causal attention not yet implemented")
def new_kernel(
self,
device: Optional[torch.device] = "cpu"
):
return orthogonal_random_matrix_(
self.kernel_size,
self.head_size,
scaling=self.ortho_scaling,
device=device
)
def forward(
self,
x: torch.Tensor,
is_query: bool,
normalize_data: bool = True,
) -> torch.Tensor:
b, h, *_ = x.shape
if normalize_data:
x_norm = 1. / (x.shape[-1] ** 0.25)
else:
x_norm = 1.
ratio = 1. / (self.omegas.shape[0] ** 0.5)
projection_matrix = repeat(self.omegas, 'j d -> b h j d', b=b, h=h)
data_dash = torch.einsum('...id,...jd->...ij', (x_norm * x), projection_matrix)
diag_x = torch.sum(x ** 2, dim=-1)
diag_x = ((diag_x / 2.0) * (x_norm ** 2)).unsqueeze(dim=-1)
if is_query:
data_dash = ratio * (
torch.exp(data_dash - diag_x - torch.max(data_dash, dim=-1, keepdim=True).values) + self.eps
)
else:
data_dash = ratio * (
torch.exp(data_dash - diag_x - torch.max(data_dash)) + self.eps
)
return data_dash
class GeneralizedRandomFeatures(RandomFourierFeatures):
"""Implements the generalized random Fourier features from Performers.
It computes φ(χ) = [f(ω_1 χ), f(ω_2 χ), ..., f(ω_n χ)] where f(.) is the | query_dimensions: int, The input query dimensions in order to sample
the noise matrix
n_dims: int, The size of the feature map (default: query_dimensions)
softmax_temp: float, A normalizer for the dot products that is
multiplied to the input features before the feature map
application (default: 1.0)
orthogonal: bool, If set to true then the random matrix should be
orthogonal which results in lower approximation variance
(default: True)
kernel_fn: callable, defines the f used for the feature map.
(default: relu)
"""
def __init__(self, query_dimensions, n_dims=None, softmax_temp=1. | passed in `kernel_fn`.
Arguments
--------- | random_line_split |
fourier.py | paper
"Rethinking Attention with Performers" https://arxiv.org/pdf/2009.14794.pdf
and the traditional random Fourier features that approximate the RBF kernel.
"""
from math import sqrt, log
from typing import Optional
from einops import repeat
import torch
from .base import Kernel
def orthogonal_random_matrix_(
num_rows: int,
num_columns: int,
scaling: float = 0,
device: Optional[torch.device] = None
):
num_full_blocks = int(num_rows / num_columns)
block_list = []
for _ in range(num_full_blocks):
q = orthogonal_matrix_chunk(num_columns, device)
block_list.append(q)
remaining_rows = num_rows - (num_full_blocks * num_columns)
if remaining_rows > 0:
q = orthogonal_matrix_chunk(num_columns, device)
block_list.append(q[:remaining_rows])
final_matrix = torch.cat(block_list)
if scaling == 0:
multiplier = torch.randn((num_rows, num_columns), device=device)\
.norm(dim=1)
elif scaling == 1:
multiplier = sqrt((float(num_columns))) * torch.ones((num_rows,), device=device)
else:
raise ValueError(f"Invalid scaling {scaling}")
return torch.diag(multiplier) @ final_matrix
def orthogonal_matrix_chunk(
cols: int,
device: torch.device = None
) -> torch.Tensor:
unstructured_block = torch.randn((cols, cols), device=device)
q, _ = torch.qr(unstructured_block.cpu(), some=True)
q = q.to(device)
return q.t()
class RandomFourierFeatures(Kernel):
"""Random Fourier Features for the RBF kernel according to [1].
[1]: "Weighted Sums of Random Kitchen Sinks: Replacing minimization with
randomization in learning" by A. Rahimi and Benjamin Recht.
Arguments
---------
hidden_size: int, The input query dimensions in order to sample
the noise matrix
softmax_temp: float, The temerature for the Gaussian kernel
approximation exp(-t * |x-y|^2)
(default: 1/sqrt(query_dimensions))
orthogonal: bool, When True the random matrix is initialized for
orthogonal random features to reduce the approximation
variance (default: False)
"""
def __init__(
self,
head_size: int,
kernel_size: Optional[int] = None,
softmax_temp: Optional[float] = None,
orthogonal: bool = False
):
super(RandomFourierFeatures, self).__init__(head_size)
assert kernel_size % 2 == 0, "kernel size not divisible by 2"
self.kernel_size = kernel_size
self.orthogonal = orthogonal
self.softmax_temp = (
1/sqrt(head_size) if softmax_temp is None
else softmax_temp
)
# Make a buffer for storing the sampled omega
self.register_buffer(
"omega",
torch.zeros(head_size, self.kernel_size//2)
)
def new_kernel(self):
if self.orthogonal:
orthogonal_random_matrix_(self.omega)
else:
self.omega.normal_()
def forward(
self,
x: torch.Tensor,
**kwargs
) -> torch.Tensor:
x = x * sqrt(self.softmax_temp)
u = x.unsqueeze(-2).matmul(self.omega).squeeze(-2)
phi = torch.cat([torch.cos(u), torch.sin(u)], dim=-1)
return phi * sqrt(2/self.kernel_size)
class SmoothedRandomFourierFeatures(RandomFourierFeatures):
"""Simply add a constant value to the dot product in order to avoid
possible numerical instabilities when the feature map is slightly
negative.
Implements K(x, y) = exp(-|x-y|^2) + s.
Arguments
---------
query_dimensions: int, The input query dimensions in order to sample
the noise matrix
n_dims: int, The size of the feature map (should be divisible by 2)
(default: query_dimensions)
softmax_temp: float, The temerature for the Gaussian kernel
approximation exp(-t * |x-y|^2)
(default: 1/sqrt(query_dimensions))
orthogonal: bool, When True the random matrix is initialized for
orthogonal random features to reduce the approximation
variance (default: False)
smoothing: float, The smoothing parameter to add to the dot product.
"""
def __init__(self, query_dimensions, n_dims=None, softmax_temp=None,
orthogonal=False, smoothing=1.0):
super(SmoothedRandomFourierFeatures, self).__init__(
query_dimensions,
n_dims=query_dimensions-1 if n_dims is None else n_dims-1,
softmax_temp=softmax_temp,
orthogonal=orthogonal,
)
self.smoothing = smoothing
def forward(self, x):
y = super().forward(x)
smoothing = torch.full(
y.shape[:-1] + (1,),
self.smoothing,
dtype=y.dtype,
device=y.device
)
return torch.cat([y, smoothing], dim=-1)
class SoftmaxKernel(Kernel):
"""Positive orthogonal random features that approximate the softmax kernel.
Basically implementation of Lemma 1 from "Rethinking Attention with
Performers".
Arguments
---------
head_size: int, The input query dimensions in order to sample
the noise matrix
kernel_size: int, The size of the feature map (should be divisible by 2)
(default: query_dimensions)
softmax_temp: float, The temerature for the softmax approximation
(default: 1/sqrt(query_dimensions))
orthogonal: bool, If set to true then the random matrix should be
orthogonal which results in lower approximation variance
(default: True)
stabilize: bool, If set to True subtract the max norm from the
exponentials to make sure that there are no infinities. It
is equivalent to a robust implementation of softmax where
the max is subtracted before the exponentiation.
(default: False)
"""
def __init__(
self,
head_size: int,
kernel_size: Optional[int] = None,
ortho_scaling: Optional[float] = 0,
causal: bool = False,
orthogonal: bool = True,
eps: float = 1e-6
):
super(SoftmaxKernel, self).__init__(head_size)
kernel_size = int(self.head_size * log(self.head_size)) if kernel_size is None else kernel_size
self.kernel_size = kernel_size
self.ortho_scaling = ortho_scaling
self.causal = causal
self.orthogonal = orthogonal
self.eps = eps
self.register_buffer(
"omegas",
self.new_kernel()
)
if self.causal:
raise NotImplementedError("linear causal attention not yet implemented")
def | (
self,
device: Optional[torch.device] = "cpu"
):
return orthogonal_random_matrix_(
self.kernel_size,
self.head_size,
scaling=self.ortho_scaling,
device=device
)
def forward(
self,
x: torch.Tensor,
is_query: bool,
normalize_data: bool = True,
) -> torch.Tensor:
b, h, *_ = x.shape
if normalize_data:
x_norm = 1. / (x.shape[-1] ** 0.25)
else:
x_norm = 1.
ratio = 1. / (self.omegas.shape[0] ** 0.5)
projection_matrix = repeat(self.omegas, 'j d -> b h j d', b=b, h=h)
data_dash = torch.einsum('...id,...jd->...ij', (x_norm * x), projection_matrix)
diag_x = torch.sum(x ** 2, dim=-1)
diag_x = ((diag_x / 2.0) * (x_norm ** 2)).unsqueeze(dim=-1)
if is_query:
data_dash = ratio * (
torch.exp(data_dash - diag_x - torch.max(data_dash, dim=-1, keepdim=True).values) + self.eps
)
else:
data_dash = ratio * (
torch.exp(data_dash - diag_x - torch.max(data_dash)) + self.eps
)
return data_dash
class GeneralizedRandomFeatures(RandomFourierFeatures):
"""Implements the generalized random Fourier features from Performers.
It computes φ(χ) = [f(ω_1 χ), f(ω_2 χ), ..., f(ω_n χ)] where f(.) is the
passed in `kernel_fn`.
Arguments
---------
query_dimensions: int, The input query dimensions in order to sample
the noise matrix
n_dims: int, The size of the feature map (default: query_dimensions)
softmax_temp: float, A normalizer for the dot products that is
multiplied to the input features before the feature map
application (default: 1.0)
orthogonal: bool, If set to true then the random matrix should be
orthogonal which results in lower approximation variance
(default: True)
kernel_fn: callable, defines the f used for the feature map.
(default: relu)
"""
def __init__(self, query_dimensions, n_dims=None, softmax_temp= | new_kernel | identifier_name |
differentiation.go | 5. Traverse the graph from output towards input. On each visit, perform the symbolic differentiation
//
// For most cases, Grad() should be used instead of Backpropagate(), as Grad() performs several checks which would be the general use case, before calling Backpropagate()
func Backpropagate(outputs, gradOutputs, wrt Nodes) (retVal Nodes, err error) {
symdiffLogf("BACKPROP START")
symdiffLogf("Outputs: %d", outputs)
symdiffLogf("gradOutputs: %d", gradOutputs)
symdiffLogf("WRT: %d", wrt)
enterLogScope()
defer leaveLogScope()
g := outputs[0].g
// this entire section about removing foreveralone nodes need a rethink
symdiffLogf("removing foreveralone nodes")
enterLogScope()
for i := 0; i < len(g.AllNodes()); i++ {
n := g.AllNodes()[i]
fr := g.From(n.ID()).Len()
to := g.To(n.ID()).Len()
if fr == 0 && to == 0 && !n.isConstant() && !n.isInput() {
g.RemoveNode(n)
symdiffLogf("removed %v(%p); %x; %s", n, n, n.ID(), n.Name())
}
}
leaveLogScope()
var sortedNodes Nodes
if sortedNodes, err = Sort(g); err != nil {
return nil, errors.Wrap(err, sortFail)
}
symdiffLogf("sorted nodes: %v", sortedNodes)
symdiffLogf("sorted nodes: %d", sortedNodes)
var affectsOutput NodeSet
var affectedByOutput NodeSet
if affectsOutput, err = forwardDiffAnalysis(outputs, sortedNodes); err != nil {
return nil, errors.Wrap(err, "Failed during forward differentiation analysis")
}
if affectedByOutput, err = backwardDiffAnalysis(wrt, sortedNodes); err != nil {
return nil, errors.Wrap(err, "Failed during forward differentiation analysis")
}
symdiffLogf("affects output: %v", affectsOutput)
symdiffLogf("affected by output : %v", affectedByOutput)
wrtSet := wrt.mapSet()
badWRTs := wrtSet.Difference(affectsOutput)
if len(badWRTs) > 0 {
return nil, SymDiffError{nodes: badWRTs.ToSlice(), err: errors.Errorf("Non Differentiable WRTs: %v", badWRTs)}
}
outputSet := outputs.mapSet()
badOutputs := outputSet.Difference(affectedByOutput)
if len(badOutputs) > 0 {
symdiffLogf("badOutputs: %#v", badOutputs)
return nil, SymDiffError{nodes: badOutputs.ToSlice(), err: errors.Errorf("Non-Differentable Outputs: %v", badOutputs)}
}
// map a node to a list of gradient terms
// these gradient terms will be summed up when we visit the node
// when iterating through the nondes in reverse topological order
nodeGradMap := make(map[*Node]Nodes)
for i, n := range outputs {
symdiffLogf("Adding outputs for %x", n.ID())
nodeGradMap[n] = Nodes{gradOutputs[i]}
}
// "active" nodes are the ones that are differentially influenced by the inputs
// and also differentiably influence the outputs. These are the nodes where we need to call the
// "pullback" function to backpropagate derivatives
activeNodes := affectsOutput.Intersect(affectedByOutput)
symdiffLogf("Active: %v", activeNodes)
symdiffLogf("Sorted: %d", sortedNodes)
symdiffLogf("nodeGradMap: %+#d", FmtNodeMap(nodeGradMap))
enterLogScope()
for _, node := range sortedNodes {
if _, ok := activeNodes[node]; !ok {
symdiffLogf("skipping %x", node.ID())
continue
}
if node.deriv != nil {
symdiffLogf("skipping %x - previously differentiated", node.ID())
nodeGradMap[node] = append(nodeGradMap[node], node.deriv)
continue
}
symdiffLogf("Working on %x %v", node.ID(), node)
enterLogScope()
// Check if there is any grads coming into this node
if len(nodeGradMap[node]) < 1 {
leaveLogScope()
return nil, SymDiffError{
single: node,
gradMap: nodeGradMap,
err: errors.New("No gradients found for node"),
}
}
// once we've reached a node, we already backpropagated from its dependents
// so we sum up the gradients
symdiffLogf("nodeGradMap[%x]: %d", node.ID(), nodeGradMap[node])
if len(nodeGradMap[node]) > 1 {
var n *Node
symdiffLogf("reduce adding")
if n, err = ReduceAdd(nodeGradMap[node], WithGroupName(gradClust)); err != nil {
leaveLogScope()
return nil, SymDiffError{
single: node,
nodes: nodeGradMap[node],
gradMap: nodeGradMap,
err: errors.Wrap(err, "ReduceAdd failed during differentiation"),
}
}
symdiffLogf("reduced to... %x", n.ID())
// node.derives = append(node.derives, n)
n.derivOf = append(n.derivOf, node)
node.deriv = n
nodeGradMap[node] = Nodes{n}
// }
} else if len(nodeGradMap[node]) == 1 {
deriv := nodeGradMap[node][0]
deriv.derivOf = append(deriv.derivOf, node)
node.deriv = deriv
}
gradNode := nodeGradMap[node][0]
if !node.isInput() {
symdiffLogf("differentiating %x (%v)", node.ID(), node.op)
enterLogScope()
var op SDOp
var childrenGrads Nodes
var ok bool
if op, ok = node.op.(SDOp); !ok {
return nil, SymDiffError{
single: node,
err: errors.New("Not a SymDifOp"),
}
}
symdiffLogf("op: %v || optype: %v || node: %v || Children: %#Y || Grad: %v", node.op, node.op.Type(), node.t, node.children, gradNode)
if childrenGrads, err = op.SymDiff(node.children, node, gradNode); err != nil {
leaveLogScope()
return nil, SymDiffError{
single: node,
grad: gradNode,
gradMap: nodeGradMap,
err: errors.Wrapf(err, ".SymDiff() failed"),
}
}
symdiffLogf("Derived(%d): %P", len(childrenGrads), childrenGrads)
leaveLogScope()
diffs := node.diffWRT()
for i, child := range node.children {
symdiffLogf("child is %v, i: %v", child, i)
differentiable := diffs[i]
childGrad := childrenGrads[i]
if differentiable {
childGrad.setGroup(gradClust)
if grads, ok := nodeGradMap[child]; ok {
grads = append(grads, childGrad)
nodeGradMap[child] = grads
} else {
nodeGradMap[child] = Nodes{childGrad}
}
} else {
symdiffLogf("Child %x is non differentiable", child.ID())
if childGrad != nil {
childGrad.setGroup(strayClust)
}
}
}
} else {
symdiffLogf("iz input")
symdiffLogf("%d ", nodeGradMap[node])
}
leaveLogScope()
}
leaveLogScope()
// only we already summed up the gradients for the input nodes, so just take
// 0th element
for _, n := range wrt {
symdiffLogf("nodeGradMap wrt: %d", nodeGradMap[n])
retVal = append(retVal, nodeGradMap[n][0])
}
return
}
// SetDerivOf is used to hack around the fundamental limitations of Gorgonia.
//
// Specifically it is used to set a node as the derivative of another node,
// used in the cuDNN version of batch norm.
//
// The cuDNN BatchNorm operation produces the derivatives for the scale and bias as a side effect
// of calculating the derivative of the input. Because Gorgonia's Ops are modelled as pure functions (and no tuples)
// this causes a bit of trouble. With the clever use of scratch space ops multireturn can be simulated.
// But this causes derivatives to not be set correctly.
func SetDerivOf(deriv, of *Node) | {
deriv.derivOf = append(deriv.derivOf, of)
of.deriv = deriv
} | identifier_body |
|
differentiation.go | list of Nodes that are affected by differentiating output.
// Given a list of WRTs, we want to find a list of nodes that will be affected when backpropagating.
func backwardDiffAnalysis(wrt, sortedNodes Nodes) (retVal NodeSet, err error) {
symdiffLogf("Backwards analysis")
enterLogScope()
defer leaveLogScope()
if !wrt.AllSameGraph() {
return nil, errors.New("The supplied output Nodes are not the same graph")
}
diffSet := wrt.mapSet()
symdiffLogf("wrt:%d diffset: %d", len(wrt), len(diffSet))
symdiffLogf("%v", diffSet)
symdiffLogf("sorted: %d", sortedNodes)
enterLogScope()
for i := len(sortedNodes) - 1; i >= 0; i-- {
n := sortedNodes[i]
symdiffLogf("working on %v. Has %d children", n, len(n.children))
var op SDOp
var ok bool
var diffs []bool
if op, ok = n.op.(SDOp); ok {
diffs = op.DiffWRT(len(n.children))
}
symdiffLogf("differentiable WRT: %v", diffs)
enterLogScope()
symdiffLogf("Children: %v", n.children)
if len(diffs) == 0 {
// check if this makes nodes unreachable. If it does, then error out
if n.isStmt {
symdiffLogf("Statement nodes are Non differentiable!")
leaveLogScope()
continue
} else if n.isInput() {
symdiffLogf("Input nodes are Non differentiable")
leaveLogScope()
continue
} else if len(n.children) == 0 {
symdiffLogf("Leaf nodes have no children")
leaveLogScope()
continue
}
g := n.g
for _, child := range n.children {
parents := graph.NodesOf(g.To(child.ID()))
if len(parents) == 1 && len(child.children) > 0 {
leaveLogScope()
return nil, errors.Errorf("Being unable to differentiate %v would leave a portion of the graph unreachable. Unable to continue", n)
}
}
symdiffLogf("SKIPPING... Non differentiable!")
leaveLogScope()
continue
}
inner:
for j, child := range n.children {
d := diffs[j]
if diffSet.Contains(child) && d {
symdiffLogf("Adding %x to differentiable set", child.ID())
diffSet.Add(n)
break inner
}
}
leaveLogScope()
}
leaveLogScope()
return diffSet, nil
}
// Backpropagate backpropagates errors by performing reverse-mode symbolic differentiation, starting from the outputs, and working its way towads the inputs.
//
// This is the rough algorithm:
// 1. Filter out nodes that are unreachable
// 2. Forwards analysis, where a list of nodes affecting the output is added to consideration
// 3. Backwards analysis, where a list of nodes affected by differentiating the output are added to the consideration
// 4. If there is a difference in both sets, it will cause an error (both sets should be the same)
// 5. Traverse the graph from output towards input. On each visit, perform the symbolic differentiation
//
// For most cases, Grad() should be used instead of Backpropagate(), as Grad() performs several checks which would be the general use case, before calling Backpropagate()
func Backpropagate(outputs, gradOutputs, wrt Nodes) (retVal Nodes, err error) {
symdiffLogf("BACKPROP START")
symdiffLogf("Outputs: %d", outputs)
symdiffLogf("gradOutputs: %d", gradOutputs)
symdiffLogf("WRT: %d", wrt)
enterLogScope()
defer leaveLogScope()
g := outputs[0].g
// this entire section about removing foreveralone nodes need a rethink
symdiffLogf("removing foreveralone nodes")
enterLogScope()
for i := 0; i < len(g.AllNodes()); i++ {
n := g.AllNodes()[i]
fr := g.From(n.ID()).Len()
to := g.To(n.ID()).Len()
if fr == 0 && to == 0 && !n.isConstant() && !n.isInput() {
g.RemoveNode(n)
symdiffLogf("removed %v(%p); %x; %s", n, n, n.ID(), n.Name())
}
}
leaveLogScope()
var sortedNodes Nodes
if sortedNodes, err = Sort(g); err != nil {
return nil, errors.Wrap(err, sortFail)
}
symdiffLogf("sorted nodes: %v", sortedNodes)
symdiffLogf("sorted nodes: %d", sortedNodes)
var affectsOutput NodeSet
var affectedByOutput NodeSet
if affectsOutput, err = forwardDiffAnalysis(outputs, sortedNodes); err != nil {
return nil, errors.Wrap(err, "Failed during forward differentiation analysis")
}
if affectedByOutput, err = backwardDiffAnalysis(wrt, sortedNodes); err != nil {
return nil, errors.Wrap(err, "Failed during forward differentiation analysis")
}
symdiffLogf("affects output: %v", affectsOutput)
symdiffLogf("affected by output : %v", affectedByOutput)
wrtSet := wrt.mapSet()
badWRTs := wrtSet.Difference(affectsOutput)
if len(badWRTs) > 0 {
return nil, SymDiffError{nodes: badWRTs.ToSlice(), err: errors.Errorf("Non Differentiable WRTs: %v", badWRTs)}
}
outputSet := outputs.mapSet()
badOutputs := outputSet.Difference(affectedByOutput)
if len(badOutputs) > 0 {
symdiffLogf("badOutputs: %#v", badOutputs)
return nil, SymDiffError{nodes: badOutputs.ToSlice(), err: errors.Errorf("Non-Differentable Outputs: %v", badOutputs)}
}
// map a node to a list of gradient terms
// these gradient terms will be summed up when we visit the node
// when iterating through the nondes in reverse topological order
nodeGradMap := make(map[*Node]Nodes)
for i, n := range outputs {
symdiffLogf("Adding outputs for %x", n.ID())
nodeGradMap[n] = Nodes{gradOutputs[i]}
}
// "active" nodes are the ones that are differentially influenced by the inputs
// and also differentiably influence the outputs. These are the nodes where we need to call the
// "pullback" function to backpropagate derivatives
activeNodes := affectsOutput.Intersect(affectedByOutput)
symdiffLogf("Active: %v", activeNodes)
symdiffLogf("Sorted: %d", sortedNodes)
symdiffLogf("nodeGradMap: %+#d", FmtNodeMap(nodeGradMap))
enterLogScope()
for _, node := range sortedNodes {
if _, ok := activeNodes[node]; !ok {
symdiffLogf("skipping %x", node.ID())
continue
}
if node.deriv != nil {
symdiffLogf("skipping %x - previously differentiated", node.ID())
nodeGradMap[node] = append(nodeGradMap[node], node.deriv)
continue
}
symdiffLogf("Working on %x %v", node.ID(), node)
enterLogScope()
// Check if there is any grads coming into this node
if len(nodeGradMap[node]) < 1 {
leaveLogScope()
return nil, SymDiffError{
single: node,
gradMap: nodeGradMap,
err: errors.New("No gradients found for node"),
}
}
// once we've reached a node, we already backpropagated from its dependents
// so we sum up the gradients
symdiffLogf("nodeGradMap[%x]: %d", node.ID(), nodeGradMap[node])
if len(nodeGradMap[node]) > 1 {
var n *Node
symdiffLogf("reduce adding")
if n, err = ReduceAdd(nodeGradMap[node], WithGroupName(gradClust)); err != nil |
symdiffLogf("reduced to... %x", n.ID())
// node.derives = append(node.derives, n)
n.derivOf = append(n.derivOf, node)
node.deriv = n
nodeGradMap[node] = Nodes{n}
// }
} else if len(nodeGradMap[node]) == 1 {
deriv := nodeGradMap[node][0]
deriv.deriv | {
leaveLogScope()
return nil, SymDiffError{
single: node,
nodes: nodeGradMap[node],
gradMap: nodeGradMap,
err: errors.Wrap(err, "ReduceAdd failed during differentiation"),
}
} | conditional_block |
differentiation.go | list of Nodes that are affected by differentiating output.
// Given a list of WRTs, we want to find a list of nodes that will be affected when backpropagating.
func backwardDiffAnalysis(wrt, sortedNodes Nodes) (retVal NodeSet, err error) {
symdiffLogf("Backwards analysis")
enterLogScope()
defer leaveLogScope()
if !wrt.AllSameGraph() {
return nil, errors.New("The supplied output Nodes are not the same graph")
}
diffSet := wrt.mapSet()
symdiffLogf("wrt:%d diffset: %d", len(wrt), len(diffSet))
symdiffLogf("%v", diffSet)
symdiffLogf("sorted: %d", sortedNodes)
enterLogScope()
for i := len(sortedNodes) - 1; i >= 0; i-- {
n := sortedNodes[i]
symdiffLogf("working on %v. Has %d children", n, len(n.children))
var op SDOp
var ok bool
var diffs []bool
if op, ok = n.op.(SDOp); ok {
diffs = op.DiffWRT(len(n.children))
}
symdiffLogf("differentiable WRT: %v", diffs)
enterLogScope()
symdiffLogf("Children: %v", n.children)
if len(diffs) == 0 {
// check if this makes nodes unreachable. If it does, then error out
if n.isStmt {
symdiffLogf("Statement nodes are Non differentiable!")
leaveLogScope()
continue
} else if n.isInput() {
symdiffLogf("Input nodes are Non differentiable")
leaveLogScope()
continue
} else if len(n.children) == 0 {
symdiffLogf("Leaf nodes have no children")
leaveLogScope()
continue
}
g := n.g
for _, child := range n.children {
parents := graph.NodesOf(g.To(child.ID()))
if len(parents) == 1 && len(child.children) > 0 {
leaveLogScope()
return nil, errors.Errorf("Being unable to differentiate %v would leave a portion of the graph unreachable. Unable to continue", n)
}
}
symdiffLogf("SKIPPING... Non differentiable!")
leaveLogScope()
continue
}
inner:
for j, child := range n.children {
d := diffs[j]
if diffSet.Contains(child) && d {
symdiffLogf("Adding %x to differentiable set", child.ID())
diffSet.Add(n)
break inner
}
}
leaveLogScope()
}
leaveLogScope()
return diffSet, nil
}
// Backpropagate backpropagates errors by performing reverse-mode symbolic differentiation, starting from the outputs, and working its way towads the inputs.
//
// This is the rough algorithm:
// 1. Filter out nodes that are unreachable
// 2. Forwards analysis, where a list of nodes affecting the output is added to consideration
// 3. Backwards analysis, where a list of nodes affected by differentiating the output are added to the consideration
// 4. If there is a difference in both sets, it will cause an error (both sets should be the same)
// 5. Traverse the graph from output towards input. On each visit, perform the symbolic differentiation
//
// For most cases, Grad() should be used instead of Backpropagate(), as Grad() performs several checks which would be the general use case, before calling Backpropagate()
func Backpropagate(outputs, gradOutputs, wrt Nodes) (retVal Nodes, err error) {
symdiffLogf("BACKPROP START")
symdiffLogf("Outputs: %d", outputs)
symdiffLogf("gradOutputs: %d", gradOutputs)
symdiffLogf("WRT: %d", wrt)
enterLogScope()
defer leaveLogScope()
g := outputs[0].g
// this entire section about removing foreveralone nodes need a rethink
symdiffLogf("removing foreveralone nodes")
enterLogScope()
for i := 0; i < len(g.AllNodes()); i++ {
n := g.AllNodes()[i]
fr := g.From(n.ID()).Len()
to := g.To(n.ID()).Len()
if fr == 0 && to == 0 && !n.isConstant() && !n.isInput() {
g.RemoveNode(n)
symdiffLogf("removed %v(%p); %x; %s", n, n, n.ID(), n.Name())
}
}
leaveLogScope()
var sortedNodes Nodes
if sortedNodes, err = Sort(g); err != nil {
return nil, errors.Wrap(err, sortFail)
}
symdiffLogf("sorted nodes: %v", sortedNodes)
symdiffLogf("sorted nodes: %d", sortedNodes)
var affectsOutput NodeSet
var affectedByOutput NodeSet
if affectsOutput, err = forwardDiffAnalysis(outputs, sortedNodes); err != nil {
return nil, errors.Wrap(err, "Failed during forward differentiation analysis")
}
if affectedByOutput, err = backwardDiffAnalysis(wrt, sortedNodes); err != nil {
return nil, errors.Wrap(err, "Failed during forward differentiation analysis")
}
symdiffLogf("affects output: %v", affectsOutput)
symdiffLogf("affected by output : %v", affectedByOutput)
wrtSet := wrt.mapSet()
badWRTs := wrtSet.Difference(affectsOutput)
if len(badWRTs) > 0 {
return nil, SymDiffError{nodes: badWRTs.ToSlice(), err: errors.Errorf("Non Differentiable WRTs: %v", badWRTs)}
}
outputSet := outputs.mapSet()
badOutputs := outputSet.Difference(affectedByOutput)
if len(badOutputs) > 0 {
symdiffLogf("badOutputs: %#v", badOutputs)
return nil, SymDiffError{nodes: badOutputs.ToSlice(), err: errors.Errorf("Non-Differentable Outputs: %v", badOutputs)}
}
// map a node to a list of gradient terms
// these gradient terms will be summed up when we visit the node
// when iterating through the nondes in reverse topological order
nodeGradMap := make(map[*Node]Nodes)
for i, n := range outputs {
symdiffLogf("Adding outputs for %x", n.ID())
nodeGradMap[n] = Nodes{gradOutputs[i]}
}
// "active" nodes are the ones that are differentially influenced by the inputs
// and also differentiably influence the outputs. These are the nodes where we need to call the
// "pullback" function to backpropagate derivatives
activeNodes := affectsOutput.Intersect(affectedByOutput)
symdiffLogf("Active: %v", activeNodes)
symdiffLogf("Sorted: %d", sortedNodes)
symdiffLogf("nodeGradMap: %+#d", FmtNodeMap(nodeGradMap))
enterLogScope()
for _, node := range sortedNodes {
if _, ok := activeNodes[node]; !ok {
symdiffLogf("skipping %x", node.ID())
continue
}
if node.deriv != nil {
symdiffLogf("skipping %x - previously differentiated", node.ID())
nodeGradMap[node] = append(nodeGradMap[node], node.deriv)
continue
}
symdiffLogf("Working on %x %v", node.ID(), node)
enterLogScope()
// Check if there is any grads coming into this node
if len(nodeGradMap[node]) < 1 {
leaveLogScope()
return nil, SymDiffError{
single: node,
gradMap: nodeGradMap,
err: errors.New("No gradients found for node"),
}
}
// once we've reached a node, we already backpropagated from its dependents
// so we sum up the gradients
symdiffLogf("nodeGradMap[%x]: %d", node.ID(), nodeGradMap[node])
if len(nodeGradMap[node]) > 1 {
var n *Node
symdiffLogf("reduce adding")
if n, err = ReduceAdd(nodeGradMap[node], WithGroupName(gradClust)); err != nil { | single: node,
nodes: nodeGradMap[node],
gradMap: nodeGradMap,
err: errors.Wrap(err, "ReduceAdd failed during differentiation"),
}
}
symdiffLogf("reduced to... %x", n.ID())
// node.derives = append(node.derives, n)
n.derivOf = append(n.derivOf, node)
node.deriv = n
nodeGradMap[node] = Nodes{n}
// }
} else if len(nodeGradMap[node]) == 1 {
deriv := nodeGradMap[node][0]
deriv.derivOf = | leaveLogScope()
return nil, SymDiffError{ | random_line_split |
differentiation.go | (outputs, sortedNodes Nodes) (retVal NodeSet, err error) {
symdiffLogf("Forward analysis. Already sorted?")
enterLogScope()
defer leaveLogScope()
if !outputs.AllSameGraph() {
return nil, errors.New("The supplied output Nodes are not the same graph")
}
diffSet := outputs.mapSet()
symdiffLogf("Diff Set: %v", diffSet)
symdiffLogf("%d", sortedNodes)
for _, n := range sortedNodes {
if diffSet.Contains(n) && !n.isInput() {
diffs := n.diffWRT()
for j, child := range n.children {
d := diffs[j]
if d {
symdiffLogf("Adding %x to differentiable set", child.ID())
diffSet.Add(child)
}
}
}
}
return diffSet, nil
}
// backwardDiffAnalysis returns a list of Nodes that are affected by differentiating output.
// Given a list of WRTs, we want to find a list of nodes that will be affected when backpropagating.
func backwardDiffAnalysis(wrt, sortedNodes Nodes) (retVal NodeSet, err error) {
symdiffLogf("Backwards analysis")
enterLogScope()
defer leaveLogScope()
if !wrt.AllSameGraph() {
return nil, errors.New("The supplied output Nodes are not the same graph")
}
diffSet := wrt.mapSet()
symdiffLogf("wrt:%d diffset: %d", len(wrt), len(diffSet))
symdiffLogf("%v", diffSet)
symdiffLogf("sorted: %d", sortedNodes)
enterLogScope()
for i := len(sortedNodes) - 1; i >= 0; i-- {
n := sortedNodes[i]
symdiffLogf("working on %v. Has %d children", n, len(n.children))
var op SDOp
var ok bool
var diffs []bool
if op, ok = n.op.(SDOp); ok {
diffs = op.DiffWRT(len(n.children))
}
symdiffLogf("differentiable WRT: %v", diffs)
enterLogScope()
symdiffLogf("Children: %v", n.children)
if len(diffs) == 0 {
// check if this makes nodes unreachable. If it does, then error out
if n.isStmt {
symdiffLogf("Statement nodes are Non differentiable!")
leaveLogScope()
continue
} else if n.isInput() {
symdiffLogf("Input nodes are Non differentiable")
leaveLogScope()
continue
} else if len(n.children) == 0 {
symdiffLogf("Leaf nodes have no children")
leaveLogScope()
continue
}
g := n.g
for _, child := range n.children {
parents := graph.NodesOf(g.To(child.ID()))
if len(parents) == 1 && len(child.children) > 0 {
leaveLogScope()
return nil, errors.Errorf("Being unable to differentiate %v would leave a portion of the graph unreachable. Unable to continue", n)
}
}
symdiffLogf("SKIPPING... Non differentiable!")
leaveLogScope()
continue
}
inner:
for j, child := range n.children {
d := diffs[j]
if diffSet.Contains(child) && d {
symdiffLogf("Adding %x to differentiable set", child.ID())
diffSet.Add(n)
break inner
}
}
leaveLogScope()
}
leaveLogScope()
return diffSet, nil
}
// Backpropagate backpropagates errors by performing reverse-mode symbolic differentiation, starting from the outputs, and working its way towads the inputs.
//
// This is the rough algorithm:
// 1. Filter out nodes that are unreachable
// 2. Forwards analysis, where a list of nodes affecting the output is added to consideration
// 3. Backwards analysis, where a list of nodes affected by differentiating the output are added to the consideration
// 4. If there is a difference in both sets, it will cause an error (both sets should be the same)
// 5. Traverse the graph from output towards input. On each visit, perform the symbolic differentiation
//
// For most cases, Grad() should be used instead of Backpropagate(), as Grad() performs several checks which would be the general use case, before calling Backpropagate()
func Backpropagate(outputs, gradOutputs, wrt Nodes) (retVal Nodes, err error) {
symdiffLogf("BACKPROP START")
symdiffLogf("Outputs: %d", outputs)
symdiffLogf("gradOutputs: %d", gradOutputs)
symdiffLogf("WRT: %d", wrt)
enterLogScope()
defer leaveLogScope()
g := outputs[0].g
// this entire section about removing foreveralone nodes need a rethink
symdiffLogf("removing foreveralone nodes")
enterLogScope()
for i := 0; i < len(g.AllNodes()); i++ {
n := g.AllNodes()[i]
fr := g.From(n.ID()).Len()
to := g.To(n.ID()).Len()
if fr == 0 && to == 0 && !n.isConstant() && !n.isInput() {
g.RemoveNode(n)
symdiffLogf("removed %v(%p); %x; %s", n, n, n.ID(), n.Name())
}
}
leaveLogScope()
var sortedNodes Nodes
if sortedNodes, err = Sort(g); err != nil {
return nil, errors.Wrap(err, sortFail)
}
symdiffLogf("sorted nodes: %v", sortedNodes)
symdiffLogf("sorted nodes: %d", sortedNodes)
var affectsOutput NodeSet
var affectedByOutput NodeSet
if affectsOutput, err = forwardDiffAnalysis(outputs, sortedNodes); err != nil {
return nil, errors.Wrap(err, "Failed during forward differentiation analysis")
}
if affectedByOutput, err = backwardDiffAnalysis(wrt, sortedNodes); err != nil {
return nil, errors.Wrap(err, "Failed during forward differentiation analysis")
}
symdiffLogf("affects output: %v", affectsOutput)
symdiffLogf("affected by output : %v", affectedByOutput)
wrtSet := wrt.mapSet()
badWRTs := wrtSet.Difference(affectsOutput)
if len(badWRTs) > 0 {
return nil, SymDiffError{nodes: badWRTs.ToSlice(), err: errors.Errorf("Non Differentiable WRTs: %v", badWRTs)}
}
outputSet := outputs.mapSet()
badOutputs := outputSet.Difference(affectedByOutput)
if len(badOutputs) > 0 {
symdiffLogf("badOutputs: %#v", badOutputs)
return nil, SymDiffError{nodes: badOutputs.ToSlice(), err: errors.Errorf("Non-Differentable Outputs: %v", badOutputs)}
}
// map a node to a list of gradient terms
// these gradient terms will be summed up when we visit the node
// when iterating through the nondes in reverse topological order
nodeGradMap := make(map[*Node]Nodes)
for i, n := range outputs {
symdiffLogf("Adding outputs for %x", n.ID())
nodeGradMap[n] = Nodes{gradOutputs[i]}
}
// "active" nodes are the ones that are differentially influenced by the inputs
// and also differentiably influence the outputs. These are the nodes where we need to call the
// "pullback" function to backpropagate derivatives
activeNodes := affectsOutput.Intersect(affectedByOutput)
symdiffLogf("Active: %v", activeNodes)
symdiffLogf("Sorted: %d", sortedNodes)
symdiffLogf("nodeGradMap: %+#d", FmtNodeMap(nodeGradMap))
enterLogScope()
for _, node := range sortedNodes {
if _, ok := activeNodes[node]; !ok {
symdiffLogf("skipping %x", node.ID())
continue
}
if node.deriv != nil {
symdiffLogf("skipping %x - previously differentiated", node.ID())
nodeGradMap[node] = append(nodeGradMap[node], node.deriv)
continue
}
symdiffLogf("Working on %x %v", node.ID(), node)
enterLogScope()
// Check if there is any grads coming into this node
if len(nodeGradMap[node]) < 1 {
leaveLogScope()
return nil, SymDiffError{
single: node,
gradMap: nodeGradMap,
err: errors.New("No gradients found for node"),
}
}
// once we've reached a node, we already backpropagated from its dependents
// so we sum up the gradients
symdiffLogf("nodeGradMap[%x]: %d", node.ID(), nodeGradMap[node])
if len(nodeGradMap[node]) > | forwardDiffAnalysis | identifier_name |
|
TypedAction.ts | .g. to a Redux `Store.dispatch`.
*
* The true benefit of TypedActions come on the Reducer-side. See
* the TypedReducer class for more on creating a TypedAction-savvy Reducer for Redux.
*
* Conforms to Flux Standard Action recommendations.
*
* @see TypedActionDef#create
*/
export interface TypedAction<T, E extends string = string> {
/**
* The type string of the action, used to uniquely identify the Action with its Definition.
*
* The TypeScript typing of this value is refined to the actual string given to `TypedAction.define()`.
*/
type: E;
/**
* The payload associated with the action, whose shape is dictated by the Definition
* that generated this action. As allowed by the payload restrictions chosen by the consumer,
* this payload could possibly be `undefined` or `null`.
*
* N.B. A NoPayloadDefinition doesn't actually define this key for generated actions, so such
* actions are usually just `{type: "..."}`.
*/
payload: T;
/**
* Optional metadata assigned to this action, which has no restrictions.
* Interesting usages of metadata:
*
* 1. To add a timestamp for when the action was first created.
* 1. To correlate a set of actions behind a single user event (such as a clickId).
* 1. To track progression of an async task, such as a {loading => success} or {loading => error} set.
* 1. To identify which actions are being triggered by a continually running job.
*
*/
meta?: any;
}
export namespace TypedAction {
/**
* **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()`
* over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35
*
* Options to TypedAction.define().
*
* @deprecated
*/
export interface DefineOptions<T> {
/**
* A function used to validate the (runtime) correctness of payloads attached to a Definition's
* actions. This can be useful to track down a noncompliant _source_ of actions,
* as otherwise debugging where incorrect actions are coming from on the Reducer side can be challenging.
*
* Not run in production.
*/
validate?: (payload: T) => boolean;
}
/**
* **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()`
* over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35
*
* One of the core functions of Redoodle, `TypedAction.define` creates a Definition
* to manage all Redux actions of a specific type string, such as `"myapp::set_foo_value"`.
*
* Each Definition also associates a payload type `T` for all of its matching actions.
* For example, the `"myapp::set_foo_value"` Action can associate a required payload shape
* `{foo: string, value: number}`, which means that all actions in the application
* with type `"myapp::set_foo_value"` *must* have payloads with a `foo` and a `value`.
*
* The syntax for invoking the function is slightly awkward, in favor of more predictable type inferencing.
* An example invocation is below; note the extra `()` after the payload type declaration in `<{}>`s.
*
*
* ```
* export const SetFooValueAction = TypedAction.define("myapp::set_foo_value")<{
* foo: string;
* value: number;
* }>();
* ```
*
*
* All Definitions for a Redux-enabled application MUST have unique strings.
*
* @deprecated
*/
export function define<E extends string>(
type: E,
): <T>(options?: DefineOptions<T>) => Definition<E, T> {
return <T>(options?: DefineOptions<T>) => {
if (
process.env.NODE_ENV !== "production" &&
options !== undefined &&
options.validate !== undefined
) {
return createDefinitionWithValidator<E, T>(type, options.validate);
} else {
return createDefinition<E, T>(type);
}
};
}
/**
* **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()`
* over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35
*
* Similar to TypedAction.define, creates a NoPayloadDefinition for the given Action type
* string, like `"example::clear_foo"`. In practice, actions without payloads are
* usually of the "clear" or "invalidate" variety.
*
* The syntax for invoking the function is slightly awkward, in favor of more predictable type inferencing.
* An example invocation is below; note the extra pair of `()`, for consistency with its sibling `define`
* function and for better future-compatibility of options.
*
*
* ```
* export const SetFooValueAction = TypedAction.defineWithoutPayload("myapp::set_foo_value")();
* ```
*
*
* All Definitions for a Redux-enabled application MUST have unique strings.
*
* @deprecated
*/
export function defineWithoutPayload<E extends string>(
type: E,
): () => NoPayloadDefinition<E> {
return () => {
return createNoPayloadDefinition<E>(type);
};
}
/**
* A central type of Redoodle, the TypedAction.Definition manages all Redux Actions
* of a specific type string, such as `"myapp::set_foo_value"`.
*
* - Definitions should be used to create Actions.
* - Definitions can be used to identify an Action, based on its own `type`.
*
* All Definitions for a Redux-enabled application MUST have unique strings.
*/
export interface Definition<E extends string, T> {
/**
* Creates an Action of this type with the given payload.
* Functionally equivalent to the explicit Definition.create().
*/
(payload: T): { type: E; payload: T };
/**
* The Type of a TypedAction refers to the physical `{type}` string
* given to matching Actions. This TypedActionString is branded
* with the payload type as well for e.g. TypedReducer type inferencing.
*/
TYPE: TypedActionString<T, E>;
/**
* Hidden field used for some workflows that need to extract the payload type back out of
* a TypedAction definition. For example, `const payload: typeof MyAction.__PAYLOAD = { ... };`
* can be used to define a payload conforming to MyAction.
*
* This value should only be used for constructing Types in TypeScript. It never holds a real value.
* Future versions of Redoodle may throw when attempting accessing this value at runtime
* to catch accidental misuse.
*/
__PAYLOAD: T;
/**
* Creates an Action of this type with the given payload.
*/
create(payload: T): { type: E; payload: T };
/**
* Creates an Action of this type with the given payload and meta.
*/
createWithMeta<M>(payload: T, meta: M): { type: E; payload: T; meta: M };
/**
* Checks whether the given Action matches this Definition, based on its own `type`.
* If so, we can safely narrow the Action's payload type based on this Definition.
*
* While this function can be used for action identification while Reducing,
* TypedReducers provide much stronger utilities when working with TypedActions.
*/
is(action: Action): action is TypedAction<T, E>;
}
/**
* @deprecated use top level PayloadOf
*/
export type PayloadOf<
D extends Definition<any, any>
> = D["TYPE"]["__type__"]["withPayload"];
/**
* A TypedAction.NoPayloadDefinition manages all Redux actions of a specific type string,
* such as `"myapp::clear_foo"`. Unlike the sibling TypedAction.Definition,
* actions matching this Definition are associated with no payload data. In practice,
* actions without payloads are usually of the "clear" or "invalidate" variety. | * All Definitions for a Redux-enabled application MUST have unique strings.
*/
export interface NoPayloadDefinition<E extends string> {
/**
* Creates an Action of this type (and no payload).
* Functionally equivalent to the explicit NoPayloadDefinition.create().
*/
(): { type: E; payload: never };
/**
* The Type of a TypedAction refers to the physical `{type}` string
* given to matching Actions. This TypedActionString is branded
* with the payload type as well for e.g. TypedReducer type inferencing.
*/
TYPE: TypedActionString<never, E>;
/**
* Creates an Action of this type (and no payload).
*/
create(): { type: E; payload: never };
/**
* Creates an Action of this type with the given meta (and no payload).
*/
createWithMeta<M>(meta: | *
* - Definitions should be used to create Actions.
* - Definitions can be used to identify an Action, based on its own `type`.
* | random_line_split |
TypedAction.ts | **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()`
* over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35
*
* Options to TypedAction.define().
*
* @deprecated
*/
export interface DefineOptions<T> {
/**
* A function used to validate the (runtime) correctness of payloads attached to a Definition's
* actions. This can be useful to track down a noncompliant _source_ of actions,
* as otherwise debugging where incorrect actions are coming from on the Reducer side can be challenging.
*
* Not run in production.
*/
validate?: (payload: T) => boolean;
}
/**
* **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()`
* over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35
*
* One of the core functions of Redoodle, `TypedAction.define` creates a Definition
* to manage all Redux actions of a specific type string, such as `"myapp::set_foo_value"`.
*
* Each Definition also associates a payload type `T` for all of its matching actions.
* For example, the `"myapp::set_foo_value"` Action can associate a required payload shape
* `{foo: string, value: number}`, which means that all actions in the application
* with type `"myapp::set_foo_value"` *must* have payloads with a `foo` and a `value`.
*
* The syntax for invoking the function is slightly awkward, in favor of more predictable type inferencing.
* An example invocation is below; note the extra `()` after the payload type declaration in `<{}>`s.
*
*
* ```
* export const SetFooValueAction = TypedAction.define("myapp::set_foo_value")<{
* foo: string;
* value: number;
* }>();
* ```
*
*
* All Definitions for a Redux-enabled application MUST have unique strings.
*
* @deprecated
*/
export function define<E extends string>(
type: E,
): <T>(options?: DefineOptions<T>) => Definition<E, T> {
return <T>(options?: DefineOptions<T>) => {
if (
process.env.NODE_ENV !== "production" &&
options !== undefined &&
options.validate !== undefined
) {
return createDefinitionWithValidator<E, T>(type, options.validate);
} else {
return createDefinition<E, T>(type);
}
};
}
/**
* **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()`
* over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35
*
* Similar to TypedAction.define, creates a NoPayloadDefinition for the given Action type
* string, like `"example::clear_foo"`. In practice, actions without payloads are
* usually of the "clear" or "invalidate" variety.
*
* The syntax for invoking the function is slightly awkward, in favor of more predictable type inferencing.
* An example invocation is below; note the extra pair of `()`, for consistency with its sibling `define`
* function and for better future-compatibility of options.
*
*
* ```
* export const SetFooValueAction = TypedAction.defineWithoutPayload("myapp::set_foo_value")();
* ```
*
*
* All Definitions for a Redux-enabled application MUST have unique strings.
*
* @deprecated
*/
export function defineWithoutPayload<E extends string>(
type: E,
): () => NoPayloadDefinition<E> {
return () => {
return createNoPayloadDefinition<E>(type);
};
}
/**
* A central type of Redoodle, the TypedAction.Definition manages all Redux Actions
* of a specific type string, such as `"myapp::set_foo_value"`.
*
* - Definitions should be used to create Actions.
* - Definitions can be used to identify an Action, based on its own `type`.
*
* All Definitions for a Redux-enabled application MUST have unique strings.
*/
export interface Definition<E extends string, T> {
/**
* Creates an Action of this type with the given payload.
* Functionally equivalent to the explicit Definition.create().
*/
(payload: T): { type: E; payload: T };
/**
* The Type of a TypedAction refers to the physical `{type}` string
* given to matching Actions. This TypedActionString is branded
* with the payload type as well for e.g. TypedReducer type inferencing.
*/
TYPE: TypedActionString<T, E>;
/**
* Hidden field used for some workflows that need to extract the payload type back out of
* a TypedAction definition. For example, `const payload: typeof MyAction.__PAYLOAD = { ... };`
* can be used to define a payload conforming to MyAction.
*
* This value should only be used for constructing Types in TypeScript. It never holds a real value.
* Future versions of Redoodle may throw when attempting accessing this value at runtime
* to catch accidental misuse.
*/
__PAYLOAD: T;
/**
* Creates an Action of this type with the given payload.
*/
create(payload: T): { type: E; payload: T };
/**
* Creates an Action of this type with the given payload and meta.
*/
createWithMeta<M>(payload: T, meta: M): { type: E; payload: T; meta: M };
/**
* Checks whether the given Action matches this Definition, based on its own `type`.
* If so, we can safely narrow the Action's payload type based on this Definition.
*
* While this function can be used for action identification while Reducing,
* TypedReducers provide much stronger utilities when working with TypedActions.
*/
is(action: Action): action is TypedAction<T, E>;
}
/**
* @deprecated use top level PayloadOf
*/
export type PayloadOf<
D extends Definition<any, any>
> = D["TYPE"]["__type__"]["withPayload"];
/**
* A TypedAction.NoPayloadDefinition manages all Redux actions of a specific type string,
* such as `"myapp::clear_foo"`. Unlike the sibling TypedAction.Definition,
* actions matching this Definition are associated with no payload data. In practice,
* actions without payloads are usually of the "clear" or "invalidate" variety.
*
* - Definitions should be used to create Actions.
* - Definitions can be used to identify an Action, based on its own `type`.
*
* All Definitions for a Redux-enabled application MUST have unique strings.
*/
export interface NoPayloadDefinition<E extends string> {
/**
* Creates an Action of this type (and no payload).
* Functionally equivalent to the explicit NoPayloadDefinition.create().
*/
(): { type: E; payload: never };
/**
* The Type of a TypedAction refers to the physical `{type}` string
* given to matching Actions. This TypedActionString is branded
* with the payload type as well for e.g. TypedReducer type inferencing.
*/
TYPE: TypedActionString<never, E>;
/**
* Creates an Action of this type (and no payload).
*/
create(): { type: E; payload: never };
/**
* Creates an Action of this type with the given meta (and no payload).
*/
createWithMeta<M>(meta: M): { type: E; payload: never; meta: M };
/**
* Checks whether the given Action matches this Definition, based on its own `type`.
* If so, we can safely narrow the Action's payload type based on this Definition.
*
* While this function can be used for action identification while Reducing,
* TypedReducers provide much stronger utilities when working with TypedActions.
*/
is(action: Action): action is TypedAction<never, E>;
}
function createDefinition<E extends string, T>(type: E): Definition<E, T> {
const create = (payload: T): { type: E; payload: T } => {
return { type, payload };
};
const createWithMeta = <M>(
payload: T,
meta: M,
): { type: E; payload: T; meta: M } => {
return { type, payload, meta };
};
const is = (action: Action): action is TypedAction<T, E> => {
return action.type === type;
};
const def = create as Definition<E, T>;
def.create = create;
def.createWithMeta = createWithMeta;
def.is = is;
def.TYPE = type as TypedActionString<T, E>;
return def;
}
function createDefinitionWithValidator<E extends string, T>(
type: E,
validate: (payload: T) => boolean,
): Definition<E, T> {
const create = (payload: T): { type: E; payload: T } => {
if (!validate(payload)) | {
throw new Error(`'${type}' validation failed`);
} | conditional_block |
|
TypedAction.ts | /**
* **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()`
* over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35
*
* One of the core functions of Redoodle, `TypedAction.define` creates a Definition
* to manage all Redux actions of a specific type string, such as `"myapp::set_foo_value"`.
*
* Each Definition also associates a payload type `T` for all of its matching actions.
* For example, the `"myapp::set_foo_value"` Action can associate a required payload shape
* `{foo: string, value: number}`, which means that all actions in the application
* with type `"myapp::set_foo_value"` *must* have payloads with a `foo` and a `value`.
*
* The syntax for invoking the function is slightly awkward, in favor of more predictable type inferencing.
* An example invocation is below; note the extra `()` after the payload type declaration in `<{}>`s.
*
*
* ```
* export const SetFooValueAction = TypedAction.define("myapp::set_foo_value")<{
* foo: string;
* value: number;
* }>();
* ```
*
*
* All Definitions for a Redux-enabled application MUST have unique strings.
*
* @deprecated
*/
export function define<E extends string>(
type: E,
): <T>(options?: DefineOptions<T>) => Definition<E, T> {
return <T>(options?: DefineOptions<T>) => {
if (
process.env.NODE_ENV !== "production" &&
options !== undefined &&
options.validate !== undefined
) {
return createDefinitionWithValidator<E, T>(type, options.validate);
} else {
return createDefinition<E, T>(type);
}
};
}
/**
* **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()`
* over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35
*
* Similar to TypedAction.define, creates a NoPayloadDefinition for the given Action type
* string, like `"example::clear_foo"`. In practice, actions without payloads are
* usually of the "clear" or "invalidate" variety.
*
* The syntax for invoking the function is slightly awkward, in favor of more predictable type inferencing.
* An example invocation is below; note the extra pair of `()`, for consistency with its sibling `define`
* function and for better future-compatibility of options.
*
*
* ```
* export const SetFooValueAction = TypedAction.defineWithoutPayload("myapp::set_foo_value")();
* ```
*
*
* All Definitions for a Redux-enabled application MUST have unique strings.
*
* @deprecated
*/
export function defineWithoutPayload<E extends string>(
type: E,
): () => NoPayloadDefinition<E> {
return () => {
return createNoPayloadDefinition<E>(type);
};
}
/**
* A central type of Redoodle, the TypedAction.Definition manages all Redux Actions
* of a specific type string, such as `"myapp::set_foo_value"`.
*
* - Definitions should be used to create Actions.
* - Definitions can be used to identify an Action, based on its own `type`.
*
* All Definitions for a Redux-enabled application MUST have unique strings.
*/
export interface Definition<E extends string, T> {
/**
* Creates an Action of this type with the given payload.
* Functionally equivalent to the explicit Definition.create().
*/
(payload: T): { type: E; payload: T };
/**
* The Type of a TypedAction refers to the physical `{type}` string
* given to matching Actions. This TypedActionString is branded
* with the payload type as well for e.g. TypedReducer type inferencing.
*/
TYPE: TypedActionString<T, E>;
/**
* Hidden field used for some workflows that need to extract the payload type back out of
* a TypedAction definition. For example, `const payload: typeof MyAction.__PAYLOAD = { ... };`
* can be used to define a payload conforming to MyAction.
*
* This value should only be used for constructing Types in TypeScript. It never holds a real value.
* Future versions of Redoodle may throw when attempting accessing this value at runtime
* to catch accidental misuse.
*/
__PAYLOAD: T;
/**
* Creates an Action of this type with the given payload.
*/
create(payload: T): { type: E; payload: T };
/**
* Creates an Action of this type with the given payload and meta.
*/
createWithMeta<M>(payload: T, meta: M): { type: E; payload: T; meta: M };
/**
* Checks whether the given Action matches this Definition, based on its own `type`.
* If so, we can safely narrow the Action's payload type based on this Definition.
*
* While this function can be used for action identification while Reducing,
* TypedReducers provide much stronger utilities when working with TypedActions.
*/
is(action: Action): action is TypedAction<T, E>;
}
/**
* @deprecated use top level PayloadOf
*/
export type PayloadOf<
D extends Definition<any, any>
> = D["TYPE"]["__type__"]["withPayload"];
/**
* A TypedAction.NoPayloadDefinition manages all Redux actions of a specific type string,
* such as `"myapp::clear_foo"`. Unlike the sibling TypedAction.Definition,
* actions matching this Definition are associated with no payload data. In practice,
* actions without payloads are usually of the "clear" or "invalidate" variety.
*
* - Definitions should be used to create Actions.
* - Definitions can be used to identify an Action, based on its own `type`.
*
* All Definitions for a Redux-enabled application MUST have unique strings.
*/
export interface NoPayloadDefinition<E extends string> {
/**
* Creates an Action of this type (and no payload).
* Functionally equivalent to the explicit NoPayloadDefinition.create().
*/
(): { type: E; payload: never };
/**
* The Type of a TypedAction refers to the physical `{type}` string
* given to matching Actions. This TypedActionString is branded
* with the payload type as well for e.g. TypedReducer type inferencing.
*/
TYPE: TypedActionString<never, E>;
/**
* Creates an Action of this type (and no payload).
*/
create(): { type: E; payload: never };
/**
* Creates an Action of this type with the given meta (and no payload).
*/
createWithMeta<M>(meta: M): { type: E; payload: never; meta: M };
/**
* Checks whether the given Action matches this Definition, based on its own `type`.
* If so, we can safely narrow the Action's payload type based on this Definition.
*
* While this function can be used for action identification while Reducing,
* TypedReducers provide much stronger utilities when working with TypedActions.
*/
is(action: Action): action is TypedAction<never, E>;
}
function createDefinition<E extends string, T>(type: E): Definition<E, T> {
const create = (payload: T): { type: E; payload: T } => {
return { type, payload };
};
const createWithMeta = <M>(
payload: T,
meta: M,
): { type: E; payload: T; meta: M } => {
return { type, payload, meta };
};
const is = (action: Action): action is TypedAction<T, E> => {
return action.type === type;
};
const def = create as Definition<E, T>;
def.create = create;
def.createWithMeta = createWithMeta;
def.is = is;
def.TYPE = type as TypedActionString<T, E>;
return def;
}
function createDefinitionWithValidator<E extends string, T>(
type: E,
validate: (payload: T) => boolean,
): Definition<E, T> {
const create = (payload: T): { type: E; payload: T } => {
if (!validate(payload)) {
throw new Error(`'${type}' validation failed`);
}
return { type, payload };
};
const createWithMeta = <M>(
payload: T,
meta: M,
): { type: E; payload: T; meta: M } => {
if (!validate(payload)) {
throw new Error(`'${type}' validation failed`);
}
return { type, payload, meta };
};
const is = (action: Action): action is TypedAction<T, E> => {
return action.type === type;
};
const def = create as Definition<E, T>;
def.create = create;
def.createWithMeta = createWithMeta;
def.is = is;
def.TYPE = type as TypedActionString<T, E>;
return def;
}
function | createNoPayloadDefinition | identifier_name |
|
TypedAction.ts | . to a Redux `Store.dispatch`.
*
* The true benefit of TypedActions come on the Reducer-side. See
* the TypedReducer class for more on creating a TypedAction-savvy Reducer for Redux.
*
* Conforms to Flux Standard Action recommendations.
*
* @see TypedActionDef#create
*/
export interface TypedAction<T, E extends string = string> {
/**
* The type string of the action, used to uniquely identify the Action with its Definition.
*
* The TypeScript typing of this value is refined to the actual string given to `TypedAction.define()`.
*/
type: E;
/**
* The payload associated with the action, whose shape is dictated by the Definition
* that generated this action. As allowed by the payload restrictions chosen by the consumer,
* this payload could possibly be `undefined` or `null`.
*
* N.B. A NoPayloadDefinition doesn't actually define this key for generated actions, so such
* actions are usually just `{type: "..."}`.
*/
payload: T;
/**
* Optional metadata assigned to this action, which has no restrictions.
* Interesting usages of metadata:
*
* 1. To add a timestamp for when the action was first created.
* 1. To correlate a set of actions behind a single user event (such as a clickId).
* 1. To track progression of an async task, such as a {loading => success} or {loading => error} set.
* 1. To identify which actions are being triggered by a continually running job.
*
*/
meta?: any;
}
export namespace TypedAction {
/**
* **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()`
* over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35
*
* Options to TypedAction.define().
*
* @deprecated
*/
export interface DefineOptions<T> {
/**
* A function used to validate the (runtime) correctness of payloads attached to a Definition's
* actions. This can be useful to track down a noncompliant _source_ of actions,
* as otherwise debugging where incorrect actions are coming from on the Reducer side can be challenging.
*
* Not run in production.
*/
validate?: (payload: T) => boolean;
}
/**
* **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()`
* over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35
*
* One of the core functions of Redoodle, `TypedAction.define` creates a Definition
* to manage all Redux actions of a specific type string, such as `"myapp::set_foo_value"`.
*
* Each Definition also associates a payload type `T` for all of its matching actions.
* For example, the `"myapp::set_foo_value"` Action can associate a required payload shape
* `{foo: string, value: number}`, which means that all actions in the application
* with type `"myapp::set_foo_value"` *must* have payloads with a `foo` and a `value`.
*
* The syntax for invoking the function is slightly awkward, in favor of more predictable type inferencing.
* An example invocation is below; note the extra `()` after the payload type declaration in `<{}>`s.
*
*
* ```
* export const SetFooValueAction = TypedAction.define("myapp::set_foo_value")<{
* foo: string;
* value: number;
* }>();
* ```
*
*
* All Definitions for a Redux-enabled application MUST have unique strings.
*
* @deprecated
*/
export function define<E extends string>(
type: E,
): <T>(options?: DefineOptions<T>) => Definition<E, T> |
/**
* **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()`
* over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35
*
* Similar to TypedAction.define, creates a NoPayloadDefinition for the given Action type
* string, like `"example::clear_foo"`. In practice, actions without payloads are
* usually of the "clear" or "invalidate" variety.
*
* The syntax for invoking the function is slightly awkward, in favor of more predictable type inferencing.
* An example invocation is below; note the extra pair of `()`, for consistency with its sibling `define`
* function and for better future-compatibility of options.
*
*
* ```
* export const SetFooValueAction = TypedAction.defineWithoutPayload("myapp::set_foo_value")();
* ```
*
*
* All Definitions for a Redux-enabled application MUST have unique strings.
*
* @deprecated
*/
export function defineWithoutPayload<E extends string>(
type: E,
): () => NoPayloadDefinition<E> {
return () => {
return createNoPayloadDefinition<E>(type);
};
}
/**
* A central type of Redoodle, the TypedAction.Definition manages all Redux Actions
* of a specific type string, such as `"myapp::set_foo_value"`.
*
* - Definitions should be used to create Actions.
* - Definitions can be used to identify an Action, based on its own `type`.
*
* All Definitions for a Redux-enabled application MUST have unique strings.
*/
export interface Definition<E extends string, T> {
/**
* Creates an Action of this type with the given payload.
* Functionally equivalent to the explicit Definition.create().
*/
(payload: T): { type: E; payload: T };
/**
* The Type of a TypedAction refers to the physical `{type}` string
* given to matching Actions. This TypedActionString is branded
* with the payload type as well for e.g. TypedReducer type inferencing.
*/
TYPE: TypedActionString<T, E>;
/**
* Hidden field used for some workflows that need to extract the payload type back out of
* a TypedAction definition. For example, `const payload: typeof MyAction.__PAYLOAD = { ... };`
* can be used to define a payload conforming to MyAction.
*
* This value should only be used for constructing Types in TypeScript. It never holds a real value.
* Future versions of Redoodle may throw when attempting accessing this value at runtime
* to catch accidental misuse.
*/
__PAYLOAD: T;
/**
* Creates an Action of this type with the given payload.
*/
create(payload: T): { type: E; payload: T };
/**
* Creates an Action of this type with the given payload and meta.
*/
createWithMeta<M>(payload: T, meta: M): { type: E; payload: T; meta: M };
/**
* Checks whether the given Action matches this Definition, based on its own `type`.
* If so, we can safely narrow the Action's payload type based on this Definition.
*
* While this function can be used for action identification while Reducing,
* TypedReducers provide much stronger utilities when working with TypedActions.
*/
is(action: Action): action is TypedAction<T, E>;
}
/**
* @deprecated use top level PayloadOf
*/
export type PayloadOf<
D extends Definition<any, any>
> = D["TYPE"]["__type__"]["withPayload"];
/**
* A TypedAction.NoPayloadDefinition manages all Redux actions of a specific type string,
* such as `"myapp::clear_foo"`. Unlike the sibling TypedAction.Definition,
* actions matching this Definition are associated with no payload data. In practice,
* actions without payloads are usually of the "clear" or "invalidate" variety.
*
* - Definitions should be used to create Actions.
* - Definitions can be used to identify an Action, based on its own `type`.
*
* All Definitions for a Redux-enabled application MUST have unique strings.
*/
export interface NoPayloadDefinition<E extends string> {
/**
* Creates an Action of this type (and no payload).
* Functionally equivalent to the explicit NoPayloadDefinition.create().
*/
(): { type: E; payload: never };
/**
* The Type of a TypedAction refers to the physical `{type}` string
* given to matching Actions. This TypedActionString is branded
* with the payload type as well for e.g. TypedReducer type inferencing.
*/
TYPE: TypedActionString<never, E>;
/**
* Creates an Action of this type (and no payload).
*/
create(): { type: E; payload: never };
/**
* Creates an Action of this type with the given meta (and no payload).
*/
createWithMeta<M>(meta | {
return <T>(options?: DefineOptions<T>) => {
if (
process.env.NODE_ENV !== "production" &&
options !== undefined &&
options.validate !== undefined
) {
return createDefinitionWithValidator<E, T>(type, options.validate);
} else {
return createDefinition<E, T>(type);
}
};
} | identifier_body |
time-sentiment.component.ts | renderedChart = false;
notDataWarn = false;
appliedFilter = false;
chartHeight = 300;
chartRange1;
chartRange2;
chartRangeFilter1;
chartRangeFilter2;
constructor(private chartService: ChartService, private _element: ElementRef) { }
ngOnInit() {
this.sentimentLineChart = dc.lineChart('#sentimentChartLine');
this.chartService.GetData().subscribe((data) => { this.data = data; });
// Crossfilter
this.chartService.getCrossfilter().subscribe((filter) => {
this.cfilter = filter;
this.setDimension();
if (this.data && this.data.length > 0) {
this.sentGroups = this.getSentGroups();
if (this.sentGroups[0]) {
this.notDataWarn = false;
this.appliedFilter = false;
this.renderChart();
} else {
this.notDataWarn = true;
}
}
});
// Collapsible view
this.chartService.GetChartMode().subscribe(mode => {
if (this.data && this.data.length > 0) {
if (mode && mode === 'small') {
this.chartHeight = 85;
this.renderChart();
} else if (mode && mode === 'big') {
this.chartHeight = 300;
this.renderChart();
}
}
});
// Gets the range through the chart service from the mainVis Component
this.chartService.getChartRange().subscribe((range) => {
if (range.chart === null) {
if (this.data && range.range) {
(this.diff_months(range.range[0], range.range[1]) < 2) ? this.notDataWarn = true : this.notDataWarn = false;
this.chartRangeFilter1 = range.range[0];
this.chartRangeFilter2 = range.range[1];
this.sentimentLineChart
.x(d3.scaleTime().domain([this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2]))
.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2)]))
.round(d3.timeMonth);
this.appliedFilter = true;
this.sentimentLineChart.redraw();
} else {
if (!dc.chartRegistry.list().some((c) => c.hasFilter())) {
this.notDataWarn = false;
this.sentimentLineChart
.x(d3.scaleTime().domain([this.chartRange1, this.chartRange2]))
.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)]));
this.appliedFilter = false;
}
}
}
});
this.renderedChart = false;
}
// Buttons and Front-End ////////////////////////////////////////////////////////////////////////////////////////////
// sets the tooltip on mouseover
setTooltipInfo(event: MouseEvent, tooltip: HTMLSpanElement) {
tooltip.style.position = 'fixed';
tooltip.style.top = (event.clientY) + 'px';
tooltip.style.left = (event.clientX - tooltip.offsetWidth - 5) + 'px';
}
// sets the crossfilter dimension
setDimension() {
this.dimension = this.cfilter.dimension((d: any) => {
const splitted = d.publishedAt.split('-');
return new Date(splitted[0] + '-' + splitted[1]);
});
}
// Time-based Stacked Chart /////////////////////////////////////////////////////////////////////////////////////////
// returns a crossfilter-group for each sentiment x
private getSentGroups(): { group: CrossFilter.Group<{}, Date, any>, sent: string}[] {
if (this.data && this.data.length < 0) { return; }
const groups: { group: CrossFilter.Group<{}, Date, any>, sent: string}[] = [];
// group by sentiment
const nested = d3.nest()
.key((d: any) => {
if (d.analysis) {
if (d.analysis.mainSentiment === 'mixed') {
return 'Mix';
} else if (d.analysis.mainSentiment === 'positive') {
return 'Pos';
} else if (d.analysis.mainSentiment === 'neutral') {
return 'Neu';
} else if (d.analysis.mainSentiment === 'negative') {
return 'Neg';
} else if (d.analysis.mainSentiment === 'na') {
return 'N/A';
}
}
})
.entries(this.data);
nested.forEach((sentiment) => {
const g = this.dimension.group().reduceSum((d: any) => {
if (d.analysis) {
let mainsentiment = '';
if (d.analysis.mainSentiment === 'mixed') {
mainsentiment = 'Mix';
} else if (d.analysis.mainSentiment === 'positive') {
mainsentiment = 'Pos';
} else if (d.analysis.mainSentiment === 'neutral') {
mainsentiment = 'Neu';
} else if (d.analysis.mainSentiment === 'negative') {
mainsentiment = 'Neg';
} else if (d.analysis.mainSentiment === 'na') {
mainsentiment = 'N/A';
}
return mainsentiment === sentiment.key;
}
});
groups.push({group: g, sent: sentiment.key });
});
return groups;
}
getStartDate(previousDate) {
const date = new Date(previousDate);
date.setDate(1);
return date;
}
// Renders line chart (aggregation)
renderChart () {
this.chartRange1 = d3.min(this.data, (d: any) => new Date(d.publishedAt));
this.chartRange1 = this.getStartDate(this.chartRange1);
this.chartRange2 = d3.max(this.data, (d: any) => new Date(d.publishedAt));
const sentGroupsOrdered = this.reorderGroups();
const chartColors = this.defineChartColors();
let firstItem = 0;
while (!sentGroupsOrdered[firstItem] && firstItem < 5) {firstItem++; }
const group1 = sentGroupsOrdered[firstItem];
this.sentimentLineChart
.renderArea(true)
.width(900)
.height(this.chartHeight)
.ordinalColors(chartColors)
.useViewBoxResizing(true)
.dimension(this.dimension)
.x(d3.scaleTime().domain([this.chartRange1, this.chartRange2]))
.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)]))
.yAxisLabel('Comments')
.interpolate('monotone')
.legend(dc.legend().x(850).y(0).itemHeight(9).gap(5))
.brushOn(false)
.group(group1.group, group1.sent)
.valueAccessor(function (d) {
return d.value;
})
.xAxis().ticks(7);
let maxSent = 0;
if (sentGroupsOrdered.length > 1) {
sentGroupsOrdered.forEach((group) => {
if (group.group === group1.group || maxSent === 4) {
return;
}
// stacks the groups
this.sentimentLineChart
.stack(group.group, group.sent, function (d) {
return d.value;
});
maxSent++;
});
}
// When filter is applied before refreshing the chart
if (this.appliedFilter) {
this.sentimentLineChart.x(d3.scaleTime().domain([this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2]));
}
// Brush: get range and send it to the other charts on brush-filtering
this.sentimentLineChart.on('filtered', (chart, filter) => {
if (filter) {
this.sentimentLineChart.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(filter[0], filter[1])]));
} else {
this.sentimentLineChart.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)]));
}
this.chartService.setChartRange({range: filter, chart: chart});
});
// Adapt chart for smaller view
(this.chartHeight < 300) ? this.sentimentLineChart.yAxis().ticks(2) : this.sentimentLineChart.yAxis().ticks(10);
(this.chartHeight < 300) ? this.sentimentLineChart.xAxisLabel('') : this.sentimentLineChart.xAxisLabel('Date');
this.sentimentLineChart.xAxis().tickFormat(d3.timeFormat('%b %Y')); // month
this.sentimentLineChart.render();
}
// Adaptable color scale
defineChartColors() {
const colorArray = [];
const sentGroupsOrdered = this.reorderGroups();
sentGroupsOrdered.forEach((g) => {
if (g.sent === 'Pos') {
colorArray.push('#4daf4a');
} else if (g.sent === 'Neu') {
colorArray.push('#666666');
} else if (g.sent === 'Neg') {
colorArray.push('#ff7f00');
} else if (g.sent === 'Mix') {
colorArray.push('#984ea3');
} else if (g.sent === 'N/A') | {
colorArray.push('#DDDDDD');
} | conditional_block |
|
time-sentiment.component.ts | styleUrls: ['./time-sentiment.component.scss']
})
export class TimeSentimentComponent implements OnInit {
aggrView = true;
compView = false;
data: any[];
cfilter: CrossFilter.CrossFilter<{}>;
dimension: CrossFilter.Dimension<{}, Date>;
sentGroups: { group: CrossFilter.Group<{}, Date, any>, sent: string}[];
sentimentLineChart: dc.LineChart;
renderedChart = false;
notDataWarn = false;
appliedFilter = false;
chartHeight = 300;
chartRange1;
chartRange2;
chartRangeFilter1;
chartRangeFilter2;
constructor(private chartService: ChartService, private _element: ElementRef) { }
ngOnInit() {
this.sentimentLineChart = dc.lineChart('#sentimentChartLine');
this.chartService.GetData().subscribe((data) => { this.data = data; });
// Crossfilter
this.chartService.getCrossfilter().subscribe((filter) => {
this.cfilter = filter;
this.setDimension();
if (this.data && this.data.length > 0) {
this.sentGroups = this.getSentGroups();
if (this.sentGroups[0]) {
this.notDataWarn = false;
this.appliedFilter = false;
this.renderChart();
} else {
this.notDataWarn = true;
}
}
});
// Collapsible view
this.chartService.GetChartMode().subscribe(mode => {
if (this.data && this.data.length > 0) {
if (mode && mode === 'small') {
this.chartHeight = 85;
this.renderChart();
} else if (mode && mode === 'big') {
this.chartHeight = 300;
this.renderChart();
}
}
});
// Gets the range through the chart service from the mainVis Component
this.chartService.getChartRange().subscribe((range) => {
if (range.chart === null) {
if (this.data && range.range) {
(this.diff_months(range.range[0], range.range[1]) < 2) ? this.notDataWarn = true : this.notDataWarn = false;
this.chartRangeFilter1 = range.range[0];
this.chartRangeFilter2 = range.range[1];
this.sentimentLineChart
.x(d3.scaleTime().domain([this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2]))
.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2)]))
.round(d3.timeMonth);
this.appliedFilter = true;
this.sentimentLineChart.redraw();
} else {
if (!dc.chartRegistry.list().some((c) => c.hasFilter())) {
this.notDataWarn = false;
this.sentimentLineChart
.x(d3.scaleTime().domain([this.chartRange1, this.chartRange2]))
.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)]));
this.appliedFilter = false;
}
}
}
});
this.renderedChart = false;
}
// Buttons and Front-End ////////////////////////////////////////////////////////////////////////////////////////////
// sets the tooltip on mouseover
setTooltipInfo(event: MouseEvent, tooltip: HTMLSpanElement) {
tooltip.style.position = 'fixed';
tooltip.style.top = (event.clientY) + 'px';
tooltip.style.left = (event.clientX - tooltip.offsetWidth - 5) + 'px';
}
// sets the crossfilter dimension
setDimension() {
this.dimension = this.cfilter.dimension((d: any) => {
const splitted = d.publishedAt.split('-');
return new Date(splitted[0] + '-' + splitted[1]);
});
}
// Time-based Stacked Chart /////////////////////////////////////////////////////////////////////////////////////////
// returns a crossfilter-group for each sentiment x
private getSentGroups(): { group: CrossFilter.Group<{}, Date, any>, sent: string}[] {
if (this.data && this.data.length < 0) { return; }
const groups: { group: CrossFilter.Group<{}, Date, any>, sent: string}[] = [];
// group by sentiment
const nested = d3.nest()
.key((d: any) => {
if (d.analysis) {
if (d.analysis.mainSentiment === 'mixed') {
return 'Mix';
} else if (d.analysis.mainSentiment === 'positive') {
return 'Pos';
} else if (d.analysis.mainSentiment === 'neutral') {
return 'Neu';
} else if (d.analysis.mainSentiment === 'negative') {
return 'Neg';
} else if (d.analysis.mainSentiment === 'na') {
return 'N/A';
}
}
})
.entries(this.data);
nested.forEach((sentiment) => {
const g = this.dimension.group().reduceSum((d: any) => {
if (d.analysis) {
let mainsentiment = '';
if (d.analysis.mainSentiment === 'mixed') {
mainsentiment = 'Mix';
} else if (d.analysis.mainSentiment === 'positive') {
mainsentiment = 'Pos';
} else if (d.analysis.mainSentiment === 'neutral') {
mainsentiment = 'Neu';
} else if (d.analysis.mainSentiment === 'negative') {
mainsentiment = 'Neg';
} else if (d.analysis.mainSentiment === 'na') {
mainsentiment = 'N/A';
}
return mainsentiment === sentiment.key;
}
});
groups.push({group: g, sent: sentiment.key });
});
return groups;
}
getStartDate(previousDate) {
const date = new Date(previousDate);
date.setDate(1);
return date;
}
// Renders line chart (aggregation)
renderChart () {
this.chartRange1 = d3.min(this.data, (d: any) => new Date(d.publishedAt));
this.chartRange1 = this.getStartDate(this.chartRange1);
this.chartRange2 = d3.max(this.data, (d: any) => new Date(d.publishedAt));
const sentGroupsOrdered = this.reorderGroups();
const chartColors = this.defineChartColors();
let firstItem = 0;
while (!sentGroupsOrdered[firstItem] && firstItem < 5) {firstItem++; }
const group1 = sentGroupsOrdered[firstItem];
this.sentimentLineChart
.renderArea(true)
.width(900)
.height(this.chartHeight)
.ordinalColors(chartColors)
.useViewBoxResizing(true)
.dimension(this.dimension)
.x(d3.scaleTime().domain([this.chartRange1, this.chartRange2]))
.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)]))
.yAxisLabel('Comments')
.interpolate('monotone')
.legend(dc.legend().x(850).y(0).itemHeight(9).gap(5))
.brushOn(false)
.group(group1.group, group1.sent)
.valueAccessor(function (d) {
return d.value;
})
.xAxis().ticks(7);
let maxSent = 0;
if (sentGroupsOrdered.length > 1) {
sentGroupsOrdered.forEach((group) => {
if (group.group === group1.group || maxSent === 4) {
return;
}
// stacks the groups
this.sentimentLineChart
.stack(group.group, group.sent, function (d) {
return d.value;
});
maxSent++;
});
}
// When filter is applied before refreshing the chart
if (this.appliedFilter) {
this.sentimentLineChart.x(d3.scaleTime().domain([this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2]));
}
// Brush: get range and send it to the other charts on brush-filtering
this.sentimentLineChart.on('filtered', (chart, filter) => {
if (filter) {
this.sentimentLineChart.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(filter[0], filter[1])]));
} else {
this.sentimentLineChart.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)]));
}
this.chartService.setChartRange({range: filter, chart: chart});
});
// Adapt chart for smaller view
(this.chartHeight < 300) ? this.sentimentLineChart.yAxis().ticks(2) : this.sentimentLineChart.yAxis().ticks(10);
(this.chartHeight < 300) ? this.sentimentLineChart.xAxisLabel('') : this.sentimentLineChart.xAxisLabel('Date');
this.sentimentLineChart.xAxis().tickFormat(d3.timeFormat('%b %Y')); // month
this.sentimentLineChart.render();
}
// Adaptable color scale
defineChartColors() {
const colorArray = [];
const sentGroupsOrdered = this.reorderGroups();
sentGroupsOrdered.forEach((g) => {
if (g.sent === 'Pos') {
colorArray.push('#4daf4a');
} else if ( | random_line_split |
||
time-sentiment.component.ts | sentimentLineChart: dc.LineChart;
renderedChart = false;
notDataWarn = false;
appliedFilter = false;
chartHeight = 300;
chartRange1;
chartRange2;
chartRangeFilter1;
chartRangeFilter2;
constructor(private chartService: ChartService, private _element: ElementRef) { }
ngOnInit() | // Collapsible view
this.chartService.GetChartMode().subscribe(mode => {
if (this.data && this.data.length > 0) {
if (mode && mode === 'small') {
this.chartHeight = 85;
this.renderChart();
} else if (mode && mode === 'big') {
this.chartHeight = 300;
this.renderChart();
}
}
});
// Gets the range through the chart service from the mainVis Component
this.chartService.getChartRange().subscribe((range) => {
if (range.chart === null) {
if (this.data && range.range) {
(this.diff_months(range.range[0], range.range[1]) < 2) ? this.notDataWarn = true : this.notDataWarn = false;
this.chartRangeFilter1 = range.range[0];
this.chartRangeFilter2 = range.range[1];
this.sentimentLineChart
.x(d3.scaleTime().domain([this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2]))
.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2)]))
.round(d3.timeMonth);
this.appliedFilter = true;
this.sentimentLineChart.redraw();
} else {
if (!dc.chartRegistry.list().some((c) => c.hasFilter())) {
this.notDataWarn = false;
this.sentimentLineChart
.x(d3.scaleTime().domain([this.chartRange1, this.chartRange2]))
.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)]));
this.appliedFilter = false;
}
}
}
});
this.renderedChart = false;
}
// Buttons and Front-End ////////////////////////////////////////////////////////////////////////////////////////////
// sets the tooltip on mouseover
setTooltipInfo(event: MouseEvent, tooltip: HTMLSpanElement) {
tooltip.style.position = 'fixed';
tooltip.style.top = (event.clientY) + 'px';
tooltip.style.left = (event.clientX - tooltip.offsetWidth - 5) + 'px';
}
// sets the crossfilter dimension
setDimension() {
this.dimension = this.cfilter.dimension((d: any) => {
const splitted = d.publishedAt.split('-');
return new Date(splitted[0] + '-' + splitted[1]);
});
}
// Time-based Stacked Chart /////////////////////////////////////////////////////////////////////////////////////////
// returns a crossfilter-group for each sentiment x
private getSentGroups(): { group: CrossFilter.Group<{}, Date, any>, sent: string}[] {
if (this.data && this.data.length < 0) { return; }
const groups: { group: CrossFilter.Group<{}, Date, any>, sent: string}[] = [];
// group by sentiment
const nested = d3.nest()
.key((d: any) => {
if (d.analysis) {
if (d.analysis.mainSentiment === 'mixed') {
return 'Mix';
} else if (d.analysis.mainSentiment === 'positive') {
return 'Pos';
} else if (d.analysis.mainSentiment === 'neutral') {
return 'Neu';
} else if (d.analysis.mainSentiment === 'negative') {
return 'Neg';
} else if (d.analysis.mainSentiment === 'na') {
return 'N/A';
}
}
})
.entries(this.data);
nested.forEach((sentiment) => {
const g = this.dimension.group().reduceSum((d: any) => {
if (d.analysis) {
let mainsentiment = '';
if (d.analysis.mainSentiment === 'mixed') {
mainsentiment = 'Mix';
} else if (d.analysis.mainSentiment === 'positive') {
mainsentiment = 'Pos';
} else if (d.analysis.mainSentiment === 'neutral') {
mainsentiment = 'Neu';
} else if (d.analysis.mainSentiment === 'negative') {
mainsentiment = 'Neg';
} else if (d.analysis.mainSentiment === 'na') {
mainsentiment = 'N/A';
}
return mainsentiment === sentiment.key;
}
});
groups.push({group: g, sent: sentiment.key });
});
return groups;
}
getStartDate(previousDate) {
const date = new Date(previousDate);
date.setDate(1);
return date;
}
// Renders line chart (aggregation)
renderChart () {
this.chartRange1 = d3.min(this.data, (d: any) => new Date(d.publishedAt));
this.chartRange1 = this.getStartDate(this.chartRange1);
this.chartRange2 = d3.max(this.data, (d: any) => new Date(d.publishedAt));
const sentGroupsOrdered = this.reorderGroups();
const chartColors = this.defineChartColors();
let firstItem = 0;
while (!sentGroupsOrdered[firstItem] && firstItem < 5) {firstItem++; }
const group1 = sentGroupsOrdered[firstItem];
this.sentimentLineChart
.renderArea(true)
.width(900)
.height(this.chartHeight)
.ordinalColors(chartColors)
.useViewBoxResizing(true)
.dimension(this.dimension)
.x(d3.scaleTime().domain([this.chartRange1, this.chartRange2]))
.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)]))
.yAxisLabel('Comments')
.interpolate('monotone')
.legend(dc.legend().x(850).y(0).itemHeight(9).gap(5))
.brushOn(false)
.group(group1.group, group1.sent)
.valueAccessor(function (d) {
return d.value;
})
.xAxis().ticks(7);
let maxSent = 0;
if (sentGroupsOrdered.length > 1) {
sentGroupsOrdered.forEach((group) => {
if (group.group === group1.group || maxSent === 4) {
return;
}
// stacks the groups
this.sentimentLineChart
.stack(group.group, group.sent, function (d) {
return d.value;
});
maxSent++;
});
}
// When filter is applied before refreshing the chart
if (this.appliedFilter) {
this.sentimentLineChart.x(d3.scaleTime().domain([this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2]));
}
// Brush: get range and send it to the other charts on brush-filtering
this.sentimentLineChart.on('filtered', (chart, filter) => {
if (filter) {
this.sentimentLineChart.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(filter[0], filter[1])]));
} else {
this.sentimentLineChart.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)]));
}
this.chartService.setChartRange({range: filter, chart: chart});
});
// Adapt chart for smaller view
(this.chartHeight < 300) ? this.sentimentLineChart.yAxis().ticks(2) : this.sentimentLineChart.yAxis().ticks(10);
(this.chartHeight < 300) ? this.sentimentLineChart.xAxisLabel('') : this.sentimentLineChart.xAxisLabel('Date');
this.sentimentLineChart.xAxis().tickFormat(d3.timeFormat('%b %Y')); // month
this.sentimentLineChart.render();
}
// Adaptable color scale
defineChartColors() {
const colorArray = [];
const sentGroupsOrdered = this.reorderGroups();
sentGroupsOrdered.forEach((g) => {
if (g.sent === 'Pos') {
colorArray.push('#4daf4a');
} else if (g.sent === 'Neu') {
colorArray.push('#666666');
} else if (g.sent === 'Neg') {
colorArray.push('#ff7f00');
} else if (g.sent === 'Mix') {
colorArray.push('#984ea3');
} else if (g.sent === 'N/A') {
| {
this.sentimentLineChart = dc.lineChart('#sentimentChartLine');
this.chartService.GetData().subscribe((data) => { this.data = data; });
// Crossfilter
this.chartService.getCrossfilter().subscribe((filter) => {
this.cfilter = filter;
this.setDimension();
if (this.data && this.data.length > 0) {
this.sentGroups = this.getSentGroups();
if (this.sentGroups[0]) {
this.notDataWarn = false;
this.appliedFilter = false;
this.renderChart();
} else {
this.notDataWarn = true;
}
}
});
| identifier_body |
time-sentiment.component.ts | implements OnInit {
aggrView = true;
compView = false;
data: any[];
cfilter: CrossFilter.CrossFilter<{}>;
dimension: CrossFilter.Dimension<{}, Date>;
sentGroups: { group: CrossFilter.Group<{}, Date, any>, sent: string}[];
sentimentLineChart: dc.LineChart;
renderedChart = false;
notDataWarn = false;
appliedFilter = false;
chartHeight = 300;
chartRange1;
chartRange2;
chartRangeFilter1;
chartRangeFilter2;
constructor(private chartService: ChartService, private _element: ElementRef) { }
ngOnInit() {
this.sentimentLineChart = dc.lineChart('#sentimentChartLine');
this.chartService.GetData().subscribe((data) => { this.data = data; });
// Crossfilter
this.chartService.getCrossfilter().subscribe((filter) => {
this.cfilter = filter;
this.setDimension();
if (this.data && this.data.length > 0) {
this.sentGroups = this.getSentGroups();
if (this.sentGroups[0]) {
this.notDataWarn = false;
this.appliedFilter = false;
this.renderChart();
} else {
this.notDataWarn = true;
}
}
});
// Collapsible view
this.chartService.GetChartMode().subscribe(mode => {
if (this.data && this.data.length > 0) {
if (mode && mode === 'small') {
this.chartHeight = 85;
this.renderChart();
} else if (mode && mode === 'big') {
this.chartHeight = 300;
this.renderChart();
}
}
});
// Gets the range through the chart service from the mainVis Component
this.chartService.getChartRange().subscribe((range) => {
if (range.chart === null) {
if (this.data && range.range) {
(this.diff_months(range.range[0], range.range[1]) < 2) ? this.notDataWarn = true : this.notDataWarn = false;
this.chartRangeFilter1 = range.range[0];
this.chartRangeFilter2 = range.range[1];
this.sentimentLineChart
.x(d3.scaleTime().domain([this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2]))
.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2)]))
.round(d3.timeMonth);
this.appliedFilter = true;
this.sentimentLineChart.redraw();
} else {
if (!dc.chartRegistry.list().some((c) => c.hasFilter())) {
this.notDataWarn = false;
this.sentimentLineChart
.x(d3.scaleTime().domain([this.chartRange1, this.chartRange2]))
.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)]));
this.appliedFilter = false;
}
}
}
});
this.renderedChart = false;
}
// Buttons and Front-End ////////////////////////////////////////////////////////////////////////////////////////////
// sets the tooltip on mouseover
setTooltipInfo(event: MouseEvent, tooltip: HTMLSpanElement) {
tooltip.style.position = 'fixed';
tooltip.style.top = (event.clientY) + 'px';
tooltip.style.left = (event.clientX - tooltip.offsetWidth - 5) + 'px';
}
// sets the crossfilter dimension
setDimension() {
this.dimension = this.cfilter.dimension((d: any) => {
const splitted = d.publishedAt.split('-');
return new Date(splitted[0] + '-' + splitted[1]);
});
}
// Time-based Stacked Chart /////////////////////////////////////////////////////////////////////////////////////////
// returns a crossfilter-group for each sentiment x
private getSentGroups(): { group: CrossFilter.Group<{}, Date, any>, sent: string}[] {
if (this.data && this.data.length < 0) { return; }
const groups: { group: CrossFilter.Group<{}, Date, any>, sent: string}[] = [];
// group by sentiment
const nested = d3.nest()
.key((d: any) => {
if (d.analysis) {
if (d.analysis.mainSentiment === 'mixed') {
return 'Mix';
} else if (d.analysis.mainSentiment === 'positive') {
return 'Pos';
} else if (d.analysis.mainSentiment === 'neutral') {
return 'Neu';
} else if (d.analysis.mainSentiment === 'negative') {
return 'Neg';
} else if (d.analysis.mainSentiment === 'na') {
return 'N/A';
}
}
})
.entries(this.data);
nested.forEach((sentiment) => {
const g = this.dimension.group().reduceSum((d: any) => {
if (d.analysis) {
let mainsentiment = '';
if (d.analysis.mainSentiment === 'mixed') {
mainsentiment = 'Mix';
} else if (d.analysis.mainSentiment === 'positive') {
mainsentiment = 'Pos';
} else if (d.analysis.mainSentiment === 'neutral') {
mainsentiment = 'Neu';
} else if (d.analysis.mainSentiment === 'negative') {
mainsentiment = 'Neg';
} else if (d.analysis.mainSentiment === 'na') {
mainsentiment = 'N/A';
}
return mainsentiment === sentiment.key;
}
});
groups.push({group: g, sent: sentiment.key });
});
return groups;
}
getStartDate(previousDate) {
const date = new Date(previousDate);
date.setDate(1);
return date;
}
// Renders line chart (aggregation)
renderChart () {
this.chartRange1 = d3.min(this.data, (d: any) => new Date(d.publishedAt));
this.chartRange1 = this.getStartDate(this.chartRange1);
this.chartRange2 = d3.max(this.data, (d: any) => new Date(d.publishedAt));
const sentGroupsOrdered = this.reorderGroups();
const chartColors = this.defineChartColors();
let firstItem = 0;
while (!sentGroupsOrdered[firstItem] && firstItem < 5) {firstItem++; }
const group1 = sentGroupsOrdered[firstItem];
this.sentimentLineChart
.renderArea(true)
.width(900)
.height(this.chartHeight)
.ordinalColors(chartColors)
.useViewBoxResizing(true)
.dimension(this.dimension)
.x(d3.scaleTime().domain([this.chartRange1, this.chartRange2]))
.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)]))
.yAxisLabel('Comments')
.interpolate('monotone')
.legend(dc.legend().x(850).y(0).itemHeight(9).gap(5))
.brushOn(false)
.group(group1.group, group1.sent)
.valueAccessor(function (d) {
return d.value;
})
.xAxis().ticks(7);
let maxSent = 0;
if (sentGroupsOrdered.length > 1) {
sentGroupsOrdered.forEach((group) => {
if (group.group === group1.group || maxSent === 4) {
return;
}
// stacks the groups
this.sentimentLineChart
.stack(group.group, group.sent, function (d) {
return d.value;
});
maxSent++;
});
}
// When filter is applied before refreshing the chart
if (this.appliedFilter) {
this.sentimentLineChart.x(d3.scaleTime().domain([this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2]));
}
// Brush: get range and send it to the other charts on brush-filtering
this.sentimentLineChart.on('filtered', (chart, filter) => {
if (filter) {
this.sentimentLineChart.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(filter[0], filter[1])]));
} else {
this.sentimentLineChart.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)]));
}
this.chartService.setChartRange({range: filter, chart: chart});
});
// Adapt chart for smaller view
(this.chartHeight < 300) ? this.sentimentLineChart.yAxis().ticks(2) : this.sentimentLineChart.yAxis().ticks(10);
(this.chartHeight < 300) ? this.sentimentLineChart.xAxisLabel('') : this.sentimentLineChart.xAxisLabel('Date');
this.sentimentLineChart.xAxis().tickFormat(d3.timeFormat('%b %Y')); // month
this.sentimentLineChart.render();
}
// Adaptable color scale
defineChartColors() {
const colorArray = [];
const sentGroupsOrdered = this.reorderGroups();
sentGroupsOrdered.forEach((g) => {
if (g.sent === 'Pos') {
colorArray.push('#4daf4a');
} else if (g.sent === 'Neu') {
colorArray.push('#6 | TimeSentimentComponent | identifier_name |
|
text.rs | /// as """.
///
pub(crate) fn escape(input: &str) -> String {
let mut result = String::with_capacity(input.len());
for c in input.chars() {
match c {
XML_ESC_AMP_CHAR => result.push_str(&to_entity(XML_ESC_AMP_CHAR)),
XML_ESC_APOS_CHAR => result.push_str(&to_entity(XML_ESC_APOS_CHAR)),
XML_ESC_GT_CHAR => result.push_str(&to_entity(XML_ESC_GT_CHAR)),
XML_ESC_LT_CHAR => result.push_str(&to_entity(XML_ESC_LT_CHAR)),
XML_ESC_QUOT_CHAR => result.push_str(&to_entity(XML_ESC_QUOT_CHAR)),
o => result.push(o),
}
}
result
}
pub(crate) fn to_entity(c: char) -> String {
format!(
"{}{}{}",
XML_NUMBERED_ENTITYREF_START, c as u16, XML_ENTITYREF_END
)
}
#[allow(dead_code)]
pub(crate) fn to_entity_hex(c: char) -> String {
format!(
"{}{:X}{}",
XML_HEX_NUMBERED_ENTITYREF_START, c as u16, XML_ENTITYREF_END
)
}
fn char_from_entity(entity: &str) -> String {
assert!(entity.starts_with("&#"));
assert!(entity.ends_with(';'));
let code_point = if &entity[2..3] == "x" {
let code_point = &entity[3..entity.len() - 1];
u32::from_str_radix(code_point, 16).unwrap()
} else {
let code_point = &entity[2..entity.len() - 1];
u32::from_str_radix(code_point, 10).unwrap()
};
let character = char::try_from(code_point).unwrap();
character.to_string()
}
///
/// From [XML 1.0 §2.2](https://www.w3.org/TR/REC-xml/#charsets)
///
/// Definition: A parsed entity contains **text**, a sequence of characters, which may represent
/// markup or character data. Definition: A **character** is an atomic unit of text as specified by
/// ISO/IEC 10646:2000. Legal characters are tab, carriage return, line feed, and the legal
/// characters of Unicode and ISO/IEC 10646. The versions of these standards cited in A.1 Normative
/// References were current at the time this document was prepared. New characters may be added to
/// these standards by amendments or new editions. Consequently, XML processors must accept any
/// character in the range specified for `Char`.
///
/// ```ebnf
/// Char ::= #x9 | #xA | #xD | [#x20-#xD7FF] | [#xE000-#xFFFD] | [#x10000-#x10FFFF]
/// /* any Unicode character, excluding the surrogate blocks, FFFE, and FFFF. */
/// ```
///
/// Document authors are encouraged to avoid "compatibility characters", as defined in section 2.3
/// of Unicode. The characters defined in the following ranges are also discouraged. They are either
/// control characters or permanently undefined Unicode characters:
///
/// ```text
/// [#x7F-#x84], [#x86-#x9F], [#xFDD0-#xFDEF],
/// [#x1FFFE-#x1FFFF], [#x2FFFE-#x2FFFF], [#x3FFFE-#x3FFFF],
/// [#x4FFFE-#x4FFFF], [#x5FFFE-#x5FFFF], [#x6FFFE-#x6FFFF],
/// [#x7FFFE-#x7FFFF], [#x8FFFE-#x8FFFF], [#x9FFFE-#x9FFFF],
/// [#xAFFFE-#xAFFFF], [#xBFFFE-#xBFFFF], [#xCFFFE-#xCFFFF],
/// [#xDFFFE-#xDFFFF], [#xEFFFE-#xEFFFF], [#xFFFFE-#xFFFFF],
/// [#x10FFFE-#x10FFFF].
/// ```
///
#[allow(dead_code)]
pub(crate) fn is_xml_10_char(c: char) -> bool {
c == '\u{0009}'
|| c == '\u{000A}'
|| c == '\u{000D}'
|| (c >= '\u{0020}' && c <= '\u{D7FF}')
|| (c >= '\u{E000}' && c <= '\u{FFFD}')
|| (c >= '\u{10000}' && c <= '\u{10FFF}')
}
#[allow(dead_code)]
pub(crate) fn is_xml_10_restricted_char(c: char) -> bool {
c == XML_ESC_AMP_CHAR
|| c == XML_ESC_APOS_CHAR
|| c == XML_ESC_GT_CHAR
|| c == XML_ESC_LT_CHAR
|| c == XML_ESC_QUOT_CHAR
}
///
/// From [XML 11 §2.2](https://www.w3.org/TR/xml11/#charsets)
///
/// ```ebnf
/// Char ::= [#x1-#xD7FF] | [#xE000-#xFFFD] | [#x10000-#x10FFFF]
/// /* any Unicode character, excluding the surrogate blocks, FFFE, and FFFF. */
/// ```
///
#[allow(dead_code)]
pub(crate) fn is_xml_11_char(c: char) -> bool {
//
// below ranges are always valid for XML 1.1 documents
// from https://en.wikipedia.org/wiki/XML#Valid_characters
//
(c >= '\u{0001}' && c <= '\u{D7FF}')
|| (c >= '\u{E000}' && c <= '\u{FFFD}')
|| (c >= '\u{10000}' && c <= '\u{10FFF}')
}
///
/// From [XML 11 §2.2](https://www.w3.org/TR/xml11/#charsets)
///
/// ```ebnf
/// RestrictedChar ::= #x1-#x8] | [#xB-#xC] | [#xE-#x1F] | [#x7F-#x84] | [#x86-#x9F]
/// ```
///
#[allow(dead_code)]
pub(crate) fn is_xml_11_restricted_char(c: char) -> bool {
//
// below ranges are always valid for XML 1.1 documents
// from https://en.wikipedia.org/wiki/XML#Valid_characters
//
(c >= '\u{01}' && c <= '\u{08}')
|| (c >= '\u{0B}' && c <= '\u{0C}')
|| (c >= '\u{0E}' && c <= '\u{1F}')
|| (c >= '\u{7F}' && c <= '\u{84}')
|| (c >= '\u{86}' && c <= '\u{9F}')
}
///
/// S (white space) consists of one or more space (#x20) characters, carriage returns, line feeds,
/// or tabs.
///
/// ```ebnf
/// S ::= (#x20 | #x9 | #xD | #xA)+
/// ```
///
/// The presence of #xD in the above production is maintained purely for backward compatibility
/// with the First Edition. As explained in 2.11 End-of-Line Handling, all #xD characters literally
/// present in an XML document are either removed or replaced by #xA characters before any other
/// processing is done. The only way to get a #xD character to match this production is to use a
/// character reference in an entity value literal.
///
#[allow(dead_code)]
pub(crate) fn is_xml_space(c: char) -> bool {
c == '\u{09}' || c == '\u{0A}' || c == '\u{0D}' || c == '\u{20}'
}
///
/// ```ebnf
/// NameStartChar ::= ":" | [A-Z] | "_" | [a-z] | [#xC0-#xD6] | [#xD8-#xF6] | [#xF8-#x2FF] |
/// [#x370-#x37D] | [#x37F-#x1FFF] | [#x200C-#x200D] | [#x2070-#x218F] |
/// [#x2C00-#x2FEF] | [#x3001-#xD7FF] | [#xF900-#xFDCF] | [#xFDF0-#xFFFD] |
/// [#x10000-#xEFFFF]
/// ```
///
#[allow(dead_code)]
pub(crate) fn is_xml_name_start_char(c: char) -> bool {
c == ':'
|| (c >= 'A' && c <= 'Z') | || c == '_'
|| (c >= 'a' && c <= 'z')
|| (c >= '\u{C0}' && c <= '\u{D6}') | random_line_split |
|
text.rs | else if let Some(a_match) = capture.name("char") {
let replacement = char_from_entity(a_match.as_str());
(a_match.start(), a_match.end(), replacement)
} else if let Some(a_match) = capture.name("char_hex") {
let replacement = char_from_entity(a_match.as_str());
(a_match.start(), a_match.end(), replacement)
} else if let Some(a_match) = capture.name("ws") {
(a_match.start(), a_match.end(), "\u{20}".to_string())
} else {
panic!("unexpected result");
};
step_2.push_str(&step_1[last_end..start]);
step_2.push_str(&replacement);
last_end = end;
}
if last_end < value.len() {
step_2.push_str(&step_1[last_end..]);
}
step_2
};
if is_cdata {
step_3
} else {
step_3.trim_matches(' ').to_string()
}
}
///
/// From XML 1.1 §2.11 [End-of-Line Handling](https://www.w3.org/TR/xml11/#sec-line-ends):
///
/// XML parsed entities are often stored in computer files which, for editing convenience, are
/// organized into lines. These lines are typically separated by some combination of the characters
/// CARRIAGE RETURN `(#xD`) and LINE FEED (`#xA`).
///
/// To simplify the tasks of applications, the XML processor must behave as if it normalized all line
/// breaks in external parsed entities (including the document entity) on input, before parsing, by
/// translating all of the following to a single `#xA` character:
///
/// * the two-character sequence `#xD` `#xA`
/// * the two-character sequence `#xD` `#x85`
/// * the single character `#x85`
/// * the single character `#x2028`
/// * any `#xD` character that is not immediately followed by `#xA` or `#x85`.
///
/// The characters `#x85` and `#x2028` cannot be reliably recognized and translated until an entity's
/// encoding declaration (if present) has been read. Therefore, it is a fatal error to use them
/// within the XML declaration or text declaration.
///
pub(crate) fn normalize_end_of_lines(value: &str) -> String {
if value.is_empty() {
value.to_string()
} else {
let line_ends = regex::Regex::new(r"\u{0D}[\u{0A}\u{85}]?|\u{85}|\u{2028}").unwrap();
line_ends.replace_all(value, "\u{0A}").to_string()
}
}
///
/// Escape character data according to XML 1.1
/// [§2.4 Character Data and Markup](https://www.w3.org/TR/xml11/#dt-chardata). This is the
/// do-everything version, not attempting to separate the rules defined below by node type.
///
/// # Specification
///
/// Text consists of intermingled character data and markup. [Definition: **Markup** takes the form
/// of start-tags, end-tags, empty-element tags, entity references, character references, comments,
/// CDATA section delimiters, document type declarations, processing instructions, XML declarations,
/// text declarations, and any white space that is at the top level of the document entity (that is,
/// outside the document element and not inside any other markup).]
///
/// [Definition: All text that is not markup constitutes the **character data** of the document].
///
/// The ampersand character (&) and the left angle bracket (<) must not appear in their literal
/// form, except when used as markup delimiters, or within a comment, a processing instruction, or
/// a CDATA section. If they are needed elsewhere, they must be escaped using either numeric
/// character references or the strings "&" and "<" respectively. The right angle bracket
/// (>) may be represented using the string ">", and must, for compatibility, be escaped using
/// either ">" or a character reference when it appears in the string "]]>" in content, when that
/// string is not marking the end of a CDATA section.
///
/// In the content of elements, character data is any string of characters which does not contain
/// the start-delimiter of any markup or the CDATA-section-close delimiter, "]]>". In a CDATA
/// section, character data is any string of characters not including the CDATA-section-close
/// delimiter.
///
/// To allow attribute values to contain both single and double quotes, the apostrophe or
/// single-quote character (') may be represented as "'", and the double-quote character (")
/// as """.
///
pub(crate) fn escape(input: &str) -> String {
let mut result = String::with_capacity(input.len());
for c in input.chars() {
match c {
XML_ESC_AMP_CHAR => result.push_str(&to_entity(XML_ESC_AMP_CHAR)),
XML_ESC_APOS_CHAR => result.push_str(&to_entity(XML_ESC_APOS_CHAR)),
XML_ESC_GT_CHAR => result.push_str(&to_entity(XML_ESC_GT_CHAR)),
XML_ESC_LT_CHAR => result.push_str(&to_entity(XML_ESC_LT_CHAR)),
XML_ESC_QUOT_CHAR => result.push_str(&to_entity(XML_ESC_QUOT_CHAR)),
o => result.push(o),
}
}
result
}
pub(crate) fn to_entity(c: char) -> String {
format!(
"{}{}{}",
XML_NUMBERED_ENTITYREF_START, c as u16, XML_ENTITYREF_END
)
}
#[allow(dead_code)]
pub(crate) fn to_entity_hex(c: char) -> String {
format!(
"{}{:X}{}",
XML_HEX_NUMBERED_ENTITYREF_START, c as u16, XML_ENTITYREF_END
)
}
fn char_from_entity(entity: &str) -> String {
assert!(entity.starts_with("&#"));
assert!(entity.ends_with(';'));
let code_point = if &entity[2..3] == "x" {
let code_point = &entity[3..entity.len() - 1];
u32::from_str_radix(code_point, 16).unwrap()
} else {
let code_point = &entity[2..entity.len() - 1];
u32::from_str_radix(code_point, 10).unwrap()
};
let character = char::try_from(code_point).unwrap();
character.to_string()
}
///
/// From [XML 1.0 §2.2](https://www.w3.org/TR/REC-xml/#charsets)
///
/// Definition: A parsed entity contains **text**, a sequence of characters, which may represent
/// markup or character data. Definition: A **character** is an atomic unit of text as specified by
/// ISO/IEC 10646:2000. Legal characters are tab, carriage return, line feed, and the legal
/// characters of Unicode and ISO/IEC 10646. The versions of these standards cited in A.1 Normative
/// References were current at the time this document was prepared. New characters may be added to
/// these standards by amendments or new editions. Consequently, XML processors must accept any
/// character in the range specified for `Char`.
///
/// ```ebnf
/// Char ::= #x9 | #xA | #xD | [#x20-#xD7FF] | [#xE000-#xFFFD] | [#x10000-#x10FFFF]
/// /* any Unicode character, excluding the surrogate blocks, FFFE, and FFFF. */
/// ```
///
/// Document authors are encouraged to avoid "compatibility characters", as defined in section 2.3
/// of Unicode. The characters defined in the following ranges are also discouraged. They are either
/// control characters or permanently undefined Unicode characters:
///
/// ```text
/// [#x7F-#x84], [#x86-#x9F], [#xFDD0-#xFDEF],
/// [#x1FFFE-#x1FFFF], [#x2FFFE-#x2FFFF], [#x3FFFE-#x3FFFF],
/// [#x4FFFE-#x4FFFF], [#x5FFFE-#x5FFFF], [#x6FFFE-#x6FFFF],
/// [#x7FFFE-#x7FFFF], [#x8FFFE-#x8FFFF], [#x9FFFE-#x9FFFF],
/// [#xAFFFE-#xAFFFF], [#xBFFFE-#xBFFFF], [#xCFFFE-#xCFFFF],
/// [#xDFFFE-#xDFFFF], [#xEFFFE-#xEFFFF], [#xFFFFE-#xFFFFF],
/// [#x10FFFE-#x10FFFF].
/// ```
///
#[allow(dead_code)]
pub(crate) fn is_xml_10 |
//
// TODO: this does not yet deal with entity references.
//
let replacement = match resolver.resolve(a_match.as_str()) {
None => panic!("unknown entity reference {}", a_match.as_str()),
Some(replacement) => {
normalize_attribute_value(&replacement, resolver, is_cdata)
}
};
(a_match.start(), a_match.end(), replacement)
} | conditional_block |
|
text.rs | xFFFFE-#xFFFFF],
/// [#x10FFFE-#x10FFFF].
/// ```
///
#[allow(dead_code)]
pub(crate) fn is_xml_10_char(c: char) -> bool {
c == '\u{0009}'
|| c == '\u{000A}'
|| c == '\u{000D}'
|| (c >= '\u{0020}' && c <= '\u{D7FF}')
|| (c >= '\u{E000}' && c <= '\u{FFFD}')
|| (c >= '\u{10000}' && c <= '\u{10FFF}')
}
#[allow(dead_code)]
pub(crate) fn is_xml_10_restricted_char(c: char) -> bool {
c == XML_ESC_AMP_CHAR
|| c == XML_ESC_APOS_CHAR
|| c == XML_ESC_GT_CHAR
|| c == XML_ESC_LT_CHAR
|| c == XML_ESC_QUOT_CHAR
}
///
/// From [XML 11 §2.2](https://www.w3.org/TR/xml11/#charsets)
///
/// ```ebnf
/// Char ::= [#x1-#xD7FF] | [#xE000-#xFFFD] | [#x10000-#x10FFFF]
/// /* any Unicode character, excluding the surrogate blocks, FFFE, and FFFF. */
/// ```
///
#[allow(dead_code)]
pub(crate) fn is_xml_11_char(c: char) -> bool {
//
// below ranges are always valid for XML 1.1 documents
// from https://en.wikipedia.org/wiki/XML#Valid_characters
//
(c >= '\u{0001}' && c <= '\u{D7FF}')
|| (c >= '\u{E000}' && c <= '\u{FFFD}')
|| (c >= '\u{10000}' && c <= '\u{10FFF}')
}
///
/// From [XML 11 §2.2](https://www.w3.org/TR/xml11/#charsets)
///
/// ```ebnf
/// RestrictedChar ::= #x1-#x8] | [#xB-#xC] | [#xE-#x1F] | [#x7F-#x84] | [#x86-#x9F]
/// ```
///
#[allow(dead_code)]
pub(crate) fn is_xml_11_restricted_char(c: char) -> bool {
//
// below ranges are always valid for XML 1.1 documents
// from https://en.wikipedia.org/wiki/XML#Valid_characters
//
(c >= '\u{01}' && c <= '\u{08}')
|| (c >= '\u{0B}' && c <= '\u{0C}')
|| (c >= '\u{0E}' && c <= '\u{1F}')
|| (c >= '\u{7F}' && c <= '\u{84}')
|| (c >= '\u{86}' && c <= '\u{9F}')
}
///
/// S (white space) consists of one or more space (#x20) characters, carriage returns, line feeds,
/// or tabs.
///
/// ```ebnf
/// S ::= (#x20 | #x9 | #xD | #xA)+
/// ```
///
/// The presence of #xD in the above production is maintained purely for backward compatibility
/// with the First Edition. As explained in 2.11 End-of-Line Handling, all #xD characters literally
/// present in an XML document are either removed or replaced by #xA characters before any other
/// processing is done. The only way to get a #xD character to match this production is to use a
/// character reference in an entity value literal.
///
#[allow(dead_code)]
pub(crate) fn is_xml_space(c: char) -> bool {
c == '\u{09}' || c == '\u{0A}' || c == '\u{0D}' || c == '\u{20}'
}
///
/// ```ebnf
/// NameStartChar ::= ":" | [A-Z] | "_" | [a-z] | [#xC0-#xD6] | [#xD8-#xF6] | [#xF8-#x2FF] |
/// [#x370-#x37D] | [#x37F-#x1FFF] | [#x200C-#x200D] | [#x2070-#x218F] |
/// [#x2C00-#x2FEF] | [#x3001-#xD7FF] | [#xF900-#xFDCF] | [#xFDF0-#xFFFD] |
/// [#x10000-#xEFFFF]
/// ```
///
#[allow(dead_code)]
pub(crate) fn is_xml_name_start_char(c: char) -> bool {
c == ':'
|| (c >= 'A' && c <= 'Z')
|| c == '_'
|| (c >= 'a' && c <= 'z')
|| (c >= '\u{C0}' && c <= '\u{D6}')
|| (c >= '\u{D8}' && c <= '\u{F6}')
|| (c >= '\u{0F8}' && c <= '\u{2FF}')
|| (c >= '\u{370}' && c <= '\u{37D}')
|| (c >= '\u{037F}' && c <= '\u{1FFF}')
|| (c >= '\u{200C}' && c <= '\u{200D}')
|| (c >= '\u{2070}' && c <= '\u{218F}')
|| (c >= '\u{2C00}' && c <= '\u{2FEF}')
|| (c >= '\u{3001}' && c <= '\u{D7FF}')
|| (c >= '\u{F900}' && c <= '\u{FDCF}')
|| (c >= '\u{FDF0}' && c <= '\u{FFFD}')
|| (c >= '\u{10000}' && c <= '\u{EFFFF}')
}
///
/// ```ebnf
/// NameChar ::= NameStartChar | "-" | "." | [0-9] | #xB7 |
/// [#x0300-#x036F] | [#x203F-#x2040]
/// ```
///
pub(crate) fn is_xml_name_char(c: char) -> bool {
is_xml_name_start_char(c)
|| c == '-'
|| c == '.'
|| (c >= '0' && c <= '9')
|| c == '\u{B7}'
|| (c >= '\u{0300}' && c <= '\u{036F}')
|| (c >= '\u{203F}' && c <= '\u{2040}')
}
///
/// ```ebnf
/// Name ::= NameStartChar (NameChar)*
/// ```
///
pub(crate) fn is_xml_name(s: &str) -> bool {
!s.is_empty() && s.starts_with(is_xml_name_start_char) && s[1..].chars().all(is_xml_name_char)
}
///
/// ```ebnf
/// Names ::= Name (#x20 Name)*
/// ```
///
#[allow(dead_code)]
pub(crate) fn is_xml_names(s: &str) -> bool {
!s.is_empty() && s.split(' ').all(is_xml_name)
}
///
/// ```ebnf
/// Nmtoken ::= (NameChar)+
/// ```
///
#[allow(dead_code)]
pub(crate) fn is_xml_nmtoken(s: &str) -> bool {
!s.is_empty() && s.chars().all(is_xml_name_char)
}
///
/// ```ebnf
/// Nmtokens ::= Nmtoken (#x20 Nmtoken)*
/// ```
///
#[allow(dead_code)]
pub(crate) fn is_xml_nmtokens(s: &str) -> bool {
!s.is_empty() && s.split(' ').all(is_xml_nmtoken)
}
// ------------------------------------------------------------------------------------------------
// Implementations
// ------------------------------------------------------------------------------------------------
impl Default for SpaceHandling {
fn default() -> Self {
SpaceHandling::Default
}
}
// ------------------------------------------------------------------------------------------------
impl Display for SpaceHandling {
fn fmt(&self, f: &mut Formatter<'_>) -> FmtResult {
write!(
f,
"{}{}{}=\"{}\"",
XML_NS_ATTRIBUTE,
XML_NS_SEPARATOR,
XML_NS_ATTR_SPACE,
match self {
SpaceHandling::Default => XML_NS_ATTR_SPACE_DEFAULT,
SpaceHandling::Preserve => XML_NS_ATTR_SPACE_PRESERVE,
}
)
}
}
// ------------------------------------------------------------------------------------------------
impl FromStr for SpaceHandling {
type Err = ();
fn from_str(s: &str) -> Result<Self, Self::Err> {
| if s == XML_NS_ATTR_SPACE_DEFAULT {
Ok(SpaceHandling::Default)
} else if s == XML_NS_ATTR_SPACE_PRESERVE {
Ok(SpaceHandling::Preserve)
} else {
Err(())
}
}
}
// | identifier_body |
|
text.rs | {
c == '\u{09}' || c == '\u{0A}' || c == '\u{0D}' || c == '\u{20}'
}
///
/// ```ebnf
/// NameStartChar ::= ":" | [A-Z] | "_" | [a-z] | [#xC0-#xD6] | [#xD8-#xF6] | [#xF8-#x2FF] |
/// [#x370-#x37D] | [#x37F-#x1FFF] | [#x200C-#x200D] | [#x2070-#x218F] |
/// [#x2C00-#x2FEF] | [#x3001-#xD7FF] | [#xF900-#xFDCF] | [#xFDF0-#xFFFD] |
/// [#x10000-#xEFFFF]
/// ```
///
#[allow(dead_code)]
pub(crate) fn is_xml_name_start_char(c: char) -> bool {
c == ':'
|| (c >= 'A' && c <= 'Z')
|| c == '_'
|| (c >= 'a' && c <= 'z')
|| (c >= '\u{C0}' && c <= '\u{D6}')
|| (c >= '\u{D8}' && c <= '\u{F6}')
|| (c >= '\u{0F8}' && c <= '\u{2FF}')
|| (c >= '\u{370}' && c <= '\u{37D}')
|| (c >= '\u{037F}' && c <= '\u{1FFF}')
|| (c >= '\u{200C}' && c <= '\u{200D}')
|| (c >= '\u{2070}' && c <= '\u{218F}')
|| (c >= '\u{2C00}' && c <= '\u{2FEF}')
|| (c >= '\u{3001}' && c <= '\u{D7FF}')
|| (c >= '\u{F900}' && c <= '\u{FDCF}')
|| (c >= '\u{FDF0}' && c <= '\u{FFFD}')
|| (c >= '\u{10000}' && c <= '\u{EFFFF}')
}
///
/// ```ebnf
/// NameChar ::= NameStartChar | "-" | "." | [0-9] | #xB7 |
/// [#x0300-#x036F] | [#x203F-#x2040]
/// ```
///
pub(crate) fn is_xml_name_char(c: char) -> bool {
is_xml_name_start_char(c)
|| c == '-'
|| c == '.'
|| (c >= '0' && c <= '9')
|| c == '\u{B7}'
|| (c >= '\u{0300}' && c <= '\u{036F}')
|| (c >= '\u{203F}' && c <= '\u{2040}')
}
///
/// ```ebnf
/// Name ::= NameStartChar (NameChar)*
/// ```
///
pub(crate) fn is_xml_name(s: &str) -> bool {
!s.is_empty() && s.starts_with(is_xml_name_start_char) && s[1..].chars().all(is_xml_name_char)
}
///
/// ```ebnf
/// Names ::= Name (#x20 Name)*
/// ```
///
#[allow(dead_code)]
pub(crate) fn is_xml_names(s: &str) -> bool {
!s.is_empty() && s.split(' ').all(is_xml_name)
}
///
/// ```ebnf
/// Nmtoken ::= (NameChar)+
/// ```
///
#[allow(dead_code)]
pub(crate) fn is_xml_nmtoken(s: &str) -> bool {
!s.is_empty() && s.chars().all(is_xml_name_char)
}
///
/// ```ebnf
/// Nmtokens ::= Nmtoken (#x20 Nmtoken)*
/// ```
///
#[allow(dead_code)]
pub(crate) fn is_xml_nmtokens(s: &str) -> bool {
!s.is_empty() && s.split(' ').all(is_xml_nmtoken)
}
// ------------------------------------------------------------------------------------------------
// Implementations
// ------------------------------------------------------------------------------------------------
impl Default for SpaceHandling {
fn default() -> Self {
SpaceHandling::Default
}
}
// ------------------------------------------------------------------------------------------------
impl Display for SpaceHandling {
fn fmt(&self, f: &mut Formatter<'_>) -> FmtResult {
write!(
f,
"{}{}{}=\"{}\"",
XML_NS_ATTRIBUTE,
XML_NS_SEPARATOR,
XML_NS_ATTR_SPACE,
match self {
SpaceHandling::Default => XML_NS_ATTR_SPACE_DEFAULT,
SpaceHandling::Preserve => XML_NS_ATTR_SPACE_PRESERVE,
}
)
}
}
// ------------------------------------------------------------------------------------------------
impl FromStr for SpaceHandling {
type Err = ();
fn from_str(s: &str) -> Result<Self, Self::Err> {
if s == XML_NS_ATTR_SPACE_DEFAULT {
Ok(SpaceHandling::Default)
} else if s == XML_NS_ATTR_SPACE_PRESERVE {
Ok(SpaceHandling::Preserve)
} else {
Err(())
}
}
}
// ------------------------------------------------------------------------------------------------
// Unit Tests
// ------------------------------------------------------------------------------------------------
#[cfg(test)]
mod tests {
use super::*;
use std::borrow::Borrow;
use std::collections::HashMap;
#[test]
fn test_space_handling_default() {
let sh = SpaceHandling::default();
assert_eq!(sh, SpaceHandling::Default);
}
#[test]
fn test_space_handling_display() {
assert_eq!(
format!("{}", SpaceHandling::Default),
format!(
"{}{}{}=\"{}\"",
XML_NS_ATTRIBUTE, XML_NS_SEPARATOR, XML_NS_ATTR_SPACE, XML_NS_ATTR_SPACE_DEFAULT
)
);
assert_eq!(
format!("{}", SpaceHandling::Preserve),
format!(
"{}{}{}=\"{}\"",
XML_NS_ATTRIBUTE, XML_NS_SEPARATOR, XML_NS_ATTR_SPACE, XML_NS_ATTR_SPACE_PRESERVE
)
);
}
#[test]
fn test_space_handling_from_str() {
assert_eq!(
SpaceHandling::from_str(XML_NS_ATTR_SPACE_DEFAULT).unwrap(),
SpaceHandling::Default
);
assert_eq!(
SpaceHandling::from_str(XML_NS_ATTR_SPACE_PRESERVE).unwrap(),
SpaceHandling::Preserve
);
assert!(SpaceHandling::from_str("").is_err());
assert!(SpaceHandling::from_str("other").is_err());
}
#[test]
fn test_end_of_line_handling() {
let input = "one\u{0D}two\u{0D}\u{0A}\u{0A}three\u{0A}\u{0D}\u{85}four\u{85}five\u{2028}";
let output = normalize_end_of_lines(&input.to_string());
assert_eq!(
output,
"one\u{0A}two\u{0A}\u{0A}three\u{0A}\u{0A}four\u{0A}five\u{0A}".to_string()
)
}
struct NoneEntityResolver {}
impl EntityResolver for NoneEntityResolver {
fn resolve(&self, name: &str) -> Option<String> {
let result: Option<String> = None;
println!("EntityResolver::resolve({:?}) -> {:?}", name, result);
result
}
}
pub(crate) fn none_entity_resolver() -> Box<dyn EntityResolver> {
let resolver = NoneEntityResolver {};
Box::new(resolver)
}
#[test]
fn test_normalize_avalue_trim() {
let resolver = none_entity_resolver();
let resolver = resolver.borrow();
assert_eq!(
normalize_attribute_value(" abc ", resolver, true),
" abc "
);
assert_eq!(normalize_attribute_value(" abc ", resolver, false), "abc");
}
struct TestResolver {
entity_map: HashMap<String, String>,
}
impl EntityResolver for TestResolver {
fn resolve(&self, entity: &str) -> Option<String> {
self.entity_map.get(entity).cloned()
}
}
impl TestResolver {
pub(crate) fn new() -> Self {
let mut new_self = Self {
entity_map: Default::default(),
};
let _safe_to_ignore = new_self
.entity_map
.insert("£".to_string(), "£".to_string());
let _safe_to_ignore = new_self
.entity_map
.insert("¥".to_string(), "¥".to_string());
let _safe_to_ignore = new_self
.entity_map
.insert("€".to_string(), "€".to_string());
let _safe_to_ignore = new_self.entity_map.insert(
"¤cy;".to_string(),
"$, £, €, and ¥".to_string(),
);
new_self
}
}
fn test_resolver() -> Box<dyn EntityResolver> {
let resolver = TestResolver::new();
Box::new(resolver)
}
#[test]
fn test_norma | lize_avalue_entity_resolver() {
| identifier_name |
|
ionic-native-map.ts | ");
let userID = localStorage.getItem("userID");
let URL = globalVars.UserAddress(userID);
// console.log(locationExists);
let data = {
alias: this.locationAlias,
address: this.address,
lat: this.lat,
long: this.lng
}
if(this.validate()){
// let locationExists: boolean = false;
// this.addressResponse.forEach(address => {
// locationExists = locationExists || (address.alias == this.locationAlias);
// console.log(address.alias, this.locationAlias);
// console.log(address.alias == this.locationAlias);
// });
// console.log('location Exists: ', locationExists);
// if(!locationExists){
this.authService.patchCall(URL, data)
.subscribe(res => {
if (res.status == 200) {
console.log(res['_body']);
}
});
}else{
this.map.setClickable(false);
let result = this.alertCntrl.openAlertDialog('Location exits', 'Please enter a location.');
// result.then(value => {
// if(value){
// this.map.setClickable(true);
// }
// })
}
// }else{
// // this.alertCntrl.openAlertDialog('Error', 'Location already Exists.')
// }
}
openAdditionalNoteDialog(myEvent) {
this.map.setClickable(false);
this.isModalVisible = this.isModalVisible ? false : true;
setTimeout(() => {
this.isModalVisible = this.isModalVisible ? false : true;
}, 200);
let popover = this.popoverCtrl.create(AdditionalNote, {}, { showBackdrop: true });
popover.present({
ev: myEvent
});
popover.onDidDismiss(data => {
if(data){
this.map.setClickable(true);
console.log(data);
this.additionalInfoText = data + "\n";
localStorage.setItem("additionalInfoText", this.additionalInfoText);
}
})
}
additionButtonClicked(myEvent) {
this.addition = this.addition ? false : true;
console.log("additionButtonClicked");
this.openAdditionalNoteDialog(myEvent);
}
locationClickedBool;
locationClicked(location) {
console.log("You have clicked on: ", location);
this.locationClickedBool = false;
this.hide = true;
if(!!location){
this.inputFieldValue = '';
if(!!location.name){
this.locationClickedBool = true;
console.log(location);
this.inputFieldValue = location.name || '';
localStorage.setItem("Location", JSON.stringify(location));
this.lat = location.geometry.location.lat;
this.lng = location.geometry.location.lng;
this.address = location.formatted_address;
this.locationAlias = location.name;
}else{
console.log('Here');
this.locationClickedBool = true;
this.inputFieldValue = location.alias || '';
localStorage.setItem("Location", JSON.stringify(location));
this.lat = location.lat;
this.lng = location.long;
this.address = location.address;
this.locationAlias = location.alias;
};
setTimeout(() => { this.available_locations = []}, 200);
}else{
console.log('Here');
this.locationClickedBool = true;
this.inputFieldValue = location.alias || '';
localStorage.setItem("Location", JSON.stringify(location));
this.lat = location.lat;
this.lng = location.long;
this.address = location.address;
this.locationAlias = location.alias;
};
setTimeout(() => { this.available_locations = []}, 200);
//gMap = new google.maps.Map(document.getElementById('map'));
// this.postion = new google.maps.LatLng(this.lat, this.lng);
// this.map.setCenter(this.postion);
// this.addMarkerMoveCamera(this.map, new LatLng(this.lat, this.lng));
// this.map.center = new google.maps.LatLng(this.lat, this.lng);
this.addMarker(this.map, new LatLng(this.lat, this.lng));
this.moveCamera(this.map, new LatLng(this.lat, this.lng));
}
validate():boolean
{
return (this.lat != null && this.lng != null && this.address != null) ? true :false;
}
startNextScreen() {
console.log("Next clicked!");
let valid:boolean = this.validate();
console.log(valid);
if(valid === true && this.locationClickedBool == true)
{
console.log(this.preGenData);
this.navCtrl.push(LaundryItems, {
preGenData: this.preGenData,
pickupDetails: {
location: {
lat: this.lat,
lng: this.lng,
address: this.address
}
},
});
}
else{
this.map.setClickable(false);
this.alertCntrl.openAlertDialog("What's missing?","No location selected.");
}
}
loadMap(){
let element: HTMLElement = ViewChild('map');
let mapOptions = {
"featureType": "all",
"elementType": "geometry",
styles: [
{ elementType: 'geometry', stylers: [{ color: '#15151b' }] },
{ elementType: 'labels.text.stroke', stylers: [{ color: '#242f3e' }] },
{ elementType: 'labels.text.fill', stylers: [{ color: '#746855' }] },
{
featureType: 'administrative',
elementType: 'labels',
stylers: [{ visibility: 'off' }]
},
{
featureType: 'poi',
elementType: 'labels',
stylers: [{ visibility: 'off' }]
},
{
featureType: 'poi.park',
elementType: 'geometry',
stylers: [{ visibility: 'off' }]
},
{
featureType: 'poi.park',
elementType: 'labels.text.fill',
stylers: [{ visibility: 'off' }]
},
{
featureType: 'road',
elementType: 'geometry',
stylers: [{ color: '#000000' }]
}
// #38414e
,
{
featureType: 'road',
elementType: 'geometry.stroke',
stylers: [{ color: '#000000' }]//212a37
},
{
featureType: 'road',
elementType: 'labels.text.fill',
stylers: [{ color: '#ffffff' }]//9ca5b3
},
{
featureType: 'road.highway',
elementType: 'geometry',
stylers: [{ color: '#000000' }]//746855
},
{
featureType: 'road.highway',
elementType: 'geometry.stroke',
stylers: [{ color: '#1f2835' }]
},
{
featureType: 'road.highway',
elementType: 'labels.text.fill',
stylers: [{ color: '#f3d19c' }]
},
{
featureType: 'transit',
elementType: 'all',
stylers: [{ visibility: 'off' }]
},
{
featureType: 'transit.station',
elementType: 'labels.text.fill',
stylers: [{ visibility: 'off' }]
},
{
featureType: 'water',
elementType: 'geometry',
stylers: [{ color: '#17263c' }]
},
{
featureType: 'water',
elementType: 'labels.text.fill',
stylers: [{ visibility: 'off' }]
},
{
featureType: 'water',
elementType: 'labels.text.stroke',
stylers: [{ visibility: 'off' }]
}
]
// mapTypeId: google.maps.MapTypeId.ROADMAP1
};
let map: GoogleMap = this.googleMaps.create(element);
map = new GoogleMap('map');
this.map = map;
// listen to MAP_READY event
// You must wait for this event to fire before adding something to the map or modifying it in anyway
map.one(GoogleMapsEvent.MAP_READY).then( () => {
console.log('Map is ready!');
// Now you can add elements to the map like the marker
map.setOptions(mapOptions);
map.setMyLocationEnabled(true);
//map.setBackgroundColor('black');
map.setPadding(0, 80, 150, 0);
this.latLng = this.getLocation(map);
map.setCompassEnabled(false);
});
}
getLocation(map: GoogleMap) {
let latLng: string;
map.getMyLocation().then(
location => {
latLng = location.latLng.lat + ',' + location.latLng.lng;
console.log("165", JSON.stringify(location.latLng));
console.log(485, ":", latLng);
this.newLocation = new LatLng(location.latLng.lat, location.latLng.lng);
// this.addMarker(map, location.latLng);
this.moveCamera(map, location.latLng);
let markerOptions: MarkerOptions = {
position: this.newLocation | };
this.addMarker(map, this.newLocation); | random_line_split |
|
ionic-native-map.ts | available_locations: Array<Object> = [];
newLocation;
marker;
constructor(public navCtrl: NavController,
public navParams: NavParams,
private googleMaps: GoogleMaps,
private platform: Platform,
private geolocation: Geolocation,
private androidPermissions: AndroidPermissions,
private alertCtrl: AlertController,
private popoverCtrl: PopoverController,
private mapService: MapService,
private authService: AuthService,
private alertCntrl: AlertDialogFactory,
private savedLocationService: SavedLocationService) {
this.token = localStorage.getItem('x-access-token');
this.userID = localStorage.getItem('userID');
this.preGenData = navParams.get('preGenData');
localStorage.setItem("additionalInfoText", "");
// setTimeout(() => {
// this.inputFieldValue = 'New Value';
// }, 3000)
}
ionViewDidLoad() {
console.log('ionViewDidLoad IonicNativeMapPage');
// this.loadMap();
setTimeout(() => {
this.loadMap();
}, 500);
}
ngAfterViewInit() | this.listenToSearchInput();
this.getMapLocation(location, this.latLng);
}
listenToSearchInput() {
this.hide = false;
let location: string;
console.log('location1:', location)
// let searchInput$ = Observable.fromEvent(this.button.nativeElement, 'keyup')
// .map(e => location = e['srcElement'].value.trim())
// .distinctUntilChanged()
// .switchMap(() => this.mapService.getJSON(location, this.latLng))
// searchInput$.subscribe(location => {
// this.available_locations = location;
// console.log(this.available_locations);
// })
}
getMapLocation(location, latLng) {
if (location) {
// let location$ = this.mapService.getJSON(location, this.latLng);
// location$.subscribe(res => console.log)
}
}
savedButtonClicked(myEvent) {
this.saved = this.saved ? false : true;
setTimeout(()=>{
this.saved = this.saved ? false : true;
}, 200);
let inputs;
this.addressResponse = inputs;
let URL = globalVars.getUsersAddress(this.userID);
this.authService.getCall(URL).
subscribe(res => {
console.log(JSON.parse(res["_body"]));
inputs = JSON.parse(res["_body"])["data"]["contact"]["address"];
console.log(inputs);
this.addressResponse = inputs;
// let result = this.alertCntrl.checkBoxAlertDialog("Saved Locations", inputs)
// console.log(result);
this.radioAlertDialog("Saved Locations", inputs)
})
}
radioAlertDialog(title: string, inputs){
this.map.setClickable(false);
let alert = this.alertCtrl.create({
title: title,
cssClass: 'alertTop'
});
inputs.forEach(input => {
alert.addInput({
type: 'radio',
label: input.alias,
value: input,
checked: false
});
});
alert.addButton({
text: 'Cancel',
handler: () => {
console.log('Cancel clicked.');
}
});
alert.addButton({
text: 'Okay',
handler: data => {
console.log('Radio data:', data);
// this.testCheckboxOpen = false;
// this.testCheckboxResult = data;
this.locationClicked(data);
}
});
alert.present();
alert.onDidDismiss((data) => {
console.log('OnDidDismiss', data);
// dataReturned = data;
this.map.setClickable(true);
return data || 'null';
});
}
saveButtonClicked() {
this.save = this.save ? false : true;
setTimeout(()=>{
this.save = this.save ? false : true;
}, 200);
console.log("saveButtonClicked");
let userID = localStorage.getItem("userID");
let URL = globalVars.UserAddress(userID);
// console.log(locationExists);
let data = {
alias: this.locationAlias,
address: this.address,
lat: this.lat,
long: this.lng
}
if(this.validate()){
// let locationExists: boolean = false;
// this.addressResponse.forEach(address => {
// locationExists = locationExists || (address.alias == this.locationAlias);
// console.log(address.alias, this.locationAlias);
// console.log(address.alias == this.locationAlias);
// });
// console.log('location Exists: ', locationExists);
// if(!locationExists){
this.authService.patchCall(URL, data)
.subscribe(res => {
if (res.status == 200) {
console.log(res['_body']);
}
});
}else{
this.map.setClickable(false);
let result = this.alertCntrl.openAlertDialog('Location exits', 'Please enter a location.');
// result.then(value => {
// if(value){
// this.map.setClickable(true);
// }
// })
}
// }else{
// // this.alertCntrl.openAlertDialog('Error', 'Location already Exists.')
// }
}
openAdditionalNoteDialog(myEvent) {
this.map.setClickable(false);
this.isModalVisible = this.isModalVisible ? false : true;
setTimeout(() => {
this.isModalVisible = this.isModalVisible ? false : true;
}, 200);
let popover = this.popoverCtrl.create(AdditionalNote, {}, { showBackdrop: true });
popover.present({
ev: myEvent
});
popover.onDidDismiss(data => {
if(data){
this.map.setClickable(true);
console.log(data);
this.additionalInfoText = data + "\n";
localStorage.setItem("additionalInfoText", this.additionalInfoText);
}
})
}
additionButtonClicked(myEvent) {
this.addition = this.addition ? false : true;
console.log("additionButtonClicked");
this.openAdditionalNoteDialog(myEvent);
}
locationClickedBool;
locationClicked(location) {
console.log("You have clicked on: ", location);
this.locationClickedBool = false;
this.hide = true;
if(!!location){
this.inputFieldValue = '';
if(!!location.name){
this.locationClickedBool = true;
console.log(location);
this.inputFieldValue = location.name || '';
localStorage.setItem("Location", JSON.stringify(location));
this.lat = location.geometry.location.lat;
this.lng = location.geometry.location.lng;
this.address = location.formatted_address;
this.locationAlias = location.name;
}else{
console.log('Here');
this.locationClickedBool = true;
this.inputFieldValue = location.alias || '';
localStorage.setItem("Location", JSON.stringify(location));
this.lat = location.lat;
this.lng = location.long;
this.address = location.address;
this.locationAlias = location.alias;
};
setTimeout(() => { this.available_locations = []}, 200);
}else{
console.log('Here');
this.locationClickedBool = true;
this.inputFieldValue = location.alias || '';
localStorage.setItem("Location", JSON.stringify(location));
this.lat = location.lat;
this.lng = location.long;
this.address = location.address;
this.locationAlias = location.alias;
};
setTimeout(() => { this.available_locations = []}, 200);
//gMap = new google.maps.Map(document.getElementById('map'));
// this.postion = new google.maps.LatLng(this.lat, this.lng);
// this.map.setCenter(this.postion);
// this.addMarkerMoveCamera(this.map, new LatLng(this.lat, this.lng));
// this.map.center = new google.maps.LatLng(this.lat, this.lng);
this.addMarker(this.map, new LatLng(this.lat, this.lng));
this.moveCamera(this.map, new LatLng(this.lat, this.lng));
}
validate():boolean
{
return (this.lat != null && this.lng != null && this.address != null) ? true :false;
}
startNextScreen() {
console.log("Next clicked!");
let valid:boolean = this.validate();
console.log(valid);
if(valid === true && this.locationClickedBool == true)
{
console.log(this.preGenData);
this.navCtrl.push(LaundryItems, {
preGenData: this.preGenData,
pickupDetails: {
location: {
lat: this.lat,
lng: this.lng,
address: this.address
}
},
});
}
else{
| {
console.log("ngAfterViewInit", this.newLocation);
// this.androidPermissions.checkPermission(this.androidPermissions.PERMISSION.ACCESS_FINE_LOCATION).then(
// success => console.log('Permission granted'),
// err => this.androidPermissions.requestPermissions(this.androidPermissions.PERMISSION.ACCESS_FINE_LOCATION)
// );
// this.geolocation.getCurrentPosition().then((resp) => {
// console.log(resp.coords.latitude);
// console.log(resp.coords.longitude);
// }).catch((error) => {
// console.log('Error getting location', error);
// });
// this.platform.ready().then(() => {
// // this.loadMap();
// });
| identifier_body |
ionic-native-map.ts | : GoogleMaps,
private platform: Platform,
private geolocation: Geolocation,
private androidPermissions: AndroidPermissions,
private alertCtrl: AlertController,
private popoverCtrl: PopoverController,
private mapService: MapService,
private authService: AuthService,
private alertCntrl: AlertDialogFactory,
private savedLocationService: SavedLocationService) {
this.token = localStorage.getItem('x-access-token');
this.userID = localStorage.getItem('userID');
this.preGenData = navParams.get('preGenData');
localStorage.setItem("additionalInfoText", "");
// setTimeout(() => {
// this.inputFieldValue = 'New Value';
// }, 3000)
}
ionViewDidLoad() {
console.log('ionViewDidLoad IonicNativeMapPage');
// this.loadMap();
setTimeout(() => {
this.loadMap();
}, 500);
}
ngAfterViewInit(){
console.log("ngAfterViewInit", this.newLocation);
// this.androidPermissions.checkPermission(this.androidPermissions.PERMISSION.ACCESS_FINE_LOCATION).then(
// success => console.log('Permission granted'),
// err => this.androidPermissions.requestPermissions(this.androidPermissions.PERMISSION.ACCESS_FINE_LOCATION)
// );
// this.geolocation.getCurrentPosition().then((resp) => {
// console.log(resp.coords.latitude);
// console.log(resp.coords.longitude);
// }).catch((error) => {
// console.log('Error getting location', error);
// });
// this.platform.ready().then(() => {
// // this.loadMap();
// });
this.listenToSearchInput();
this.getMapLocation(location, this.latLng);
}
listenToSearchInput() {
this.hide = false;
let location: string;
console.log('location1:', location)
// let searchInput$ = Observable.fromEvent(this.button.nativeElement, 'keyup')
// .map(e => location = e['srcElement'].value.trim())
// .distinctUntilChanged()
// .switchMap(() => this.mapService.getJSON(location, this.latLng))
// searchInput$.subscribe(location => {
// this.available_locations = location;
// console.log(this.available_locations);
// })
}
getMapLocation(location, latLng) {
if (location) {
// let location$ = this.mapService.getJSON(location, this.latLng);
// location$.subscribe(res => console.log)
}
}
savedButtonClicked(myEvent) {
this.saved = this.saved ? false : true;
setTimeout(()=>{
this.saved = this.saved ? false : true;
}, 200);
let inputs;
this.addressResponse = inputs;
let URL = globalVars.getUsersAddress(this.userID);
this.authService.getCall(URL).
subscribe(res => {
console.log(JSON.parse(res["_body"]));
inputs = JSON.parse(res["_body"])["data"]["contact"]["address"];
console.log(inputs);
this.addressResponse = inputs;
// let result = this.alertCntrl.checkBoxAlertDialog("Saved Locations", inputs)
// console.log(result);
this.radioAlertDialog("Saved Locations", inputs)
})
}
radioAlertDialog(title: string, inputs){
this.map.setClickable(false);
let alert = this.alertCtrl.create({
title: title,
cssClass: 'alertTop'
});
inputs.forEach(input => {
alert.addInput({
type: 'radio',
label: input.alias,
value: input,
checked: false
});
});
alert.addButton({
text: 'Cancel',
handler: () => {
console.log('Cancel clicked.');
}
});
alert.addButton({
text: 'Okay',
handler: data => {
console.log('Radio data:', data);
// this.testCheckboxOpen = false;
// this.testCheckboxResult = data;
this.locationClicked(data);
}
});
alert.present();
alert.onDidDismiss((data) => {
console.log('OnDidDismiss', data);
// dataReturned = data;
this.map.setClickable(true);
return data || 'null';
});
}
saveButtonClicked() {
this.save = this.save ? false : true;
setTimeout(()=>{
this.save = this.save ? false : true;
}, 200);
console.log("saveButtonClicked");
let userID = localStorage.getItem("userID");
let URL = globalVars.UserAddress(userID);
// console.log(locationExists);
let data = {
alias: this.locationAlias,
address: this.address,
lat: this.lat,
long: this.lng
}
if(this.validate()){
// let locationExists: boolean = false;
// this.addressResponse.forEach(address => {
// locationExists = locationExists || (address.alias == this.locationAlias);
// console.log(address.alias, this.locationAlias);
// console.log(address.alias == this.locationAlias);
// });
// console.log('location Exists: ', locationExists);
// if(!locationExists){
this.authService.patchCall(URL, data)
.subscribe(res => {
if (res.status == 200) {
console.log(res['_body']);
}
});
}else{
this.map.setClickable(false);
let result = this.alertCntrl.openAlertDialog('Location exits', 'Please enter a location.');
// result.then(value => {
// if(value){
// this.map.setClickable(true);
// }
// })
}
// }else{
// // this.alertCntrl.openAlertDialog('Error', 'Location already Exists.')
// }
}
openAdditionalNoteDialog(myEvent) {
this.map.setClickable(false);
this.isModalVisible = this.isModalVisible ? false : true;
setTimeout(() => {
this.isModalVisible = this.isModalVisible ? false : true;
}, 200);
let popover = this.popoverCtrl.create(AdditionalNote, {}, { showBackdrop: true });
popover.present({
ev: myEvent
});
popover.onDidDismiss(data => {
if(data){
this.map.setClickable(true);
console.log(data);
this.additionalInfoText = data + "\n";
localStorage.setItem("additionalInfoText", this.additionalInfoText);
}
})
}
additionButtonClicked(myEvent) {
this.addition = this.addition ? false : true;
console.log("additionButtonClicked");
this.openAdditionalNoteDialog(myEvent);
}
locationClickedBool;
locationClicked(location) {
console.log("You have clicked on: ", location);
this.locationClickedBool = false;
this.hide = true;
if(!!location){
this.inputFieldValue = '';
if(!!location.name){
this.locationClickedBool = true;
console.log(location);
this.inputFieldValue = location.name || '';
localStorage.setItem("Location", JSON.stringify(location));
this.lat = location.geometry.location.lat;
this.lng = location.geometry.location.lng;
this.address = location.formatted_address;
this.locationAlias = location.name;
}else{
console.log('Here');
this.locationClickedBool = true;
this.inputFieldValue = location.alias || '';
localStorage.setItem("Location", JSON.stringify(location));
this.lat = location.lat;
this.lng = location.long;
this.address = location.address;
this.locationAlias = location.alias;
};
setTimeout(() => { this.available_locations = []}, 200);
}else{
console.log('Here');
this.locationClickedBool = true;
this.inputFieldValue = location.alias || '';
localStorage.setItem("Location", JSON.stringify(location));
this.lat = location.lat;
this.lng = location.long;
this.address = location.address;
this.locationAlias = location.alias;
};
setTimeout(() => { this.available_locations = []}, 200);
//gMap = new google.maps.Map(document.getElementById('map'));
// this.postion = new google.maps.LatLng(this.lat, this.lng);
// this.map.setCenter(this.postion);
// this.addMarkerMoveCamera(this.map, new LatLng(this.lat, this.lng));
// this.map.center = new google.maps.LatLng(this.lat, this.lng);
this.addMarker(this.map, new LatLng(this.lat, this.lng));
this.moveCamera(this.map, new LatLng(this.lat, this.lng));
}
validate():boolean
{
return (this.lat != null && this.lng != null && this.address != null) ? true :false;
}
startNextScreen() {
console.log("Next clicked!");
let valid:boolean = this.validate();
console.log(valid);
if(valid === true && this.locationClickedBool == true)
{
console.log(this.preGenData);
this.navCtrl.push(LaundryItems, {
preGenData: this.preGenData,
pickupDetails: {
location: {
lat: this.lat,
lng: this.lng,
address: this.address
}
},
});
}
else{
this.map.setClickable(false);
this.alertCntrl.openAlertDialog("What's missing?","No location selected.");
}
}
| loadMap | identifier_name |
|
ionic-native-map.ts | available_locations: Array<Object> = [];
newLocation;
marker;
constructor(public navCtrl: NavController,
public navParams: NavParams,
private googleMaps: GoogleMaps,
private platform: Platform,
private geolocation: Geolocation,
private androidPermissions: AndroidPermissions,
private alertCtrl: AlertController,
private popoverCtrl: PopoverController,
private mapService: MapService,
private authService: AuthService,
private alertCntrl: AlertDialogFactory,
private savedLocationService: SavedLocationService) {
this.token = localStorage.getItem('x-access-token');
this.userID = localStorage.getItem('userID');
this.preGenData = navParams.get('preGenData');
localStorage.setItem("additionalInfoText", "");
// setTimeout(() => {
// this.inputFieldValue = 'New Value';
// }, 3000)
}
ionViewDidLoad() {
console.log('ionViewDidLoad IonicNativeMapPage');
// this.loadMap();
setTimeout(() => {
this.loadMap();
}, 500);
}
ngAfterViewInit(){
console.log("ngAfterViewInit", this.newLocation);
// this.androidPermissions.checkPermission(this.androidPermissions.PERMISSION.ACCESS_FINE_LOCATION).then(
// success => console.log('Permission granted'),
// err => this.androidPermissions.requestPermissions(this.androidPermissions.PERMISSION.ACCESS_FINE_LOCATION)
// );
// this.geolocation.getCurrentPosition().then((resp) => {
// console.log(resp.coords.latitude);
// console.log(resp.coords.longitude);
// }).catch((error) => {
// console.log('Error getting location', error);
// });
// this.platform.ready().then(() => {
// // this.loadMap();
// });
this.listenToSearchInput();
this.getMapLocation(location, this.latLng);
}
listenToSearchInput() {
this.hide = false;
let location: string;
console.log('location1:', location)
// let searchInput$ = Observable.fromEvent(this.button.nativeElement, 'keyup')
// .map(e => location = e['srcElement'].value.trim())
// .distinctUntilChanged()
// .switchMap(() => this.mapService.getJSON(location, this.latLng))
// searchInput$.subscribe(location => {
// this.available_locations = location;
// console.log(this.available_locations);
// })
}
getMapLocation(location, latLng) {
if (location) |
}
savedButtonClicked(myEvent) {
this.saved = this.saved ? false : true;
setTimeout(()=>{
this.saved = this.saved ? false : true;
}, 200);
let inputs;
this.addressResponse = inputs;
let URL = globalVars.getUsersAddress(this.userID);
this.authService.getCall(URL).
subscribe(res => {
console.log(JSON.parse(res["_body"]));
inputs = JSON.parse(res["_body"])["data"]["contact"]["address"];
console.log(inputs);
this.addressResponse = inputs;
// let result = this.alertCntrl.checkBoxAlertDialog("Saved Locations", inputs)
// console.log(result);
this.radioAlertDialog("Saved Locations", inputs)
})
}
radioAlertDialog(title: string, inputs){
this.map.setClickable(false);
let alert = this.alertCtrl.create({
title: title,
cssClass: 'alertTop'
});
inputs.forEach(input => {
alert.addInput({
type: 'radio',
label: input.alias,
value: input,
checked: false
});
});
alert.addButton({
text: 'Cancel',
handler: () => {
console.log('Cancel clicked.');
}
});
alert.addButton({
text: 'Okay',
handler: data => {
console.log('Radio data:', data);
// this.testCheckboxOpen = false;
// this.testCheckboxResult = data;
this.locationClicked(data);
}
});
alert.present();
alert.onDidDismiss((data) => {
console.log('OnDidDismiss', data);
// dataReturned = data;
this.map.setClickable(true);
return data || 'null';
});
}
saveButtonClicked() {
this.save = this.save ? false : true;
setTimeout(()=>{
this.save = this.save ? false : true;
}, 200);
console.log("saveButtonClicked");
let userID = localStorage.getItem("userID");
let URL = globalVars.UserAddress(userID);
// console.log(locationExists);
let data = {
alias: this.locationAlias,
address: this.address,
lat: this.lat,
long: this.lng
}
if(this.validate()){
// let locationExists: boolean = false;
// this.addressResponse.forEach(address => {
// locationExists = locationExists || (address.alias == this.locationAlias);
// console.log(address.alias, this.locationAlias);
// console.log(address.alias == this.locationAlias);
// });
// console.log('location Exists: ', locationExists);
// if(!locationExists){
this.authService.patchCall(URL, data)
.subscribe(res => {
if (res.status == 200) {
console.log(res['_body']);
}
});
}else{
this.map.setClickable(false);
let result = this.alertCntrl.openAlertDialog('Location exits', 'Please enter a location.');
// result.then(value => {
// if(value){
// this.map.setClickable(true);
// }
// })
}
// }else{
// // this.alertCntrl.openAlertDialog('Error', 'Location already Exists.')
// }
}
openAdditionalNoteDialog(myEvent) {
this.map.setClickable(false);
this.isModalVisible = this.isModalVisible ? false : true;
setTimeout(() => {
this.isModalVisible = this.isModalVisible ? false : true;
}, 200);
let popover = this.popoverCtrl.create(AdditionalNote, {}, { showBackdrop: true });
popover.present({
ev: myEvent
});
popover.onDidDismiss(data => {
if(data){
this.map.setClickable(true);
console.log(data);
this.additionalInfoText = data + "\n";
localStorage.setItem("additionalInfoText", this.additionalInfoText);
}
})
}
additionButtonClicked(myEvent) {
this.addition = this.addition ? false : true;
console.log("additionButtonClicked");
this.openAdditionalNoteDialog(myEvent);
}
locationClickedBool;
locationClicked(location) {
console.log("You have clicked on: ", location);
this.locationClickedBool = false;
this.hide = true;
if(!!location){
this.inputFieldValue = '';
if(!!location.name){
this.locationClickedBool = true;
console.log(location);
this.inputFieldValue = location.name || '';
localStorage.setItem("Location", JSON.stringify(location));
this.lat = location.geometry.location.lat;
this.lng = location.geometry.location.lng;
this.address = location.formatted_address;
this.locationAlias = location.name;
}else{
console.log('Here');
this.locationClickedBool = true;
this.inputFieldValue = location.alias || '';
localStorage.setItem("Location", JSON.stringify(location));
this.lat = location.lat;
this.lng = location.long;
this.address = location.address;
this.locationAlias = location.alias;
};
setTimeout(() => { this.available_locations = []}, 200);
}else{
console.log('Here');
this.locationClickedBool = true;
this.inputFieldValue = location.alias || '';
localStorage.setItem("Location", JSON.stringify(location));
this.lat = location.lat;
this.lng = location.long;
this.address = location.address;
this.locationAlias = location.alias;
};
setTimeout(() => { this.available_locations = []}, 200);
//gMap = new google.maps.Map(document.getElementById('map'));
// this.postion = new google.maps.LatLng(this.lat, this.lng);
// this.map.setCenter(this.postion);
// this.addMarkerMoveCamera(this.map, new LatLng(this.lat, this.lng));
// this.map.center = new google.maps.LatLng(this.lat, this.lng);
this.addMarker(this.map, new LatLng(this.lat, this.lng));
this.moveCamera(this.map, new LatLng(this.lat, this.lng));
}
validate():boolean
{
return (this.lat != null && this.lng != null && this.address != null) ? true :false;
}
startNextScreen() {
console.log("Next clicked!");
let valid:boolean = this.validate();
console.log(valid);
if(valid === true && this.locationClickedBool == true)
{
console.log(this.preGenData);
this.navCtrl.push(LaundryItems, {
preGenData: this.preGenData,
pickupDetails: {
location: {
lat: this.lat,
lng: this.lng,
address: this.address
}
},
});
}
else{ | {
// let location$ = this.mapService.getJSON(location, this.latLng);
// location$.subscribe(res => console.log)
} | conditional_block |
lib.rs | " BASIS,
// See the License for the specific language governing permissions and
// limitations under the License.
//
// Parts of this work are derived from the `protoc-rust-grpc` crate by
// Stepan Koltsov <[email protected]>.
//
// Copyright 2016, Stepan Koltsov <[email protected]>.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// See the License for the specific language governing permissions and
// limitations under the License.
#![deny(warnings)]
#![warn(missing_docs)]
//! An API for programmatically invoking the grpcio gRPC compiler in the same vein as the
//! [rust-protoc-grpc](https://crates.io/crates/protoc-rust-grpc) crate from Stepan Koltsov.
extern crate grpcio_compiler;
#[macro_use]
extern crate anyhow;
extern crate tempfile;
extern crate protobuf;
extern crate protobuf_codegen;
extern crate protoc;
use std::convert::AsRef;
use std::fs::File;
use std::io::{Read, Write};
use std::iter::Iterator;
use std::path::{Path, PathBuf};
use std::vec::Vec;
use anyhow::Context;
use tempfile::NamedTempFile;
use protobuf::{compiler_plugin, descriptor, Message};
use protobuf_codegen::Customize;
use protoc::{DescriptorSetOutArgs, Protoc};
/// Custom error type used throughout this crate.
pub type CompileError = ::anyhow::Error;
/// Custom result type used throughout this crate.
pub type CompileResult<T> = Result<T, CompileError>;
fn stringify_paths<Paths>(paths: Paths) -> CompileResult<Vec<String>>
where
Paths: IntoIterator,
Paths::Item: AsRef<Path>,
{
paths
.into_iter()
.map(|input| match input.as_ref().to_str() {
Some(s) => Ok(s.to_owned()),
None => Err(format_err!(
"failed to convert {:?} to string",
input.as_ref()
)),
})
.collect()
}
fn write_out_generated_files<P>(
generation_results: Vec<compiler_plugin::GenResult>,
output_dir: P,
) -> CompileResult<()>
where
P: AsRef<Path>,
{
for result in generation_results {
let file = output_dir.as_ref().join(result.name);
File::create(&file)
.context(format!("failed to create {:?}", &file))?
.write_all(&result.content)
.context(format!("failed to write {:?}", &file))?;
}
Ok(())
}
fn absolutize<P>(path: P) -> CompileResult<PathBuf>
where
P: AsRef<Path>,
{
let p = path.as_ref();
if p.is_relative() {
match std::env::current_dir() {
Ok(cwd) => Ok(cwd.join(p)),
Err(err) => Err(format_err!(
"Failed to determine CWD needed to absolutize a relative path: {:?}",
err
)),
}
} else {
Ok(PathBuf::from(p))
}
}
fn normalize<Paths, Bases>(
paths: Paths,
bases: Bases,
) -> CompileResult<(Vec<PathBuf>, Vec<PathBuf>, Vec<PathBuf>)>
where
Paths: IntoIterator,
Paths::Item: AsRef<Path>,
Bases: IntoIterator,
Bases::Item: AsRef<Path>,
{
let absolutized_bases = bases
.into_iter()
.map(absolutize)
.collect::<CompileResult<Vec<PathBuf>>>()?;
// We deal with the following cases:
// a.) absolute paths
// b.) paths relative to CWD
// c.) paths relative to bases
//
// We take the strategy of transforming the relative path cases (b & c) into absolute paths (a)
// and use the strip_prefix API from there.
let absolutized_paths = paths
.into_iter()
.map(|p| {
let rel_path = p.as_ref().to_path_buf();
let absolute_path = absolutize(&rel_path)?;
Ok((rel_path, absolute_path))
})
// TODO(John Sirois): Use `.flatten()` pending https://github.com/rust-lang/rust/issues/48213
.flat_map(|r: CompileResult<(PathBuf, PathBuf)>| r)
.map(|(rel_path, abs_path)| {
if abs_path.exists() {
// Case a or b.
Ok(abs_path)
} else {
// Case c.
for b in &absolutized_bases {
let absolutized_path = b.join(&rel_path);
if absolutized_path.exists() {
return Ok(absolutized_path);
}
}
Err(format_err!(
"Failed to find the absolute path of input {:?}",
rel_path
))
}
})
.collect::<CompileResult<Vec<PathBuf>>>()?;
let relativized_paths: Vec<PathBuf> = absolutized_paths
.iter()
.map(|p| {
for b in &absolutized_bases {
if let Ok(rel_path) = p.strip_prefix(&b) {
return Ok(PathBuf::from(rel_path));
}
}
Err(format_err!(
"The input path {:?} is not contained by any of the include paths {:?}",
p,
absolutized_bases
))
})
.collect::<CompileResult<Vec<PathBuf>>>()?;
Ok((absolutized_bases, absolutized_paths, relativized_paths))
}
/// Compiles a list a gRPC definitions to rust modules.
///
/// # Arguments
///
/// * `inputs` - A list of protobuf definition paths to compile. Paths can be specified as absolute,
/// relative to the CWD or relative to one of the `includes` paths. Note that the directory each
/// member of `inputs` is found under must be included in the `includes` parameter.
/// * `includes` - A list of of include directory paths to pass to `protoc`. Include paths can be
/// specified either as absolute or relative to the CWD. Note that the directory each member of
/// `inputs` is found under must be included in this parameter.
/// * `output` - Directory to place the generated rust modules into.
/// * `customizations` - An Option<protobuf_codegen::Customize> allowing customization options to be
/// passed to protobuf_codegen
pub fn compile_grpc_protos<Inputs, Includes, Output>(
inputs: Inputs,
includes: Includes,
output: Output,
customizations: Option<Customize>,
) -> CompileResult<()>
where
Inputs: IntoIterator,
Inputs::Item: AsRef<Path>,
Includes: IntoIterator,
Includes::Item: AsRef<Path>,
Output: AsRef<Path>,
{
let protoc = Protoc::from_env_path();
protoc
.check()
.context("failed to find `protoc`, `protoc` must be availabe in `PATH`")?;
let (absolutized_includes, absolutized_paths, relativized_inputs) =
normalize(inputs, includes)?;
let stringified_inputs_absolute = stringify_paths(absolutized_paths)?;
let stringified_inputs = stringify_paths(relativized_inputs)?;
let stringified_includes = stringify_paths(absolutized_includes)?;
let descriptor_set = NamedTempFile::new()?;
protoc
.write_descriptor_set(DescriptorSetOutArgs {
out: match descriptor_set.as_ref().to_str() {
Some(s) => s,
None => bail!("failed to convert descriptor set path to string"),
},
input: stringified_inputs_absolute
.iter()
.map(String::as_str)
.collect::<Vec<&str>>()
.as_slice(),
includes: stringified_includes
.iter()
.map(String::as_str)
.collect::<Vec<&str>>()
.as_slice(),
include_imports: true,
})
.context("failed to write descriptor set")?;
let mut serialized_descriptor_set = Vec::new();
File::open(&descriptor_set)
.context("failed to open descriptor set")?
.read_to_end(&mut serialized_descriptor_set)
.context("failed to read descriptor set")?;
let descriptor_set =
descriptor::FileDescriptorSet::parse_from_bytes(&serialized_descriptor_set)
.context("failed to parse descriptor set")?;
let customize = customizations.unwrap_or_default();
write_out_generated_files(
grpcio_compiler::codegen::gen(descriptor_set.get_file(), stringified_inputs.as_slice()),
&output,
)
.context("failed to write generated grpc definitions")?;
write_out_generated_files(
protobuf_codegen::gen(
descriptor_set.get_file(),
stringified_inputs.as_slice(),
&customize,
),
&output,
)
.context("failed to write out generated protobuf definitions")?;
Ok(())
}
#[cfg(test)]
mod tests {
use super::*;
use std::path::PathBuf;
use tempfile::tempdir;
fn | <Input, Output>(input: Input, expected_outputs: Output)
where
Input: AsRef<Path>,
| assert_compile_grpc_protos | identifier_name |
lib.rs | " BASIS,
// See the License for the specific language governing permissions and
// limitations under the License.
//
// Parts of this work are derived from the `protoc-rust-grpc` crate by
// Stepan Koltsov <[email protected]>.
//
// Copyright 2016, Stepan Koltsov <[email protected]>.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// See the License for the specific language governing permissions and
// limitations under the License.
#![deny(warnings)]
#![warn(missing_docs)]
//! An API for programmatically invoking the grpcio gRPC compiler in the same vein as the
//! [rust-protoc-grpc](https://crates.io/crates/protoc-rust-grpc) crate from Stepan Koltsov.
extern crate grpcio_compiler;
#[macro_use]
extern crate anyhow;
extern crate tempfile;
extern crate protobuf;
extern crate protobuf_codegen;
extern crate protoc;
use std::convert::AsRef;
use std::fs::File;
use std::io::{Read, Write};
use std::iter::Iterator;
use std::path::{Path, PathBuf};
use std::vec::Vec;
use anyhow::Context;
use tempfile::NamedTempFile;
use protobuf::{compiler_plugin, descriptor, Message};
use protobuf_codegen::Customize;
use protoc::{DescriptorSetOutArgs, Protoc};
/// Custom error type used throughout this crate.
pub type CompileError = ::anyhow::Error;
/// Custom result type used throughout this crate.
pub type CompileResult<T> = Result<T, CompileError>;
fn stringify_paths<Paths>(paths: Paths) -> CompileResult<Vec<String>>
where
Paths: IntoIterator,
Paths::Item: AsRef<Path>,
{
paths
.into_iter()
.map(|input| match input.as_ref().to_str() {
Some(s) => Ok(s.to_owned()),
None => Err(format_err!(
"failed to convert {:?} to string",
input.as_ref()
)),
})
.collect()
}
fn write_out_generated_files<P>(
generation_results: Vec<compiler_plugin::GenResult>,
output_dir: P,
) -> CompileResult<()>
where
P: AsRef<Path>,
{
for result in generation_results {
let file = output_dir.as_ref().join(result.name);
File::create(&file)
.context(format!("failed to create {:?}", &file))?
.write_all(&result.content)
.context(format!("failed to write {:?}", &file))?;
}
Ok(())
}
fn absolutize<P>(path: P) -> CompileResult<PathBuf>
where
P: AsRef<Path>,
|
fn normalize<Paths, Bases>(
paths: Paths,
bases: Bases,
) -> CompileResult<(Vec<PathBuf>, Vec<PathBuf>, Vec<PathBuf>)>
where
Paths: IntoIterator,
Paths::Item: AsRef<Path>,
Bases: IntoIterator,
Bases::Item: AsRef<Path>,
{
let absolutized_bases = bases
.into_iter()
.map(absolutize)
.collect::<CompileResult<Vec<PathBuf>>>()?;
// We deal with the following cases:
// a.) absolute paths
// b.) paths relative to CWD
// c.) paths relative to bases
//
// We take the strategy of transforming the relative path cases (b & c) into absolute paths (a)
// and use the strip_prefix API from there.
let absolutized_paths = paths
.into_iter()
.map(|p| {
let rel_path = p.as_ref().to_path_buf();
let absolute_path = absolutize(&rel_path)?;
Ok((rel_path, absolute_path))
})
// TODO(John Sirois): Use `.flatten()` pending https://github.com/rust-lang/rust/issues/48213
.flat_map(|r: CompileResult<(PathBuf, PathBuf)>| r)
.map(|(rel_path, abs_path)| {
if abs_path.exists() {
// Case a or b.
Ok(abs_path)
} else {
// Case c.
for b in &absolutized_bases {
let absolutized_path = b.join(&rel_path);
if absolutized_path.exists() {
return Ok(absolutized_path);
}
}
Err(format_err!(
"Failed to find the absolute path of input {:?}",
rel_path
))
}
})
.collect::<CompileResult<Vec<PathBuf>>>()?;
let relativized_paths: Vec<PathBuf> = absolutized_paths
.iter()
.map(|p| {
for b in &absolutized_bases {
if let Ok(rel_path) = p.strip_prefix(&b) {
return Ok(PathBuf::from(rel_path));
}
}
Err(format_err!(
"The input path {:?} is not contained by any of the include paths {:?}",
p,
absolutized_bases
))
})
.collect::<CompileResult<Vec<PathBuf>>>()?;
Ok((absolutized_bases, absolutized_paths, relativized_paths))
}
/// Compiles a list a gRPC definitions to rust modules.
///
/// # Arguments
///
/// * `inputs` - A list of protobuf definition paths to compile. Paths can be specified as absolute,
/// relative to the CWD or relative to one of the `includes` paths. Note that the directory each
/// member of `inputs` is found under must be included in the `includes` parameter.
/// * `includes` - A list of of include directory paths to pass to `protoc`. Include paths can be
/// specified either as absolute or relative to the CWD. Note that the directory each member of
/// `inputs` is found under must be included in this parameter.
/// * `output` - Directory to place the generated rust modules into.
/// * `customizations` - An Option<protobuf_codegen::Customize> allowing customization options to be
/// passed to protobuf_codegen
pub fn compile_grpc_protos<Inputs, Includes, Output>(
inputs: Inputs,
includes: Includes,
output: Output,
customizations: Option<Customize>,
) -> CompileResult<()>
where
Inputs: IntoIterator,
Inputs::Item: AsRef<Path>,
Includes: IntoIterator,
Includes::Item: AsRef<Path>,
Output: AsRef<Path>,
{
let protoc = Protoc::from_env_path();
protoc
.check()
.context("failed to find `protoc`, `protoc` must be availabe in `PATH`")?;
let (absolutized_includes, absolutized_paths, relativized_inputs) =
normalize(inputs, includes)?;
let stringified_inputs_absolute = stringify_paths(absolutized_paths)?;
let stringified_inputs = stringify_paths(relativized_inputs)?;
let stringified_includes = stringify_paths(absolutized_includes)?;
let descriptor_set = NamedTempFile::new()?;
protoc
.write_descriptor_set(DescriptorSetOutArgs {
out: match descriptor_set.as_ref().to_str() {
Some(s) => s,
None => bail!("failed to convert descriptor set path to string"),
},
input: stringified_inputs_absolute
.iter()
.map(String::as_str)
.collect::<Vec<&str>>()
.as_slice(),
includes: stringified_includes
.iter()
.map(String::as_str)
.collect::<Vec<&str>>()
.as_slice(),
include_imports: true,
})
.context("failed to write descriptor set")?;
let mut serialized_descriptor_set = Vec::new();
File::open(&descriptor_set)
.context("failed to open descriptor set")?
.read_to_end(&mut serialized_descriptor_set)
.context("failed to read descriptor set")?;
let descriptor_set =
descriptor::FileDescriptorSet::parse_from_bytes(&serialized_descriptor_set)
.context("failed to parse descriptor set")?;
let customize = customizations.unwrap_or_default();
write_out_generated_files(
grpcio_compiler::codegen::gen(descriptor_set.get_file(), stringified_inputs.as_slice()),
&output,
)
.context("failed to write generated grpc definitions")?;
write_out_generated_files(
protobuf_codegen::gen(
descriptor_set.get_file(),
stringified_inputs.as_slice(),
&customize,
),
&output,
)
.context("failed to write out generated protobuf definitions")?;
Ok(())
}
#[cfg(test)]
mod tests {
use super::*;
use std::path::PathBuf;
use tempfile::tempdir;
fn assert_compile_grpc_protos<Input, Output>(input: Input, expected_outputs: Output)
where
Input: AsRef<Path>,
| {
let p = path.as_ref();
if p.is_relative() {
match std::env::current_dir() {
Ok(cwd) => Ok(cwd.join(p)),
Err(err) => Err(format_err!(
"Failed to determine CWD needed to absolutize a relative path: {:?}",
err
)),
}
} else {
Ok(PathBuf::from(p))
}
} | identifier_body |
lib.rs | IS" BASIS,
// See the License for the specific language governing permissions and
// limitations under the License.
//
// Parts of this work are derived from the `protoc-rust-grpc` crate by
// Stepan Koltsov <[email protected]>.
//
// Copyright 2016, Stepan Koltsov <[email protected]>.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// See the License for the specific language governing permissions and
// limitations under the License.
#![deny(warnings)]
#![warn(missing_docs)]
//! An API for programmatically invoking the grpcio gRPC compiler in the same vein as the
//! [rust-protoc-grpc](https://crates.io/crates/protoc-rust-grpc) crate from Stepan Koltsov.
extern crate grpcio_compiler;
#[macro_use]
extern crate anyhow;
extern crate tempfile;
extern crate protobuf;
extern crate protobuf_codegen;
extern crate protoc;
use std::convert::AsRef;
use std::fs::File;
use std::io::{Read, Write};
use std::iter::Iterator;
use std::path::{Path, PathBuf};
use std::vec::Vec;
use anyhow::Context;
use tempfile::NamedTempFile;
use protobuf::{compiler_plugin, descriptor, Message};
use protobuf_codegen::Customize;
use protoc::{DescriptorSetOutArgs, Protoc};
/// Custom error type used throughout this crate.
pub type CompileError = ::anyhow::Error;
/// Custom result type used throughout this crate.
pub type CompileResult<T> = Result<T, CompileError>;
fn stringify_paths<Paths>(paths: Paths) -> CompileResult<Vec<String>>
where
Paths: IntoIterator,
Paths::Item: AsRef<Path>,
{
paths
.into_iter()
.map(|input| match input.as_ref().to_str() {
Some(s) => Ok(s.to_owned()),
None => Err(format_err!(
"failed to convert {:?} to string",
input.as_ref()
)),
})
.collect()
}
fn write_out_generated_files<P>(
generation_results: Vec<compiler_plugin::GenResult>,
output_dir: P,
) -> CompileResult<()>
where
P: AsRef<Path>,
{
for result in generation_results {
let file = output_dir.as_ref().join(result.name);
File::create(&file)
.context(format!("failed to create {:?}", &file))?
.write_all(&result.content)
.context(format!("failed to write {:?}", &file))?;
}
Ok(())
}
fn absolutize<P>(path: P) -> CompileResult<PathBuf>
where
P: AsRef<Path>,
{
let p = path.as_ref();
if p.is_relative() {
match std::env::current_dir() {
Ok(cwd) => Ok(cwd.join(p)),
Err(err) => Err(format_err!(
"Failed to determine CWD needed to absolutize a relative path: {:?}",
err
)),
}
} else {
Ok(PathBuf::from(p))
}
}
fn normalize<Paths, Bases>(
paths: Paths,
bases: Bases,
) -> CompileResult<(Vec<PathBuf>, Vec<PathBuf>, Vec<PathBuf>)>
where
Paths: IntoIterator,
Paths::Item: AsRef<Path>,
Bases: IntoIterator,
Bases::Item: AsRef<Path>,
{
let absolutized_bases = bases
.into_iter()
.map(absolutize)
.collect::<CompileResult<Vec<PathBuf>>>()?;
// We deal with the following cases:
// a.) absolute paths
// b.) paths relative to CWD
// c.) paths relative to bases
//
// We take the strategy of transforming the relative path cases (b & c) into absolute paths (a)
// and use the strip_prefix API from there.
let absolutized_paths = paths
.into_iter()
.map(|p| {
let rel_path = p.as_ref().to_path_buf();
let absolute_path = absolutize(&rel_path)?;
Ok((rel_path, absolute_path))
})
// TODO(John Sirois): Use `.flatten()` pending https://github.com/rust-lang/rust/issues/48213
.flat_map(|r: CompileResult<(PathBuf, PathBuf)>| r)
.map(|(rel_path, abs_path)| {
if abs_path.exists() {
// Case a or b.
Ok(abs_path)
} else {
// Case c.
for b in &absolutized_bases {
let absolutized_path = b.join(&rel_path);
if absolutized_path.exists() {
return Ok(absolutized_path);
}
}
Err(format_err!(
"Failed to find the absolute path of input {:?}",
rel_path
))
}
})
.collect::<CompileResult<Vec<PathBuf>>>()?;
let relativized_paths: Vec<PathBuf> = absolutized_paths
.iter()
.map(|p| {
for b in &absolutized_bases {
if let Ok(rel_path) = p.strip_prefix(&b) {
return Ok(PathBuf::from(rel_path));
}
}
Err(format_err!(
"The input path {:?} is not contained by any of the include paths {:?}",
p,
absolutized_bases
))
})
.collect::<CompileResult<Vec<PathBuf>>>()?;
Ok((absolutized_bases, absolutized_paths, relativized_paths))
}
/// Compiles a list a gRPC definitions to rust modules.
///
/// # Arguments
///
/// * `inputs` - A list of protobuf definition paths to compile. Paths can be specified as absolute,
/// relative to the CWD or relative to one of the `includes` paths. Note that the directory each
/// member of `inputs` is found under must be included in the `includes` parameter.
/// * `includes` - A list of of include directory paths to pass to `protoc`. Include paths can be
/// specified either as absolute or relative to the CWD. Note that the directory each member of
/// `inputs` is found under must be included in this parameter.
/// * `output` - Directory to place the generated rust modules into.
/// * `customizations` - An Option<protobuf_codegen::Customize> allowing customization options to be
/// passed to protobuf_codegen
pub fn compile_grpc_protos<Inputs, Includes, Output>(
inputs: Inputs,
includes: Includes,
output: Output,
customizations: Option<Customize>,
) -> CompileResult<()>
where
Inputs: IntoIterator,
Inputs::Item: AsRef<Path>,
Includes: IntoIterator,
Includes::Item: AsRef<Path>,
Output: AsRef<Path>,
{
let protoc = Protoc::from_env_path();
protoc
.check()
.context("failed to find `protoc`, `protoc` must be availabe in `PATH`")?;
let (absolutized_includes, absolutized_paths, relativized_inputs) =
normalize(inputs, includes)?;
let stringified_inputs_absolute = stringify_paths(absolutized_paths)?;
let stringified_inputs = stringify_paths(relativized_inputs)?;
let stringified_includes = stringify_paths(absolutized_includes)?;
let descriptor_set = NamedTempFile::new()?;
protoc
.write_descriptor_set(DescriptorSetOutArgs {
out: match descriptor_set.as_ref().to_str() {
Some(s) => s,
None => bail!("failed to convert descriptor set path to string"),
},
input: stringified_inputs_absolute | .as_slice(),
includes: stringified_includes
.iter()
.map(String::as_str)
.collect::<Vec<&str>>()
.as_slice(),
include_imports: true,
})
.context("failed to write descriptor set")?;
let mut serialized_descriptor_set = Vec::new();
File::open(&descriptor_set)
.context("failed to open descriptor set")?
.read_to_end(&mut serialized_descriptor_set)
.context("failed to read descriptor set")?;
let descriptor_set =
descriptor::FileDescriptorSet::parse_from_bytes(&serialized_descriptor_set)
.context("failed to parse descriptor set")?;
let customize = customizations.unwrap_or_default();
write_out_generated_files(
grpcio_compiler::codegen::gen(descriptor_set.get_file(), stringified_inputs.as_slice()),
&output,
)
.context("failed to write generated grpc definitions")?;
write_out_generated_files(
protobuf_codegen::gen(
descriptor_set.get_file(),
stringified_inputs.as_slice(),
&customize,
),
&output,
)
.context("failed to write out generated protobuf definitions")?;
Ok(())
}
#[cfg(test)]
mod tests {
use super::*;
use std::path::PathBuf;
use tempfile::tempdir;
fn assert_compile_grpc_protos<Input, Output>(input: Input, expected_outputs: Output)
where
Input: AsRef<Path>,
Output | .iter()
.map(String::as_str)
.collect::<Vec<&str>>() | random_line_split |
baselines_utils.py | class SetFromFlat(object):
def __init__(self, var_list, dtype=tf.float32):
assigns = []
shapes = list(map(var_shape, var_list))
total_size = np.sum([intprod(shape) for shape in shapes])
self.theta = theta = tf.placeholder(dtype, [total_size])
start = 0
assigns = []
for (shape, v) in zip(shapes, var_list):
size = intprod(shape)
assigns.append(tf.assign(v, tf.reshape(theta[start:start + size], shape)))
start += size
self.op = tf.group(*assigns)
def __call__(self, theta):
tf.get_default_session().run(self.op, feed_dict={self.theta: theta})
class GetFlat(object):
def __init__(self, var_list):
self.op = tf.concat(axis=0, values=[tf.reshape(v, [numel(v)]) for v in var_list])
def __call__(self):
return tf.get_default_session().run(self.op)
_PLACEHOLDER_CACHE = {} # name -> (placeholder, dtype, shape)
def get_placeholder(name, dtype, shape):
if name in _PLACEHOLDER_CACHE:
out, dtype1, shape1 = _PLACEHOLDER_CACHE[name]
assert dtype1 == dtype and shape1 == shape
return out
else:
out = tf.placeholder(dtype=dtype, shape=shape, name=name)
_PLACEHOLDER_CACHE[name] = (out, dtype, shape)
return out
def get_placeholder_cached(name):
return _PLACEHOLDER_CACHE[name][0]
def flattenallbut0(x):
return tf.reshape(x, [-1, intprod(x.get_shape().as_list()[1:])])
# ================================================================
# Diagnostics
# ================================================================
def display_var_info(vars):
from baselines import logger
count_params = 0
for v in vars:
name = v.name
if "/Adam" in name or "beta1_power" in name or "beta2_power" in name: continue
count_params += np.prod(v.shape.as_list())
if "/b:" in name: continue # Wx+b, bias is not interesting to look at => count params, but not print
logger.info(" %s%s%s" % (name, " "*(55-len(name)), str(v.shape)))
logger.info("Total model parameters: %0.1f million" % (count_params*1e-6))
class LinearSchedule(object):
def __init__(self, schedule_timesteps, final_p, initial_p=1.0):
"""Linear interpolation between initial_p and final_p over
schedule_timesteps. After this many timesteps pass final_p is
returned.
Parameters
----------
schedule_timesteps: int
Number of timesteps for which to linearly anneal initial_p
to final_p
initial_p: float
initial output value
final_p: float
final output value
"""
self.schedule_timesteps = schedule_timesteps
self.final_p = final_p
self.initial_p = initial_p
def value(self, t):
"""See Schedule.value"""
fraction = min(float(t) / self.schedule_timesteps, 1.0)
return self.initial_p + fraction * (self.final_p - self.initial_p)
# deep q
import os
import tensorflow as tf
# ================================================================
# Saving variables
# ================================================================
def load_state(fname):
saver = tf.train.Saver()
saver.restore(tf.get_default_session(), fname)
def save_state(fname):
os.makedirs(os.path.dirname(fname), exist_ok=True)
saver = tf.train.Saver()
saver.save(tf.get_default_session(), fname)
# ================================================================
# Placeholders
# ================================================================
class TfInput(object):
def __init__(self, name="(unnamed)"):
"""Generalized Tensorflow placeholder. The main differences are:
- possibly uses multiple placeholders internally and returns multiple values
- can apply light postprocessing to the value feed to placeholder.
"""
self.name = name
def get(self):
"""Return the tf variable(s) representing the possibly postprocessed value
of placeholder(s).
"""
raise NotImplemented()
def make_feed_dict(data):
"""Given data input it to the placeholder(s)."""
raise NotImplemented()
class PlaceholderTfInput(TfInput):
def __init__(self, placeholder):
"""Wrapper for regular tensorflow placeholder."""
super().__init__(placeholder.name)
self._placeholder = placeholder
def get(self):
return self._placeholder
def make_feed_dict(self, data):
return {self._placeholder: data}
class BatchInput(PlaceholderTfInput):
def __init__(self, shape, dtype=tf.float32, name=None):
"""Creates a placeholder for a batch of tensors of a given shape and dtype
Parameters
----------
shape: [int]
shape of a single elemenet of the batch
dtype: tf.dtype
number representation used for tensor contents
name: str
name of the underlying placeholder
"""
super().__init__(tf.placeholder(dtype, [None] + list(shape), name=name))
class Uint8Input(PlaceholderTfInput):
def __init__(self, shape, name=None):
"""Takes input in uint8 format which is cast to float32 and divided by 255
before passing it to the model.
On GPU this ensures lower data transfer times.
Parameters
----------
shape: [int]
shape of the tensor.
name: str
name of the underlying placeholder
"""
super().__init__(tf.placeholder(tf.uint8, [None] + list(shape), name=name))
self._shape = shape
self._output = tf.cast(super().get(), tf.float32) / 255.0
def get(self):
return self._output
# Tree
import operator
class SegmentTree(object):
def __init__(self, capacity, operation, neutral_element):
"""Build a Segment Tree data structure.
https://en.wikipedia.org/wiki/Segment_tree
Can be used as regular array, but with two
important differences:
a) setting item's value is slightly slower.
It is O(lg capacity) instead of O(1).
b) user has access to an efficient `reduce`
operation which reduces `operation` over
a contiguous subsequence of items in the
array.
Paramters
---------
capacity: int
Total size of the array - must be a power of two.
operation: lambda obj, obj -> obj
and operation for combining elements (eg. sum, max)
must for a mathematical group together with the set of
possible values for array elements.
neutral_element: obj
neutral element for the operation above. eg. float('-inf')
for max and 0 for sum.
"""
assert capacity > 0 and capacity & (capacity - 1) == 0, "capacity must be positive and a power of 2."
self._capacity = capacity
self._value = [neutral_element for _ in range(2 * capacity)]
self._operation = operation
def _reduce_helper(self, start, end, node, node_start, node_end):
if start == node_start and end == node_end:
return self._value[node]
mid = (node_start + node_end) // 2
if end <= mid:
return self._reduce_helper(start, end, 2 * node, node_start, mid)
else:
if mid + 1 <= start:
return self._reduce_helper(start, end, 2 * node + 1, mid + 1, node_end)
else:
return self._operation(
self._reduce_helper(start, mid, 2 * node, node_start, mid),
self._reduce_helper(mid + 1, end, 2 * node + 1, mid + 1, node_end)
)
def reduce(self, start=0, end=None):
"""Returns result of applying `self.operation`
to a contiguous subsequence of the array.
self.operation(arr[start], operation(arr[start+1], operation(... arr[end])))
Parameters
----------
start: int
beginning of the subsequence
end: int
end of the subsequences
Returns
-------
reduced: obj
result of reducing self.operation over the specified range of array elements.
"""
if end is None:
end = self._capacity
if end < 0:
end += self._capacity
end -= 1
return self._reduce_helper(start, end, 1, 0, self._capacity - 1)
def __setitem__(self, idx, val):
# index of the leaf
idx += self._capacity
self._value[idx] = val
idx //= 2
while idx >= 1:
self._value[idx] = self._operation(
self._value[2 * idx],
self._value[2 * idx + 1]
)
idx //= 2
def __getitem__(self, idx):
assert 0 <= idx < self._capacity
return self._value[self._capacity + idx]
class SumSegmentTree(SegmentTree):
def __init__(self, capacity):
| super(SumSegmentTree, self).__init__(
capacity=capacity,
operation=operator.add,
neutral_element=0.0
) | identifier_body |
|
baselines_utils.py | else:
feed_dict[inpt] = value
def __call__(self, *args):
assert len(args) <= len(self.inputs), "Too many arguments provided"
feed_dict = {}
# Update the args
for inpt, value in zip(self.inputs, args):
self._feed_input(feed_dict, inpt, value)
# Update feed dict with givens.
for inpt in self.givens:
feed_dict[inpt] = feed_dict.get(inpt, self.givens[inpt])
results = tf.get_default_session().run(self.outputs_update, feed_dict=feed_dict)[:-1]
return results
# ================================================================
# Flat vectors
# ================================================================
def var_shape(x):
out = x.get_shape().as_list()
assert all(isinstance(a, int) for a in out), \
"shape function assumes that shape is fully known"
return out
def numel(x):
return intprod(var_shape(x))
def intprod(x):
return int(np.prod(x))
def flatgrad(loss, var_list, clip_norm=None):
grads = tf.gradients(loss, var_list)
if clip_norm is not None:
grads = [tf.clip_by_norm(grad, clip_norm=clip_norm) for grad in grads]
return tf.concat(axis=0, values=[
tf.reshape(grad if grad is not None else tf.zeros_like(v), [numel(v)])
for (v, grad) in zip(var_list, grads)
])
class SetFromFlat(object):
def __init__(self, var_list, dtype=tf.float32):
assigns = []
shapes = list(map(var_shape, var_list))
total_size = np.sum([intprod(shape) for shape in shapes])
self.theta = theta = tf.placeholder(dtype, [total_size])
start = 0
assigns = []
for (shape, v) in zip(shapes, var_list):
size = intprod(shape)
assigns.append(tf.assign(v, tf.reshape(theta[start:start + size], shape)))
start += size
self.op = tf.group(*assigns)
def __call__(self, theta):
tf.get_default_session().run(self.op, feed_dict={self.theta: theta})
class GetFlat(object):
def __init__(self, var_list):
self.op = tf.concat(axis=0, values=[tf.reshape(v, [numel(v)]) for v in var_list])
def __call__(self):
return tf.get_default_session().run(self.op)
_PLACEHOLDER_CACHE = {} # name -> (placeholder, dtype, shape)
def get_placeholder(name, dtype, shape):
if name in _PLACEHOLDER_CACHE:
out, dtype1, shape1 = _PLACEHOLDER_CACHE[name]
assert dtype1 == dtype and shape1 == shape
return out
else:
out = tf.placeholder(dtype=dtype, shape=shape, name=name)
_PLACEHOLDER_CACHE[name] = (out, dtype, shape)
return out
def get_placeholder_cached(name):
return _PLACEHOLDER_CACHE[name][0]
def flattenallbut0(x):
return tf.reshape(x, [-1, intprod(x.get_shape().as_list()[1:])])
# ================================================================
# Diagnostics
# ================================================================
def display_var_info(vars):
from baselines import logger
count_params = 0
for v in vars:
name = v.name
if "/Adam" in name or "beta1_power" in name or "beta2_power" in name: continue
count_params += np.prod(v.shape.as_list())
if "/b:" in name: continue # Wx+b, bias is not interesting to look at => count params, but not print
logger.info(" %s%s%s" % (name, " "*(55-len(name)), str(v.shape)))
logger.info("Total model parameters: %0.1f million" % (count_params*1e-6))
class LinearSchedule(object):
def __init__(self, schedule_timesteps, final_p, initial_p=1.0):
"""Linear interpolation between initial_p and final_p over
schedule_timesteps. After this many timesteps pass final_p is
returned.
Parameters
----------
schedule_timesteps: int
Number of timesteps for which to linearly anneal initial_p
to final_p
initial_p: float
initial output value
final_p: float
final output value
"""
self.schedule_timesteps = schedule_timesteps
self.final_p = final_p
self.initial_p = initial_p
def value(self, t):
"""See Schedule.value"""
fraction = min(float(t) / self.schedule_timesteps, 1.0)
return self.initial_p + fraction * (self.final_p - self.initial_p)
# deep q
import os
import tensorflow as tf
# ================================================================
# Saving variables
# ================================================================
def load_state(fname):
saver = tf.train.Saver()
saver.restore(tf.get_default_session(), fname)
def save_state(fname):
os.makedirs(os.path.dirname(fname), exist_ok=True)
saver = tf.train.Saver()
saver.save(tf.get_default_session(), fname)
# ================================================================
# Placeholders
# ================================================================
class TfInput(object):
def __init__(self, name="(unnamed)"):
"""Generalized Tensorflow placeholder. The main differences are:
- possibly uses multiple placeholders internally and returns multiple values
- can apply light postprocessing to the value feed to placeholder.
"""
self.name = name
def get(self):
"""Return the tf variable(s) representing the possibly postprocessed value
of placeholder(s).
"""
raise NotImplemented()
def make_feed_dict(data):
"""Given data input it to the placeholder(s)."""
raise NotImplemented()
class PlaceholderTfInput(TfInput):
def __init__(self, placeholder):
"""Wrapper for regular tensorflow placeholder."""
super().__init__(placeholder.name)
self._placeholder = placeholder
def get(self):
return self._placeholder
def make_feed_dict(self, data):
return {self._placeholder: data}
class BatchInput(PlaceholderTfInput):
def __init__(self, shape, dtype=tf.float32, name=None):
"""Creates a placeholder for a batch of tensors of a given shape and dtype
Parameters
----------
shape: [int]
shape of a single elemenet of the batch
dtype: tf.dtype
number representation used for tensor contents
name: str
name of the underlying placeholder
"""
super().__init__(tf.placeholder(dtype, [None] + list(shape), name=name))
class Uint8Input(PlaceholderTfInput):
def __init__(self, shape, name=None):
"""Takes input in uint8 format which is cast to float32 and divided by 255
before passing it to the model.
On GPU this ensures lower data transfer times.
Parameters
----------
shape: [int]
shape of the tensor.
name: str
name of the underlying placeholder
"""
super().__init__(tf.placeholder(tf.uint8, [None] + list(shape), name=name))
self._shape = shape
self._output = tf.cast(super().get(), tf.float32) / 255.0
def get(self):
return self._output
# Tree
import operator
class SegmentTree(object):
def __init__(self, capacity, operation, neutral_element):
"""Build a Segment Tree data structure.
https://en.wikipedia.org/wiki/Segment_tree
Can be used as regular array, but with two
important differences:
a) setting item's value is slightly slower.
It is O(lg capacity) instead of O(1).
b) user has access to an efficient `reduce`
operation which reduces `operation` over
a contiguous subsequence of items in the
array.
Paramters
---------
capacity: int
Total size of the array - must be a power of two.
operation: lambda obj, obj -> obj
and operation for combining elements (eg. sum, max)
must for a mathematical group together with the set of
possible values for array elements.
neutral_element: obj
neutral element for the operation above. eg. float('-inf')
for max and 0 for sum.
"""
assert capacity > 0 and capacity & (capacity - 1) == 0, "capacity must be positive and a power of 2."
self._capacity = capacity
self._value = [neutral_element for _ in range(2 * capacity)]
self._operation = operation
def _reduce_helper(self, start, end, node, node_start, node_end):
if start == node_start and end == node_end:
return self._value[node]
mid = (node_start + node_end) // 2
if end <= mid:
return self._reduce_helper(start, end, 2 * node, node_start, mid)
else:
if mid + 1 <= start:
return self._reduce_helper(start, end, 2 * node + 1, mid + 1, node_end)
else:
return self._operation(
self._reduce_helper(start, mid, 2 * node, node_start, mid),
self._reduce_helper(mid + 1, end, 2 * node + 1, mid + 1, node_end) | )
def reduce(self, start=0, end=None): | random_line_split |
|
baselines_utils.py | summary_tag=None):
with tf.variable_scope(name):
stride_shape = [1, stride[0], stride[1], 1]
filter_shape = [filter_size[0], filter_size[1], int(x.get_shape()[3]), num_filters]
# there are "num input feature maps * filter height * filter width"
# inputs to each hidden unit
fan_in = intprod(filter_shape[:3])
# each unit in the lower layer receives a gradient from:
# "num output feature maps * filter height * filter width" /
# pooling size
fan_out = intprod(filter_shape[:2]) * num_filters
# initialize weights with random weights
w_bound = np.sqrt(6. / (fan_in + fan_out))
w = tf.get_variable("W", filter_shape, dtype, tf.random_uniform_initializer(-w_bound, w_bound),
collections=collections)
b = tf.get_variable("b", [1, 1, 1, num_filters], initializer=tf.zeros_initializer(),
collections=collections)
if summary_tag is not None:
tf.summary.image(summary_tag,
tf.transpose(tf.reshape(w, [filter_size[0], filter_size[1], -1, 1]),
[2, 0, 1, 3]),
max_images=10)
return tf.nn.conv2d(x, w, stride_shape, pad) + b
# ================================================================
# Theano-like Function
# ================================================================
def function(inputs, outputs, updates=None, givens=None):
"""Just like Theano function. Take a bunch of tensorflow placeholders and expressions
computed based on those placeholders and produces f(inputs) -> outputs. Function f takes
values to be fed to the input's placeholders and produces the values of the expressions
in outputs.
Input values can be passed in the same order as inputs or can be provided as kwargs based
on placeholder name (passed to constructor or accessible via placeholder.op.name).
Example:
x = tf.placeholder(tf.int32, (), name="x")
y = tf.placeholder(tf.int32, (), name="y")
z = 3 * x + 2 * y
lin = function([x, y], z, givens={y: 0})
with single_threaded_session():
initialize()
assert lin(2) == 6
assert lin(x=3) == 9
assert lin(2, 2) == 10
assert lin(x=2, y=3) == 12
Parameters
----------
inputs: [tf.placeholder, tf.constant, or object with make_feed_dict method]
list of input arguments
outputs: [tf.Variable] or tf.Variable
list of outputs or a single output to be returned from function. Returned
value will also have the same shape.
"""
if isinstance(outputs, list):
return _Function(inputs, outputs, updates, givens=givens)
elif isinstance(outputs, (dict, collections.OrderedDict)):
f = _Function(inputs, outputs.values(), updates, givens=givens)
return lambda *args, **kwargs: type(outputs)(zip(outputs.keys(), f(*args, **kwargs)))
else:
f = _Function(inputs, [outputs], updates, givens=givens)
return lambda *args, **kwargs: f(*args, **kwargs)[0]
class _Function(object):
def __init__(self, inputs, outputs, updates, givens):
for inpt in inputs:
if not hasattr(inpt, 'make_feed_dict') and not (type(inpt) is tf.Tensor and len(inpt.op.inputs) == 0):
assert False, "inputs should all be placeholders, constants, or have a make_feed_dict method"
self.inputs = inputs
updates = updates or []
self.update_group = tf.group(*updates)
self.outputs_update = list(outputs) + [self.update_group]
self.givens = {} if givens is None else givens
def _feed_input(self, feed_dict, inpt, value):
if hasattr(inpt, 'make_feed_dict'):
feed_dict.update(inpt.make_feed_dict(value))
else:
feed_dict[inpt] = value
def __call__(self, *args):
assert len(args) <= len(self.inputs), "Too many arguments provided"
feed_dict = {}
# Update the args
for inpt, value in zip(self.inputs, args):
self._feed_input(feed_dict, inpt, value)
# Update feed dict with givens.
for inpt in self.givens:
feed_dict[inpt] = feed_dict.get(inpt, self.givens[inpt])
results = tf.get_default_session().run(self.outputs_update, feed_dict=feed_dict)[:-1]
return results
# ================================================================
# Flat vectors
# ================================================================
def | (x):
out = x.get_shape().as_list()
assert all(isinstance(a, int) for a in out), \
"shape function assumes that shape is fully known"
return out
def numel(x):
return intprod(var_shape(x))
def intprod(x):
return int(np.prod(x))
def flatgrad(loss, var_list, clip_norm=None):
grads = tf.gradients(loss, var_list)
if clip_norm is not None:
grads = [tf.clip_by_norm(grad, clip_norm=clip_norm) for grad in grads]
return tf.concat(axis=0, values=[
tf.reshape(grad if grad is not None else tf.zeros_like(v), [numel(v)])
for (v, grad) in zip(var_list, grads)
])
class SetFromFlat(object):
def __init__(self, var_list, dtype=tf.float32):
assigns = []
shapes = list(map(var_shape, var_list))
total_size = np.sum([intprod(shape) for shape in shapes])
self.theta = theta = tf.placeholder(dtype, [total_size])
start = 0
assigns = []
for (shape, v) in zip(shapes, var_list):
size = intprod(shape)
assigns.append(tf.assign(v, tf.reshape(theta[start:start + size], shape)))
start += size
self.op = tf.group(*assigns)
def __call__(self, theta):
tf.get_default_session().run(self.op, feed_dict={self.theta: theta})
class GetFlat(object):
def __init__(self, var_list):
self.op = tf.concat(axis=0, values=[tf.reshape(v, [numel(v)]) for v in var_list])
def __call__(self):
return tf.get_default_session().run(self.op)
_PLACEHOLDER_CACHE = {} # name -> (placeholder, dtype, shape)
def get_placeholder(name, dtype, shape):
if name in _PLACEHOLDER_CACHE:
out, dtype1, shape1 = _PLACEHOLDER_CACHE[name]
assert dtype1 == dtype and shape1 == shape
return out
else:
out = tf.placeholder(dtype=dtype, shape=shape, name=name)
_PLACEHOLDER_CACHE[name] = (out, dtype, shape)
return out
def get_placeholder_cached(name):
return _PLACEHOLDER_CACHE[name][0]
def flattenallbut0(x):
return tf.reshape(x, [-1, intprod(x.get_shape().as_list()[1:])])
# ================================================================
# Diagnostics
# ================================================================
def display_var_info(vars):
from baselines import logger
count_params = 0
for v in vars:
name = v.name
if "/Adam" in name or "beta1_power" in name or "beta2_power" in name: continue
count_params += np.prod(v.shape.as_list())
if "/b:" in name: continue # Wx+b, bias is not interesting to look at => count params, but not print
logger.info(" %s%s%s" % (name, " "*(55-len(name)), str(v.shape)))
logger.info("Total model parameters: %0.1f million" % (count_params*1e-6))
class LinearSchedule(object):
def __init__(self, schedule_timesteps, final_p, initial_p=1.0):
"""Linear interpolation between initial_p and final_p over
schedule_timesteps. After this many timesteps pass final_p is
returned.
Parameters
----------
schedule_timesteps: int
Number of timesteps for which to linearly anneal initial_p
to final_p
initial_p: float
initial output value
final_p: float
final output value
"""
self.schedule_timesteps = schedule_timesteps
self.final_p = final_p
self.initial_p = initial_p
def value(self, t):
"""See Schedule.value"""
fraction = min(float(t) / self.schedule_timesteps, 1.0)
return self.initial_p + fraction * (self.final_p - self.initial_p)
# deep q
import os
import tensorflow as tf
# ================================================================
# Saving variables
# ================================================================
def load_state(fname):
saver = tf.train.Saver()
saver.restore(tf.get_default_session(), fname)
def save_state(fname):
os.makedirs(os.path.dirname(fname), exist_ok=True)
saver = tf.train.Saver()
saver.save(tf.get_default_session(), fname)
# ================================================================
# Placeholders
# ================================================================
class TfInput(object):
def __init__(self, name="(unnamed) | var_shape | identifier_name |
baselines_utils.py | summary_tag=None):
with tf.variable_scope(name):
stride_shape = [1, stride[0], stride[1], 1]
filter_shape = [filter_size[0], filter_size[1], int(x.get_shape()[3]), num_filters]
# there are "num input feature maps * filter height * filter width"
# inputs to each hidden unit
fan_in = intprod(filter_shape[:3])
# each unit in the lower layer receives a gradient from:
# "num output feature maps * filter height * filter width" /
# pooling size
fan_out = intprod(filter_shape[:2]) * num_filters
# initialize weights with random weights
w_bound = np.sqrt(6. / (fan_in + fan_out))
w = tf.get_variable("W", filter_shape, dtype, tf.random_uniform_initializer(-w_bound, w_bound),
collections=collections)
b = tf.get_variable("b", [1, 1, 1, num_filters], initializer=tf.zeros_initializer(),
collections=collections)
if summary_tag is not None:
tf.summary.image(summary_tag,
tf.transpose(tf.reshape(w, [filter_size[0], filter_size[1], -1, 1]),
[2, 0, 1, 3]),
max_images=10)
return tf.nn.conv2d(x, w, stride_shape, pad) + b
# ================================================================
# Theano-like Function
# ================================================================
def function(inputs, outputs, updates=None, givens=None):
"""Just like Theano function. Take a bunch of tensorflow placeholders and expressions
computed based on those placeholders and produces f(inputs) -> outputs. Function f takes
values to be fed to the input's placeholders and produces the values of the expressions
in outputs.
Input values can be passed in the same order as inputs or can be provided as kwargs based
on placeholder name (passed to constructor or accessible via placeholder.op.name).
Example:
x = tf.placeholder(tf.int32, (), name="x")
y = tf.placeholder(tf.int32, (), name="y")
z = 3 * x + 2 * y
lin = function([x, y], z, givens={y: 0})
with single_threaded_session():
initialize()
assert lin(2) == 6
assert lin(x=3) == 9
assert lin(2, 2) == 10
assert lin(x=2, y=3) == 12
Parameters
----------
inputs: [tf.placeholder, tf.constant, or object with make_feed_dict method]
list of input arguments
outputs: [tf.Variable] or tf.Variable
list of outputs or a single output to be returned from function. Returned
value will also have the same shape.
"""
if isinstance(outputs, list):
return _Function(inputs, outputs, updates, givens=givens)
elif isinstance(outputs, (dict, collections.OrderedDict)):
f = _Function(inputs, outputs.values(), updates, givens=givens)
return lambda *args, **kwargs: type(outputs)(zip(outputs.keys(), f(*args, **kwargs)))
else:
f = _Function(inputs, [outputs], updates, givens=givens)
return lambda *args, **kwargs: f(*args, **kwargs)[0]
class _Function(object):
def __init__(self, inputs, outputs, updates, givens):
for inpt in inputs:
if not hasattr(inpt, 'make_feed_dict') and not (type(inpt) is tf.Tensor and len(inpt.op.inputs) == 0):
assert False, "inputs should all be placeholders, constants, or have a make_feed_dict method"
self.inputs = inputs
updates = updates or []
self.update_group = tf.group(*updates)
self.outputs_update = list(outputs) + [self.update_group]
self.givens = {} if givens is None else givens
def _feed_input(self, feed_dict, inpt, value):
if hasattr(inpt, 'make_feed_dict'):
feed_dict.update(inpt.make_feed_dict(value))
else:
feed_dict[inpt] = value
def __call__(self, *args):
assert len(args) <= len(self.inputs), "Too many arguments provided"
feed_dict = {}
# Update the args
for inpt, value in zip(self.inputs, args):
self._feed_input(feed_dict, inpt, value)
# Update feed dict with givens.
for inpt in self.givens:
feed_dict[inpt] = feed_dict.get(inpt, self.givens[inpt])
results = tf.get_default_session().run(self.outputs_update, feed_dict=feed_dict)[:-1]
return results
# ================================================================
# Flat vectors
# ================================================================
def var_shape(x):
out = x.get_shape().as_list()
assert all(isinstance(a, int) for a in out), \
"shape function assumes that shape is fully known"
return out
def numel(x):
return intprod(var_shape(x))
def intprod(x):
return int(np.prod(x))
def flatgrad(loss, var_list, clip_norm=None):
grads = tf.gradients(loss, var_list)
if clip_norm is not None:
grads = [tf.clip_by_norm(grad, clip_norm=clip_norm) for grad in grads]
return tf.concat(axis=0, values=[
tf.reshape(grad if grad is not None else tf.zeros_like(v), [numel(v)])
for (v, grad) in zip(var_list, grads)
])
class SetFromFlat(object):
def __init__(self, var_list, dtype=tf.float32):
assigns = []
shapes = list(map(var_shape, var_list))
total_size = np.sum([intprod(shape) for shape in shapes])
self.theta = theta = tf.placeholder(dtype, [total_size])
start = 0
assigns = []
for (shape, v) in zip(shapes, var_list):
|
self.op = tf.group(*assigns)
def __call__(self, theta):
tf.get_default_session().run(self.op, feed_dict={self.theta: theta})
class GetFlat(object):
def __init__(self, var_list):
self.op = tf.concat(axis=0, values=[tf.reshape(v, [numel(v)]) for v in var_list])
def __call__(self):
return tf.get_default_session().run(self.op)
_PLACEHOLDER_CACHE = {} # name -> (placeholder, dtype, shape)
def get_placeholder(name, dtype, shape):
if name in _PLACEHOLDER_CACHE:
out, dtype1, shape1 = _PLACEHOLDER_CACHE[name]
assert dtype1 == dtype and shape1 == shape
return out
else:
out = tf.placeholder(dtype=dtype, shape=shape, name=name)
_PLACEHOLDER_CACHE[name] = (out, dtype, shape)
return out
def get_placeholder_cached(name):
return _PLACEHOLDER_CACHE[name][0]
def flattenallbut0(x):
return tf.reshape(x, [-1, intprod(x.get_shape().as_list()[1:])])
# ================================================================
# Diagnostics
# ================================================================
def display_var_info(vars):
from baselines import logger
count_params = 0
for v in vars:
name = v.name
if "/Adam" in name or "beta1_power" in name or "beta2_power" in name: continue
count_params += np.prod(v.shape.as_list())
if "/b:" in name: continue # Wx+b, bias is not interesting to look at => count params, but not print
logger.info(" %s%s%s" % (name, " "*(55-len(name)), str(v.shape)))
logger.info("Total model parameters: %0.1f million" % (count_params*1e-6))
class LinearSchedule(object):
def __init__(self, schedule_timesteps, final_p, initial_p=1.0):
"""Linear interpolation between initial_p and final_p over
schedule_timesteps. After this many timesteps pass final_p is
returned.
Parameters
----------
schedule_timesteps: int
Number of timesteps for which to linearly anneal initial_p
to final_p
initial_p: float
initial output value
final_p: float
final output value
"""
self.schedule_timesteps = schedule_timesteps
self.final_p = final_p
self.initial_p = initial_p
def value(self, t):
"""See Schedule.value"""
fraction = min(float(t) / self.schedule_timesteps, 1.0)
return self.initial_p + fraction * (self.final_p - self.initial_p)
# deep q
import os
import tensorflow as tf
# ================================================================
# Saving variables
# ================================================================
def load_state(fname):
saver = tf.train.Saver()
saver.restore(tf.get_default_session(), fname)
def save_state(fname):
os.makedirs(os.path.dirname(fname), exist_ok=True)
saver = tf.train.Saver()
saver.save(tf.get_default_session(), fname)
# ================================================================
# Placeholders
# ================================================================
class TfInput(object):
def __init__(self, name="(unnamed)"):
| size = intprod(shape)
assigns.append(tf.assign(v, tf.reshape(theta[start:start + size], shape)))
start += size | conditional_block |
encoder.rs | ::Error> for EncodingError {
fn from(err: io::Error) -> EncodingError {
EncodingError::IoError(err)
}
}
impl From<EncodingError> for io::Error {
fn from(err: EncodingError) -> io::Error {
io::Error::new(io::ErrorKind::Other, (&err as &error::Error).description())
}
}
/// PNG Encoder
pub struct Encoder<W: Write> {
w: W,
info: Info,
}
impl<W: Write> Encoder<W> {
pub fn new(w: W, width: u32, height: u32) -> Encoder<W> {
let mut info = Info::default();
info.width = width;
info.height = height;
Encoder { w: w, info: info }
}
pub fn new_animated_with_frame_rate(w: W, width: u32, height: u32, frames: u32, delay_num: u16, delay_den: u16) -> Result<Encoder<W>> {
let mut enc = Encoder::new_animated(w, width, height, frames)?;
let mut frame_ctl = enc.info.frame_control.unwrap();
frame_ctl.delay_num = delay_num;
frame_ctl.delay_den = delay_den;
enc.info.frame_control = Some(frame_ctl);
Ok(enc)
}
pub fn new_animated(w: W, width: u32, height: u32, frames: u32) -> Result<Encoder<W>> {
if frames > 0 | else {
Err(EncodingError::Format("invalid number of frames for an animated PNG".into()))
}
}
pub fn write_header(self) -> Result<Writer<W>> {
Writer::new(self.w, self.info).init()
}
}
impl<W: Write> HasParameters for Encoder<W> {}
impl<W: Write> Parameter<Encoder<W>> for ColorType {
fn set_param(self, this: &mut Encoder<W>) {
this.info.color_type = self
}
}
impl<W: Write> Parameter<Encoder<W>> for BitDepth {
fn set_param(self, this: &mut Encoder<W>) {
this.info.bit_depth = self
}
}
/// PNG writer
pub struct Writer<W: Write> {
w: W,
info: Info,
separate_default_image: bool,
}
impl<W: Write> Writer<W> {
fn new(w: W, info: Info) -> Writer<W> {
let w = Writer { w: w, info: info, separate_default_image: false };
w
}
fn init(mut self) -> Result<Self> {
try!(self.w.write(&[137, 80, 78, 71, 13, 10, 26, 10]));
let mut data = [0; 13];
try!((&mut data[..]).write_be(self.info.width));
try!((&mut data[4..]).write_be(self.info.height));
data[8] = self.info.bit_depth as u8;
data[9] = self.info.color_type as u8;
data[12] = if self.info.interlaced { 1 } else { 0 };
try!(self.write_chunk(chunk::IHDR, &data));
match self.info {
Info { animation_control: Some(anim_ctl), frame_control: Some(_), ..} => {
let mut data = [0; 8];
try!((&mut data[..]).write_be(anim_ctl.num_frames));
try!((&mut data[4..]).write_be(anim_ctl.num_plays));
try!(self.write_chunk(chunk::acTL, &data));
}
_ => {}
};
Ok(self)
}
pub fn write_chunk_with_fields(&mut self, name: [u8; 4], data: &[u8], fields: Option<&[u8]>) -> Result<()> {
self.w.write_be(data.len() as u32 + (if fields.is_some() { fields.unwrap().len() as u32 } else { 0 }))?;
self.w.write(&name)?;
if fields.is_some() { try!(self.w.write(fields.unwrap())); }
self.w.write(data)?;
let mut crc = Crc32::new();
crc.update(&name);
if fields.is_some() { crc.update(fields.unwrap()); }
crc.update(data);
self.w.write_be(crc.checksum())?;
Ok(())
}
pub fn write_chunk(&mut self, name: [u8; 4], data: &[u8]) -> Result<()> {
self.write_chunk_with_fields(name, data, None)
}
/// Writes the image data.
pub fn write_image_data(&mut self, data: &[u8]) -> Result<()> {
let zlib = self.get_image_data(data)?;
self.write_chunk(chunk::IDAT, &try!(zlib.finish()))
}
fn get_image_data(&mut self, data: &[u8]) -> Result<deflate::write::ZlibEncoder<Vec<u8>>> {
let bpp = self.info.bytes_per_pixel();
let in_len = self.info.raw_row_length() - 1;
let mut prev = vec![0; in_len];
let mut current = vec![0; in_len];
let data_size = in_len * self.info.height as usize;
if data.len() < data_size || data_size == 0 {
return Err(EncodingError::Format("not enough image data provided".into()));
}
let mut zlib = deflate::write::ZlibEncoder::new(Vec::new(), deflate::Compression::Fast);
let filter_method = FilterType::Sub;
for line in data.chunks(in_len) {
current.copy_from_slice(&line);
try!(zlib.write_all(&[filter_method as u8]));
filter(filter_method, bpp, &prev, &mut current);
try!(zlib.write_all(¤t));
mem::swap(&mut prev, &mut current);
}
Ok(zlib)
}
pub fn write_separate_default_image(&mut self, data: &[u8]) -> Result<()> {
match self.info {
Info { animation_control: Some(_), frame_control: Some(frame_control), ..} => {
if frame_control.sequence_number != 0 {
Err(EncodingError::Format("separate default image provided after frame sequence has begun".into()))
} else if self.separate_default_image {
Err(EncodingError::Format("default image already written".into()))
} else {
self.separate_default_image = true;
self.write_image_data(data)
}
}
_ => {
Err(EncodingError::Format("default image provided for a non-animated PNG".into()))
}
}
}
#[allow(non_snake_case)]
fn write_fcTL(&mut self) -> Result<()> {
let frame_ctl = self.info.frame_control.ok_or(EncodingError::Format("cannot write fcTL for a non-animated PNG".into()))?;
let mut data = [0u8; 26];
(&mut data[..]).write_be(frame_ctl.sequence_number)?;
(&mut data[4..]).write_be(frame_ctl.width)?;
(&mut data[8..]).write_be(frame_ctl.height)?;
(&mut data[12..]).write_be(frame_ctl.x_offset)?;
(&mut data[16..]).write_be(frame_ctl.y_offset)?;
(&mut data[20..]).write_be(frame_ctl.delay_num)?;
(&mut data[22..]).write_be(frame_ctl.delay_den)?;
data[24] = frame_ctl.dispose_op as u8;
data[25] = frame_ctl.blend_op as u8;
self.write_chunk(chunk::fcTL, &data)
}
#[allow(non_snake_case)]
fn write_fdAT(&mut self, data: &[u8]) -> Result<()> {
// println!("Writing fdAT:{:?}", self.info.frame_control.unwrap().sequence_number+1);
let zlib = self.get_image_data(data)?;
let mut data = [0u8; 4];
(&mut data[..]).write_be(self.info.frame_control
.ok_or(EncodingError::Format("cannot write fdAT for a non-animated PNG".into()))?.sequence_number+1u32)?;
self.write_chunk_with_fields(chunk::fdAT, &zlib.finish()?, Some(&data))
}
pub fn write_frame(&mut self, data: &[u8]) -> Result<()> {
// println!("{:?}", self.info.frame_control.unwrap().sequence_number);
match self.info {
Info { animation_control: Some(AnimationControl { num_frames: 0, ..}) , frame_control: Some(_), ..} => {
Err(EncodingError::Format("exceeded number of frames specified".into()))
},
Info { animation_control: Some(anim_ctl), frame_control: Some(frame_control), ..} => {
match frame_control.sequence_number {
0 => {
let ret = if self.separate_default_image { // If we've already written the default image we can write frames the normal way
| {
let mut encoder = Encoder::new(w, width, height);
let animation_ctl = AnimationControl { num_frames: frames, num_plays: 0 };
let mut frame_ctl = FrameControl::default();
frame_ctl.width = width;
frame_ctl.height = height;
encoder.info.animation_control = Some(animation_ctl);
encoder.info.frame_control = Some(frame_ctl);
Ok(encoder)
} | conditional_block |
encoder.rs | use self::EncodingError::*;
match *self {
IoError(ref err) => err.description(),
Format(ref desc) => &desc,
}
}
}
impl fmt::Display for EncodingError {
fn fmt(&self, fmt: &mut fmt::Formatter) -> result::Result<(), fmt::Error> {
write!(fmt, "{}", (self as &error::Error).description())
}
}
impl From<io::Error> for EncodingError {
fn from(err: io::Error) -> EncodingError {
EncodingError::IoError(err)
}
}
impl From<EncodingError> for io::Error {
fn from(err: EncodingError) -> io::Error {
io::Error::new(io::ErrorKind::Other, (&err as &error::Error).description())
}
}
/// PNG Encoder
pub struct Encoder<W: Write> {
w: W,
info: Info,
}
impl<W: Write> Encoder<W> {
pub fn new(w: W, width: u32, height: u32) -> Encoder<W> {
let mut info = Info::default();
info.width = width;
info.height = height;
Encoder { w: w, info: info }
}
pub fn new_animated_with_frame_rate(w: W, width: u32, height: u32, frames: u32, delay_num: u16, delay_den: u16) -> Result<Encoder<W>> {
let mut enc = Encoder::new_animated(w, width, height, frames)?;
let mut frame_ctl = enc.info.frame_control.unwrap();
frame_ctl.delay_num = delay_num;
frame_ctl.delay_den = delay_den;
enc.info.frame_control = Some(frame_ctl);
Ok(enc)
}
pub fn new_animated(w: W, width: u32, height: u32, frames: u32) -> Result<Encoder<W>> {
if frames > 0 {
let mut encoder = Encoder::new(w, width, height);
let animation_ctl = AnimationControl { num_frames: frames, num_plays: 0 };
let mut frame_ctl = FrameControl::default();
frame_ctl.width = width;
frame_ctl.height = height;
encoder.info.animation_control = Some(animation_ctl);
encoder.info.frame_control = Some(frame_ctl);
Ok(encoder)
} else {
Err(EncodingError::Format("invalid number of frames for an animated PNG".into()))
}
}
pub fn write_header(self) -> Result<Writer<W>> {
Writer::new(self.w, self.info).init()
}
}
impl<W: Write> HasParameters for Encoder<W> {}
impl<W: Write> Parameter<Encoder<W>> for ColorType {
fn set_param(self, this: &mut Encoder<W>) {
this.info.color_type = self
}
}
impl<W: Write> Parameter<Encoder<W>> for BitDepth {
fn set_param(self, this: &mut Encoder<W>) {
this.info.bit_depth = self
}
}
/// PNG writer
pub struct Writer<W: Write> {
w: W,
info: Info,
separate_default_image: bool,
}
impl<W: Write> Writer<W> {
fn new(w: W, info: Info) -> Writer<W> {
let w = Writer { w: w, info: info, separate_default_image: false };
w
}
fn init(mut self) -> Result<Self> {
try!(self.w.write(&[137, 80, 78, 71, 13, 10, 26, 10]));
let mut data = [0; 13];
try!((&mut data[..]).write_be(self.info.width));
try!((&mut data[4..]).write_be(self.info.height));
data[8] = self.info.bit_depth as u8;
data[9] = self.info.color_type as u8;
data[12] = if self.info.interlaced { 1 } else { 0 };
try!(self.write_chunk(chunk::IHDR, &data));
match self.info {
Info { animation_control: Some(anim_ctl), frame_control: Some(_), ..} => {
let mut data = [0; 8];
try!((&mut data[..]).write_be(anim_ctl.num_frames));
try!((&mut data[4..]).write_be(anim_ctl.num_plays));
try!(self.write_chunk(chunk::acTL, &data));
}
_ => {}
};
Ok(self)
}
pub fn write_chunk_with_fields(&mut self, name: [u8; 4], data: &[u8], fields: Option<&[u8]>) -> Result<()> {
self.w.write_be(data.len() as u32 + (if fields.is_some() { fields.unwrap().len() as u32 } else { 0 }))?;
self.w.write(&name)?;
if fields.is_some() { try!(self.w.write(fields.unwrap())); }
self.w.write(data)?;
let mut crc = Crc32::new();
crc.update(&name);
if fields.is_some() { crc.update(fields.unwrap()); }
crc.update(data);
self.w.write_be(crc.checksum())?;
Ok(())
}
pub fn write_chunk(&mut self, name: [u8; 4], data: &[u8]) -> Result<()> {
self.write_chunk_with_fields(name, data, None)
}
/// Writes the image data.
pub fn write_image_data(&mut self, data: &[u8]) -> Result<()> {
let zlib = self.get_image_data(data)?;
self.write_chunk(chunk::IDAT, &try!(zlib.finish()))
}
fn get_image_data(&mut self, data: &[u8]) -> Result<deflate::write::ZlibEncoder<Vec<u8>>> {
let bpp = self.info.bytes_per_pixel();
let in_len = self.info.raw_row_length() - 1;
let mut prev = vec![0; in_len];
let mut current = vec![0; in_len];
let data_size = in_len * self.info.height as usize;
if data.len() < data_size || data_size == 0 {
return Err(EncodingError::Format("not enough image data provided".into()));
}
let mut zlib = deflate::write::ZlibEncoder::new(Vec::new(), deflate::Compression::Fast);
let filter_method = FilterType::Sub;
for line in data.chunks(in_len) {
current.copy_from_slice(&line);
try!(zlib.write_all(&[filter_method as u8]));
filter(filter_method, bpp, &prev, &mut current);
try!(zlib.write_all(¤t));
mem::swap(&mut prev, &mut current);
}
Ok(zlib)
}
pub fn write_separate_default_image(&mut self, data: &[u8]) -> Result<()> {
match self.info {
Info { animation_control: Some(_), frame_control: Some(frame_control), ..} => {
if frame_control.sequence_number != 0 {
Err(EncodingError::Format("separate default image provided after frame sequence has begun".into()))
} else if self.separate_default_image {
Err(EncodingError::Format("default image already written".into()))
} else {
self.separate_default_image = true;
self.write_image_data(data)
}
}
_ => {
Err(EncodingError::Format("default image provided for a non-animated PNG".into()))
}
}
}
#[allow(non_snake_case)]
fn write_fcTL(&mut self) -> Result<()> {
let frame_ctl = self.info.frame_control.ok_or(EncodingError::Format("cannot write fcTL for a non-animated PNG".into()))?;
let mut data = [0u8; 26];
(&mut data[..]).write_be(frame_ctl.sequence_number)?;
(&mut data[4..]).write_be(frame_ctl.width)?;
(&mut data[8..]).write_be(frame_ctl.height)?;
(&mut data[12..]).write_be(frame_ctl.x_offset)?;
(&mut data[16..]).write_be(frame_ctl.y_offset)?;
(&mut data[20..]).write_be(frame_ctl.delay_num)?;
(&mut data[22..]).write_be(frame_ctl.delay_den)?;
data[24] = frame_ctl.dispose_op as u8;
data[25] = frame_ctl.blend_op as u8;
self.write_chunk(chunk::fcTL, &data)
}
#[allow(non_snake_case)]
fn write_fdAT(&mut self, data: &[u8]) -> Result<()> {
// println!("Writing fdAT:{:?}", self.info.frame_control.unwrap().sequence_number+1);
let zlib = self.get_image_data(data)?;
let mut data = [0u8; 4];
(&mut data[..]).write_be(self.info.frame_control
.ok_or(EncodingError::Format("cannot write fdAT for a non-animated PNG".into()))?.sequence_number+1u32)?;
self.write_chunk_with_fields(chunk::fdAT, &zlib.finish()?, Some(&data))
}
pub fn write_frame(&mut self, data: &[u8]) -> Result<()> {
// println!("{:?}", self.info.frame_control.unwrap().sequence_number);
match self.info {
Info { animation_control: Some(AnimationControl | fn description(&self) -> &str { | random_line_split |
|
encoder.rs | >> {
if frames > 0 {
let mut encoder = Encoder::new(w, width, height);
let animation_ctl = AnimationControl { num_frames: frames, num_plays: 0 };
let mut frame_ctl = FrameControl::default();
frame_ctl.width = width;
frame_ctl.height = height;
encoder.info.animation_control = Some(animation_ctl);
encoder.info.frame_control = Some(frame_ctl);
Ok(encoder)
} else {
Err(EncodingError::Format("invalid number of frames for an animated PNG".into()))
}
}
pub fn write_header(self) -> Result<Writer<W>> {
Writer::new(self.w, self.info).init()
}
}
impl<W: Write> HasParameters for Encoder<W> {}
impl<W: Write> Parameter<Encoder<W>> for ColorType {
fn set_param(self, this: &mut Encoder<W>) {
this.info.color_type = self
}
}
impl<W: Write> Parameter<Encoder<W>> for BitDepth {
fn set_param(self, this: &mut Encoder<W>) {
this.info.bit_depth = self
}
}
/// PNG writer
pub struct Writer<W: Write> {
w: W,
info: Info,
separate_default_image: bool,
}
impl<W: Write> Writer<W> {
fn new(w: W, info: Info) -> Writer<W> {
let w = Writer { w: w, info: info, separate_default_image: false };
w
}
fn init(mut self) -> Result<Self> {
try!(self.w.write(&[137, 80, 78, 71, 13, 10, 26, 10]));
let mut data = [0; 13];
try!((&mut data[..]).write_be(self.info.width));
try!((&mut data[4..]).write_be(self.info.height));
data[8] = self.info.bit_depth as u8;
data[9] = self.info.color_type as u8;
data[12] = if self.info.interlaced { 1 } else { 0 };
try!(self.write_chunk(chunk::IHDR, &data));
match self.info {
Info { animation_control: Some(anim_ctl), frame_control: Some(_), ..} => {
let mut data = [0; 8];
try!((&mut data[..]).write_be(anim_ctl.num_frames));
try!((&mut data[4..]).write_be(anim_ctl.num_plays));
try!(self.write_chunk(chunk::acTL, &data));
}
_ => {}
};
Ok(self)
}
pub fn write_chunk_with_fields(&mut self, name: [u8; 4], data: &[u8], fields: Option<&[u8]>) -> Result<()> {
self.w.write_be(data.len() as u32 + (if fields.is_some() { fields.unwrap().len() as u32 } else { 0 }))?;
self.w.write(&name)?;
if fields.is_some() { try!(self.w.write(fields.unwrap())); }
self.w.write(data)?;
let mut crc = Crc32::new();
crc.update(&name);
if fields.is_some() { crc.update(fields.unwrap()); }
crc.update(data);
self.w.write_be(crc.checksum())?;
Ok(())
}
pub fn write_chunk(&mut self, name: [u8; 4], data: &[u8]) -> Result<()> {
self.write_chunk_with_fields(name, data, None)
}
/// Writes the image data.
pub fn write_image_data(&mut self, data: &[u8]) -> Result<()> {
let zlib = self.get_image_data(data)?;
self.write_chunk(chunk::IDAT, &try!(zlib.finish()))
}
fn get_image_data(&mut self, data: &[u8]) -> Result<deflate::write::ZlibEncoder<Vec<u8>>> {
let bpp = self.info.bytes_per_pixel();
let in_len = self.info.raw_row_length() - 1;
let mut prev = vec![0; in_len];
let mut current = vec![0; in_len];
let data_size = in_len * self.info.height as usize;
if data.len() < data_size || data_size == 0 {
return Err(EncodingError::Format("not enough image data provided".into()));
}
let mut zlib = deflate::write::ZlibEncoder::new(Vec::new(), deflate::Compression::Fast);
let filter_method = FilterType::Sub;
for line in data.chunks(in_len) {
current.copy_from_slice(&line);
try!(zlib.write_all(&[filter_method as u8]));
filter(filter_method, bpp, &prev, &mut current);
try!(zlib.write_all(¤t));
mem::swap(&mut prev, &mut current);
}
Ok(zlib)
}
pub fn write_separate_default_image(&mut self, data: &[u8]) -> Result<()> {
match self.info {
Info { animation_control: Some(_), frame_control: Some(frame_control), ..} => {
if frame_control.sequence_number != 0 {
Err(EncodingError::Format("separate default image provided after frame sequence has begun".into()))
} else if self.separate_default_image {
Err(EncodingError::Format("default image already written".into()))
} else {
self.separate_default_image = true;
self.write_image_data(data)
}
}
_ => {
Err(EncodingError::Format("default image provided for a non-animated PNG".into()))
}
}
}
#[allow(non_snake_case)]
fn write_fcTL(&mut self) -> Result<()> {
let frame_ctl = self.info.frame_control.ok_or(EncodingError::Format("cannot write fcTL for a non-animated PNG".into()))?;
let mut data = [0u8; 26];
(&mut data[..]).write_be(frame_ctl.sequence_number)?;
(&mut data[4..]).write_be(frame_ctl.width)?;
(&mut data[8..]).write_be(frame_ctl.height)?;
(&mut data[12..]).write_be(frame_ctl.x_offset)?;
(&mut data[16..]).write_be(frame_ctl.y_offset)?;
(&mut data[20..]).write_be(frame_ctl.delay_num)?;
(&mut data[22..]).write_be(frame_ctl.delay_den)?;
data[24] = frame_ctl.dispose_op as u8;
data[25] = frame_ctl.blend_op as u8;
self.write_chunk(chunk::fcTL, &data)
}
#[allow(non_snake_case)]
fn write_fdAT(&mut self, data: &[u8]) -> Result<()> {
// println!("Writing fdAT:{:?}", self.info.frame_control.unwrap().sequence_number+1);
let zlib = self.get_image_data(data)?;
let mut data = [0u8; 4];
(&mut data[..]).write_be(self.info.frame_control
.ok_or(EncodingError::Format("cannot write fdAT for a non-animated PNG".into()))?.sequence_number+1u32)?;
self.write_chunk_with_fields(chunk::fdAT, &zlib.finish()?, Some(&data))
}
pub fn write_frame(&mut self, data: &[u8]) -> Result<()> {
// println!("{:?}", self.info.frame_control.unwrap().sequence_number);
match self.info {
Info { animation_control: Some(AnimationControl { num_frames: 0, ..}) , frame_control: Some(_), ..} => {
Err(EncodingError::Format("exceeded number of frames specified".into()))
},
Info { animation_control: Some(anim_ctl), frame_control: Some(frame_control), ..} => {
match frame_control.sequence_number {
0 => {
let ret = if self.separate_default_image { // If we've already written the default image we can write frames the normal way
// fcTL + fdAT
self.write_fcTL().and(self.write_fdAT(data))
} else { // If not we'll have to set the first frame to be both:
// fcTL + first frame (IDAT)
self.write_fcTL().and(self.write_image_data(data))
};
let mut fc = self.info.frame_control.unwrap();
fc.inc_seq_num(1);
self.info.frame_control = Some(fc);
ret
},
x if x == 2 * anim_ctl.num_frames - 1 => {
// println!("We're done, boss");
// This is the last frame:
// Do the usual and also set AnimationControl to no remaining frames:
let ret = self.write_fcTL().and(self.write_fdAT(data));
let mut fc = self.info.frame_control.unwrap();
fc.set_seq_num(0);
self.info.frame_control = Some(fc);
ret
},
_ => {
// Usual case:
// fcTL + fdAT
// println!("Buisness as usual");
let ret = self.write_fcTL().and(self.write_fdAT(data));
let mut fc = self.info.frame_control.unwrap();
fc.inc_seq_num(2);
self.info.frame_control = Some(fc);
ret
}
}
},
_ => {
Err(EncodingError::Format("frame provided for a non-animated PNG".into()))
}
}
}
}
impl<W: Write> Drop for Writer<W> {
fn | drop | identifier_name |
|
encoder.rs | ::Error> for EncodingError {
fn from(err: io::Error) -> EncodingError {
EncodingError::IoError(err)
}
}
impl From<EncodingError> for io::Error {
fn from(err: EncodingError) -> io::Error |
}
/// PNG Encoder
pub struct Encoder<W: Write> {
w: W,
info: Info,
}
impl<W: Write> Encoder<W> {
pub fn new(w: W, width: u32, height: u32) -> Encoder<W> {
let mut info = Info::default();
info.width = width;
info.height = height;
Encoder { w: w, info: info }
}
pub fn new_animated_with_frame_rate(w: W, width: u32, height: u32, frames: u32, delay_num: u16, delay_den: u16) -> Result<Encoder<W>> {
let mut enc = Encoder::new_animated(w, width, height, frames)?;
let mut frame_ctl = enc.info.frame_control.unwrap();
frame_ctl.delay_num = delay_num;
frame_ctl.delay_den = delay_den;
enc.info.frame_control = Some(frame_ctl);
Ok(enc)
}
pub fn new_animated(w: W, width: u32, height: u32, frames: u32) -> Result<Encoder<W>> {
if frames > 0 {
let mut encoder = Encoder::new(w, width, height);
let animation_ctl = AnimationControl { num_frames: frames, num_plays: 0 };
let mut frame_ctl = FrameControl::default();
frame_ctl.width = width;
frame_ctl.height = height;
encoder.info.animation_control = Some(animation_ctl);
encoder.info.frame_control = Some(frame_ctl);
Ok(encoder)
} else {
Err(EncodingError::Format("invalid number of frames for an animated PNG".into()))
}
}
pub fn write_header(self) -> Result<Writer<W>> {
Writer::new(self.w, self.info).init()
}
}
impl<W: Write> HasParameters for Encoder<W> {}
impl<W: Write> Parameter<Encoder<W>> for ColorType {
fn set_param(self, this: &mut Encoder<W>) {
this.info.color_type = self
}
}
impl<W: Write> Parameter<Encoder<W>> for BitDepth {
fn set_param(self, this: &mut Encoder<W>) {
this.info.bit_depth = self
}
}
/// PNG writer
pub struct Writer<W: Write> {
w: W,
info: Info,
separate_default_image: bool,
}
impl<W: Write> Writer<W> {
fn new(w: W, info: Info) -> Writer<W> {
let w = Writer { w: w, info: info, separate_default_image: false };
w
}
fn init(mut self) -> Result<Self> {
try!(self.w.write(&[137, 80, 78, 71, 13, 10, 26, 10]));
let mut data = [0; 13];
try!((&mut data[..]).write_be(self.info.width));
try!((&mut data[4..]).write_be(self.info.height));
data[8] = self.info.bit_depth as u8;
data[9] = self.info.color_type as u8;
data[12] = if self.info.interlaced { 1 } else { 0 };
try!(self.write_chunk(chunk::IHDR, &data));
match self.info {
Info { animation_control: Some(anim_ctl), frame_control: Some(_), ..} => {
let mut data = [0; 8];
try!((&mut data[..]).write_be(anim_ctl.num_frames));
try!((&mut data[4..]).write_be(anim_ctl.num_plays));
try!(self.write_chunk(chunk::acTL, &data));
}
_ => {}
};
Ok(self)
}
pub fn write_chunk_with_fields(&mut self, name: [u8; 4], data: &[u8], fields: Option<&[u8]>) -> Result<()> {
self.w.write_be(data.len() as u32 + (if fields.is_some() { fields.unwrap().len() as u32 } else { 0 }))?;
self.w.write(&name)?;
if fields.is_some() { try!(self.w.write(fields.unwrap())); }
self.w.write(data)?;
let mut crc = Crc32::new();
crc.update(&name);
if fields.is_some() { crc.update(fields.unwrap()); }
crc.update(data);
self.w.write_be(crc.checksum())?;
Ok(())
}
pub fn write_chunk(&mut self, name: [u8; 4], data: &[u8]) -> Result<()> {
self.write_chunk_with_fields(name, data, None)
}
/// Writes the image data.
pub fn write_image_data(&mut self, data: &[u8]) -> Result<()> {
let zlib = self.get_image_data(data)?;
self.write_chunk(chunk::IDAT, &try!(zlib.finish()))
}
fn get_image_data(&mut self, data: &[u8]) -> Result<deflate::write::ZlibEncoder<Vec<u8>>> {
let bpp = self.info.bytes_per_pixel();
let in_len = self.info.raw_row_length() - 1;
let mut prev = vec![0; in_len];
let mut current = vec![0; in_len];
let data_size = in_len * self.info.height as usize;
if data.len() < data_size || data_size == 0 {
return Err(EncodingError::Format("not enough image data provided".into()));
}
let mut zlib = deflate::write::ZlibEncoder::new(Vec::new(), deflate::Compression::Fast);
let filter_method = FilterType::Sub;
for line in data.chunks(in_len) {
current.copy_from_slice(&line);
try!(zlib.write_all(&[filter_method as u8]));
filter(filter_method, bpp, &prev, &mut current);
try!(zlib.write_all(¤t));
mem::swap(&mut prev, &mut current);
}
Ok(zlib)
}
pub fn write_separate_default_image(&mut self, data: &[u8]) -> Result<()> {
match self.info {
Info { animation_control: Some(_), frame_control: Some(frame_control), ..} => {
if frame_control.sequence_number != 0 {
Err(EncodingError::Format("separate default image provided after frame sequence has begun".into()))
} else if self.separate_default_image {
Err(EncodingError::Format("default image already written".into()))
} else {
self.separate_default_image = true;
self.write_image_data(data)
}
}
_ => {
Err(EncodingError::Format("default image provided for a non-animated PNG".into()))
}
}
}
#[allow(non_snake_case)]
fn write_fcTL(&mut self) -> Result<()> {
let frame_ctl = self.info.frame_control.ok_or(EncodingError::Format("cannot write fcTL for a non-animated PNG".into()))?;
let mut data = [0u8; 26];
(&mut data[..]).write_be(frame_ctl.sequence_number)?;
(&mut data[4..]).write_be(frame_ctl.width)?;
(&mut data[8..]).write_be(frame_ctl.height)?;
(&mut data[12..]).write_be(frame_ctl.x_offset)?;
(&mut data[16..]).write_be(frame_ctl.y_offset)?;
(&mut data[20..]).write_be(frame_ctl.delay_num)?;
(&mut data[22..]).write_be(frame_ctl.delay_den)?;
data[24] = frame_ctl.dispose_op as u8;
data[25] = frame_ctl.blend_op as u8;
self.write_chunk(chunk::fcTL, &data)
}
#[allow(non_snake_case)]
fn write_fdAT(&mut self, data: &[u8]) -> Result<()> {
// println!("Writing fdAT:{:?}", self.info.frame_control.unwrap().sequence_number+1);
let zlib = self.get_image_data(data)?;
let mut data = [0u8; 4];
(&mut data[..]).write_be(self.info.frame_control
.ok_or(EncodingError::Format("cannot write fdAT for a non-animated PNG".into()))?.sequence_number+1u32)?;
self.write_chunk_with_fields(chunk::fdAT, &zlib.finish()?, Some(&data))
}
pub fn write_frame(&mut self, data: &[u8]) -> Result<()> {
// println!("{:?}", self.info.frame_control.unwrap().sequence_number);
match self.info {
Info { animation_control: Some(AnimationControl { num_frames: 0, ..}) , frame_control: Some(_), ..} => {
Err(EncodingError::Format("exceeded number of frames specified".into()))
},
Info { animation_control: Some(anim_ctl), frame_control: Some(frame_control), ..} => {
match frame_control.sequence_number {
0 => {
let ret = if self.separate_default_image { // If we've already written the default image we can write frames the normal way
| {
io::Error::new(io::ErrorKind::Other, (&err as &error::Error).description())
} | identifier_body |
conn.rs | Ok(t) => t,
Err(e) => return Err(e)
};
Ok(Connection {
addr: addr,
host: host,
sock: Some(Plain(sock)),
name: name,
term: t
})
}
pub fn status(&mut self) {
self.send_handshake(false);
// Send the status request
self.write_packet(Packet::new_out(0x0));
// and read back the response
let (packet_id, mut packet) = self.read_packet();
// Make sure we got the right response
assert_eq!(packet_id, 0x0);
// Get the JSON
let json = ExtraJSON::new(json::from_str(packet.read_string()).unwrap());
println!("Minecraft Server Status [{}:{}]", self.host, self.addr.port);
println!("Version: {}", json["version"]["name"].string());
println!("Protocol: {}", json["version"]["protocol"].as_int());
println!("Description: {}", json["description"].string());
println!("Players: ({}/{})", json["players"]["online"].as_int(), json["players"]["max"].as_int());
let players = json["players"]["sample"].list();
for player in players.iter() {
println!("\t{} ({})", player["name"].string(), player["id"].string());
}
}
pub fn run(mut self) {
// If the server is in online-mode
// we need to do authentication and
// enable encryption
self.login();
// Get a port to read messages from stdin
let msgs = self.read_messages();
// Yay, all good.
// Now we just loop and read in all the packets we can
// We don't actually do anything for most of them except
// for chat and keep alives.
loop {
// Got a message in the queue to send?
'msg: loop {
match msgs.try_recv() {
comm::Data(msg) => {
if msg.is_empty() {
continue;
} else if msg.len() > 100 {
println!("Message too long.");
continue;
}
// Send the message!
let mut p = Packet::new_out(0x1);
p.write_string(msg);
self.write_packet(p);
}
comm::Empty => break 'msg,
comm::Disconnected => fail!("input stream disconnected")
}
}
// Read in and handle a packet
let (packet_id, mut packet) = self.read_packet();
self.handle_message(packet_id, &mut packet);
}
}
fn handle_message(&mut self, packet_id: i32, packet: &mut packet::InPacket) {
// Keep Alive
if packet_id == 0x0 | else if packet_id == 0x2 {
let json = packet.read_string();
debug!("Got chat message: {}", json);
// Let's wrap up the Json so that we can
// deal with it more easily
let j = json::from_str(json).unwrap();
let j = ExtraJSON::new(j);
let ty = j["translate"].string();
// Player Chat
if "chat.type.text" == ty {
let user = j["with"][0]["text"].string();
let msg = j["with"][1].string();
self.term.attr(term::attr::ForegroundColor(term::color::BRIGHT_GREEN));
write!(&mut self.term as &mut Writer, "<{}> ", user);
self.term.reset();
self.term.write(msg.as_bytes());
self.term.write(bytes!("\n"));
// Server Message
} else if "chat.type.announcement" == ty {
let msg = j["with"][1]["extra"].list_map(|x| x.string()).concat();
self.term.attr(term::attr::ForegroundColor(term::color::BRIGHT_YELLOW));
self.term.write(bytes!("[Server] "));
self.term.reset();
self.term.write(msg.as_bytes());
self.term.write(bytes!("\n"));
}
}
}
fn login(&mut self) {
self.send_handshake(true);
self.send_username();
// Read the next packet and find out whether we need
// to do authentication and encryption
let (mut packet_id, mut packet) = self.read_packet();
debug!("Packet ID: {}", packet_id);
if packet_id == 0x1 {
// Encryption Request
// online-mode = true
self.enable_encryption(&mut packet);
// Read the next packet...
let (pi, p) = self.read_packet();
packet_id = pi;
packet = p;
}
if packet_id == 0x0 {
// Disconnect
let reason = packet.read_string();
debug!("Reason: {}", reason);
fail!("Received disconnect.");
}
// Login Success
assert_eq!(packet_id, 0x2);
let uuid = packet.read_string();
let username = packet.read_string();
debug!("UUID: {}", uuid);
debug!("Username: {}", username);
}
fn enable_encryption(&mut self, packet: &mut packet::InPacket) {
// Get all the data from the Encryption Request packet
let server_id = packet.read_string();
let key_len = packet.read_be_i16().unwrap();
let public_key = packet.read_exact(key_len as uint).unwrap();
let token_len = packet.read_be_i16().unwrap();
let verify_token = packet.read_exact(token_len as uint).unwrap();
// Server's public key
let pk = crypto::RSAPublicKey::from_bytes(public_key.as_slice()).unwrap();
// Generate random 16 byte key
let mut key = [0u8, ..16];
rand::task_rng().fill_bytes(key);
// Encrypt shared secret with server's public key
let ekey = pk.encrypt(key).unwrap();
// Encrypt verify token with server's public key
let etoken = pk.encrypt(verify_token.as_slice()).unwrap();
// Generate the server id hash
let mut sha1 = crypto::SHA1::new();
sha1.update(server_id.as_bytes());
sha1.update(key);
sha1.update(public_key.as_slice());
let hash = sha1.special_digest();
debug!("Hash: {}", hash);
// Do client auth
self.authenticate(hash);
// Create Encryption Response Packet
let mut erp = Packet::new_out(0x1);
// Write encrypted shared secret
erp.write_be_i16(ekey.len() as i16);
erp.write(ekey);
// Write encrypted verify token
erp.write_be_i16(etoken.len() as i16);
erp.write(etoken);
// Send
self.write_packet(erp);
// Create AES cipher with shared secret
let aes = crypto::AES::new(key.to_owned(), key.to_owned()).unwrap();
// Get the plain TCP stream
let sock = match self.sock.take_unwrap() {
Plain(s) => s,
_ => fail!("Expected plain socket!")
};
// and wwrap it in an AES Stream
let sock = crypto::AesStream::new(sock, aes);
// and put the new encrypted stream back
// everything form this point is encrypted
self.sock = Some(Encrypted(sock));
}
fn authenticate(&mut self, hash: ~str) {
let url = ~"https://authserver.mojang.com/authenticate";
let c = process::ProcessConfig {
program: "/usr/bin/curl",
args: &[~"-d", ~"@-", ~"-H", ~"Content-Type:application/json", url],
env: None,
cwd: None,
stdin: process::CreatePipe(true, false),
stdout: process::CreatePipe(false, true),
.. process::ProcessConfig::new()
};
let mut p = process::Process::configure(c).unwrap();
// write json to stdin and close it
write!(p.stdin.get_mut_ref() as &mut Writer, r#"
\{
"agent": \{
"name": "Minecraft",
"version": 1
\},
"username": "{}",
"password": "{}"
\}"#, "USER", "PASS"); // XXX: Don't hardcode these...
p.stdin = None;
// read response
let out = p.wait_with_output().output;
let out = str::from_utf8_owned(out.move_iter().collect()).unwrap();
debug!("Got - {}", out);
let json = ExtraJSON::new(json::from_str(out).unwrap());
let token = json["accessToken"].string();
let profile = json["selectedProfile"]["id"].string();
let url = ~"https://sessionserver.mojang.com/session/minecraft/join";
let c = process::ProcessConfig {
program: "/usr/bin/curl",
args: &[~"-d", ~"@-", ~"-H", ~"Content-Type:application/json", url],
env: None,
cwd: None,
stdin: process::CreatePipe(true, false),
stdout: process::CreatePipe(false, true),
.. process::ProcessConfig::new()
};
let | {
let x = packet.read_be_i32().unwrap();
// Need to respond
let mut resp = Packet::new_out(0x0);
resp.write_be_i32(x);
self.write_packet(resp);
// Chat Message
} | conditional_block |
conn.rs | "].list_map(|x| x.string()).concat();
self.term.attr(term::attr::ForegroundColor(term::color::BRIGHT_YELLOW));
self.term.write(bytes!("[Server] "));
self.term.reset();
self.term.write(msg.as_bytes());
self.term.write(bytes!("\n"));
}
}
}
fn login(&mut self) {
self.send_handshake(true);
self.send_username();
// Read the next packet and find out whether we need
// to do authentication and encryption
let (mut packet_id, mut packet) = self.read_packet();
debug!("Packet ID: {}", packet_id);
if packet_id == 0x1 {
// Encryption Request
// online-mode = true
self.enable_encryption(&mut packet);
// Read the next packet...
let (pi, p) = self.read_packet();
packet_id = pi;
packet = p;
}
if packet_id == 0x0 {
// Disconnect
let reason = packet.read_string();
debug!("Reason: {}", reason);
fail!("Received disconnect.");
}
// Login Success
assert_eq!(packet_id, 0x2);
let uuid = packet.read_string();
let username = packet.read_string();
debug!("UUID: {}", uuid);
debug!("Username: {}", username);
}
fn enable_encryption(&mut self, packet: &mut packet::InPacket) {
// Get all the data from the Encryption Request packet
let server_id = packet.read_string();
let key_len = packet.read_be_i16().unwrap();
let public_key = packet.read_exact(key_len as uint).unwrap();
let token_len = packet.read_be_i16().unwrap();
let verify_token = packet.read_exact(token_len as uint).unwrap();
// Server's public key
let pk = crypto::RSAPublicKey::from_bytes(public_key.as_slice()).unwrap();
// Generate random 16 byte key
let mut key = [0u8, ..16];
rand::task_rng().fill_bytes(key);
// Encrypt shared secret with server's public key
let ekey = pk.encrypt(key).unwrap();
// Encrypt verify token with server's public key
let etoken = pk.encrypt(verify_token.as_slice()).unwrap();
// Generate the server id hash
let mut sha1 = crypto::SHA1::new();
sha1.update(server_id.as_bytes());
sha1.update(key);
sha1.update(public_key.as_slice());
let hash = sha1.special_digest();
debug!("Hash: {}", hash);
// Do client auth
self.authenticate(hash);
// Create Encryption Response Packet
let mut erp = Packet::new_out(0x1);
// Write encrypted shared secret
erp.write_be_i16(ekey.len() as i16);
erp.write(ekey);
// Write encrypted verify token
erp.write_be_i16(etoken.len() as i16);
erp.write(etoken);
// Send
self.write_packet(erp);
// Create AES cipher with shared secret
let aes = crypto::AES::new(key.to_owned(), key.to_owned()).unwrap();
// Get the plain TCP stream
let sock = match self.sock.take_unwrap() {
Plain(s) => s,
_ => fail!("Expected plain socket!")
};
// and wwrap it in an AES Stream
let sock = crypto::AesStream::new(sock, aes);
// and put the new encrypted stream back
// everything form this point is encrypted
self.sock = Some(Encrypted(sock));
}
fn authenticate(&mut self, hash: ~str) {
let url = ~"https://authserver.mojang.com/authenticate";
let c = process::ProcessConfig {
program: "/usr/bin/curl",
args: &[~"-d", ~"@-", ~"-H", ~"Content-Type:application/json", url],
env: None,
cwd: None,
stdin: process::CreatePipe(true, false),
stdout: process::CreatePipe(false, true),
.. process::ProcessConfig::new()
};
let mut p = process::Process::configure(c).unwrap();
// write json to stdin and close it
write!(p.stdin.get_mut_ref() as &mut Writer, r#"
\{
"agent": \{
"name": "Minecraft",
"version": 1
\},
"username": "{}",
"password": "{}"
\}"#, "USER", "PASS"); // XXX: Don't hardcode these...
p.stdin = None;
// read response
let out = p.wait_with_output().output;
let out = str::from_utf8_owned(out.move_iter().collect()).unwrap();
debug!("Got - {}", out);
let json = ExtraJSON::new(json::from_str(out).unwrap());
let token = json["accessToken"].string();
let profile = json["selectedProfile"]["id"].string();
let url = ~"https://sessionserver.mojang.com/session/minecraft/join";
let c = process::ProcessConfig {
program: "/usr/bin/curl",
args: &[~"-d", ~"@-", ~"-H", ~"Content-Type:application/json", url],
env: None,
cwd: None,
stdin: process::CreatePipe(true, false),
stdout: process::CreatePipe(false, true),
.. process::ProcessConfig::new()
};
let mut p = process::Process::configure(c).unwrap();
// write json to stdin and close it
write!(p.stdin.get_mut_ref() as &mut Writer, r#"
\{
"accessToken": "{}",
"selectedProfile": "{}",
"serverId": "{}"
\}"#, token, profile, hash);
p.stdin = None;
// read response
let out = p.wait_with_output().output;
let out = str::from_utf8_owned(out.move_iter().collect()).unwrap();
debug!("Got - {}", out);
}
fn read_messages(&self) -> Receiver<~str> {
let (chan, port) = comm::channel();
spawn(proc() {
println!("Type message and then [ENTER] to send:");
let mut stdin = BufferedReader::new(io::stdin());
for line in stdin.lines() {
chan.send(line.unwrap().trim().to_owned());
}
});
port
}
fn write_packet(&mut self, p: packet::OutPacket) {
// Get the actual buffer
let buf = p.buf();
// Write out the packet length
self.sock.write_varint(buf.len() as i32);
// and the actual payload
self.sock.write(buf.as_slice());
}
fn read_packet(&mut self) -> (i32, packet::InPacket) {
// Read the packet length
let len = self.sock.read_varint();
// Now the payload
let buf = self.sock.read_exact(len as uint).unwrap();
let mut p = Packet::new_in(buf);
// Get the packet id
let id = p.read_varint();
(id, p)
}
fn send_handshake(&mut self, login: bool) {
let mut p = Packet::new_out(0x0);
// Protocol Version
p.write_varint(4);
// Server host
p.write_string(self.host);
// Server port
p.write_be_u16(self.addr.port);
// State
// 1 - status, 2 - login
p.write_varint(if login { 2 } else { 1 });
self.write_packet(p);
}
fn send_username(&mut self) {
let mut p = Packet::new_out(0x0);
p.write_string(self.name);
self.write_packet(p);
}
}
impl Reader for Sock {
fn read(&mut self, buf: &mut [u8]) -> io::IoResult<uint> {
match *self {
Plain(ref mut s) => s.read(buf),
Encrypted(ref mut s) => s.read(buf)
}
}
}
impl Writer for Sock {
fn write(&mut self, buf: &[u8]) -> io::IoResult<()> {
match *self {
Plain(ref mut s) => s.write(buf),
Encrypted(ref mut s) => s.write(buf)
}
}
fn flush(&mut self) -> io::IoResult<()> {
match *self {
Plain(ref mut s) => s.flush(),
Encrypted(ref mut s) => s.flush()
}
}
}
impl Reader for Option<Sock> {
fn read(&mut self, buf: &mut [u8]) -> io::IoResult<uint> {
match *self {
Some(ref mut s) => s.read(buf),
None => Err(io::standard_error(io::OtherIoError))
}
}
}
impl Writer for Option<Sock> {
fn write(&mut self, buf: &[u8]) -> io::IoResult<()> {
match *self {
Some(ref mut s) => s.write(buf),
None => Err(io::standard_error(io::OtherIoError))
}
}
fn flush(&mut self) -> io::IoResult<()> | {
match *self {
Some(ref mut s) => s.flush(),
None => Err(io::standard_error(io::OtherIoError))
}
} | identifier_body |
|
conn.rs | {
addr: SocketAddr,
host: ~str,
sock: Option<Sock>,
name: ~str,
term: term::Terminal<LineBufferedWriter<StdWriter>>
}
impl Connection {
pub fn new(name: ~str, host: ~str, port: u16) -> Result<Connection, ~str> {
// Resolve host
let addr = match addrinfo::get_host_addresses(host) {
Ok(a) => a[0],
Err(e) => return Err(e.to_str())
};
let addr = SocketAddr { ip: addr, port: port };
debug!("Connecting to server at {}.", addr.to_str());
let sock = TcpStream::connect(addr);
let sock = match sock {
Ok(s) => s,
Err(e) => return Err(format!("{} - {}", e.kind.to_str(), e.desc))
};
debug!("Successfully connected to server.");
let t = match term::Terminal::new(io::stdout()) {
Ok(t) => t,
Err(e) => return Err(e)
};
Ok(Connection {
addr: addr,
host: host,
sock: Some(Plain(sock)),
name: name,
term: t
})
}
pub fn status(&mut self) {
self.send_handshake(false);
// Send the status request
self.write_packet(Packet::new_out(0x0));
// and read back the response
let (packet_id, mut packet) = self.read_packet();
// Make sure we got the right response
assert_eq!(packet_id, 0x0);
// Get the JSON
let json = ExtraJSON::new(json::from_str(packet.read_string()).unwrap());
println!("Minecraft Server Status [{}:{}]", self.host, self.addr.port);
println!("Version: {}", json["version"]["name"].string());
println!("Protocol: {}", json["version"]["protocol"].as_int());
println!("Description: {}", json["description"].string());
println!("Players: ({}/{})", json["players"]["online"].as_int(), json["players"]["max"].as_int());
let players = json["players"]["sample"].list();
for player in players.iter() {
println!("\t{} ({})", player["name"].string(), player["id"].string());
}
}
pub fn run(mut self) {
// If the server is in online-mode
// we need to do authentication and
// enable encryption
self.login();
// Get a port to read messages from stdin
let msgs = self.read_messages();
// Yay, all good.
// Now we just loop and read in all the packets we can
// We don't actually do anything for most of them except
// for chat and keep alives.
loop {
// Got a message in the queue to send?
'msg: loop {
match msgs.try_recv() {
comm::Data(msg) => {
if msg.is_empty() {
continue;
} else if msg.len() > 100 {
println!("Message too long.");
continue;
}
// Send the message!
let mut p = Packet::new_out(0x1);
p.write_string(msg);
self.write_packet(p);
}
comm::Empty => break 'msg,
comm::Disconnected => fail!("input stream disconnected")
}
}
// Read in and handle a packet
let (packet_id, mut packet) = self.read_packet();
self.handle_message(packet_id, &mut packet);
}
}
fn handle_message(&mut self, packet_id: i32, packet: &mut packet::InPacket) {
// Keep Alive
if packet_id == 0x0 {
let x = packet.read_be_i32().unwrap();
// Need to respond
let mut resp = Packet::new_out(0x0);
resp.write_be_i32(x);
self.write_packet(resp);
// Chat Message
} else if packet_id == 0x2 {
let json = packet.read_string();
debug!("Got chat message: {}", json);
// Let's wrap up the Json so that we can
// deal with it more easily
let j = json::from_str(json).unwrap();
let j = ExtraJSON::new(j);
let ty = j["translate"].string();
// Player Chat
if "chat.type.text" == ty {
let user = j["with"][0]["text"].string();
let msg = j["with"][1].string();
self.term.attr(term::attr::ForegroundColor(term::color::BRIGHT_GREEN));
write!(&mut self.term as &mut Writer, "<{}> ", user);
self.term.reset();
self.term.write(msg.as_bytes());
self.term.write(bytes!("\n"));
// Server Message
} else if "chat.type.announcement" == ty {
let msg = j["with"][1]["extra"].list_map(|x| x.string()).concat();
self.term.attr(term::attr::ForegroundColor(term::color::BRIGHT_YELLOW));
self.term.write(bytes!("[Server] "));
self.term.reset();
self.term.write(msg.as_bytes());
self.term.write(bytes!("\n"));
}
}
}
fn login(&mut self) {
self.send_handshake(true);
self.send_username();
// Read the next packet and find out whether we need
// to do authentication and encryption
let (mut packet_id, mut packet) = self.read_packet();
debug!("Packet ID: {}", packet_id);
if packet_id == 0x1 {
// Encryption Request
// online-mode = true
self.enable_encryption(&mut packet);
// Read the next packet...
let (pi, p) = self.read_packet();
packet_id = pi;
packet = p;
}
if packet_id == 0x0 {
// Disconnect
let reason = packet.read_string();
debug!("Reason: {}", reason);
fail!("Received disconnect.");
}
// Login Success
assert_eq!(packet_id, 0x2);
let uuid = packet.read_string();
let username = packet.read_string();
debug!("UUID: {}", uuid);
debug!("Username: {}", username);
}
fn enable_encryption(&mut self, packet: &mut packet::InPacket) {
// Get all the data from the Encryption Request packet
let server_id = packet.read_string();
let key_len = packet.read_be_i16().unwrap();
let public_key = packet.read_exact(key_len as uint).unwrap();
let token_len = packet.read_be_i16().unwrap();
let verify_token = packet.read_exact(token_len as uint).unwrap();
// Server's public key
let pk = crypto::RSAPublicKey::from_bytes(public_key.as_slice()).unwrap();
// Generate random 16 byte key
let mut key = [0u8, ..16];
rand::task_rng().fill_bytes(key);
// Encrypt shared secret with server's public key
let ekey = pk.encrypt(key).unwrap();
// Encrypt verify token with server's public key
let etoken = pk.encrypt(verify_token.as_slice()).unwrap();
// Generate the server id hash
let mut sha1 = crypto::SHA1::new();
sha1.update(server_id.as_bytes());
sha1.update(key);
sha1.update(public_key.as_slice());
let hash = sha1.special_digest();
debug!("Hash: {}", hash);
// Do client auth
self.authenticate(hash);
// Create Encryption Response Packet
let mut erp = Packet::new_out(0x1);
// Write encrypted shared secret
erp.write_be_i16(ekey.len() as i16);
erp.write(ekey);
// Write encrypted verify token
erp.write_be_i16(etoken.len() as i16);
erp.write(etoken);
// Send
self.write_packet(erp);
// Create AES cipher with shared secret
let aes = crypto::AES::new(key.to_owned(), key.to_owned()).unwrap();
// Get the plain TCP stream
let sock = match self.sock.take_unwrap() {
Plain(s) => s,
_ => fail!("Expected plain socket!")
};
// and wwrap it in an AES Stream
let sock = crypto::AesStream::new(sock, aes);
// and put the new encrypted stream back
// everything form this point is encrypted
self.sock = Some(Encrypted(sock));
}
fn authenticate(&mut self, hash: ~str) {
let url = ~"https://authserver.mojang.com/authenticate";
let c = process::ProcessConfig {
program: "/usr/bin/curl",
args: &[~"-d", ~"@-", ~"-H", ~"Content-Type:application/json", url],
env: None,
cwd: None,
stdin: process::CreatePipe(true, false),
stdout: process::CreatePipe(false, true),
.. process::ProcessConfig::new()
};
let mut p = process::Process::configure(c).unwrap();
// write json to stdin and close it
write!(p.stdin.get_mut_ref() as &mut Writer, r#"
\{
"agent": \{
"name": "Minecraft",
"version": 1
\},
"username": "{}",
"password": "{}"
| Connection | identifier_name |
|
conn.rs |
})
}
pub fn status(&mut self) {
self.send_handshake(false);
// Send the status request
self.write_packet(Packet::new_out(0x0));
// and read back the response
let (packet_id, mut packet) = self.read_packet();
// Make sure we got the right response
assert_eq!(packet_id, 0x0);
// Get the JSON
let json = ExtraJSON::new(json::from_str(packet.read_string()).unwrap());
println!("Minecraft Server Status [{}:{}]", self.host, self.addr.port);
println!("Version: {}", json["version"]["name"].string());
println!("Protocol: {}", json["version"]["protocol"].as_int());
println!("Description: {}", json["description"].string());
println!("Players: ({}/{})", json["players"]["online"].as_int(), json["players"]["max"].as_int());
let players = json["players"]["sample"].list();
for player in players.iter() {
println!("\t{} ({})", player["name"].string(), player["id"].string());
}
}
pub fn run(mut self) {
// If the server is in online-mode
// we need to do authentication and
// enable encryption
self.login();
// Get a port to read messages from stdin
let msgs = self.read_messages();
// Yay, all good.
// Now we just loop and read in all the packets we can
// We don't actually do anything for most of them except
// for chat and keep alives.
loop {
// Got a message in the queue to send?
'msg: loop {
match msgs.try_recv() {
comm::Data(msg) => {
if msg.is_empty() {
continue;
} else if msg.len() > 100 {
println!("Message too long.");
continue;
}
// Send the message!
let mut p = Packet::new_out(0x1);
p.write_string(msg);
self.write_packet(p);
}
comm::Empty => break 'msg,
comm::Disconnected => fail!("input stream disconnected")
}
}
// Read in and handle a packet
let (packet_id, mut packet) = self.read_packet();
self.handle_message(packet_id, &mut packet);
}
}
fn handle_message(&mut self, packet_id: i32, packet: &mut packet::InPacket) {
// Keep Alive
if packet_id == 0x0 {
let x = packet.read_be_i32().unwrap();
// Need to respond
let mut resp = Packet::new_out(0x0);
resp.write_be_i32(x);
self.write_packet(resp);
// Chat Message
} else if packet_id == 0x2 {
let json = packet.read_string();
debug!("Got chat message: {}", json);
// Let's wrap up the Json so that we can
// deal with it more easily
let j = json::from_str(json).unwrap();
let j = ExtraJSON::new(j);
let ty = j["translate"].string();
// Player Chat
if "chat.type.text" == ty {
let user = j["with"][0]["text"].string();
let msg = j["with"][1].string();
self.term.attr(term::attr::ForegroundColor(term::color::BRIGHT_GREEN));
write!(&mut self.term as &mut Writer, "<{}> ", user);
self.term.reset();
self.term.write(msg.as_bytes());
self.term.write(bytes!("\n"));
// Server Message
} else if "chat.type.announcement" == ty {
let msg = j["with"][1]["extra"].list_map(|x| x.string()).concat();
self.term.attr(term::attr::ForegroundColor(term::color::BRIGHT_YELLOW));
self.term.write(bytes!("[Server] "));
self.term.reset();
self.term.write(msg.as_bytes());
self.term.write(bytes!("\n"));
}
}
}
fn login(&mut self) {
self.send_handshake(true);
self.send_username();
// Read the next packet and find out whether we need
// to do authentication and encryption
let (mut packet_id, mut packet) = self.read_packet();
debug!("Packet ID: {}", packet_id);
if packet_id == 0x1 {
// Encryption Request
// online-mode = true
self.enable_encryption(&mut packet);
// Read the next packet...
let (pi, p) = self.read_packet();
packet_id = pi;
packet = p;
}
if packet_id == 0x0 {
// Disconnect
let reason = packet.read_string();
debug!("Reason: {}", reason);
fail!("Received disconnect.");
}
// Login Success
assert_eq!(packet_id, 0x2);
let uuid = packet.read_string();
let username = packet.read_string();
debug!("UUID: {}", uuid);
debug!("Username: {}", username);
}
fn enable_encryption(&mut self, packet: &mut packet::InPacket) {
// Get all the data from the Encryption Request packet
let server_id = packet.read_string();
let key_len = packet.read_be_i16().unwrap();
let public_key = packet.read_exact(key_len as uint).unwrap();
let token_len = packet.read_be_i16().unwrap();
let verify_token = packet.read_exact(token_len as uint).unwrap();
// Server's public key
let pk = crypto::RSAPublicKey::from_bytes(public_key.as_slice()).unwrap();
// Generate random 16 byte key
let mut key = [0u8, ..16];
rand::task_rng().fill_bytes(key);
// Encrypt shared secret with server's public key
let ekey = pk.encrypt(key).unwrap();
// Encrypt verify token with server's public key
let etoken = pk.encrypt(verify_token.as_slice()).unwrap();
// Generate the server id hash
let mut sha1 = crypto::SHA1::new();
sha1.update(server_id.as_bytes());
sha1.update(key);
sha1.update(public_key.as_slice());
let hash = sha1.special_digest();
debug!("Hash: {}", hash);
// Do client auth
self.authenticate(hash);
// Create Encryption Response Packet
let mut erp = Packet::new_out(0x1);
// Write encrypted shared secret
erp.write_be_i16(ekey.len() as i16);
erp.write(ekey);
// Write encrypted verify token
erp.write_be_i16(etoken.len() as i16);
erp.write(etoken);
// Send
self.write_packet(erp);
// Create AES cipher with shared secret
let aes = crypto::AES::new(key.to_owned(), key.to_owned()).unwrap();
// Get the plain TCP stream
let sock = match self.sock.take_unwrap() {
Plain(s) => s,
_ => fail!("Expected plain socket!")
};
// and wwrap it in an AES Stream
let sock = crypto::AesStream::new(sock, aes);
// and put the new encrypted stream back
// everything form this point is encrypted
self.sock = Some(Encrypted(sock));
}
fn authenticate(&mut self, hash: ~str) {
let url = ~"https://authserver.mojang.com/authenticate";
let c = process::ProcessConfig {
program: "/usr/bin/curl",
args: &[~"-d", ~"@-", ~"-H", ~"Content-Type:application/json", url],
env: None,
cwd: None,
stdin: process::CreatePipe(true, false),
stdout: process::CreatePipe(false, true),
.. process::ProcessConfig::new()
};
let mut p = process::Process::configure(c).unwrap();
// write json to stdin and close it
write!(p.stdin.get_mut_ref() as &mut Writer, r#"
\{
"agent": \{
"name": "Minecraft",
"version": 1
\},
"username": "{}",
"password": "{}"
\}"#, "USER", "PASS"); // XXX: Don't hardcode these...
p.stdin = None;
// read response
let out = p.wait_with_output().output;
let out = str::from_utf8_owned(out.move_iter().collect()).unwrap();
debug!("Got - {}", out);
let json = ExtraJSON::new(json::from_str(out).unwrap());
let token = json["accessToken"].string();
let profile = json["selectedProfile"]["id"].string();
let url = ~"https://sessionserver.mojang.com/session/minecraft/join";
let c = process::ProcessConfig {
program: "/usr/bin/curl",
args: &[~"-d", ~"@-", ~"-H", ~"Content-Type:application/json", url],
env: None,
cwd: None,
stdin: process::CreatePipe(true, false),
stdout: process::CreatePipe(false, true),
.. process::ProcessConfig::new()
};
let mut p = process::Process::configure(c).unwrap(); |
// write json to stdin and close it
write!(p.stdin.get_mut_ref() as &mut Writer, r#"
\{
"accessToken": "{}", | random_line_split |
|
prefix_code.rs | 0f0f) << 4;
v = (v & 0xcccc) >> 2 | (v & 0x3333) << 2;
v = (v & 0xaaaa) >> 1 | (v & 0x5555) << 1;
v
}
pub fn generate_codes(sizes: &[u8], codes: &mut [PrefixCode]) -> bool {
let mut num_codes: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1];
let mut next_code: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1];
for &s in sizes.iter() {
num_codes[s as uint] += 1;
}
let mut code = 0u32;
for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) {
next_code[i] = code;
code += num_codes[i];
code <<= 1;
}
if code != (1 << (MAX_EXPECTED_CODE_SIZE + 1)) {
let mut t = 0u32;
for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) {
t += num_codes[i];
if t > 1 {
//return false; // Error, sizes don't add up
fail!("Code sizes don't add up");
}
}
}
for i in range(0, sizes.len()) {
let c = sizes[i];
let code = next_code[c as uint];
next_code[c as uint] += 1;
let rev_code = reverse_u16(code) >> (16 - c as uint);
codes[i] = PrefixCode::new(rev_code, c);
}
true
}
pub fn generate_codes_for_decode(
sizes: &[u8],
codes: &mut [PrefixCode],
dec_first_offset: &mut [u16, ..17],
dec_max_code: &mut [u32, ..18],
dec_offset_to_sym: &mut [u16],
decoder_table: &mut [u16],
max_code_size: u32) -> bool {
let mut num_codes: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1];
let mut next_code: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1];
for &s in sizes.iter() {
num_codes[s as uint] += 1;
}
let mut code = 0u32;
let mut offset = 0u32;
for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) {
next_code[i] = code;
dec_first_offset[i] = offset as u16 - code as u16;
code += num_codes[i];
dec_max_code[i] = code << (16 - i);
code <<= 1;
offset += num_codes[i];
}
dec_max_code[17] = 0x10000;
if code != (1 << (MAX_EXPECTED_CODE_SIZE + 1)) {
let mut t = 0u32;
for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) {
t += num_codes[i];
if t > 1 |
}
}
for p in decoder_table.mut_iter() {
*p = 0xffff;
}
for i in range(0, sizes.len()) {
let s = sizes[i] as uint;
let code = next_code[s];
next_code[s] += 1;
let offset = (code as u16 + dec_first_offset[s]) as uint;
dec_offset_to_sym[offset] = i as u16;
let rev_code = reverse_u16(code) >> (16 - s);
codes[i] = PrefixCode::new(rev_code, s as u8);
if s as u32 <= max_code_size {
let step = 1 << s;
let code = rev_code;
for p in range_step(code, 1 << max_code_size as uint, step) {
decoder_table[p as uint] = i as u16;
}
}
}
true
}
pub fn generate_decoder_table(codes: &[PrefixCode], decoder_table: &mut [u16], max_code_size: u32) {
assert!(decoder_table.len() == (1 << max_code_size as uint));
for p in decoder_table.mut_iter() {
*p = 0xffff;
}
for i in range(0, codes.len()) {
if codes[i].size() as u32 <= max_code_size {
assert!(codes[i].size() > 0);
let step = 1 << codes[i].size() as uint;
let code = codes[i].code();
for p in range_step(code, 1 << max_code_size as uint, step) {
decoder_table[p as uint] = i as u16;
}
}
}
}
static POLAR_MAX_SYMBOLS: u32 = 256;
pub fn polar_code_lengths(symbols: &[OrdFreq], sizes: &mut [u8]) -> u32 {
unsafe {
let mut tmp_freq: [u32, ..POLAR_MAX_SYMBOLS] = mem::uninitialized();
let mut orig_total_freq = 0;
let mut cur_total = 0;
let mut start_index = 0;
let mut max_code_size = 0;
let num_syms = symbols.len() as u32;
for i in range(0, symbols.len()) {
let sym_freq = symbols[symbols.len() - 1 - i].freq();
//let sym_freq = symbols[i].freq();
let sym_len = 31 - ctlz32(sym_freq);
let adjusted_sym_freq = 1 << sym_len as uint;
orig_total_freq += sym_freq;
tmp_freq[i] = adjusted_sym_freq;
cur_total += adjusted_sym_freq;
}
let mut tree_total = 1 << (31 - ctlz32(orig_total_freq)) as uint;
if tree_total < orig_total_freq {
tree_total <<= 1;
}
while cur_total < tree_total && start_index < num_syms {
let mut i = start_index;
while i < num_syms {
let freq = tmp_freq[i as uint];
if cur_total + freq <= tree_total {
tmp_freq[i as uint] += freq;
cur_total += freq;
if cur_total == tree_total {
break;
}
} else {
start_index = i + 1;
}
i += 1;
}
}
assert_eq!(cur_total, tree_total);
let tree_total_bits = 32 - ctlz32(tree_total);
for i in range(0, symbols.len()) {
let codesize = tree_total_bits - (32 - ctlz32(tmp_freq[i]));
max_code_size = max(max_code_size, codesize);
sizes[symbols[symbols.len() - 1 - i].sym() as uint] = codesize as u8;
//sizes[symbols[i].sym() as uint] = codesize as u8;
}
max_code_size
}
}
pub trait PrefixModel {
fn incr(&mut self, sym: u32);
fn update(&mut self, for_encoding: bool);
fn write<BW: ::prefix_code::BitWriter>(&mut self, bw: &mut BW, sym: u32);
fn read<BR: ::prefix_code::BitReader>(&mut self, br: &mut BR) -> u32;
}
pub trait BitWriter {
fn push_bits_uni(&mut self, bits: u32, count: u32);
}
pub trait BitReader {
fn pull_bits_uni(&mut self, count: u32) -> u32;
fn peek_bits_uni16(&self) -> u16;
fn skip_bits_uni(&mut self, count: u32);
}
#[deriving(Copy)]
pub struct Foo {
f: [u32, ..256]
}
impl Clone for Foo {
fn clone(&self) -> Foo {
Foo {
f: self.f
}
}
}
macro_rules! define_polar_model(
($name: ident, $symbol_count: expr) => {
//#[deriving(Clone)]
pub struct $name {
freq: [u32, ..$symbol_count],
codes: [::prefix_code::PrefixCode, ..$symbol_count],
decoder_table: [u16, ..(1 << 9)],
sum: u32,
next_rebuild: u32,
dec_max_code: [u32, ..18],
dec_first_offset: [u16, ..17],
dec_offset_to_sym: [u16, ..$symbol_count]
| {
//return false; // Error, sizes don't add up
fail!("Code sizes don't add up");
} | conditional_block |
prefix_code.rs | (self) -> u16 {
self.s
}
}
pub fn sort_symbols2<'a>(mut first: &'a mut [OrdFreq], mut second: &'a mut [OrdFreq]) -> &'a mut [OrdFreq] {
let mut hist = [0u32, ..256 * 2];
for &s in first.iter() {
let f = s.freq();
hist[ (f & 0xff) as uint] += 1;
hist[256 + ((f >> 8) & 0xff) as uint] += 1;
}
let num_syms = first.len();
// If all radix-1 digits are zero, we only need one pass
let passes = if hist[256] == num_syms as u32 { 1 } else { 2 };
for pass in range(0, passes) {
let c = &mut first[0] as *mut _;
let n = &mut second[0] as *mut _;
let histp = &mut hist[pass << 8] as *mut _;
let mut offsets: [u32, ..256] = unsafe { mem::uninitialized() };
let mut cur_ofs = 0;
for i in range_step(0u, 256, 2) {
offsets[i] = cur_ofs;
cur_ofs += unsafe { *histp.offset(i as int) };
offsets[i + 1] = cur_ofs;
cur_ofs += unsafe { *histp.offset(i as int + 1) };
}
let pass_shift = pass << 3;
let mut p = c;
let endp = unsafe { c.offset(num_syms as int) };
while p != endp {
let mut f = unsafe { *p }.freq();
f = (f >> pass_shift) & 0xff;
let dst_offset = offsets[f as uint];
offsets[f as uint] += 1;
unsafe {
*n.offset(dst_offset as int) = *p;
p = p.offset(1);
}
}
mem::swap(&mut first, &mut second);
}
let mut prev = 0;
for i in range(0, num_syms) {
assert!(first[i].freq() >= prev);
prev = first[i].freq();
}
first
}
#[deriving(Clone)]
pub struct PrefixCode(u32);
impl PrefixCode {
#[inline]
pub fn new(code: u32, size: u8) -> PrefixCode {
PrefixCode(code + (size as u32 << 16))
}
pub fn code(self) -> u32 {
let PrefixCode(v) = self;
v & 0xffff
}
pub fn size(self) -> u32 {
let PrefixCode(v) = self;
v >> 16
}
}
#[inline]
pub fn reverse_u16(mut v: u32) -> u32 {
v = (v & 0xff00) >> 8 | (v & 0x00ff) << 8;
v = (v & 0xf0f0) >> 4 | (v & 0x0f0f) << 4;
v = (v & 0xcccc) >> 2 | (v & 0x3333) << 2;
v = (v & 0xaaaa) >> 1 | (v & 0x5555) << 1;
v
}
pub fn generate_codes(sizes: &[u8], codes: &mut [PrefixCode]) -> bool {
let mut num_codes: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1];
let mut next_code: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1];
for &s in sizes.iter() {
num_codes[s as uint] += 1;
}
let mut code = 0u32;
for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) {
next_code[i] = code;
code += num_codes[i];
code <<= 1;
}
if code != (1 << (MAX_EXPECTED_CODE_SIZE + 1)) {
let mut t = 0u32;
for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) {
t += num_codes[i];
if t > 1 {
//return false; // Error, sizes don't add up
fail!("Code sizes don't add up");
}
}
}
for i in range(0, sizes.len()) {
let c = sizes[i];
let code = next_code[c as uint];
next_code[c as uint] += 1;
let rev_code = reverse_u16(code) >> (16 - c as uint);
codes[i] = PrefixCode::new(rev_code, c);
}
true
}
pub fn generate_codes_for_decode(
sizes: &[u8],
codes: &mut [PrefixCode],
dec_first_offset: &mut [u16, ..17],
dec_max_code: &mut [u32, ..18],
dec_offset_to_sym: &mut [u16],
decoder_table: &mut [u16],
max_code_size: u32) -> bool {
let mut num_codes: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1];
let mut next_code: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1];
for &s in sizes.iter() {
num_codes[s as uint] += 1;
}
let mut code = 0u32;
let mut offset = 0u32;
for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) {
next_code[i] = code;
dec_first_offset[i] = offset as u16 - code as u16;
code += num_codes[i];
dec_max_code[i] = code << (16 - i);
code <<= 1;
offset += num_codes[i];
}
dec_max_code[17] = 0x10000;
if code != (1 << (MAX_EXPECTED_CODE_SIZE + 1)) {
let mut t = 0u32;
for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) {
t += num_codes[i];
if t > 1 {
//return false; // Error, sizes don't add up
fail!("Code sizes don't add up");
}
}
}
for p in decoder_table.mut_iter() {
*p = 0xffff;
}
for i in range(0, sizes.len()) {
let s = sizes[i] as uint;
let code = next_code[s];
next_code[s] += 1;
let offset = (code as u16 + dec_first_offset[s]) as uint;
dec_offset_to_sym[offset] = i as u16;
let rev_code = reverse_u16(code) >> (16 - s);
codes[i] = PrefixCode::new(rev_code, s as u8);
if s as u32 <= max_code_size {
let step = 1 << s;
let code = rev_code;
for p in range_step(code, 1 << max_code_size as uint, step) {
decoder_table[p as uint] = i as u16;
}
}
}
true
}
pub fn generate_decoder_table(codes: &[PrefixCode], decoder_table: &mut [u16], max_code_size: u32) {
assert!(decoder_table.len() == (1 << max_code_size as uint));
for p in decoder_table.mut_iter() {
*p = 0xffff;
}
for i in range(0, codes.len()) {
if codes[i].size() as u32 <= max_code_size {
assert!(codes[i].size() > 0);
let step = 1 << codes[i].size() as uint;
let code = codes[i].code();
for p in range_step(code, 1 << max_code_size as uint, step) {
decoder_table[p as uint] = i as u16;
}
}
}
}
static POLAR_MAX_SYMBOLS: u32 = 256;
pub fn polar_code_lengths(symbols: &[OrdFreq], sizes: &mut [u8]) -> u32 {
unsafe {
let mut tmp_freq: [u32, ..POLAR_MAX_SYMBOLS] = mem::uninitialized();
let mut orig_total_freq = 0;
let mut cur_total = 0;
let mut start_index = 0;
let mut max_code_size = 0;
let num_syms = symbols.len() as u32;
for i in range(0, symbols.len()) {
let sym_freq = symbols[symbols.len() - 1 - i].freq();
| sym | identifier_name |
|
prefix_code.rs | 0f0f) << 4;
v = (v & 0xcccc) >> 2 | (v & 0x3333) << 2;
v = (v & 0xaaaa) >> 1 | (v & 0x5555) << 1;
v
}
pub fn generate_codes(sizes: &[u8], codes: &mut [PrefixCode]) -> bool {
let mut num_codes: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1];
let mut next_code: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1];
for &s in sizes.iter() {
num_codes[s as uint] += 1;
}
let mut code = 0u32;
for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) {
next_code[i] = code;
code += num_codes[i];
code <<= 1;
}
if code != (1 << (MAX_EXPECTED_CODE_SIZE + 1)) {
let mut t = 0u32;
for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) {
t += num_codes[i];
if t > 1 {
//return false; // Error, sizes don't add up
fail!("Code sizes don't add up");
}
}
}
for i in range(0, sizes.len()) {
let c = sizes[i];
let code = next_code[c as uint];
next_code[c as uint] += 1;
let rev_code = reverse_u16(code) >> (16 - c as uint);
codes[i] = PrefixCode::new(rev_code, c);
}
true
}
pub fn generate_codes_for_decode(
sizes: &[u8],
codes: &mut [PrefixCode],
dec_first_offset: &mut [u16, ..17],
dec_max_code: &mut [u32, ..18],
dec_offset_to_sym: &mut [u16],
decoder_table: &mut [u16],
max_code_size: u32) -> bool {
let mut num_codes: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1]; |
let mut code = 0u32;
let mut offset = 0u32;
for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) {
next_code[i] = code;
dec_first_offset[i] = offset as u16 - code as u16;
code += num_codes[i];
dec_max_code[i] = code << (16 - i);
code <<= 1;
offset += num_codes[i];
}
dec_max_code[17] = 0x10000;
if code != (1 << (MAX_EXPECTED_CODE_SIZE + 1)) {
let mut t = 0u32;
for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) {
t += num_codes[i];
if t > 1 {
//return false; // Error, sizes don't add up
fail!("Code sizes don't add up");
}
}
}
for p in decoder_table.mut_iter() {
*p = 0xffff;
}
for i in range(0, sizes.len()) {
let s = sizes[i] as uint;
let code = next_code[s];
next_code[s] += 1;
let offset = (code as u16 + dec_first_offset[s]) as uint;
dec_offset_to_sym[offset] = i as u16;
let rev_code = reverse_u16(code) >> (16 - s);
codes[i] = PrefixCode::new(rev_code, s as u8);
if s as u32 <= max_code_size {
let step = 1 << s;
let code = rev_code;
for p in range_step(code, 1 << max_code_size as uint, step) {
decoder_table[p as uint] = i as u16;
}
}
}
true
}
pub fn generate_decoder_table(codes: &[PrefixCode], decoder_table: &mut [u16], max_code_size: u32) {
assert!(decoder_table.len() == (1 << max_code_size as uint));
for p in decoder_table.mut_iter() {
*p = 0xffff;
}
for i in range(0, codes.len()) {
if codes[i].size() as u32 <= max_code_size {
assert!(codes[i].size() > 0);
let step = 1 << codes[i].size() as uint;
let code = codes[i].code();
for p in range_step(code, 1 << max_code_size as uint, step) {
decoder_table[p as uint] = i as u16;
}
}
}
}
static POLAR_MAX_SYMBOLS: u32 = 256;
pub fn polar_code_lengths(symbols: &[OrdFreq], sizes: &mut [u8]) -> u32 {
unsafe {
let mut tmp_freq: [u32, ..POLAR_MAX_SYMBOLS] = mem::uninitialized();
let mut orig_total_freq = 0;
let mut cur_total = 0;
let mut start_index = 0;
let mut max_code_size = 0;
let num_syms = symbols.len() as u32;
for i in range(0, symbols.len()) {
let sym_freq = symbols[symbols.len() - 1 - i].freq();
//let sym_freq = symbols[i].freq();
let sym_len = 31 - ctlz32(sym_freq);
let adjusted_sym_freq = 1 << sym_len as uint;
orig_total_freq += sym_freq;
tmp_freq[i] = adjusted_sym_freq;
cur_total += adjusted_sym_freq;
}
let mut tree_total = 1 << (31 - ctlz32(orig_total_freq)) as uint;
if tree_total < orig_total_freq {
tree_total <<= 1;
}
while cur_total < tree_total && start_index < num_syms {
let mut i = start_index;
while i < num_syms {
let freq = tmp_freq[i as uint];
if cur_total + freq <= tree_total {
tmp_freq[i as uint] += freq;
cur_total += freq;
if cur_total == tree_total {
break;
}
} else {
start_index = i + 1;
}
i += 1;
}
}
assert_eq!(cur_total, tree_total);
let tree_total_bits = 32 - ctlz32(tree_total);
for i in range(0, symbols.len()) {
let codesize = tree_total_bits - (32 - ctlz32(tmp_freq[i]));
max_code_size = max(max_code_size, codesize);
sizes[symbols[symbols.len() - 1 - i].sym() as uint] = codesize as u8;
//sizes[symbols[i].sym() as uint] = codesize as u8;
}
max_code_size
}
}
pub trait PrefixModel {
fn incr(&mut self, sym: u32);
fn update(&mut self, for_encoding: bool);
fn write<BW: ::prefix_code::BitWriter>(&mut self, bw: &mut BW, sym: u32);
fn read<BR: ::prefix_code::BitReader>(&mut self, br: &mut BR) -> u32;
}
pub trait BitWriter {
fn push_bits_uni(&mut self, bits: u32, count: u32);
}
pub trait BitReader {
fn pull_bits_uni(&mut self, count: u32) -> u32;
fn peek_bits_uni16(&self) -> u16;
fn skip_bits_uni(&mut self, count: u32);
}
#[deriving(Copy)]
pub struct Foo {
f: [u32, ..256]
}
impl Clone for Foo {
fn clone(&self) -> Foo {
Foo {
f: self.f
}
}
}
macro_rules! define_polar_model(
($name: ident, $symbol_count: expr) => {
//#[deriving(Clone)]
pub struct $name {
freq: [u32, ..$symbol_count],
codes: [::prefix_code::PrefixCode, ..$symbol_count],
decoder_table: [u16, ..(1 << 9)],
sum: u32,
next_rebuild: u32,
dec_max_code: [u32, ..18],
dec_first_offset: [u16, ..17],
dec_offset_to_sym: [u16, ..$symbol_count]
}
| let mut next_code: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1];
for &s in sizes.iter() {
num_codes[s as uint] += 1;
} | random_line_split |
fma_utils.py | r.raise_for_status()
tracks = []
artists = []
date_created = []
for track in r.json()['aTracks']:
|
return tracks, artists, date_created
def _get_data(self, dataset, fma_id, fields=None):
url = self.BASE_URL + dataset + 's.json?'
url += dataset + '_id=' + str(fma_id) + '&api_key=' + self.api_key
# print(url)
r = requests.get(url)
r.raise_for_status()
if r.json()['errors']:
raise Exception(r.json()['errors'])
data = r.json()['dataset'][0]
r_id = data[dataset + '_id']
if r_id != str(fma_id):
raise Exception('The received id {} does not correspond to'
'the requested one {}'.format(r_id, fma_id))
if fields is None:
return data
if type(fields) is list:
ret = {}
for field in fields:
ret[field] = data[field]
return ret
else:
return data[fields]
def get_track(self, track_id, fields=None):
return self._get_data('track', track_id, fields)
def get_album(self, album_id, fields=None):
return self._get_data('album', album_id, fields)
def get_artist(self, artist_id, fields=None):
return self._get_data('artist', artist_id, fields)
def get_all(self, dataset, id_range):
index = dataset + '_id'
id_ = 2 if dataset is 'track' else 1
row = self._get_data(dataset, id_)
df = pd.DataFrame(columns=row.keys())
df.set_index(index, inplace=True)
not_found_ids = []
for id_ in id_range:
try:
row = self._get_data(dataset, id_)
except:
not_found_ids.append(id_)
continue
row.pop(index)
df.loc[id_] = row
return df, not_found_ids
def download_track(self, track_file, path):
url = 'https://files.freemusicarchive.org/' + track_file
r = requests.get(url, stream=True)
r.raise_for_status()
with open(path, 'wb') as f:
shutil.copyfileobj(r.raw, f)
def get_track_genres(self, track_id):
genres = self.get_track(track_id, 'track_genres')
genre_ids = []
genre_titles = []
for genre in genres:
genre_ids.append(genre['genre_id'])
genre_titles.append(genre['genre_title'])
return genre_ids, genre_titles
def get_all_genres(self):
df = pd.DataFrame(columns=['genre_parent_id', 'genre_title',
'genre_handle', 'genre_color'])
df.index.rename('genre_id', inplace=True)
page = 1
while True:
url = self.BASE_URL + 'genres.json?limit=50'
url += '&page={}&api_key={}'.format(page, self.api_key)
r = requests.get(url)
for genre in r.json()['dataset']:
genre_id = int(genre.pop(df.index.name))
df.loc[genre_id] = genre
assert (r.json()['page'] == str(page))
page += 1
if page > r.json()['total_pages']:
break
return df
class Genres:
def __init__(self, genres_df):
self.df = genres_df
def create_tree(self, roots, depth=None):
if type(roots) is not list:
roots = [roots]
graph = pydot.Dot(graph_type='digraph', strict=True)
def create_node(genre_id):
title = self.df.at[genre_id, 'title']
ntracks = self.df.at[genre_id, '#tracks']
#name = self.df.at[genre_id, 'title'] + '\n' + str(genre_id)
name = '"{}\n{} / {}"'.format(title, genre_id, ntracks)
return pydot.Node(name)
def create_tree(root_id, node_p, depth):
if depth == 0:
return
children = self.df[self.df['parent'] == root_id]
for child in children.iterrows():
genre_id = child[0]
node_c = create_node(genre_id)
graph.add_edge(pydot.Edge(node_p, node_c))
create_tree(genre_id, node_c,
depth-1 if depth is not None else None)
for root in roots:
node_p = create_node(root)
graph.add_node(node_p)
create_tree(root, node_p, depth)
return graph
def find_roots(self):
roots = []
for gid, row in self.df.iterrows():
parent = row['parent']
title = row['title']
if parent == 0:
roots.append(gid)
elif parent not in self.df.index:
msg = '{} ({}) has parent {} which is missing'.format(
gid, title, parent)
raise RuntimeError(msg)
return roots
def load(filepath):
filename = os.path.basename(filepath)
if 'features' in filename:
return pd.read_csv(filepath, index_col=0, header=[0, 1, 2])
if 'echonest' in filename:
return pd.read_csv(filepath, index_col=0, header=[0, 1, 2])
if 'genres' in filename:
return pd.read_csv(filepath, index_col=0)
if 'tracks' in filename:
tracks = pd.read_csv(filepath, index_col=0, header=[0, 1])
COLUMNS = [('track', 'tags'), ('album', 'tags'), ('artist', 'tags'),
('track', 'genres'), ('track', 'genres_all')]
for column in COLUMNS:
tracks[column] = tracks[column].map(ast.literal_eval)
COLUMNS = [('track', 'date_created'), ('track', 'date_recorded'),
('album', 'date_created'), ('album', 'date_released'),
('artist', 'date_created'), ('artist', 'active_year_begin'),
('artist', 'active_year_end')]
for column in COLUMNS:
tracks[column] = pd.to_datetime(tracks[column])
SUBSETS = ('small', 'medium', 'large')
tracks['set', 'subset'] = tracks['set', 'subset'].astype(
'category', categories=SUBSETS, ordered=True)
COLUMNS = [('track', 'genre_top'), ('track', 'license'),
('album', 'type'), ('album', 'information'),
('artist', 'bio')]
for column in COLUMNS:
tracks[column] = tracks[column].astype('category')
return tracks
def get_audio_path(audio_dir, track_id):
tid_str = '{:06d}'.format(track_id)
return os.path.join(audio_dir, tid_str[:3], tid_str + '.mp3')
class Loader:
def load(self, filepath):
raise NotImplemented()
class RawAudioLoader(Loader):
def __init__(self, sampling_rate=SAMPLING_RATE):
self.sampling_rate = sampling_rate
self.shape = (NB_AUDIO_SAMPLES * sampling_rate // SAMPLING_RATE, )
def load(self, filepath):
return self._load(filepath)[:self.shape[0]]
class LibrosaLoader(RawAudioLoader):
def _load(self, filepath):
import librosa
sr = self.sampling_rate if self.sampling_rate != SAMPLING_RATE else None
# kaiser_fast is 3x faster than kaiser_best
#x, sr = librosa.load(filepath, sr=sr, res_type='kaiser_fast')
x, sr = librosa.load(filepath, sr=sr)
return x
class AudioreadLoader(RawAudioLoader):
def _load(self, filepath):
import audioread
a = audioread.audio_open(filepath)
a.read_data()
class PydubLoader(RawAudioLoader):
def _load(self, filepath):
from pydub import AudioSegment
song = AudioSegment.from_file(filepath)
song = song.set_channels(1)
x = song.get_array_of_samples()
# print(filepath) if song.channels != 2 else None
return np.array(x)
class FfmpegLoader(RawAudioLoader):
def _load(self, filepath):
"""Fastest and less CPU intensive loading method."""
import subprocess as sp
command = ['ffmpeg',
'-i', filepath,
'-f', 's16le',
'-acodec', 'pcm_s16le',
'-ac', '1'] # channels: 2 for stereo, 1 for mono
if self.sampling_rate != SAMPLING_RATE:
command.extend(['-ar', str(self.sampling_rate)])
command.append('-')
# 30s at 44.1 kHz ~= 1.3e6
proc = sp.run(command, stdout=sp.PIPE, bufsize=10**7, stderr=sp.DEVNULL, check=True)
return np.fromstring(proc.stdout, dtype="int16")
def build_sample_loader(audio_dir, Y, loader):
class SampleLoader:
def __init__(self, tids, batch_size=4):
self.lock1 = multiprocessing.Lock()
| tracks.append(track['track_id'])
artists.append(track['artist_name'])
date_created.append(track['track_date_created']) | conditional_block |
fma_utils.py | r.raise_for_status()
tracks = []
artists = []
date_created = []
for track in r.json()['aTracks']:
tracks.append(track['track_id'])
artists.append(track['artist_name'])
date_created.append(track['track_date_created'])
return tracks, artists, date_created
def _get_data(self, dataset, fma_id, fields=None):
url = self.BASE_URL + dataset + 's.json?'
url += dataset + '_id=' + str(fma_id) + '&api_key=' + self.api_key
# print(url)
r = requests.get(url)
r.raise_for_status()
if r.json()['errors']:
raise Exception(r.json()['errors'])
data = r.json()['dataset'][0]
r_id = data[dataset + '_id']
if r_id != str(fma_id):
raise Exception('The received id {} does not correspond to'
'the requested one {}'.format(r_id, fma_id))
if fields is None:
return data
if type(fields) is list:
ret = {}
for field in fields:
ret[field] = data[field]
return ret
else:
return data[fields]
def get_track(self, track_id, fields=None):
return self._get_data('track', track_id, fields)
def get_album(self, album_id, fields=None):
return self._get_data('album', album_id, fields)
def get_artist(self, artist_id, fields=None):
return self._get_data('artist', artist_id, fields)
def get_all(self, dataset, id_range):
index = dataset + '_id'
id_ = 2 if dataset is 'track' else 1
row = self._get_data(dataset, id_)
df = pd.DataFrame(columns=row.keys())
df.set_index(index, inplace=True)
not_found_ids = []
for id_ in id_range:
try:
row = self._get_data(dataset, id_)
except:
not_found_ids.append(id_)
continue
row.pop(index)
df.loc[id_] = row
return df, not_found_ids
def download_track(self, track_file, path):
url = 'https://files.freemusicarchive.org/' + track_file
r = requests.get(url, stream=True)
r.raise_for_status()
with open(path, 'wb') as f:
shutil.copyfileobj(r.raw, f)
def get_track_genres(self, track_id):
genres = self.get_track(track_id, 'track_genres')
genre_ids = []
genre_titles = []
for genre in genres:
genre_ids.append(genre['genre_id'])
genre_titles.append(genre['genre_title'])
return genre_ids, genre_titles
def get_all_genres(self):
df = pd.DataFrame(columns=['genre_parent_id', 'genre_title',
'genre_handle', 'genre_color'])
df.index.rename('genre_id', inplace=True)
page = 1
while True:
url = self.BASE_URL + 'genres.json?limit=50'
url += '&page={}&api_key={}'.format(page, self.api_key)
r = requests.get(url)
for genre in r.json()['dataset']:
genre_id = int(genre.pop(df.index.name))
df.loc[genre_id] = genre
assert (r.json()['page'] == str(page))
page += 1
if page > r.json()['total_pages']:
break
return df
class Genres:
def __init__(self, genres_df):
self.df = genres_df
def create_tree(self, roots, depth=None):
if type(roots) is not list:
roots = [roots]
graph = pydot.Dot(graph_type='digraph', strict=True)
def create_node(genre_id):
title = self.df.at[genre_id, 'title']
ntracks = self.df.at[genre_id, '#tracks']
#name = self.df.at[genre_id, 'title'] + '\n' + str(genre_id)
name = '"{}\n{} / {}"'.format(title, genre_id, ntracks)
return pydot.Node(name)
def create_tree(root_id, node_p, depth):
if depth == 0:
return
children = self.df[self.df['parent'] == root_id]
for child in children.iterrows():
genre_id = child[0]
node_c = create_node(genre_id)
graph.add_edge(pydot.Edge(node_p, node_c))
create_tree(genre_id, node_c,
depth-1 if depth is not None else None)
for root in roots:
node_p = create_node(root)
graph.add_node(node_p)
create_tree(root, node_p, depth)
return graph
def find_roots(self):
roots = []
for gid, row in self.df.iterrows():
parent = row['parent']
title = row['title']
if parent == 0:
roots.append(gid)
elif parent not in self.df.index:
msg = '{} ({}) has parent {} which is missing'.format(
gid, title, parent)
raise RuntimeError(msg)
return roots
def load(filepath):
filename = os.path.basename(filepath)
if 'features' in filename:
return pd.read_csv(filepath, index_col=0, header=[0, 1, 2])
if 'echonest' in filename:
return pd.read_csv(filepath, index_col=0, header=[0, 1, 2])
if 'genres' in filename:
return pd.read_csv(filepath, index_col=0)
if 'tracks' in filename:
tracks = pd.read_csv(filepath, index_col=0, header=[0, 1])
COLUMNS = [('track', 'tags'), ('album', 'tags'), ('artist', 'tags'),
('track', 'genres'), ('track', 'genres_all')]
for column in COLUMNS:
tracks[column] = tracks[column].map(ast.literal_eval)
COLUMNS = [('track', 'date_created'), ('track', 'date_recorded'),
('album', 'date_created'), ('album', 'date_released'),
('artist', 'date_created'), ('artist', 'active_year_begin'),
('artist', 'active_year_end')]
for column in COLUMNS:
tracks[column] = pd.to_datetime(tracks[column])
SUBSETS = ('small', 'medium', 'large')
tracks['set', 'subset'] = tracks['set', 'subset'].astype(
'category', categories=SUBSETS, ordered=True)
COLUMNS = [('track', 'genre_top'), ('track', 'license'),
('album', 'type'), ('album', 'information'),
('artist', 'bio')]
for column in COLUMNS:
tracks[column] = tracks[column].astype('category')
return tracks
def get_audio_path(audio_dir, track_id):
tid_str = '{:06d}'.format(track_id)
return os.path.join(audio_dir, tid_str[:3], tid_str + '.mp3')
class | :
def load(self, filepath):
raise NotImplemented()
class RawAudioLoader(Loader):
def __init__(self, sampling_rate=SAMPLING_RATE):
self.sampling_rate = sampling_rate
self.shape = (NB_AUDIO_SAMPLES * sampling_rate // SAMPLING_RATE, )
def load(self, filepath):
return self._load(filepath)[:self.shape[0]]
class LibrosaLoader(RawAudioLoader):
def _load(self, filepath):
import librosa
sr = self.sampling_rate if self.sampling_rate != SAMPLING_RATE else None
# kaiser_fast is 3x faster than kaiser_best
#x, sr = librosa.load(filepath, sr=sr, res_type='kaiser_fast')
x, sr = librosa.load(filepath, sr=sr)
return x
class AudioreadLoader(RawAudioLoader):
def _load(self, filepath):
import audioread
a = audioread.audio_open(filepath)
a.read_data()
class PydubLoader(RawAudioLoader):
def _load(self, filepath):
from pydub import AudioSegment
song = AudioSegment.from_file(filepath)
song = song.set_channels(1)
x = song.get_array_of_samples()
# print(filepath) if song.channels != 2 else None
return np.array(x)
class FfmpegLoader(RawAudioLoader):
def _load(self, filepath):
"""Fastest and less CPU intensive loading method."""
import subprocess as sp
command = ['ffmpeg',
'-i', filepath,
'-f', 's16le',
'-acodec', 'pcm_s16le',
'-ac', '1'] # channels: 2 for stereo, 1 for mono
if self.sampling_rate != SAMPLING_RATE:
command.extend(['-ar', str(self.sampling_rate)])
command.append('-')
# 30s at 44.1 kHz ~= 1.3e6
proc = sp.run(command, stdout=sp.PIPE, bufsize=10**7, stderr=sp.DEVNULL, check=True)
return np.fromstring(proc.stdout, dtype="int16")
def build_sample_loader(audio_dir, Y, loader):
class SampleLoader:
def __init__(self, tids, batch_size=4):
self.lock1 = multiprocessing.Lock()
| Loader | identifier_name |
fma_utils.py | )
r.raise_for_status()
tracks = []
artists = []
date_created = []
for track in r.json()['aTracks']:
tracks.append(track['track_id'])
artists.append(track['artist_name'])
date_created.append(track['track_date_created'])
return tracks, artists, date_created
def _get_data(self, dataset, fma_id, fields=None):
| return data[fields]
def get_track(self, track_id, fields=None):
return self._get_data('track', track_id, fields)
def get_album(self, album_id, fields=None):
return self._get_data('album', album_id, fields)
def get_artist(self, artist_id, fields=None):
return self._get_data('artist', artist_id, fields)
def get_all(self, dataset, id_range):
index = dataset + '_id'
id_ = 2 if dataset is 'track' else 1
row = self._get_data(dataset, id_)
df = pd.DataFrame(columns=row.keys())
df.set_index(index, inplace=True)
not_found_ids = []
for id_ in id_range:
try:
row = self._get_data(dataset, id_)
except:
not_found_ids.append(id_)
continue
row.pop(index)
df.loc[id_] = row
return df, not_found_ids
def download_track(self, track_file, path):
url = 'https://files.freemusicarchive.org/' + track_file
r = requests.get(url, stream=True)
r.raise_for_status()
with open(path, 'wb') as f:
shutil.copyfileobj(r.raw, f)
def get_track_genres(self, track_id):
genres = self.get_track(track_id, 'track_genres')
genre_ids = []
genre_titles = []
for genre in genres:
genre_ids.append(genre['genre_id'])
genre_titles.append(genre['genre_title'])
return genre_ids, genre_titles
def get_all_genres(self):
df = pd.DataFrame(columns=['genre_parent_id', 'genre_title',
'genre_handle', 'genre_color'])
df.index.rename('genre_id', inplace=True)
page = 1
while True:
url = self.BASE_URL + 'genres.json?limit=50'
url += '&page={}&api_key={}'.format(page, self.api_key)
r = requests.get(url)
for genre in r.json()['dataset']:
genre_id = int(genre.pop(df.index.name))
df.loc[genre_id] = genre
assert (r.json()['page'] == str(page))
page += 1
if page > r.json()['total_pages']:
break
return df
class Genres:
def __init__(self, genres_df):
self.df = genres_df
def create_tree(self, roots, depth=None):
if type(roots) is not list:
roots = [roots]
graph = pydot.Dot(graph_type='digraph', strict=True)
def create_node(genre_id):
title = self.df.at[genre_id, 'title']
ntracks = self.df.at[genre_id, '#tracks']
#name = self.df.at[genre_id, 'title'] + '\n' + str(genre_id)
name = '"{}\n{} / {}"'.format(title, genre_id, ntracks)
return pydot.Node(name)
def create_tree(root_id, node_p, depth):
if depth == 0:
return
children = self.df[self.df['parent'] == root_id]
for child in children.iterrows():
genre_id = child[0]
node_c = create_node(genre_id)
graph.add_edge(pydot.Edge(node_p, node_c))
create_tree(genre_id, node_c,
depth-1 if depth is not None else None)
for root in roots:
node_p = create_node(root)
graph.add_node(node_p)
create_tree(root, node_p, depth)
return graph
def find_roots(self):
roots = []
for gid, row in self.df.iterrows():
parent = row['parent']
title = row['title']
if parent == 0:
roots.append(gid)
elif parent not in self.df.index:
msg = '{} ({}) has parent {} which is missing'.format(
gid, title, parent)
raise RuntimeError(msg)
return roots
def load(filepath):
filename = os.path.basename(filepath)
if 'features' in filename:
return pd.read_csv(filepath, index_col=0, header=[0, 1, 2])
if 'echonest' in filename:
return pd.read_csv(filepath, index_col=0, header=[0, 1, 2])
if 'genres' in filename:
return pd.read_csv(filepath, index_col=0)
if 'tracks' in filename:
tracks = pd.read_csv(filepath, index_col=0, header=[0, 1])
COLUMNS = [('track', 'tags'), ('album', 'tags'), ('artist', 'tags'),
('track', 'genres'), ('track', 'genres_all')]
for column in COLUMNS:
tracks[column] = tracks[column].map(ast.literal_eval)
COLUMNS = [('track', 'date_created'), ('track', 'date_recorded'),
('album', 'date_created'), ('album', 'date_released'),
('artist', 'date_created'), ('artist', 'active_year_begin'),
('artist', 'active_year_end')]
for column in COLUMNS:
tracks[column] = pd.to_datetime(tracks[column])
SUBSETS = ('small', 'medium', 'large')
tracks['set', 'subset'] = tracks['set', 'subset'].astype(
'category', categories=SUBSETS, ordered=True)
COLUMNS = [('track', 'genre_top'), ('track', 'license'),
('album', 'type'), ('album', 'information'),
('artist', 'bio')]
for column in COLUMNS:
tracks[column] = tracks[column].astype('category')
return tracks
def get_audio_path(audio_dir, track_id):
tid_str = '{:06d}'.format(track_id)
return os.path.join(audio_dir, tid_str[:3], tid_str + '.mp3')
class Loader:
def load(self, filepath):
raise NotImplemented()
class RawAudioLoader(Loader):
def __init__(self, sampling_rate=SAMPLING_RATE):
self.sampling_rate = sampling_rate
self.shape = (NB_AUDIO_SAMPLES * sampling_rate // SAMPLING_RATE, )
def load(self, filepath):
return self._load(filepath)[:self.shape[0]]
class LibrosaLoader(RawAudioLoader):
def _load(self, filepath):
import librosa
sr = self.sampling_rate if self.sampling_rate != SAMPLING_RATE else None
# kaiser_fast is 3x faster than kaiser_best
#x, sr = librosa.load(filepath, sr=sr, res_type='kaiser_fast')
x, sr = librosa.load(filepath, sr=sr)
return x
class AudioreadLoader(RawAudioLoader):
def _load(self, filepath):
import audioread
a = audioread.audio_open(filepath)
a.read_data()
class PydubLoader(RawAudioLoader):
def _load(self, filepath):
from pydub import AudioSegment
song = AudioSegment.from_file(filepath)
song = song.set_channels(1)
x = song.get_array_of_samples()
# print(filepath) if song.channels != 2 else None
return np.array(x)
class FfmpegLoader(RawAudioLoader):
def _load(self, filepath):
"""Fastest and less CPU intensive loading method."""
import subprocess as sp
command = ['ffmpeg',
'-i', filepath,
'-f', 's16le',
'-acodec', 'pcm_s16le',
'-ac', '1'] # channels: 2 for stereo, 1 for mono
if self.sampling_rate != SAMPLING_RATE:
command.extend(['-ar', str(self.sampling_rate)])
command.append('-')
# 30s at 44.1 kHz ~= 1.3e6
proc = sp.run(command, stdout=sp.PIPE, bufsize=10**7, stderr=sp.DEVNULL, check=True)
return np.fromstring(proc.stdout, dtype="int16")
def build_sample_loader(audio_dir, Y, loader):
class SampleLoader:
def __init__(self, tids, batch_size=4):
self.lock1 = multiprocessing.Lock()
| url = self.BASE_URL + dataset + 's.json?'
url += dataset + '_id=' + str(fma_id) + '&api_key=' + self.api_key
# print(url)
r = requests.get(url)
r.raise_for_status()
if r.json()['errors']:
raise Exception(r.json()['errors'])
data = r.json()['dataset'][0]
r_id = data[dataset + '_id']
if r_id != str(fma_id):
raise Exception('The received id {} does not correspond to'
'the requested one {}'.format(r_id, fma_id))
if fields is None:
return data
if type(fields) is list:
ret = {}
for field in fields:
ret[field] = data[field]
return ret
else: | identifier_body |
fma_utils.py | )
r.raise_for_status()
tracks = []
artists = []
date_created = []
for track in r.json()['aTracks']:
tracks.append(track['track_id'])
artists.append(track['artist_name'])
date_created.append(track['track_date_created'])
return tracks, artists, date_created
def _get_data(self, dataset, fma_id, fields=None):
url = self.BASE_URL + dataset + 's.json?'
url += dataset + '_id=' + str(fma_id) + '&api_key=' + self.api_key
# print(url)
r = requests.get(url)
r.raise_for_status()
if r.json()['errors']:
raise Exception(r.json()['errors'])
data = r.json()['dataset'][0]
r_id = data[dataset + '_id']
if r_id != str(fma_id):
raise Exception('The received id {} does not correspond to'
'the requested one {}'.format(r_id, fma_id))
if fields is None:
return data
if type(fields) is list:
ret = {}
for field in fields:
ret[field] = data[field]
return ret
else:
return data[fields]
def get_track(self, track_id, fields=None):
return self._get_data('track', track_id, fields)
def get_album(self, album_id, fields=None):
return self._get_data('album', album_id, fields)
def get_artist(self, artist_id, fields=None):
return self._get_data('artist', artist_id, fields)
def get_all(self, dataset, id_range):
index = dataset + '_id'
id_ = 2 if dataset is 'track' else 1
row = self._get_data(dataset, id_)
df = pd.DataFrame(columns=row.keys())
df.set_index(index, inplace=True)
not_found_ids = []
for id_ in id_range:
try:
row = self._get_data(dataset, id_)
except:
not_found_ids.append(id_)
continue
row.pop(index)
df.loc[id_] = row
return df, not_found_ids
def download_track(self, track_file, path):
url = 'https://files.freemusicarchive.org/' + track_file
r = requests.get(url, stream=True)
r.raise_for_status()
with open(path, 'wb') as f:
shutil.copyfileobj(r.raw, f)
def get_track_genres(self, track_id):
genres = self.get_track(track_id, 'track_genres')
genre_ids = []
genre_titles = []
for genre in genres:
genre_ids.append(genre['genre_id'])
genre_titles.append(genre['genre_title'])
return genre_ids, genre_titles
def get_all_genres(self):
df = pd.DataFrame(columns=['genre_parent_id', 'genre_title',
'genre_handle', 'genre_color'])
df.index.rename('genre_id', inplace=True)
page = 1
while True:
url = self.BASE_URL + 'genres.json?limit=50'
url += '&page={}&api_key={}'.format(page, self.api_key)
r = requests.get(url)
for genre in r.json()['dataset']:
genre_id = int(genre.pop(df.index.name))
df.loc[genre_id] = genre
assert (r.json()['page'] == str(page))
page += 1
if page > r.json()['total_pages']:
break
return df
class Genres:
def __init__(self, genres_df):
self.df = genres_df
def create_tree(self, roots, depth=None):
if type(roots) is not list:
roots = [roots]
graph = pydot.Dot(graph_type='digraph', strict=True)
def create_node(genre_id):
title = self.df.at[genre_id, 'title']
ntracks = self.df.at[genre_id, '#tracks'] | def create_tree(root_id, node_p, depth):
if depth == 0:
return
children = self.df[self.df['parent'] == root_id]
for child in children.iterrows():
genre_id = child[0]
node_c = create_node(genre_id)
graph.add_edge(pydot.Edge(node_p, node_c))
create_tree(genre_id, node_c,
depth-1 if depth is not None else None)
for root in roots:
node_p = create_node(root)
graph.add_node(node_p)
create_tree(root, node_p, depth)
return graph
def find_roots(self):
roots = []
for gid, row in self.df.iterrows():
parent = row['parent']
title = row['title']
if parent == 0:
roots.append(gid)
elif parent not in self.df.index:
msg = '{} ({}) has parent {} which is missing'.format(
gid, title, parent)
raise RuntimeError(msg)
return roots
def load(filepath):
filename = os.path.basename(filepath)
if 'features' in filename:
return pd.read_csv(filepath, index_col=0, header=[0, 1, 2])
if 'echonest' in filename:
return pd.read_csv(filepath, index_col=0, header=[0, 1, 2])
if 'genres' in filename:
return pd.read_csv(filepath, index_col=0)
if 'tracks' in filename:
tracks = pd.read_csv(filepath, index_col=0, header=[0, 1])
COLUMNS = [('track', 'tags'), ('album', 'tags'), ('artist', 'tags'),
('track', 'genres'), ('track', 'genres_all')]
for column in COLUMNS:
tracks[column] = tracks[column].map(ast.literal_eval)
COLUMNS = [('track', 'date_created'), ('track', 'date_recorded'),
('album', 'date_created'), ('album', 'date_released'),
('artist', 'date_created'), ('artist', 'active_year_begin'),
('artist', 'active_year_end')]
for column in COLUMNS:
tracks[column] = pd.to_datetime(tracks[column])
SUBSETS = ('small', 'medium', 'large')
tracks['set', 'subset'] = tracks['set', 'subset'].astype(
'category', categories=SUBSETS, ordered=True)
COLUMNS = [('track', 'genre_top'), ('track', 'license'),
('album', 'type'), ('album', 'information'),
('artist', 'bio')]
for column in COLUMNS:
tracks[column] = tracks[column].astype('category')
return tracks
def get_audio_path(audio_dir, track_id):
tid_str = '{:06d}'.format(track_id)
return os.path.join(audio_dir, tid_str[:3], tid_str + '.mp3')
class Loader:
def load(self, filepath):
raise NotImplemented()
class RawAudioLoader(Loader):
def __init__(self, sampling_rate=SAMPLING_RATE):
self.sampling_rate = sampling_rate
self.shape = (NB_AUDIO_SAMPLES * sampling_rate // SAMPLING_RATE, )
def load(self, filepath):
return self._load(filepath)[:self.shape[0]]
class LibrosaLoader(RawAudioLoader):
def _load(self, filepath):
import librosa
sr = self.sampling_rate if self.sampling_rate != SAMPLING_RATE else None
# kaiser_fast is 3x faster than kaiser_best
#x, sr = librosa.load(filepath, sr=sr, res_type='kaiser_fast')
x, sr = librosa.load(filepath, sr=sr)
return x
class AudioreadLoader(RawAudioLoader):
def _load(self, filepath):
import audioread
a = audioread.audio_open(filepath)
a.read_data()
class PydubLoader(RawAudioLoader):
def _load(self, filepath):
from pydub import AudioSegment
song = AudioSegment.from_file(filepath)
song = song.set_channels(1)
x = song.get_array_of_samples()
# print(filepath) if song.channels != 2 else None
return np.array(x)
class FfmpegLoader(RawAudioLoader):
def _load(self, filepath):
"""Fastest and less CPU intensive loading method."""
import subprocess as sp
command = ['ffmpeg',
'-i', filepath,
'-f', 's16le',
'-acodec', 'pcm_s16le',
'-ac', '1'] # channels: 2 for stereo, 1 for mono
if self.sampling_rate != SAMPLING_RATE:
command.extend(['-ar', str(self.sampling_rate)])
command.append('-')
# 30s at 44.1 kHz ~= 1.3e6
proc = sp.run(command, stdout=sp.PIPE, bufsize=10**7, stderr=sp.DEVNULL, check=True)
return np.fromstring(proc.stdout, dtype="int16")
def build_sample_loader(audio_dir, Y, loader):
class SampleLoader:
def __init__(self, tids, batch_size=4):
self.lock1 = multiprocessing.Lock()
| #name = self.df.at[genre_id, 'title'] + '\n' + str(genre_id)
name = '"{}\n{} / {}"'.format(title, genre_id, ntracks)
return pydot.Node(name)
| random_line_split |
main.rs | .x = 5;
*r.y = 6;
}
assert_eq!(5, point.x);
assert_eq!(6, point.y);
point = Point { x: 0, ..point};
assert_eq!(6, point.y);
struct Color(i32, i32, i32);
let black = Color(17, 0, 0);
let Color(r, _, _) = black;
println!("{}", r);
enum Message {
Quit,
ChangeColor(i32, i32, i32),
Move { x: i32, y: i32 },
Write(String),
}
let v = vec!["Hello".to_string(), "World".to_string()];
let v1: Vec<Message> = v.into_iter().map(Message::Write).collect();
let x = 5;
match x {
1 => println!("one"),
2 => println!("two"),
3 => println!("three"),
4 => println!("four"),
5 => println!("five"),
6 | 7 => println!("six or seven"),
_ => println!("something else"),
}
let number = match x {
1 => "one",
2 => "two",
3 => "three",
4 => "four",
5 => "five",
_ => "something else",
};
let message = Message::Quit;
match message {
Message::Quit => println!("quit"),
Message::ChangeColor(r, g, b) => println!("color"),
Message::Move { x, y: new_name_for_y } => println!("move"),
Message::Write(s) => println!("write"),
};
let x = 1;
let c = 'c';
match c {
x => println!("x: {} c: {}", x, c),
}
println!("x: {}", x);
let origin = Point { x: 0, y: 0 };
let Point { x, y } = origin;
let tuple = (5, String::from("five"));
let (x, _) = tuple;
//string is not moved thanks to _
println!("Tuple is: {:?}", tuple);
let (x, ..) = tuple;
let mut x = 5;
match x {
ref name @ 1 ... 5 if *name < 5 => println!("one through four {}", name),
ref name @ 1 ... 5 if *name >= 5 => println!("five {}", name),
ref mut mr => println!("Got a mutable reference to {}", mr),
}
struct Circle {
x: f64,
y: f64,
radius: f64,
}
| impl Circle {
fn area(&self) -> f64 {
std::f64::consts::PI * (self.radius * self.radius)
}
fn reference(&self) -> &Circle{
println!("taking self by reference!");
self
}
fn mutable_reference(&mut self) {
println!("taking self by mutable reference!");
}
fn takes_ownership(self) {
println!("taking ownership of self!");
}
fn new(x: f64, y: f64, radius: f64) -> Circle {
Circle {
x: x,
y: y,
radius: radius,
}
}
}
struct CircleBuilder {
x: f64,
y: f64,
radius: f64,
}
let mut c = Circle { x: 0.0, y: 0.0, radius: 2.0 };
c = Circle::new(0.0, 0.0, 2.0);
println!("{}", c.reference().area());
impl CircleBuilder {
fn new() -> CircleBuilder {
CircleBuilder { x: 0.0, y: 0.0, radius: 1.0, }
}
fn x(&mut self, coordinate: f64) -> &mut CircleBuilder {
self.x = coordinate;
self
}
fn y(&mut self, coordinate: f64) -> &mut CircleBuilder {
self.y = coordinate;
self
}
fn radius(&mut self, radius: f64) -> &mut CircleBuilder {
self.radius = radius;
self
}
fn finalize(&self) -> Circle {
Circle { x: self.x, y: self.y, radius: self.radius }
}
}
c = CircleBuilder::new().x(1.0)
.y(2.0)
.radius(2.0)
.finalize();;
println!("{}", c.reference().area());
let greeting = "Hello there."; // greeting: &'static str
let mut s = "Hello".to_string(); // mut s: String
fn takes_slice(slice: &str) {
println!("Got: {}", slice);
}
takes_slice(&s);
for c in s.chars() {
print!("{}, ", c);
}
let c = s.chars().nth(0);
let sl = {
let tmp = &s[0..5];
println!("{}", tmp);
};
let mut concat = s + "foo";
println!("{}", concat);
let concat2 = "bar".to_string() + &concat;
println!("{}", concat2);
let x: Option<i32> = Some(5);
fn takes_anything<T>(x: T) {
// Do something with `x`.
}
takes_anything(concat2);
struct PointGeneric<T> {
x: T,
y: T,
}
impl<T> PointGeneric<T> {
fn swap(&mut self) {
std::mem::swap(&mut self.x, &mut self.y);
}
}
let int_origin = PointGeneric { x: 0, y: 0 };
let float_origin = PointGeneric { x: 0.0, y: 0.0 };
trait HasArea {
fn area(&self) -> f64;
fn is_larger(&self, &Self) -> bool;
}
impl HasArea for Circle {
fn area(&self) -> f64 {
std::f64::consts::PI * (self.radius * self.radius)
}
fn is_larger(&self, other: &Self) -> bool {
self.area() > other.area()
}
}
use std::fmt::Debug;
fn print_area<T: HasArea>(shape: T) {
println!("This shape has an area of {}", shape.area());
}
fn test <T: HasArea + Debug>(){
}
fn test2 <T>() where T : HasArea + Debug{
}
let c = Circle {
x: 0.0f64,
y: 0.0f64,
radius: 1.0f64,
};
print_area(c);
trait bar : HasArea {
fn is_valid(&self) -> bool;
fn is_invalid(&self) -> bool { !self.is_valid() }
}
#[derive(Debug)]
struct deriving;
impl Drop for Circle {
fn drop(&mut self) {
println!("Dropping!");
}
}
fn main() {
let x = Circle { x: 0.0, y: 0.0, radius: 2.0 };
// Do stuff.
}
let option: Option<i32> = Some(5);
match option {
Some(x) => { println!("match!"); },
None => {},
}
if option.is_some() {
let x = option.unwrap();
println!("match!");
}
if let Some(x) = option {
println!("match!");
}
trait FooBar {
fn method(&self) -> String;
}
impl FooBar for u8 {
fn method(&self) -> String { format!("u8: {}", *self) }
}
impl FooBar for String {
fn method(&self) -> String { format!("string: {}", *self) }
}
fn do_something<T: FooBar>(x: T) {
x.method();
}
let x = 5u8;
let y = "Hello".to_string();
do_something(x);
do_something(y);
fn do_something2(x: &FooBar) {
x.method();
}
let x = 5u8;
//casting
do_something2(&x as &FooBar);
//coercing
do_something2(&x);
let add = |x| x + 1;
println!("{}", add(2));
let mut num = 5;
{
let mut add_num = |x: i32| num += x;
add_num(5);
}
assert_eq!(10, num);
//move closure
let mut num = 5;
{
let mut add_num = move |x: i32| num += x;
add_num(5);
}
assert_eq!(5, num);
fn call_with_one<F>(closure : F) -> i32
where F: Fn(i32) -> i32{
closure(1)
}
let answer = call_with_one(|x| x + 2);
assert_eq!(3, answer);
fn call_with_one2(some_closure: &Fn(i32) -> i32) -> i32 {
some_closure(1)
}
let | random_line_split |
|
main.rs | .x = 5;
*r.y = 6;
}
assert_eq!(5, point.x);
assert_eq!(6, point.y);
point = Point { x: 0, ..point};
assert_eq!(6, point.y);
struct Color(i32, i32, i32);
let black = Color(17, 0, 0);
let Color(r, _, _) = black;
println!("{}", r);
enum Message {
Quit,
ChangeColor(i32, i32, i32),
Move { x: i32, y: i32 },
Write(String),
}
let v = vec!["Hello".to_string(), "World".to_string()];
let v1: Vec<Message> = v.into_iter().map(Message::Write).collect();
let x = 5;
match x {
1 => println!("one"),
2 => println!("two"),
3 => println!("three"),
4 => println!("four"),
5 => println!("five"),
6 | 7 => println!("six or seven"),
_ => println!("something else"),
}
let number = match x {
1 => "one",
2 => "two",
3 => "three",
4 => "four",
5 => "five",
_ => "something else",
};
let message = Message::Quit;
match message {
Message::Quit => println!("quit"),
Message::ChangeColor(r, g, b) => println!("color"),
Message::Move { x, y: new_name_for_y } => println!("move"),
Message::Write(s) => println!("write"),
};
let x = 1;
let c = 'c';
match c {
x => println!("x: {} c: {}", x, c),
}
println!("x: {}", x);
let origin = Point { x: 0, y: 0 };
let Point { x, y } = origin;
let tuple = (5, String::from("five"));
let (x, _) = tuple;
//string is not moved thanks to _
println!("Tuple is: {:?}", tuple);
let (x, ..) = tuple;
let mut x = 5;
match x {
ref name @ 1 ... 5 if *name < 5 => println!("one through four {}", name),
ref name @ 1 ... 5 if *name >= 5 => println!("five {}", name),
ref mut mr => println!("Got a mutable reference to {}", mr),
}
struct Circle {
x: f64,
y: f64,
radius: f64,
}
impl Circle {
fn area(&self) -> f64 {
std::f64::consts::PI * (self.radius * self.radius)
}
fn reference(&self) -> &Circle{
println!("taking self by reference!");
self
}
fn mutable_reference(&mut self) {
println!("taking self by mutable reference!");
}
fn takes_ownership(self) {
println!("taking ownership of self!");
}
fn new(x: f64, y: f64, radius: f64) -> Circle {
Circle {
x: x,
y: y,
radius: radius,
}
}
}
struct CircleBuilder {
x: f64,
y: f64,
radius: f64,
}
let mut c = Circle { x: 0.0, y: 0.0, radius: 2.0 };
c = Circle::new(0.0, 0.0, 2.0);
println!("{}", c.reference().area());
impl CircleBuilder {
fn new() -> CircleBuilder {
CircleBuilder { x: 0.0, y: 0.0, radius: 1.0, }
}
fn x(&mut self, coordinate: f64) -> &mut CircleBuilder {
self.x = coordinate;
self
}
fn y(&mut self, coordinate: f64) -> &mut CircleBuilder {
self.y = coordinate;
self
}
fn radius(&mut self, radius: f64) -> &mut CircleBuilder {
self.radius = radius;
self
}
fn finalize(&self) -> Circle {
Circle { x: self.x, y: self.y, radius: self.radius }
}
}
c = CircleBuilder::new().x(1.0)
.y(2.0)
.radius(2.0)
.finalize();;
println!("{}", c.reference().area());
let greeting = "Hello there."; // greeting: &'static str
let mut s = "Hello".to_string(); // mut s: String
fn takes_slice(slice: &str) {
println!("Got: {}", slice);
}
takes_slice(&s);
for c in s.chars() {
print!("{}, ", c);
}
let c = s.chars().nth(0);
let sl = {
let tmp = &s[0..5];
println!("{}", tmp);
};
let mut concat = s + "foo";
println!("{}", concat);
let concat2 = "bar".to_string() + &concat;
println!("{}", concat2);
let x: Option<i32> = Some(5);
fn takes_anything<T>(x: T) {
// Do something with `x`.
}
takes_anything(concat2);
struct PointGeneric<T> {
x: T,
y: T,
}
impl<T> PointGeneric<T> {
fn swap(&mut self) {
std::mem::swap(&mut self.x, &mut self.y);
}
}
let int_origin = PointGeneric { x: 0, y: 0 };
let float_origin = PointGeneric { x: 0.0, y: 0.0 };
trait HasArea {
fn area(&self) -> f64;
fn is_larger(&self, &Self) -> bool;
}
impl HasArea for Circle {
fn area(&self) -> f64 {
std::f64::consts::PI * (self.radius * self.radius)
}
fn is_larger(&self, other: &Self) -> bool {
self.area() > other.area()
}
}
use std::fmt::Debug;
fn print_area<T: HasArea>(shape: T) {
println!("This shape has an area of {}", shape.area());
}
fn test <T: HasArea + Debug>(){
}
fn test2 <T>() where T : HasArea + Debug{
}
let c = Circle {
x: 0.0f64,
y: 0.0f64,
radius: 1.0f64,
};
print_area(c);
trait bar : HasArea {
fn is_valid(&self) -> bool;
fn is_invalid(&self) -> bool { !self.is_valid() }
}
#[derive(Debug)]
struct deriving;
impl Drop for Circle {
fn drop(&mut self) {
println!("Dropping!");
}
}
fn main() {
let x = Circle { x: 0.0, y: 0.0, radius: 2.0 };
// Do stuff.
}
let option: Option<i32> = Some(5);
match option {
Some(x) => { println!("match!"); },
None => {},
}
if option.is_some() {
let x = option.unwrap();
println!("match!");
}
if let Some(x) = option {
println!("match!");
}
trait FooBar {
fn method(&self) -> String;
}
impl FooBar for u8 {
fn method(&self) -> String { format!("u8: {}", *self) }
}
impl FooBar for String {
fn | (&self) -> String { format!("string: {}", *self) }
}
fn do_something<T: FooBar>(x: T) {
x.method();
}
let x = 5u8;
let y = "Hello".to_string();
do_something(x);
do_something(y);
fn do_something2(x: &FooBar) {
x.method();
}
let x = 5u8;
//casting
do_something2(&x as &FooBar);
//coercing
do_something2(&x);
let add = |x| x + 1;
println!("{}", add(2));
let mut num = 5;
{
let mut add_num = |x: i32| num += x;
add_num(5);
}
assert_eq!(10, num);
//move closure
let mut num = 5;
{
let mut add_num = move |x: i32| num += x;
add_num(5);
}
assert_eq!(5, num);
fn call_with_one<F>(closure : F) -> i32
where F: Fn(i32) -> i32{
closure(1)
}
let answer = call_with_one(|x| x + 2);
assert_eq!(3, answer);
fn call_with_one2(some_closure: &Fn(i32) -> i32) -> i32 {
some_closure(1)
}
| method | identifier_name |
main.rs | !("taking self by mutable reference!");
}
fn takes_ownership(self) {
println!("taking ownership of self!");
}
fn new(x: f64, y: f64, radius: f64) -> Circle {
Circle {
x: x,
y: y,
radius: radius,
}
}
}
struct CircleBuilder {
x: f64,
y: f64,
radius: f64,
}
let mut c = Circle { x: 0.0, y: 0.0, radius: 2.0 };
c = Circle::new(0.0, 0.0, 2.0);
println!("{}", c.reference().area());
impl CircleBuilder {
fn new() -> CircleBuilder {
CircleBuilder { x: 0.0, y: 0.0, radius: 1.0, }
}
fn x(&mut self, coordinate: f64) -> &mut CircleBuilder {
self.x = coordinate;
self
}
fn y(&mut self, coordinate: f64) -> &mut CircleBuilder {
self.y = coordinate;
self
}
fn radius(&mut self, radius: f64) -> &mut CircleBuilder {
self.radius = radius;
self
}
fn finalize(&self) -> Circle {
Circle { x: self.x, y: self.y, radius: self.radius }
}
}
c = CircleBuilder::new().x(1.0)
.y(2.0)
.radius(2.0)
.finalize();;
println!("{}", c.reference().area());
let greeting = "Hello there."; // greeting: &'static str
let mut s = "Hello".to_string(); // mut s: String
fn takes_slice(slice: &str) {
println!("Got: {}", slice);
}
takes_slice(&s);
for c in s.chars() {
print!("{}, ", c);
}
let c = s.chars().nth(0);
let sl = {
let tmp = &s[0..5];
println!("{}", tmp);
};
let mut concat = s + "foo";
println!("{}", concat);
let concat2 = "bar".to_string() + &concat;
println!("{}", concat2);
let x: Option<i32> = Some(5);
fn takes_anything<T>(x: T) {
// Do something with `x`.
}
takes_anything(concat2);
struct PointGeneric<T> {
x: T,
y: T,
}
impl<T> PointGeneric<T> {
fn swap(&mut self) {
std::mem::swap(&mut self.x, &mut self.y);
}
}
let int_origin = PointGeneric { x: 0, y: 0 };
let float_origin = PointGeneric { x: 0.0, y: 0.0 };
trait HasArea {
fn area(&self) -> f64;
fn is_larger(&self, &Self) -> bool;
}
impl HasArea for Circle {
fn area(&self) -> f64 {
std::f64::consts::PI * (self.radius * self.radius)
}
fn is_larger(&self, other: &Self) -> bool {
self.area() > other.area()
}
}
use std::fmt::Debug;
fn print_area<T: HasArea>(shape: T) {
println!("This shape has an area of {}", shape.area());
}
fn test <T: HasArea + Debug>(){
}
fn test2 <T>() where T : HasArea + Debug{
}
let c = Circle {
x: 0.0f64,
y: 0.0f64,
radius: 1.0f64,
};
print_area(c);
trait bar : HasArea {
fn is_valid(&self) -> bool;
fn is_invalid(&self) -> bool { !self.is_valid() }
}
#[derive(Debug)]
struct deriving;
impl Drop for Circle {
fn drop(&mut self) {
println!("Dropping!");
}
}
fn main() {
let x = Circle { x: 0.0, y: 0.0, radius: 2.0 };
// Do stuff.
}
let option: Option<i32> = Some(5);
match option {
Some(x) => { println!("match!"); },
None => {},
}
if option.is_some() {
let x = option.unwrap();
println!("match!");
}
if let Some(x) = option {
println!("match!");
}
trait FooBar {
fn method(&self) -> String;
}
impl FooBar for u8 {
fn method(&self) -> String { format!("u8: {}", *self) }
}
impl FooBar for String {
fn method(&self) -> String { format!("string: {}", *self) }
}
fn do_something<T: FooBar>(x: T) {
x.method();
}
let x = 5u8;
let y = "Hello".to_string();
do_something(x);
do_something(y);
fn do_something2(x: &FooBar) {
x.method();
}
let x = 5u8;
//casting
do_something2(&x as &FooBar);
//coercing
do_something2(&x);
let add = |x| x + 1;
println!("{}", add(2));
let mut num = 5;
{
let mut add_num = |x: i32| num += x;
add_num(5);
}
assert_eq!(10, num);
//move closure
let mut num = 5;
{
let mut add_num = move |x: i32| num += x;
add_num(5);
}
assert_eq!(5, num);
fn call_with_one<F>(closure : F) -> i32
where F: Fn(i32) -> i32{
closure(1)
}
let answer = call_with_one(|x| x + 2);
assert_eq!(3, answer);
fn call_with_one2(some_closure: &Fn(i32) -> i32) -> i32 {
some_closure(1)
}
let answer = call_with_one2(&|x| x + 2);
assert_eq!(3, answer);
fn call_with_ref<F>(some_closure:F) -> i32
where F: for<'a> Fn(&'a i32) -> i32 {
let value = 0;
some_closure(&value)
}
fn add_one(i: i32) -> i32 {
i + 1
}
let f = add_one;
call_with_one2(&f);
fn factory() -> Box<Fn(i32) -> i32> {
let num = 5;
Box::new(move |x| x + num)
}
let f = factory();
let answer = f(1);
assert_eq!(6, answer);
trait Foo2 {
fn f(&self);
}
trait Bar2 {
fn f(&self);
}
struct Baz;
impl Foo2 for Baz {
fn f(&self) { println!("Baz’s impl of Foo"); }
}
impl Bar2 for Baz {
fn f(&self) { println!("Baz’s impl of Bar"); }
}
let b = Baz;
Foo2::f(&b);
Bar2::f(&b);
println!("Hello in English: {}", phrases::english::greetings::hello());
println!("Hello in English: {}", hi());
//inline, several memory address
//better than static
const TOTO: i32 = 12;
//same address for all use
static mut TOTO2: i32 = 12;
unsafe {
TOTO2 = 2;
}
#[test]
fn check() {
assert_eq!(2, 1 + 1);
}
#[cfg(target_os = "macos")]
mod macos_only {
}
type mytype = String;
let s:mytype = "toto".to_string();
use std::result;
enum ConcreteError {
Foo,
Bar,
}
type Result<T> = result::Result<T, ConcreteError>;
let casty = TOTO as i64;
use std::mem;
unsafe {
let a = [0u8, 1u8, 0u8, 0u8];
let b = mem::transmute::<[u8; 4], u32>(a);
println!("{}", b);
}
trait Graph {
type N;
type E;
fn has_edge(&self, &Self::N, &Self::N) -> bool;
fn edges(&self, &Self::N) -> Vec<Self::E>;
}
struct Node;
struct Edge;
struct MyGraph;
impl Graph for MyGraph {
type N = Node;
type E = Edge;
fn has_edge(&self, n1: &Node, n2: &Node) -> bool {
true
}
fn edges(&self, n: &Node) -> Vec<Edge> {
| Vec::new()
}
}
| identifier_body |
|
static.go | {
return err
}
return nil
}
// ValidateHeader validates Header.
// Returns the first error encountered.
func ValidateHeader(header *Header) error {
if header == nil {
return consts.ErrNilHeader
}
tokenType := header.TokenTyp
if tokenType < NoType || tokenType > Jet {
return consts.ErrUnknownTokenType
}
alg := header.Alg
if alg < NoAlg || alg > Hs512 {
return consts.ErrUnknownAlgorithm
}
return nil
}
// ValidateBody validates Body.
// Checks if token string has expired.
// Returns the first error encountered.
func ValidateBody(body *Body) error {
if body == nil {
return consts.ErrNilBody
}
if err := validation.ValidateUserUUID(body.UUID); err != nil {
return err
}
permission := body.Permission
if permission < NoPermission || permission > Admin {
return consts.ErrUnknownPermission
}
if isExpired(body.ExpirationTimestamp) {
return consts.ErrExpiredBody
}
return nil
}
// ValidateSecret checks if the secret is still valid and has not expired.
// Returns an error if the Secret is not valid and has expired.
func ValidateSecret(secret *pbauth.Secret) error {
if secret == nil {
return consts.ErrNilSecret
}
if strings.TrimSpace(secret.Key) == "" {
return consts.ErrEmptySecret
}
createTime := secret.CreatedTimestamp
if createTime == 0 || createTime > time.Now().UTC().Unix() {
return consts.ErrInvalidSecretCreateTimestamp
}
if isExpired(secret.ExpirationTimestamp) {
return consts.ErrExpiredSecret
}
return nil
}
func isExpired(timestamp int64) bool {
if timestamp <= 0 || time.Now().UTC().Unix() >= timestamp {
return true
}
return false
}
// NewToken generates token string using a header, body, and secret.
// Return error if an error exists during signing.
func NewToken(header *Header, body *Body, secret *pbauth.Secret) (string, error) {
if err := ValidateHeader(header); err != nil {
return "", err
}
if err := ValidateBody(body); err != nil |
if err := ValidateSecret(secret); err != nil {
return "", err
}
if body.Permission == Admin && header.Alg != Hs512 {
return "", consts.ErrInvalidPermission
}
// Currently supports JWT, JET
if header.TokenTyp != Jwt && header.TokenTyp != Jet {
return "", consts.ErrUnknownTokenType
}
tokenString, err := getTokenSignature(header, body, secret)
if err != nil {
return "", err
}
return tokenString, nil
}
// getTokenSignature gets the token signature using the encoded header, body, and secret key.
// Return error if an error exists during signing.
func getTokenSignature(header *Header, body *Body, secret *pbauth.Secret) (string, error) {
if err := ValidateHeader(header); err != nil {
return "", err
}
if err := ValidateBody(body); err != nil {
return "", err
}
if err := ValidateSecret(secret); err != nil {
return "", err
}
if body.Permission == Admin && header.Alg != Hs512 {
return "", consts.ErrInvalidPermission
}
if header.TokenTyp != Jwt && header.TokenTyp != Jet {
return "", consts.ErrUnknownTokenType
}
// Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)>
// 1. Encode the header
encodedHeader, err := base64Encode(header)
if err != nil {
return "", err
}
// 2. Encode the body
encodedBody, err := base64Encode(body)
if err != nil {
return "", err
}
// 3. Build <encoded header>.<encoded body>
// 4. Build <hashed(<encoded header>.<encoded body>)>
// 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)>
return buildTokenSignature(encodedHeader, encodedBody, header.Alg, secret)
}
// buildTokenSignature builds the token signature using the encoded header, body, selected algorithm, and secret key.
// Return error if an error exists during signing.
func buildTokenSignature(encodedHeader string, encodedBody string, alg Algorithm, secret *pbauth.Secret) (string, error) {
if strings.TrimSpace(encodedHeader) == "" {
return "", consts.ErrInvalidEncodedHeader
}
if strings.TrimSpace(encodedBody) == "" {
return "", consts.ErrInvalidEncodedBody
}
if err := ValidateSecret(secret); err != nil {
return "", err
}
// 3. Build <encoded header>.<encoded body>
var bufferHeaderBody bytes.Buffer
bufferHeaderBody.WriteString(encodedHeader)
bufferHeaderBody.WriteString(".")
bufferHeaderBody.WriteString(encodedBody)
encodedHeaderBody := bufferHeaderBody.String()
// 4. Build <hashed(<encoded header>.<encoded body>)>
encodedSignature, err := hashSignature(alg, encodedHeaderBody, secret)
if err != nil {
return "", nil
}
// 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)>
var bufferTokenSignature bytes.Buffer
bufferTokenSignature.WriteString(encodedHeaderBody)
bufferTokenSignature.WriteString(".")
bufferTokenSignature.WriteString(encodedSignature)
return bufferTokenSignature.String(), nil
}
// base64Encode takes in a interface and encodes it as a string.
// Returns a base 64 encoded string or error during marshalling.
func base64Encode(src interface{}) (string, error) {
if src == nil {
return "", consts.ErrNilInterface
}
srcMarshal, err := json.Marshal(src)
if err != nil {
return "", err
}
srcString := string(srcMarshal)
// TODO maybe use Trim
return strings.TrimRight(base64.URLEncoding.EncodeToString([]byte(srcString)), "="), nil
}
// base64Encode takes in a base 64 encoded string.
// Returns the actual string or an error of it fails to decode the string.
func base64Decode(src string) (string, error) {
if strings.TrimSpace(src) == "" {
return "", consts.ErrEmptyString
}
if l := len(src) % 4; l > 0 {
src += strings.Repeat("=", 4-l)
}
decoded, err := base64.URLEncoding.DecodeString(src)
if err != nil {
errMsg := fmt.Errorf("decoding error %s", err)
return "", errMsg
}
return string(decoded), nil
}
// hashSignature generates a HMAC hash of a string using a secret
func hashSignature(alg Algorithm, signatureValue string, secret *pbauth.Secret) (string, error) {
if strings.TrimSpace(signatureValue) == "" {
return "", consts.ErrInvalidSignatureValue
}
if err := ValidateSecret(secret); err != nil {
return "", err
}
key := []byte(secret.Key)
var h hash.Hash
switch alg {
case Hs256:
h = hmac.New(sha256.New, key)
case Hs512:
h = hmac.New(sha512.New, key)
default:
return "", consts.ErrNoHashAlgorithm
}
h.Write([]byte(signatureValue))
return base64.URLEncoding.EncodeToString(h.Sum(nil)), nil
}
// isEquivalentHash validates a hash against a value
func isEquivalentHash(alg Algorithm, signatureValue string, secret *pbauth.Secret, hashedValue string) bool {
if err := ValidateSecret(secret); err != nil {
return false
}
/*
hashSignature cannot be reversed all you can do is hash the same character and compare it with a hashed value.
If it evaluates to true, then the character is a what is in the hash.
The isValidHash function only hashes the value with the secret and compared it with the hash.
*/
actualHashedValue, err := hashSignature(alg, signatureValue, secret)
if err != nil {
return false
}
return hashedValue == actualHashedValue
}
// ExtractUUID takes in a token string and extracts the UUID from the body.
// Returns the uuid or an empty string due to an error.
func ExtractUUID(tokenString string) string {
tokenSignature := strings.Split(tokenString, ".")
if len(tokenSignature) != 3 {
return ""
}
decodedBody, err := base64Decode(tokenSignature[1])
if err != nil {
return ""
}
body := &Body{}
if err := json.Unmarshal([]byte(decodedBody), body); err != nil {
return ""
}
if body == nil {
return ""
}
if err := validation.ValidateUserUUID(body.UUID); err != nil {
return ""
}
return body.UUID
}
// GenerateSecretKey generates a base64 URL-safe string
// built from securely generated random bytes.
// Number of bytes is determined by tokenSize.
// Return error if system's secure random number generator fails.
func GenerateSecretKey(tokenSize int) (string, error) {
if tokenSize <= 0 {
return "", consts.ErrInvalidTokenSize
}
keyGenLocker.Lock()
defer keyGenLocker.Unlock()
randomBytes := make([]byte, tokenSize)
_, err := cryptorand.Read(randomBytes)
if err != nil {
return "", err
}
return base64.URLEncoding.EncodeToString(randomBytes), nil
}
// GenerateExpirationTimestamp returns | {
return "", err
} | conditional_block |
static.go | {
return err
}
return nil
}
// ValidateHeader validates Header.
// Returns the first error encountered.
func ValidateHeader(header *Header) error {
if header == nil {
return consts.ErrNilHeader
}
tokenType := header.TokenTyp
if tokenType < NoType || tokenType > Jet {
return consts.ErrUnknownTokenType
}
alg := header.Alg
if alg < NoAlg || alg > Hs512 {
return consts.ErrUnknownAlgorithm
}
return nil
}
// ValidateBody validates Body.
// Checks if token string has expired.
// Returns the first error encountered.
func ValidateBody(body *Body) error {
if body == nil {
return consts.ErrNilBody
}
if err := validation.ValidateUserUUID(body.UUID); err != nil {
return err
}
permission := body.Permission
if permission < NoPermission || permission > Admin {
return consts.ErrUnknownPermission
}
if isExpired(body.ExpirationTimestamp) {
return consts.ErrExpiredBody
}
return nil
}
// ValidateSecret checks if the secret is still valid and has not expired.
// Returns an error if the Secret is not valid and has expired.
func ValidateSecret(secret *pbauth.Secret) error {
if secret == nil {
return consts.ErrNilSecret
}
if strings.TrimSpace(secret.Key) == "" {
return consts.ErrEmptySecret
}
createTime := secret.CreatedTimestamp
if createTime == 0 || createTime > time.Now().UTC().Unix() {
return consts.ErrInvalidSecretCreateTimestamp
}
if isExpired(secret.ExpirationTimestamp) {
return consts.ErrExpiredSecret
}
return nil
}
func isExpired(timestamp int64) bool {
if timestamp <= 0 || time.Now().UTC().Unix() >= timestamp {
return true
}
return false
}
// NewToken generates token string using a header, body, and secret.
// Return error if an error exists during signing.
func NewToken(header *Header, body *Body, secret *pbauth.Secret) (string, error) {
if err := ValidateHeader(header); err != nil {
return "", err
}
if err := ValidateBody(body); err != nil {
return "", err
}
if err := ValidateSecret(secret); err != nil {
return "", err
}
if body.Permission == Admin && header.Alg != Hs512 {
return "", consts.ErrInvalidPermission
}
// Currently supports JWT, JET
if header.TokenTyp != Jwt && header.TokenTyp != Jet {
return "", consts.ErrUnknownTokenType
}
tokenString, err := getTokenSignature(header, body, secret)
if err != nil {
return "", err
}
return tokenString, nil
}
// getTokenSignature gets the token signature using the encoded header, body, and secret key.
// Return error if an error exists during signing.
func getTokenSignature(header *Header, body *Body, secret *pbauth.Secret) (string, error) {
if err := ValidateHeader(header); err != nil {
return "", err
}
if err := ValidateBody(body); err != nil {
return "", err
}
if err := ValidateSecret(secret); err != nil {
return "", err
}
if body.Permission == Admin && header.Alg != Hs512 {
return "", consts.ErrInvalidPermission
}
if header.TokenTyp != Jwt && header.TokenTyp != Jet {
return "", consts.ErrUnknownTokenType
}
// Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)>
// 1. Encode the header
encodedHeader, err := base64Encode(header)
if err != nil {
return "", err
}
// 2. Encode the body
encodedBody, err := base64Encode(body)
if err != nil {
return "", err
}
// 3. Build <encoded header>.<encoded body>
// 4. Build <hashed(<encoded header>.<encoded body>)>
// 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)>
return buildTokenSignature(encodedHeader, encodedBody, header.Alg, secret)
}
// buildTokenSignature builds the token signature using the encoded header, body, selected algorithm, and secret key.
// Return error if an error exists during signing.
func buildTokenSignature(encodedHeader string, encodedBody string, alg Algorithm, secret *pbauth.Secret) (string, error) {
if strings.TrimSpace(encodedHeader) == "" {
return "", consts.ErrInvalidEncodedHeader
}
if strings.TrimSpace(encodedBody) == "" {
return "", consts.ErrInvalidEncodedBody
}
if err := ValidateSecret(secret); err != nil {
return "", err
}
// 3. Build <encoded header>.<encoded body>
var bufferHeaderBody bytes.Buffer
bufferHeaderBody.WriteString(encodedHeader)
bufferHeaderBody.WriteString(".")
bufferHeaderBody.WriteString(encodedBody) | encodedSignature, err := hashSignature(alg, encodedHeaderBody, secret)
if err != nil {
return "", nil
}
// 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)>
var bufferTokenSignature bytes.Buffer
bufferTokenSignature.WriteString(encodedHeaderBody)
bufferTokenSignature.WriteString(".")
bufferTokenSignature.WriteString(encodedSignature)
return bufferTokenSignature.String(), nil
}
// base64Encode takes in a interface and encodes it as a string.
// Returns a base 64 encoded string or error during marshalling.
func base64Encode(src interface{}) (string, error) {
if src == nil {
return "", consts.ErrNilInterface
}
srcMarshal, err := json.Marshal(src)
if err != nil {
return "", err
}
srcString := string(srcMarshal)
// TODO maybe use Trim
return strings.TrimRight(base64.URLEncoding.EncodeToString([]byte(srcString)), "="), nil
}
// base64Encode takes in a base 64 encoded string.
// Returns the actual string or an error of it fails to decode the string.
func base64Decode(src string) (string, error) {
if strings.TrimSpace(src) == "" {
return "", consts.ErrEmptyString
}
if l := len(src) % 4; l > 0 {
src += strings.Repeat("=", 4-l)
}
decoded, err := base64.URLEncoding.DecodeString(src)
if err != nil {
errMsg := fmt.Errorf("decoding error %s", err)
return "", errMsg
}
return string(decoded), nil
}
// hashSignature generates a HMAC hash of a string using a secret
func hashSignature(alg Algorithm, signatureValue string, secret *pbauth.Secret) (string, error) {
if strings.TrimSpace(signatureValue) == "" {
return "", consts.ErrInvalidSignatureValue
}
if err := ValidateSecret(secret); err != nil {
return "", err
}
key := []byte(secret.Key)
var h hash.Hash
switch alg {
case Hs256:
h = hmac.New(sha256.New, key)
case Hs512:
h = hmac.New(sha512.New, key)
default:
return "", consts.ErrNoHashAlgorithm
}
h.Write([]byte(signatureValue))
return base64.URLEncoding.EncodeToString(h.Sum(nil)), nil
}
// isEquivalentHash validates a hash against a value
func isEquivalentHash(alg Algorithm, signatureValue string, secret *pbauth.Secret, hashedValue string) bool {
if err := ValidateSecret(secret); err != nil {
return false
}
/*
hashSignature cannot be reversed all you can do is hash the same character and compare it with a hashed value.
If it evaluates to true, then the character is a what is in the hash.
The isValidHash function only hashes the value with the secret and compared it with the hash.
*/
actualHashedValue, err := hashSignature(alg, signatureValue, secret)
if err != nil {
return false
}
return hashedValue == actualHashedValue
}
// ExtractUUID takes in a token string and extracts the UUID from the body.
// Returns the uuid or an empty string due to an error.
func ExtractUUID(tokenString string) string {
tokenSignature := strings.Split(tokenString, ".")
if len(tokenSignature) != 3 {
return ""
}
decodedBody, err := base64Decode(tokenSignature[1])
if err != nil {
return ""
}
body := &Body{}
if err := json.Unmarshal([]byte(decodedBody), body); err != nil {
return ""
}
if body == nil {
return ""
}
if err := validation.ValidateUserUUID(body.UUID); err != nil {
return ""
}
return body.UUID
}
// GenerateSecretKey generates a base64 URL-safe string
// built from securely generated random bytes.
// Number of bytes is determined by tokenSize.
// Return error if system's secure random number generator fails.
func GenerateSecretKey(tokenSize int) (string, error) {
if tokenSize <= 0 {
return "", consts.ErrInvalidTokenSize
}
keyGenLocker.Lock()
defer keyGenLocker.Unlock()
randomBytes := make([]byte, tokenSize)
_, err := cryptorand.Read(randomBytes)
if err != nil {
return "", err
}
return base64.URLEncoding.EncodeToString(randomBytes), nil
}
// GenerateExpirationTimestamp returns the expiration | encodedHeaderBody := bufferHeaderBody.String()
// 4. Build <hashed(<encoded header>.<encoded body>)> | random_line_split |
static.go | {
return err
}
return nil
}
// ValidateHeader validates Header.
// Returns the first error encountered.
func ValidateHeader(header *Header) error {
if header == nil {
return consts.ErrNilHeader
}
tokenType := header.TokenTyp
if tokenType < NoType || tokenType > Jet {
return consts.ErrUnknownTokenType
}
alg := header.Alg
if alg < NoAlg || alg > Hs512 {
return consts.ErrUnknownAlgorithm
}
return nil
}
// ValidateBody validates Body.
// Checks if token string has expired.
// Returns the first error encountered.
func ValidateBody(body *Body) error {
if body == nil {
return consts.ErrNilBody
}
if err := validation.ValidateUserUUID(body.UUID); err != nil {
return err
}
permission := body.Permission
if permission < NoPermission || permission > Admin {
return consts.ErrUnknownPermission
}
if isExpired(body.ExpirationTimestamp) {
return consts.ErrExpiredBody
}
return nil
}
// ValidateSecret checks if the secret is still valid and has not expired.
// Returns an error if the Secret is not valid and has expired.
func ValidateSecret(secret *pbauth.Secret) error |
func isExpired(timestamp int64) bool {
if timestamp <= 0 || time.Now().UTC().Unix() >= timestamp {
return true
}
return false
}
// NewToken generates token string using a header, body, and secret.
// Return error if an error exists during signing.
func NewToken(header *Header, body *Body, secret *pbauth.Secret) (string, error) {
if err := ValidateHeader(header); err != nil {
return "", err
}
if err := ValidateBody(body); err != nil {
return "", err
}
if err := ValidateSecret(secret); err != nil {
return "", err
}
if body.Permission == Admin && header.Alg != Hs512 {
return "", consts.ErrInvalidPermission
}
// Currently supports JWT, JET
if header.TokenTyp != Jwt && header.TokenTyp != Jet {
return "", consts.ErrUnknownTokenType
}
tokenString, err := getTokenSignature(header, body, secret)
if err != nil {
return "", err
}
return tokenString, nil
}
// getTokenSignature gets the token signature using the encoded header, body, and secret key.
// Return error if an error exists during signing.
func getTokenSignature(header *Header, body *Body, secret *pbauth.Secret) (string, error) {
if err := ValidateHeader(header); err != nil {
return "", err
}
if err := ValidateBody(body); err != nil {
return "", err
}
if err := ValidateSecret(secret); err != nil {
return "", err
}
if body.Permission == Admin && header.Alg != Hs512 {
return "", consts.ErrInvalidPermission
}
if header.TokenTyp != Jwt && header.TokenTyp != Jet {
return "", consts.ErrUnknownTokenType
}
// Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)>
// 1. Encode the header
encodedHeader, err := base64Encode(header)
if err != nil {
return "", err
}
// 2. Encode the body
encodedBody, err := base64Encode(body)
if err != nil {
return "", err
}
// 3. Build <encoded header>.<encoded body>
// 4. Build <hashed(<encoded header>.<encoded body>)>
// 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)>
return buildTokenSignature(encodedHeader, encodedBody, header.Alg, secret)
}
// buildTokenSignature builds the token signature using the encoded header, body, selected algorithm, and secret key.
// Return error if an error exists during signing.
func buildTokenSignature(encodedHeader string, encodedBody string, alg Algorithm, secret *pbauth.Secret) (string, error) {
if strings.TrimSpace(encodedHeader) == "" {
return "", consts.ErrInvalidEncodedHeader
}
if strings.TrimSpace(encodedBody) == "" {
return "", consts.ErrInvalidEncodedBody
}
if err := ValidateSecret(secret); err != nil {
return "", err
}
// 3. Build <encoded header>.<encoded body>
var bufferHeaderBody bytes.Buffer
bufferHeaderBody.WriteString(encodedHeader)
bufferHeaderBody.WriteString(".")
bufferHeaderBody.WriteString(encodedBody)
encodedHeaderBody := bufferHeaderBody.String()
// 4. Build <hashed(<encoded header>.<encoded body>)>
encodedSignature, err := hashSignature(alg, encodedHeaderBody, secret)
if err != nil {
return "", nil
}
// 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)>
var bufferTokenSignature bytes.Buffer
bufferTokenSignature.WriteString(encodedHeaderBody)
bufferTokenSignature.WriteString(".")
bufferTokenSignature.WriteString(encodedSignature)
return bufferTokenSignature.String(), nil
}
// base64Encode takes in a interface and encodes it as a string.
// Returns a base 64 encoded string or error during marshalling.
func base64Encode(src interface{}) (string, error) {
if src == nil {
return "", consts.ErrNilInterface
}
srcMarshal, err := json.Marshal(src)
if err != nil {
return "", err
}
srcString := string(srcMarshal)
// TODO maybe use Trim
return strings.TrimRight(base64.URLEncoding.EncodeToString([]byte(srcString)), "="), nil
}
// base64Encode takes in a base 64 encoded string.
// Returns the actual string or an error of it fails to decode the string.
func base64Decode(src string) (string, error) {
if strings.TrimSpace(src) == "" {
return "", consts.ErrEmptyString
}
if l := len(src) % 4; l > 0 {
src += strings.Repeat("=", 4-l)
}
decoded, err := base64.URLEncoding.DecodeString(src)
if err != nil {
errMsg := fmt.Errorf("decoding error %s", err)
return "", errMsg
}
return string(decoded), nil
}
// hashSignature generates a HMAC hash of a string using a secret
func hashSignature(alg Algorithm, signatureValue string, secret *pbauth.Secret) (string, error) {
if strings.TrimSpace(signatureValue) == "" {
return "", consts.ErrInvalidSignatureValue
}
if err := ValidateSecret(secret); err != nil {
return "", err
}
key := []byte(secret.Key)
var h hash.Hash
switch alg {
case Hs256:
h = hmac.New(sha256.New, key)
case Hs512:
h = hmac.New(sha512.New, key)
default:
return "", consts.ErrNoHashAlgorithm
}
h.Write([]byte(signatureValue))
return base64.URLEncoding.EncodeToString(h.Sum(nil)), nil
}
// isEquivalentHash validates a hash against a value
func isEquivalentHash(alg Algorithm, signatureValue string, secret *pbauth.Secret, hashedValue string) bool {
if err := ValidateSecret(secret); err != nil {
return false
}
/*
hashSignature cannot be reversed all you can do is hash the same character and compare it with a hashed value.
If it evaluates to true, then the character is a what is in the hash.
The isValidHash function only hashes the value with the secret and compared it with the hash.
*/
actualHashedValue, err := hashSignature(alg, signatureValue, secret)
if err != nil {
return false
}
return hashedValue == actualHashedValue
}
// ExtractUUID takes in a token string and extracts the UUID from the body.
// Returns the uuid or an empty string due to an error.
func ExtractUUID(tokenString string) string {
tokenSignature := strings.Split(tokenString, ".")
if len(tokenSignature) != 3 {
return ""
}
decodedBody, err := base64Decode(tokenSignature[1])
if err != nil {
return ""
}
body := &Body{}
if err := json.Unmarshal([]byte(decodedBody), body); err != nil {
return ""
}
if body == nil {
return ""
}
if err := validation.ValidateUserUUID(body.UUID); err != nil {
return ""
}
return body.UUID
}
// GenerateSecretKey generates a base64 URL-safe string
// built from securely generated random bytes.
// Number of bytes is determined by tokenSize.
// Return error if system's secure random number generator fails.
func GenerateSecretKey(tokenSize int) (string, error) {
if tokenSize <= 0 {
return "", consts.ErrInvalidTokenSize
}
keyGenLocker.Lock()
defer keyGenLocker.Unlock()
randomBytes := make([]byte, tokenSize)
_, err := cryptorand.Read(randomBytes)
if err != nil {
return "", err
}
return base64.URLEncoding.EncodeToString(randomBytes), nil
}
// GenerateExpirationTimestamp returns | {
if secret == nil {
return consts.ErrNilSecret
}
if strings.TrimSpace(secret.Key) == "" {
return consts.ErrEmptySecret
}
createTime := secret.CreatedTimestamp
if createTime == 0 || createTime > time.Now().UTC().Unix() {
return consts.ErrInvalidSecretCreateTimestamp
}
if isExpired(secret.ExpirationTimestamp) {
return consts.ErrExpiredSecret
}
return nil
} | identifier_body |
static.go | {
return err
}
return nil
}
// ValidateHeader validates Header.
// Returns the first error encountered.
func ValidateHeader(header *Header) error {
if header == nil {
return consts.ErrNilHeader
}
tokenType := header.TokenTyp
if tokenType < NoType || tokenType > Jet {
return consts.ErrUnknownTokenType
}
alg := header.Alg
if alg < NoAlg || alg > Hs512 {
return consts.ErrUnknownAlgorithm
}
return nil
}
// ValidateBody validates Body.
// Checks if token string has expired.
// Returns the first error encountered.
func ValidateBody(body *Body) error {
if body == nil {
return consts.ErrNilBody
}
if err := validation.ValidateUserUUID(body.UUID); err != nil {
return err
}
permission := body.Permission
if permission < NoPermission || permission > Admin {
return consts.ErrUnknownPermission
}
if isExpired(body.ExpirationTimestamp) {
return consts.ErrExpiredBody
}
return nil
}
// ValidateSecret checks if the secret is still valid and has not expired.
// Returns an error if the Secret is not valid and has expired.
func ValidateSecret(secret *pbauth.Secret) error {
if secret == nil {
return consts.ErrNilSecret
}
if strings.TrimSpace(secret.Key) == "" {
return consts.ErrEmptySecret
}
createTime := secret.CreatedTimestamp
if createTime == 0 || createTime > time.Now().UTC().Unix() {
return consts.ErrInvalidSecretCreateTimestamp
}
if isExpired(secret.ExpirationTimestamp) {
return consts.ErrExpiredSecret
}
return nil
}
func isExpired(timestamp int64) bool {
if timestamp <= 0 || time.Now().UTC().Unix() >= timestamp {
return true
}
return false
}
// NewToken generates token string using a header, body, and secret.
// Return error if an error exists during signing.
func NewToken(header *Header, body *Body, secret *pbauth.Secret) (string, error) {
if err := ValidateHeader(header); err != nil {
return "", err
}
if err := ValidateBody(body); err != nil {
return "", err
}
if err := ValidateSecret(secret); err != nil {
return "", err
}
if body.Permission == Admin && header.Alg != Hs512 {
return "", consts.ErrInvalidPermission
}
// Currently supports JWT, JET
if header.TokenTyp != Jwt && header.TokenTyp != Jet {
return "", consts.ErrUnknownTokenType
}
tokenString, err := getTokenSignature(header, body, secret)
if err != nil {
return "", err
}
return tokenString, nil
}
// getTokenSignature gets the token signature using the encoded header, body, and secret key.
// Return error if an error exists during signing.
func | (header *Header, body *Body, secret *pbauth.Secret) (string, error) {
if err := ValidateHeader(header); err != nil {
return "", err
}
if err := ValidateBody(body); err != nil {
return "", err
}
if err := ValidateSecret(secret); err != nil {
return "", err
}
if body.Permission == Admin && header.Alg != Hs512 {
return "", consts.ErrInvalidPermission
}
if header.TokenTyp != Jwt && header.TokenTyp != Jet {
return "", consts.ErrUnknownTokenType
}
// Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)>
// 1. Encode the header
encodedHeader, err := base64Encode(header)
if err != nil {
return "", err
}
// 2. Encode the body
encodedBody, err := base64Encode(body)
if err != nil {
return "", err
}
// 3. Build <encoded header>.<encoded body>
// 4. Build <hashed(<encoded header>.<encoded body>)>
// 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)>
return buildTokenSignature(encodedHeader, encodedBody, header.Alg, secret)
}
// buildTokenSignature builds the token signature using the encoded header, body, selected algorithm, and secret key.
// Return error if an error exists during signing.
func buildTokenSignature(encodedHeader string, encodedBody string, alg Algorithm, secret *pbauth.Secret) (string, error) {
if strings.TrimSpace(encodedHeader) == "" {
return "", consts.ErrInvalidEncodedHeader
}
if strings.TrimSpace(encodedBody) == "" {
return "", consts.ErrInvalidEncodedBody
}
if err := ValidateSecret(secret); err != nil {
return "", err
}
// 3. Build <encoded header>.<encoded body>
var bufferHeaderBody bytes.Buffer
bufferHeaderBody.WriteString(encodedHeader)
bufferHeaderBody.WriteString(".")
bufferHeaderBody.WriteString(encodedBody)
encodedHeaderBody := bufferHeaderBody.String()
// 4. Build <hashed(<encoded header>.<encoded body>)>
encodedSignature, err := hashSignature(alg, encodedHeaderBody, secret)
if err != nil {
return "", nil
}
// 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)>
var bufferTokenSignature bytes.Buffer
bufferTokenSignature.WriteString(encodedHeaderBody)
bufferTokenSignature.WriteString(".")
bufferTokenSignature.WriteString(encodedSignature)
return bufferTokenSignature.String(), nil
}
// base64Encode takes in a interface and encodes it as a string.
// Returns a base 64 encoded string or error during marshalling.
func base64Encode(src interface{}) (string, error) {
if src == nil {
return "", consts.ErrNilInterface
}
srcMarshal, err := json.Marshal(src)
if err != nil {
return "", err
}
srcString := string(srcMarshal)
// TODO maybe use Trim
return strings.TrimRight(base64.URLEncoding.EncodeToString([]byte(srcString)), "="), nil
}
// base64Encode takes in a base 64 encoded string.
// Returns the actual string or an error of it fails to decode the string.
func base64Decode(src string) (string, error) {
if strings.TrimSpace(src) == "" {
return "", consts.ErrEmptyString
}
if l := len(src) % 4; l > 0 {
src += strings.Repeat("=", 4-l)
}
decoded, err := base64.URLEncoding.DecodeString(src)
if err != nil {
errMsg := fmt.Errorf("decoding error %s", err)
return "", errMsg
}
return string(decoded), nil
}
// hashSignature generates a HMAC hash of a string using a secret
func hashSignature(alg Algorithm, signatureValue string, secret *pbauth.Secret) (string, error) {
if strings.TrimSpace(signatureValue) == "" {
return "", consts.ErrInvalidSignatureValue
}
if err := ValidateSecret(secret); err != nil {
return "", err
}
key := []byte(secret.Key)
var h hash.Hash
switch alg {
case Hs256:
h = hmac.New(sha256.New, key)
case Hs512:
h = hmac.New(sha512.New, key)
default:
return "", consts.ErrNoHashAlgorithm
}
h.Write([]byte(signatureValue))
return base64.URLEncoding.EncodeToString(h.Sum(nil)), nil
}
// isEquivalentHash validates a hash against a value
func isEquivalentHash(alg Algorithm, signatureValue string, secret *pbauth.Secret, hashedValue string) bool {
if err := ValidateSecret(secret); err != nil {
return false
}
/*
hashSignature cannot be reversed all you can do is hash the same character and compare it with a hashed value.
If it evaluates to true, then the character is a what is in the hash.
The isValidHash function only hashes the value with the secret and compared it with the hash.
*/
actualHashedValue, err := hashSignature(alg, signatureValue, secret)
if err != nil {
return false
}
return hashedValue == actualHashedValue
}
// ExtractUUID takes in a token string and extracts the UUID from the body.
// Returns the uuid or an empty string due to an error.
func ExtractUUID(tokenString string) string {
tokenSignature := strings.Split(tokenString, ".")
if len(tokenSignature) != 3 {
return ""
}
decodedBody, err := base64Decode(tokenSignature[1])
if err != nil {
return ""
}
body := &Body{}
if err := json.Unmarshal([]byte(decodedBody), body); err != nil {
return ""
}
if body == nil {
return ""
}
if err := validation.ValidateUserUUID(body.UUID); err != nil {
return ""
}
return body.UUID
}
// GenerateSecretKey generates a base64 URL-safe string
// built from securely generated random bytes.
// Number of bytes is determined by tokenSize.
// Return error if system's secure random number generator fails.
func GenerateSecretKey(tokenSize int) (string, error) {
if tokenSize <= 0 {
return "", consts.ErrInvalidTokenSize
}
keyGenLocker.Lock()
defer keyGenLocker.Unlock()
randomBytes := make([]byte, tokenSize)
_, err := cryptorand.Read(randomBytes)
if err != nil {
return "", err
}
return base64.URLEncoding.EncodeToString(randomBytes), nil
}
// GenerateExpirationTimestamp returns the | getTokenSignature | identifier_name |
reader.rs | Context {
fn validate(&self) -> Result<()> {
if self.max_concurrency == 0 {
return Err(Error::msg("max concurrency must be at least 1."));
}
Ok(())
}
}
impl ReaderContext {
fn default_reader_threads() -> usize {
1
}
}
/// A given query payload that describes how the reader should
/// search the index.
#[derive(Debug, Deserialize)]
pub struct QueryPayload {
/// The query(s) itself.
query: QuerySelector,
/// The amount of results to limit by.
#[serde(default = "QueryPayload::default_limit")]
limit: usize,
/// The amount of documents to skip before getting the results.
#[serde(default)]
offset: usize,
/// A specified field to order results by, this defaults to the
/// score of the indexed documents (relevancy).
order_by: Option<String>,
/// How to sort the data (asc/desc).
#[serde(default)]
sort: Sort,
}
impl QueryPayload {
fn default_limit() -> usize {
20
}
}
/// What order to sort the returned data.
#[derive(Debug, Copy, Clone, Deserialize)]
#[serde(rename_all = "lowercase")]
pub enum Sort {
/// Sort the data in ascending order.
Asc,
/// Sort the data in descending order. (Default)
Desc,
}
|
#[derive(Debug, Serialize)]
pub struct QueryResults {
/// The retrieved documents.
pub(crate) hits: Vec<DocumentHit>,
/// The total amount of documents matching the search
count: usize,
/// The amount of time taken to search in seconds.
time_taken: f32,
}
impl QueryResults {
#[inline]
pub fn len(&self) -> usize {
self.hits.len()
}
}
/// Attaches an order by clause to the collector.
///
/// This collected the values with be returned in the order according to the
/// given field value.
fn order_and_search<R: AsScore + tantivy::fastfield::FastValue>(
searcher: &Searcher,
field: Field,
query: &dyn Query,
collector: TopDocs,
executor: &Executor,
) -> Result<(Vec<(R, DocAddress)>, usize)> {
let collector = collector.order_by_fast_field(field);
searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)
}
/// Performs the search operation and processes the returned results.
fn process_search<S: AsScore>(
searcher: &Searcher,
schema: &Schema,
top_docs: Vec<(S, DocAddress)>,
) -> Result<Vec<DocumentHit>> {
let mut hits = Vec::with_capacity(top_docs.len());
for (ratio, ref_address) in top_docs {
let retrieved_doc = searcher.doc(ref_address)?;
let mut doc = schema.to_named_doc(&retrieved_doc);
let id = doc.0
.remove("_id")
.ok_or_else(|| Error::msg("document has been missed labeled (missing primary key '_id'), the dataset is invalid"))?;
if let Value::U64(v) = id[0] {
hits.push(DocumentHit {
doc,
document_id: v,
score: ratio.as_score(),
});
} else {
return Err(Error::msg("document has been missed labeled (missing identifier tag), the dataset is invalid"));
}
}
Ok(hits)
}
/// Orders the search results by the given field with a given sort (ASC, DESC)
///
/// This function is super messy just because of all the type inference
/// so any contributions to clean this up would be very appreciated.
fn order_or_sort(
sort: Sort,
field: Field,
query: &dyn Query,
schema: &Schema,
searcher: &Searcher,
collector: TopDocs,
executor: &Executor,
) -> Result<(Vec<DocumentHit>, usize)> {
let field_type = schema.get_field_entry(field).field_type();
if let Sort::Desc = sort {
return match field_type {
FieldType::I64(_) => {
let out: (Vec<(i64, DocAddress)>, usize) =
order_and_search(&searcher, field, query, collector, executor)?;
Ok((process_search(&searcher, schema, out.0)?, out.1))
},
FieldType::U64(_) => {
let out: (Vec<(u64, DocAddress)>, usize) =
order_and_search(&searcher, field, query, collector, executor)?;
Ok((process_search(&searcher, schema, out.0)?, out.1))
},
FieldType::F64(_) => {
let out: (Vec<(f64, DocAddress)>, usize) =
order_and_search(&searcher, field, query, collector, executor)?;
Ok((process_search(&searcher, schema, out.0)?, out.1))
},
FieldType::Date(_) => {
let out: (Vec<(DateTime, DocAddress)>, usize) =
order_and_search(&searcher, field, query, collector, executor)?;
Ok((process_search(&searcher, schema, out.0)?, out.1))
},
_ => Err(Error::msg("field is not a fast field")),
};
}
let out = match field_type {
FieldType::I64(_) => {
let collector =
collector.custom_score(move |segment_reader: &SegmentReader| {
let reader = segment_reader
.fast_fields()
.i64(field)
.expect("field exists");
move |doc: DocId| {
let value: i64 = reader.get(doc);
std::cmp::Reverse(value)
}
});
let out: (Vec<(Reverse<i64>, DocAddress)>, usize) = searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)?;
(process_search(&searcher, schema, out.0)?, out.1)
},
FieldType::U64(_) => {
let collector =
collector.custom_score(move |segment_reader: &SegmentReader| {
let reader = segment_reader
.fast_fields()
.u64(field)
.expect("field exists");
move |doc: DocId| {
let value: u64 = reader.get(doc);
std::cmp::Reverse(value)
}
});
let out: (Vec<(Reverse<u64>, DocAddress)>, usize) = searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)?;
(process_search(&searcher, schema, out.0)?, out.1)
},
FieldType::F64(_) => {
let collector =
collector.custom_score(move |segment_reader: &SegmentReader| {
let reader = segment_reader
.fast_fields()
.f64(field)
.expect("field exists");
move |doc: DocId| {
let value: f64 = reader.get(doc);
std::cmp::Reverse(value)
}
});
let out: (Vec<(Reverse<f64>, DocAddress)>, usize) = searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)?;
(process_search(&searcher, schema, out.0)?, out.1)
},
FieldType::Date(_) => {
let collector =
collector.custom_score(move |segment_reader: &SegmentReader| {
let reader = segment_reader
.fast_fields()
.date(field)
.expect("field exists");
move |doc: DocId| {
let value: DateTime = reader.get(doc);
std::cmp::Reverse(value)
}
});
let out: (Vec<(Reverse<DateTime>, DocAddress)>, usize) = searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)?;
(process_search(&searcher, schema, out.0)?, out.1)
},
_ => return Err(Error::msg("field is not a fast field")),
};
return Ok(out);
}
/// The reader of the given index.
///
/// This manages all searches on the index which encompasses the concurrency
/// limiters and thread pool execution.
///
/// Each index should only have on `Reader` instance.
pub(crate) struct Reader {
/// The executor pool.
pool: crate::ReaderExecutor,
/// The query factory system.
query_handler: QueryBuilder,
}
impl Reader {
/// Creates a new reader from the given index context.
pub(crate) async fn create(ctx: &IndexContext) -> Result<Self> {
let reader: IndexReader = ctx
.index
.reader_builder()
.reload_policy(ReloadPolicy::OnCommit)
.num_searchers(ctx.reader_ctx.max_concurrency)
.try_into()?;
info!(
"[ READER @ {} ] index reader created with reload policy=OnCommit, num_searchers={}",
&ctx.name,
ctx.reader_ctx.max_concurrency,
);
let pool = {
let pool = SearcherExecutorPool::create(
reader,
ctx.reader_ctx.reader_threads,
ctx.reader_ctx.max_concurrency,
)
.await?;
Arc::new(pool)
};
info!(
"[ READER @ {} ] | impl Default for Sort {
fn default() -> Self {
Self::Desc
}
} | random_line_split |
reader.rs | Context {
fn validate(&self) -> Result<()> {
if self.max_concurrency == 0 {
return Err(Error::msg("max concurrency must be at least 1."));
}
Ok(())
}
}
impl ReaderContext {
fn default_reader_threads() -> usize {
1
}
}
/// A given query payload that describes how the reader should
/// search the index.
#[derive(Debug, Deserialize)]
pub struct QueryPayload {
/// The query(s) itself.
query: QuerySelector,
/// The amount of results to limit by.
#[serde(default = "QueryPayload::default_limit")]
limit: usize,
/// The amount of documents to skip before getting the results.
#[serde(default)]
offset: usize,
/// A specified field to order results by, this defaults to the
/// score of the indexed documents (relevancy).
order_by: Option<String>,
/// How to sort the data (asc/desc).
#[serde(default)]
sort: Sort,
}
impl QueryPayload {
fn default_limit() -> usize {
20
}
}
/// What order to sort the returned data.
#[derive(Debug, Copy, Clone, Deserialize)]
#[serde(rename_all = "lowercase")]
pub enum Sort {
/// Sort the data in ascending order.
Asc,
/// Sort the data in descending order. (Default)
Desc,
}
impl Default for Sort {
fn default() -> Self {
Self::Desc
}
}
#[derive(Debug, Serialize)]
pub struct QueryResults {
/// The retrieved documents.
pub(crate) hits: Vec<DocumentHit>,
/// The total amount of documents matching the search
count: usize,
/// The amount of time taken to search in seconds.
time_taken: f32,
}
impl QueryResults {
#[inline]
pub fn len(&self) -> usize {
self.hits.len()
}
}
/// Attaches an order by clause to the collector.
///
/// This collected the values with be returned in the order according to the
/// given field value.
fn order_and_search<R: AsScore + tantivy::fastfield::FastValue>(
searcher: &Searcher,
field: Field,
query: &dyn Query,
collector: TopDocs,
executor: &Executor,
) -> Result<(Vec<(R, DocAddress)>, usize)> {
let collector = collector.order_by_fast_field(field);
searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)
}
/// Performs the search operation and processes the returned results.
fn process_search<S: AsScore>(
searcher: &Searcher,
schema: &Schema,
top_docs: Vec<(S, DocAddress)>,
) -> Result<Vec<DocumentHit>> {
let mut hits = Vec::with_capacity(top_docs.len());
for (ratio, ref_address) in top_docs {
let retrieved_doc = searcher.doc(ref_address)?;
let mut doc = schema.to_named_doc(&retrieved_doc);
let id = doc.0
.remove("_id")
.ok_or_else(|| Error::msg("document has been missed labeled (missing primary key '_id'), the dataset is invalid"))?;
if let Value::U64(v) = id[0] {
hits.push(DocumentHit {
doc,
document_id: v,
score: ratio.as_score(),
});
} else {
return Err(Error::msg("document has been missed labeled (missing identifier tag), the dataset is invalid"));
}
}
Ok(hits)
}
/// Orders the search results by the given field with a given sort (ASC, DESC)
///
/// This function is super messy just because of all the type inference
/// so any contributions to clean this up would be very appreciated.
fn order_or_sort(
sort: Sort,
field: Field,
query: &dyn Query,
schema: &Schema,
searcher: &Searcher,
collector: TopDocs,
executor: &Executor,
) -> Result<(Vec<DocumentHit>, usize)> {
let field_type = schema.get_field_entry(field).field_type();
if let Sort::Desc = sort {
return match field_type {
FieldType::I64(_) => {
let out: (Vec<(i64, DocAddress)>, usize) =
order_and_search(&searcher, field, query, collector, executor)?;
Ok((process_search(&searcher, schema, out.0)?, out.1))
},
FieldType::U64(_) => {
let out: (Vec<(u64, DocAddress)>, usize) =
order_and_search(&searcher, field, query, collector, executor)?;
Ok((process_search(&searcher, schema, out.0)?, out.1))
},
FieldType::F64(_) => {
let out: (Vec<(f64, DocAddress)>, usize) =
order_and_search(&searcher, field, query, collector, executor)?;
Ok((process_search(&searcher, schema, out.0)?, out.1))
},
FieldType::Date(_) => | ,
_ => Err(Error::msg("field is not a fast field")),
};
}
let out = match field_type {
FieldType::I64(_) => {
let collector =
collector.custom_score(move |segment_reader: &SegmentReader| {
let reader = segment_reader
.fast_fields()
.i64(field)
.expect("field exists");
move |doc: DocId| {
let value: i64 = reader.get(doc);
std::cmp::Reverse(value)
}
});
let out: (Vec<(Reverse<i64>, DocAddress)>, usize) = searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)?;
(process_search(&searcher, schema, out.0)?, out.1)
},
FieldType::U64(_) => {
let collector =
collector.custom_score(move |segment_reader: &SegmentReader| {
let reader = segment_reader
.fast_fields()
.u64(field)
.expect("field exists");
move |doc: DocId| {
let value: u64 = reader.get(doc);
std::cmp::Reverse(value)
}
});
let out: (Vec<(Reverse<u64>, DocAddress)>, usize) = searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)?;
(process_search(&searcher, schema, out.0)?, out.1)
},
FieldType::F64(_) => {
let collector =
collector.custom_score(move |segment_reader: &SegmentReader| {
let reader = segment_reader
.fast_fields()
.f64(field)
.expect("field exists");
move |doc: DocId| {
let value: f64 = reader.get(doc);
std::cmp::Reverse(value)
}
});
let out: (Vec<(Reverse<f64>, DocAddress)>, usize) = searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)?;
(process_search(&searcher, schema, out.0)?, out.1)
},
FieldType::Date(_) => {
let collector =
collector.custom_score(move |segment_reader: &SegmentReader| {
let reader = segment_reader
.fast_fields()
.date(field)
.expect("field exists");
move |doc: DocId| {
let value: DateTime = reader.get(doc);
std::cmp::Reverse(value)
}
});
let out: (Vec<(Reverse<DateTime>, DocAddress)>, usize) = searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)?;
(process_search(&searcher, schema, out.0)?, out.1)
},
_ => return Err(Error::msg("field is not a fast field")),
};
return Ok(out);
}
/// The reader of the given index.
///
/// This manages all searches on the index which encompasses the concurrency
/// limiters and thread pool execution.
///
/// Each index should only have on `Reader` instance.
pub(crate) struct Reader {
/// The executor pool.
pool: crate::ReaderExecutor,
/// The query factory system.
query_handler: QueryBuilder,
}
impl Reader {
/// Creates a new reader from the given index context.
pub(crate) async fn create(ctx: &IndexContext) -> Result<Self> {
let reader: IndexReader = ctx
.index
.reader_builder()
.reload_policy(ReloadPolicy::OnCommit)
.num_searchers(ctx.reader_ctx.max_concurrency)
.try_into()?;
info!(
"[ READER @ {} ] index reader created with reload policy=OnCommit, num_searchers={}",
&ctx.name,
ctx.reader_ctx.max_concurrency,
);
let pool = {
let pool = SearcherExecutorPool::create(
reader,
ctx.reader_ctx.reader_threads,
ctx.reader_ctx.max_concurrency,
)
.await?;
Arc::new(pool)
};
info!(
"[ READER @ {} | {
let out: (Vec<(DateTime, DocAddress)>, usize) =
order_and_search(&searcher, field, query, collector, executor)?;
Ok((process_search(&searcher, schema, out.0)?, out.1))
} | conditional_block |
reader.rs | Context {
fn | (&self) -> Result<()> {
if self.max_concurrency == 0 {
return Err(Error::msg("max concurrency must be at least 1."));
}
Ok(())
}
}
impl ReaderContext {
fn default_reader_threads() -> usize {
1
}
}
/// A given query payload that describes how the reader should
/// search the index.
#[derive(Debug, Deserialize)]
pub struct QueryPayload {
/// The query(s) itself.
query: QuerySelector,
/// The amount of results to limit by.
#[serde(default = "QueryPayload::default_limit")]
limit: usize,
/// The amount of documents to skip before getting the results.
#[serde(default)]
offset: usize,
/// A specified field to order results by, this defaults to the
/// score of the indexed documents (relevancy).
order_by: Option<String>,
/// How to sort the data (asc/desc).
#[serde(default)]
sort: Sort,
}
impl QueryPayload {
fn default_limit() -> usize {
20
}
}
/// What order to sort the returned data.
#[derive(Debug, Copy, Clone, Deserialize)]
#[serde(rename_all = "lowercase")]
pub enum Sort {
/// Sort the data in ascending order.
Asc,
/// Sort the data in descending order. (Default)
Desc,
}
impl Default for Sort {
fn default() -> Self {
Self::Desc
}
}
#[derive(Debug, Serialize)]
pub struct QueryResults {
/// The retrieved documents.
pub(crate) hits: Vec<DocumentHit>,
/// The total amount of documents matching the search
count: usize,
/// The amount of time taken to search in seconds.
time_taken: f32,
}
impl QueryResults {
#[inline]
pub fn len(&self) -> usize {
self.hits.len()
}
}
/// Attaches an order by clause to the collector.
///
/// This collected the values with be returned in the order according to the
/// given field value.
fn order_and_search<R: AsScore + tantivy::fastfield::FastValue>(
searcher: &Searcher,
field: Field,
query: &dyn Query,
collector: TopDocs,
executor: &Executor,
) -> Result<(Vec<(R, DocAddress)>, usize)> {
let collector = collector.order_by_fast_field(field);
searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)
}
/// Performs the search operation and processes the returned results.
fn process_search<S: AsScore>(
searcher: &Searcher,
schema: &Schema,
top_docs: Vec<(S, DocAddress)>,
) -> Result<Vec<DocumentHit>> {
let mut hits = Vec::with_capacity(top_docs.len());
for (ratio, ref_address) in top_docs {
let retrieved_doc = searcher.doc(ref_address)?;
let mut doc = schema.to_named_doc(&retrieved_doc);
let id = doc.0
.remove("_id")
.ok_or_else(|| Error::msg("document has been missed labeled (missing primary key '_id'), the dataset is invalid"))?;
if let Value::U64(v) = id[0] {
hits.push(DocumentHit {
doc,
document_id: v,
score: ratio.as_score(),
});
} else {
return Err(Error::msg("document has been missed labeled (missing identifier tag), the dataset is invalid"));
}
}
Ok(hits)
}
/// Orders the search results by the given field with a given sort (ASC, DESC)
///
/// This function is super messy just because of all the type inference
/// so any contributions to clean this up would be very appreciated.
fn order_or_sort(
sort: Sort,
field: Field,
query: &dyn Query,
schema: &Schema,
searcher: &Searcher,
collector: TopDocs,
executor: &Executor,
) -> Result<(Vec<DocumentHit>, usize)> {
let field_type = schema.get_field_entry(field).field_type();
if let Sort::Desc = sort {
return match field_type {
FieldType::I64(_) => {
let out: (Vec<(i64, DocAddress)>, usize) =
order_and_search(&searcher, field, query, collector, executor)?;
Ok((process_search(&searcher, schema, out.0)?, out.1))
},
FieldType::U64(_) => {
let out: (Vec<(u64, DocAddress)>, usize) =
order_and_search(&searcher, field, query, collector, executor)?;
Ok((process_search(&searcher, schema, out.0)?, out.1))
},
FieldType::F64(_) => {
let out: (Vec<(f64, DocAddress)>, usize) =
order_and_search(&searcher, field, query, collector, executor)?;
Ok((process_search(&searcher, schema, out.0)?, out.1))
},
FieldType::Date(_) => {
let out: (Vec<(DateTime, DocAddress)>, usize) =
order_and_search(&searcher, field, query, collector, executor)?;
Ok((process_search(&searcher, schema, out.0)?, out.1))
},
_ => Err(Error::msg("field is not a fast field")),
};
}
let out = match field_type {
FieldType::I64(_) => {
let collector =
collector.custom_score(move |segment_reader: &SegmentReader| {
let reader = segment_reader
.fast_fields()
.i64(field)
.expect("field exists");
move |doc: DocId| {
let value: i64 = reader.get(doc);
std::cmp::Reverse(value)
}
});
let out: (Vec<(Reverse<i64>, DocAddress)>, usize) = searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)?;
(process_search(&searcher, schema, out.0)?, out.1)
},
FieldType::U64(_) => {
let collector =
collector.custom_score(move |segment_reader: &SegmentReader| {
let reader = segment_reader
.fast_fields()
.u64(field)
.expect("field exists");
move |doc: DocId| {
let value: u64 = reader.get(doc);
std::cmp::Reverse(value)
}
});
let out: (Vec<(Reverse<u64>, DocAddress)>, usize) = searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)?;
(process_search(&searcher, schema, out.0)?, out.1)
},
FieldType::F64(_) => {
let collector =
collector.custom_score(move |segment_reader: &SegmentReader| {
let reader = segment_reader
.fast_fields()
.f64(field)
.expect("field exists");
move |doc: DocId| {
let value: f64 = reader.get(doc);
std::cmp::Reverse(value)
}
});
let out: (Vec<(Reverse<f64>, DocAddress)>, usize) = searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)?;
(process_search(&searcher, schema, out.0)?, out.1)
},
FieldType::Date(_) => {
let collector =
collector.custom_score(move |segment_reader: &SegmentReader| {
let reader = segment_reader
.fast_fields()
.date(field)
.expect("field exists");
move |doc: DocId| {
let value: DateTime = reader.get(doc);
std::cmp::Reverse(value)
}
});
let out: (Vec<(Reverse<DateTime>, DocAddress)>, usize) = searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)?;
(process_search(&searcher, schema, out.0)?, out.1)
},
_ => return Err(Error::msg("field is not a fast field")),
};
return Ok(out);
}
/// The reader of the given index.
///
/// This manages all searches on the index which encompasses the concurrency
/// limiters and thread pool execution.
///
/// Each index should only have on `Reader` instance.
pub(crate) struct Reader {
/// The executor pool.
pool: crate::ReaderExecutor,
/// The query factory system.
query_handler: QueryBuilder,
}
impl Reader {
/// Creates a new reader from the given index context.
pub(crate) async fn create(ctx: &IndexContext) -> Result<Self> {
let reader: IndexReader = ctx
.index
.reader_builder()
.reload_policy(ReloadPolicy::OnCommit)
.num_searchers(ctx.reader_ctx.max_concurrency)
.try_into()?;
info!(
"[ READER @ {} ] index reader created with reload policy=OnCommit, num_searchers={}",
&ctx.name,
ctx.reader_ctx.max_concurrency,
);
let pool = {
let pool = SearcherExecutorPool::create(
reader,
ctx.reader_ctx.reader_threads,
ctx.reader_ctx.max_concurrency,
)
.await?;
Arc::new(pool)
};
info!(
"[ READER @ {} ] | validate | identifier_name |
reader.rs | pub(crate) hits: Vec<DocumentHit>,
/// The total amount of documents matching the search
count: usize,
/// The amount of time taken to search in seconds.
time_taken: f32,
}
impl QueryResults {
#[inline]
pub fn len(&self) -> usize {
self.hits.len()
}
}
/// Attaches an order by clause to the collector.
///
/// This collected the values with be returned in the order according to the
/// given field value.
fn order_and_search<R: AsScore + tantivy::fastfield::FastValue>(
searcher: &Searcher,
field: Field,
query: &dyn Query,
collector: TopDocs,
executor: &Executor,
) -> Result<(Vec<(R, DocAddress)>, usize)> {
let collector = collector.order_by_fast_field(field);
searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)
}
/// Performs the search operation and processes the returned results.
fn process_search<S: AsScore>(
searcher: &Searcher,
schema: &Schema,
top_docs: Vec<(S, DocAddress)>,
) -> Result<Vec<DocumentHit>> {
let mut hits = Vec::with_capacity(top_docs.len());
for (ratio, ref_address) in top_docs {
let retrieved_doc = searcher.doc(ref_address)?;
let mut doc = schema.to_named_doc(&retrieved_doc);
let id = doc.0
.remove("_id")
.ok_or_else(|| Error::msg("document has been missed labeled (missing primary key '_id'), the dataset is invalid"))?;
if let Value::U64(v) = id[0] {
hits.push(DocumentHit {
doc,
document_id: v,
score: ratio.as_score(),
});
} else {
return Err(Error::msg("document has been missed labeled (missing identifier tag), the dataset is invalid"));
}
}
Ok(hits)
}
/// Orders the search results by the given field with a given sort (ASC, DESC)
///
/// This function is super messy just because of all the type inference
/// so any contributions to clean this up would be very appreciated.
fn order_or_sort(
sort: Sort,
field: Field,
query: &dyn Query,
schema: &Schema,
searcher: &Searcher,
collector: TopDocs,
executor: &Executor,
) -> Result<(Vec<DocumentHit>, usize)> {
let field_type = schema.get_field_entry(field).field_type();
if let Sort::Desc = sort {
return match field_type {
FieldType::I64(_) => {
let out: (Vec<(i64, DocAddress)>, usize) =
order_and_search(&searcher, field, query, collector, executor)?;
Ok((process_search(&searcher, schema, out.0)?, out.1))
},
FieldType::U64(_) => {
let out: (Vec<(u64, DocAddress)>, usize) =
order_and_search(&searcher, field, query, collector, executor)?;
Ok((process_search(&searcher, schema, out.0)?, out.1))
},
FieldType::F64(_) => {
let out: (Vec<(f64, DocAddress)>, usize) =
order_and_search(&searcher, field, query, collector, executor)?;
Ok((process_search(&searcher, schema, out.0)?, out.1))
},
FieldType::Date(_) => {
let out: (Vec<(DateTime, DocAddress)>, usize) =
order_and_search(&searcher, field, query, collector, executor)?;
Ok((process_search(&searcher, schema, out.0)?, out.1))
},
_ => Err(Error::msg("field is not a fast field")),
};
}
let out = match field_type {
FieldType::I64(_) => {
let collector =
collector.custom_score(move |segment_reader: &SegmentReader| {
let reader = segment_reader
.fast_fields()
.i64(field)
.expect("field exists");
move |doc: DocId| {
let value: i64 = reader.get(doc);
std::cmp::Reverse(value)
}
});
let out: (Vec<(Reverse<i64>, DocAddress)>, usize) = searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)?;
(process_search(&searcher, schema, out.0)?, out.1)
},
FieldType::U64(_) => {
let collector =
collector.custom_score(move |segment_reader: &SegmentReader| {
let reader = segment_reader
.fast_fields()
.u64(field)
.expect("field exists");
move |doc: DocId| {
let value: u64 = reader.get(doc);
std::cmp::Reverse(value)
}
});
let out: (Vec<(Reverse<u64>, DocAddress)>, usize) = searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)?;
(process_search(&searcher, schema, out.0)?, out.1)
},
FieldType::F64(_) => {
let collector =
collector.custom_score(move |segment_reader: &SegmentReader| {
let reader = segment_reader
.fast_fields()
.f64(field)
.expect("field exists");
move |doc: DocId| {
let value: f64 = reader.get(doc);
std::cmp::Reverse(value)
}
});
let out: (Vec<(Reverse<f64>, DocAddress)>, usize) = searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)?;
(process_search(&searcher, schema, out.0)?, out.1)
},
FieldType::Date(_) => {
let collector =
collector.custom_score(move |segment_reader: &SegmentReader| {
let reader = segment_reader
.fast_fields()
.date(field)
.expect("field exists");
move |doc: DocId| {
let value: DateTime = reader.get(doc);
std::cmp::Reverse(value)
}
});
let out: (Vec<(Reverse<DateTime>, DocAddress)>, usize) = searcher
.search_with_executor(query, &(collector, Count), executor)
.map_err(Error::from)?;
(process_search(&searcher, schema, out.0)?, out.1)
},
_ => return Err(Error::msg("field is not a fast field")),
};
return Ok(out);
}
/// The reader of the given index.
///
/// This manages all searches on the index which encompasses the concurrency
/// limiters and thread pool execution.
///
/// Each index should only have on `Reader` instance.
pub(crate) struct Reader {
/// The executor pool.
pool: crate::ReaderExecutor,
/// The query factory system.
query_handler: QueryBuilder,
}
impl Reader {
/// Creates a new reader from the given index context.
pub(crate) async fn create(ctx: &IndexContext) -> Result<Self> {
let reader: IndexReader = ctx
.index
.reader_builder()
.reload_policy(ReloadPolicy::OnCommit)
.num_searchers(ctx.reader_ctx.max_concurrency)
.try_into()?;
info!(
"[ READER @ {} ] index reader created with reload policy=OnCommit, num_searchers={}",
&ctx.name,
ctx.reader_ctx.max_concurrency,
);
let pool = {
let pool = SearcherExecutorPool::create(
reader,
ctx.reader_ctx.reader_threads,
ctx.reader_ctx.max_concurrency,
)
.await?;
Arc::new(pool)
};
info!(
"[ READER @ {} ] executor pool has successfully started! max_concurrency={}, total_threads={}",
&ctx.name,
ctx.reader_ctx.max_concurrency,
ctx.reader_ctx.max_concurrency * ctx.reader_ctx.reader_threads
);
let query_ctx = ctx.query_ctx.clone();
let query_handler = QueryBuilder::new(
query_ctx,
ctx.stop_words.clone(),
ctx.correction_manager.clone(),
&ctx.index,
pool.clone(),
);
info!(
"[ QUERY-BUILDER @ {} ] query builder constructed with config: fast-fuzzy={} strip-stop-words={}.",
&ctx.name,
ctx.query_ctx.use_fast_fuzzy,
ctx.query_ctx.strip_stop_words,
);
Ok(Self {
pool,
query_handler,
})
}
/// Gets a singular document from the given id.
///
/// If no document is found an error is raised without context.
pub(crate) async fn get_document(&self, id: DocumentId) -> Result<DocumentHit> | {
let id_field = self.query_handler.id_field();
let document = self
.pool
.spawn(move |searcher, executor| {
let qry = TermQuery::new(
Term::from_field_u64(id_field, id),
IndexRecordOption::Basic,
);
let mut results = searcher.search_with_executor(
&qry,
&TopDocs::with_limit(1),
executor,
)?;
if results.len() == 0 {
return Err(Error::msg(format!(
"no document exists with id: '{}'",
id
))); | identifier_body |
|
LibphysGRU.py | d)
self.identity = tf.eye(Sd)
return E, U, W, V, b, c, [None, None]
def GRUnn(self, out_prev, x_t):
E, U, W, V, b, c = self.E, self.U, self.W, self.V, self.b, self.c
Hd, Sd, Bd = self.signal2model.hidden_dim, self.signal2model.signal_dim, tf.shape(x_t)[0]
coversion_ones = tf.ones((1, Bd), dtype=tf.float32, name="conversion_matrix")
# s_prev, o_prev, l_prev = out_prev
s_prev, o_prev = out_prev
def GRU(last_input, gru_params):
s_g_prev, u, w, b = gru_params
z = tf.nn.sigmoid(tf.matmul(u[0], last_input) +
tf.matmul(w[0], s_g_prev) +
tf.matmul(tf.reshape(b[0], (Hd, 1)), coversion_ones))
r = tf.nn.sigmoid(tf.matmul(u[1], last_input) + tf.matmul(w[1], s_g_prev) +
tf.matmul(tf.reshape(b[1], (Hd, 1)), coversion_ones))
value = tf.matmul(u[2], last_input) + tf.matmul(w[2], s_g_prev * r) + \
tf.matmul(tf.reshape(b[2], (Hd, 1)), coversion_ones)
s_candidate = tf.nn.tanh(value)
output = tf.add(((tf.ones_like(z) - z) * s_candidate), (z * s_g_prev), name="out_GRU")
return output
# x_e -> (Hd x Mb)
x_e = tf.gather(self.E, x_t, axis=1)
s_t_ = []
s_t_.append(GRU(x_e, [s_prev[0], U[0], W[0], b[0]]))
s_t_.append(GRU(s_t_[0], [s_prev[1], U[1], W[1], b[1]]))
s_t_.append(GRU(s_t_[1], [s_prev[2], U[2], W[2], b[2]]))
s_t = tf.stack(s_t_)
# tf.scan(GRU, (s_prev, self.U, self.W, self.b), initializer=x_e, parallel_iterations=1, name="states")
logits = tf.matmul(self.V, s_t[-1]) + tf.matmul(tf.reshape(self.c, (Sd, 1)), coversion_ones)
o_t = tf.nn.softmax(logits, axis=2)
return [s_t, o_t]#, logits]
def feed_forward_predict(self, x_batch):
initial_s = tf.zeros((self.signal2model.n_grus, self.signal2model.hidden_dim, tf.shape(x_batch)[1]), dtype=np.float32)
initial_out = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32)
# initial_l = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32)
# x_batch = (N x Bd) - N (samples); Bd - Batch dimension
# [s, o, l] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out, initial_l], parallel_iterations=1,
# name="network_output")
[_, o] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out], parallel_iterations=1,
name="network_output")
return o
def feed_forward_predict_with_states(self, x_batch):
initial_s = tf.zeros((self.signal2model.n_grus, self.signal2model.hidden_dim, tf.shape(x_batch)[1]),
dtype=np.float32)
initial_out = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32)
# initial_l = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32)
# x_batch = (N x Bd) - N (samples); Bd - Batch dimension | # [s, o, l] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out, initial_l], parallel_iterations=1,
# name="network_output")
[s, o] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out], parallel_iterations=1,
name="network_output")
return [s, o]
def calculate_predictions(self):
# MAP NOT WORKING:
# shape(X)[0] -> Windows
# shape(X)[1] -> Samples
# n_batches = int(signal2model.batch_size / self.signal2model.mini_batch_size)
# N = tf.shape(self.X)[1]
# print(X)
# get the matrices from E with tf.gather(E, X, axis=1, name="X_e")
# transpose these matrices for (batch_size, HD, N)
# reshape to enter map, where each minibatch is entered at the same time (n_batches, mini_batch, HD, N)
# transpose to enter the DNN inside -> (n_batches, N, mini_batch)
return self.feed_forward_predict(tf.transpose(self.X))
def to_one_hot_vector_in_mini_batches(self, matrix):
return self.get_one_hot(matrix)
def get_one_hot(self, columns):
return tf.gather(self.identity, columns)
def calculate_cross_entropy(self):
return None
# logits = tf.transpose(self.logits, perm=[2, 0, 1])
# n_batches = int(self.signal2model.batch_size / self.signal2model.mini_batch_size)
# y = tf.reshape(self.Y, (n_batches, self.signal2model.mini_batch_size, tf.shape(self.Y)[1]))
# self.full_loss = tf.losses.sparse_softmax_cross_entropy_with_logits(labels=y, logits=logits)
# return self.full_loss
def calculate_mse(self):
_y = self.to_one_hot_vector_in_mini_batches(self.Y)
return tf.reduce_mean(tf.subtract(tf.transpose(self.out, perm=[2, 0, 1]), _y) ** 2, axis=2, name="mse")
def calculate_mse_vector_loss(self, x, y):
with tf.variable_scope('vector_loss'):
return tf.reduce_mean(self.calculate_minibatch_mse(x, y), axis=0, name="vector_loss")
def calculate_mse_loss(self):
return tf.reduce_mean(self.calculate_minibatch_mse(), axis=0, name="loss")
@property
def loss_op(self):
""" An Operation that takes one optimization step. """
return self.loss
def init_optimizer(self):
trainables = self.parameters
grads = tf.gradients(self.loss, trainables)
grad_var_pairs = zip(grads, trainables)
# grads, _ = tf.clip_by_global_norm(grads, clip_norm=1.0)
# self.learning_rate = tf.train.exponential_decay(
# self.signal2model.learning_rate_val, self.epoch, self.signal2model.count_to_break_max,
# self.signal2model.decay, staircase=True)
# with tf.device('/gpu:1'):
optimizer = tf.train.RMSPropOptimizer(self.learning_rate_gpu)
# optimizer = tf.train.AdamOptimizer(self.learning_rate)
self._optimize_op = optimizer.apply_gradients(grad_var_pairs)
@property
def optimize_op(self):
""" An Operation that takes one optimization step. """
return self._optimize_op
def train(self, X, Y, signal2model=None):
self.batch_size += np.shape(X)[0]
self.init_time = time.time()
plt.ion()
if signal2model is not None:
self.signal2model = signal2model
plt.ion()
condition_not_met = True
history = []
self.epoch = 0
self.loss_history = []
tf.summary.scalar('loss', self.loss)
merged = tf.summary.merge_all()
train_writer = tf.summary.FileWriter('train',
self.session.graph)
# run_options = tf.RunOptions(report_tensor_allocations_upon_oom=True)
# db_url = 'postgres://belo:passsword@localhost/postgres'
# experiments, steps, model_params = labnotebook.initialize(db_url)
# model_desc = {'loss': 0.}
# experiment = labnotebook.start_experiment(model_desc=model_desc)
tf.global_variables_initializer()
while condition_not_met:
self.epoch += 1
# tic = time.time()
random_indexes = np.random.permutation(self.signal2model.batch_size)
groups = np.reshape(random_indexes,
(int(self.signal2model.batch_size/self.signal2model.mini_batch_size),
self.signal2model.mini_batch_size))
for group in groups:
dictionary = self.shuffle(X, Y, group)
op, group_loss = self.session.run(
[self.optimize_op, self.loss_op],
feed_dict=dictionary)#, options=run_options)
# labnotebook.step_experiment(experiment,
# timestep=str(self.epoch),
# trainacc=0,
# valacc=0,
# trainloss=str(group_loss))
# print("toc: {0} secs".format(time | random_line_split |
|
LibphysGRU.py | , U, W, V, b, c, [None, None]
def GRUnn(self, out_prev, x_t):
E, U, W, V, b, c = self.E, self.U, self.W, self.V, self.b, self.c
Hd, Sd, Bd = self.signal2model.hidden_dim, self.signal2model.signal_dim, tf.shape(x_t)[0]
coversion_ones = tf.ones((1, Bd), dtype=tf.float32, name="conversion_matrix")
# s_prev, o_prev, l_prev = out_prev
s_prev, o_prev = out_prev
def GRU(last_input, gru_params):
s_g_prev, u, w, b = gru_params
z = tf.nn.sigmoid(tf.matmul(u[0], last_input) +
tf.matmul(w[0], s_g_prev) +
tf.matmul(tf.reshape(b[0], (Hd, 1)), coversion_ones))
r = tf.nn.sigmoid(tf.matmul(u[1], last_input) + tf.matmul(w[1], s_g_prev) +
tf.matmul(tf.reshape(b[1], (Hd, 1)), coversion_ones))
value = tf.matmul(u[2], last_input) + tf.matmul(w[2], s_g_prev * r) + \
tf.matmul(tf.reshape(b[2], (Hd, 1)), coversion_ones)
s_candidate = tf.nn.tanh(value)
output = tf.add(((tf.ones_like(z) - z) * s_candidate), (z * s_g_prev), name="out_GRU")
return output
# x_e -> (Hd x Mb)
x_e = tf.gather(self.E, x_t, axis=1)
s_t_ = []
s_t_.append(GRU(x_e, [s_prev[0], U[0], W[0], b[0]]))
s_t_.append(GRU(s_t_[0], [s_prev[1], U[1], W[1], b[1]]))
s_t_.append(GRU(s_t_[1], [s_prev[2], U[2], W[2], b[2]]))
s_t = tf.stack(s_t_)
# tf.scan(GRU, (s_prev, self.U, self.W, self.b), initializer=x_e, parallel_iterations=1, name="states")
logits = tf.matmul(self.V, s_t[-1]) + tf.matmul(tf.reshape(self.c, (Sd, 1)), coversion_ones)
o_t = tf.nn.softmax(logits, axis=2)
return [s_t, o_t]#, logits]
def feed_forward_predict(self, x_batch):
initial_s = tf.zeros((self.signal2model.n_grus, self.signal2model.hidden_dim, tf.shape(x_batch)[1]), dtype=np.float32)
initial_out = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32)
# initial_l = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32)
# x_batch = (N x Bd) - N (samples); Bd - Batch dimension
# [s, o, l] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out, initial_l], parallel_iterations=1,
# name="network_output")
[_, o] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out], parallel_iterations=1,
name="network_output")
return o
def feed_forward_predict_with_states(self, x_batch):
initial_s = tf.zeros((self.signal2model.n_grus, self.signal2model.hidden_dim, tf.shape(x_batch)[1]),
dtype=np.float32)
initial_out = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32)
# initial_l = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32)
# x_batch = (N x Bd) - N (samples); Bd - Batch dimension
# [s, o, l] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out, initial_l], parallel_iterations=1,
# name="network_output")
[s, o] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out], parallel_iterations=1,
name="network_output")
return [s, o]
def calculate_predictions(self):
# MAP NOT WORKING:
# shape(X)[0] -> Windows
# shape(X)[1] -> Samples
# n_batches = int(signal2model.batch_size / self.signal2model.mini_batch_size)
# N = tf.shape(self.X)[1]
# print(X)
# get the matrices from E with tf.gather(E, X, axis=1, name="X_e")
# transpose these matrices for (batch_size, HD, N)
# reshape to enter map, where each minibatch is entered at the same time (n_batches, mini_batch, HD, N)
# transpose to enter the DNN inside -> (n_batches, N, mini_batch)
return self.feed_forward_predict(tf.transpose(self.X))
def to_one_hot_vector_in_mini_batches(self, matrix):
return self.get_one_hot(matrix)
def get_one_hot(self, columns):
return tf.gather(self.identity, columns)
def calculate_cross_entropy(self):
return None
# logits = tf.transpose(self.logits, perm=[2, 0, 1])
# n_batches = int(self.signal2model.batch_size / self.signal2model.mini_batch_size)
# y = tf.reshape(self.Y, (n_batches, self.signal2model.mini_batch_size, tf.shape(self.Y)[1]))
# self.full_loss = tf.losses.sparse_softmax_cross_entropy_with_logits(labels=y, logits=logits)
# return self.full_loss
def calculate_mse(self):
_y = self.to_one_hot_vector_in_mini_batches(self.Y)
return tf.reduce_mean(tf.subtract(tf.transpose(self.out, perm=[2, 0, 1]), _y) ** 2, axis=2, name="mse")
def calculate_mse_vector_loss(self, x, y):
with tf.variable_scope('vector_loss'):
return tf.reduce_mean(self.calculate_minibatch_mse(x, y), axis=0, name="vector_loss")
def calculate_mse_loss(self):
return tf.reduce_mean(self.calculate_minibatch_mse(), axis=0, name="loss")
@property
def loss_op(self):
""" An Operation that takes one optimization step. """
return self.loss
def init_optimizer(self):
trainables = self.parameters
grads = tf.gradients(self.loss, trainables)
grad_var_pairs = zip(grads, trainables)
# grads, _ = tf.clip_by_global_norm(grads, clip_norm=1.0)
# self.learning_rate = tf.train.exponential_decay(
# self.signal2model.learning_rate_val, self.epoch, self.signal2model.count_to_break_max,
# self.signal2model.decay, staircase=True)
# with tf.device('/gpu:1'):
optimizer = tf.train.RMSPropOptimizer(self.learning_rate_gpu)
# optimizer = tf.train.AdamOptimizer(self.learning_rate)
self._optimize_op = optimizer.apply_gradients(grad_var_pairs)
@property
def optimize_op(self):
""" An Operation that takes one optimization step. """
return self._optimize_op
def train(self, X, Y, signal2model=None):
self.batch_size += np.shape(X)[0]
self.init_time = time.time()
plt.ion()
if signal2model is not None:
self.signal2model = signal2model
plt.ion()
condition_not_met = True
history = []
self.epoch = 0
self.loss_history = []
tf.summary.scalar('loss', self.loss)
merged = tf.summary.merge_all()
train_writer = tf.summary.FileWriter('train',
self.session.graph)
# run_options = tf.RunOptions(report_tensor_allocations_upon_oom=True)
# db_url = 'postgres://belo:passsword@localhost/postgres'
# experiments, steps, model_params = labnotebook.initialize(db_url)
# model_desc = {'loss': 0.}
# experiment = labnotebook.start_experiment(model_desc=model_desc)
tf.global_variables_initializer()
while condition_not_met:
| self.epoch += 1
# tic = time.time()
random_indexes = np.random.permutation(self.signal2model.batch_size)
groups = np.reshape(random_indexes,
(int(self.signal2model.batch_size/self.signal2model.mini_batch_size),
self.signal2model.mini_batch_size))
for group in groups:
dictionary = self.shuffle(X, Y, group)
op, group_loss = self.session.run(
[self.optimize_op, self.loss_op],
feed_dict=dictionary)#, options=run_options)
# labnotebook.step_experiment(experiment,
# timestep=str(self.epoch),
# trainacc=0,
# valacc=0,
# trainloss=str(group_loss))
# print("toc: {0} secs".format(time.time()-tic))
# new_tic = time.time() | conditional_block |
|
LibphysGRU.py | )
self.identity = tf.eye(Sd)
return E, U, W, V, b, c, [None, None]
def GRUnn(self, out_prev, x_t):
E, U, W, V, b, c = self.E, self.U, self.W, self.V, self.b, self.c
Hd, Sd, Bd = self.signal2model.hidden_dim, self.signal2model.signal_dim, tf.shape(x_t)[0]
coversion_ones = tf.ones((1, Bd), dtype=tf.float32, name="conversion_matrix")
# s_prev, o_prev, l_prev = out_prev
s_prev, o_prev = out_prev
def GRU(last_input, gru_params):
s_g_prev, u, w, b = gru_params
z = tf.nn.sigmoid(tf.matmul(u[0], last_input) +
tf.matmul(w[0], s_g_prev) +
tf.matmul(tf.reshape(b[0], (Hd, 1)), coversion_ones))
r = tf.nn.sigmoid(tf.matmul(u[1], last_input) + tf.matmul(w[1], s_g_prev) +
tf.matmul(tf.reshape(b[1], (Hd, 1)), coversion_ones))
value = tf.matmul(u[2], last_input) + tf.matmul(w[2], s_g_prev * r) + \
tf.matmul(tf.reshape(b[2], (Hd, 1)), coversion_ones)
s_candidate = tf.nn.tanh(value)
output = tf.add(((tf.ones_like(z) - z) * s_candidate), (z * s_g_prev), name="out_GRU")
return output
# x_e -> (Hd x Mb)
x_e = tf.gather(self.E, x_t, axis=1)
s_t_ = []
s_t_.append(GRU(x_e, [s_prev[0], U[0], W[0], b[0]]))
s_t_.append(GRU(s_t_[0], [s_prev[1], U[1], W[1], b[1]]))
s_t_.append(GRU(s_t_[1], [s_prev[2], U[2], W[2], b[2]]))
s_t = tf.stack(s_t_)
# tf.scan(GRU, (s_prev, self.U, self.W, self.b), initializer=x_e, parallel_iterations=1, name="states")
logits = tf.matmul(self.V, s_t[-1]) + tf.matmul(tf.reshape(self.c, (Sd, 1)), coversion_ones)
o_t = tf.nn.softmax(logits, axis=2)
return [s_t, o_t]#, logits]
def feed_forward_predict(self, x_batch):
initial_s = tf.zeros((self.signal2model.n_grus, self.signal2model.hidden_dim, tf.shape(x_batch)[1]), dtype=np.float32)
initial_out = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32)
# initial_l = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32)
# x_batch = (N x Bd) - N (samples); Bd - Batch dimension
# [s, o, l] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out, initial_l], parallel_iterations=1,
# name="network_output")
[_, o] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out], parallel_iterations=1,
name="network_output")
return o
def feed_forward_predict_with_states(self, x_batch):
initial_s = tf.zeros((self.signal2model.n_grus, self.signal2model.hidden_dim, tf.shape(x_batch)[1]),
dtype=np.float32)
initial_out = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32)
# initial_l = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32)
# x_batch = (N x Bd) - N (samples); Bd - Batch dimension
# [s, o, l] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out, initial_l], parallel_iterations=1,
# name="network_output")
[s, o] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out], parallel_iterations=1,
name="network_output")
return [s, o]
def calculate_predictions(self):
# MAP NOT WORKING:
# shape(X)[0] -> Windows
# shape(X)[1] -> Samples
# n_batches = int(signal2model.batch_size / self.signal2model.mini_batch_size)
# N = tf.shape(self.X)[1]
# print(X)
# get the matrices from E with tf.gather(E, X, axis=1, name="X_e")
# transpose these matrices for (batch_size, HD, N)
# reshape to enter map, where each minibatch is entered at the same time (n_batches, mini_batch, HD, N)
# transpose to enter the DNN inside -> (n_batches, N, mini_batch)
return self.feed_forward_predict(tf.transpose(self.X))
def to_one_hot_vector_in_mini_batches(self, matrix):
return self.get_one_hot(matrix)
def get_one_hot(self, columns):
return tf.gather(self.identity, columns)
def calculate_cross_entropy(self):
return None
# logits = tf.transpose(self.logits, perm=[2, 0, 1])
# n_batches = int(self.signal2model.batch_size / self.signal2model.mini_batch_size)
# y = tf.reshape(self.Y, (n_batches, self.signal2model.mini_batch_size, tf.shape(self.Y)[1]))
# self.full_loss = tf.losses.sparse_softmax_cross_entropy_with_logits(labels=y, logits=logits)
# return self.full_loss
def calculate_mse(self):
_y = self.to_one_hot_vector_in_mini_batches(self.Y)
return tf.reduce_mean(tf.subtract(tf.transpose(self.out, perm=[2, 0, 1]), _y) ** 2, axis=2, name="mse")
def calculate_mse_vector_loss(self, x, y):
with tf.variable_scope('vector_loss'):
return tf.reduce_mean(self.calculate_minibatch_mse(x, y), axis=0, name="vector_loss")
def calculate_mse_loss(self):
return tf.reduce_mean(self.calculate_minibatch_mse(), axis=0, name="loss")
@property
def loss_op(self):
""" An Operation that takes one optimization step. """
return self.loss
def init_optimizer(self):
|
@property
def optimize_op(self):
""" An Operation that takes one optimization step. """
return self._optimize_op
def train(self, X, Y, signal2model=None):
self.batch_size += np.shape(X)[0]
self.init_time = time.time()
plt.ion()
if signal2model is not None:
self.signal2model = signal2model
plt.ion()
condition_not_met = True
history = []
self.epoch = 0
self.loss_history = []
tf.summary.scalar('loss', self.loss)
merged = tf.summary.merge_all()
train_writer = tf.summary.FileWriter('train',
self.session.graph)
# run_options = tf.RunOptions(report_tensor_allocations_upon_oom=True)
# db_url = 'postgres://belo:passsword@localhost/postgres'
# experiments, steps, model_params = labnotebook.initialize(db_url)
# model_desc = {'loss': 0.}
# experiment = labnotebook.start_experiment(model_desc=model_desc)
tf.global_variables_initializer()
while condition_not_met:
self.epoch += 1
# tic = time.time()
random_indexes = np.random.permutation(self.signal2model.batch_size)
groups = np.reshape(random_indexes,
(int(self.signal2model.batch_size/self.signal2model.mini_batch_size),
self.signal2model.mini_batch_size))
for group in groups:
dictionary = self.shuffle(X, Y, group)
op, group_loss = self.session.run(
[self.optimize_op, self.loss_op],
feed_dict=dictionary)#, options=run_options)
# labnotebook.step_experiment(experiment,
# timestep=str(self.epoch),
# trainacc=0,
# valacc=0,
# trainloss=str(group_loss))
# print("toc: {0} secs".format | trainables = self.parameters
grads = tf.gradients(self.loss, trainables)
grad_var_pairs = zip(grads, trainables)
# grads, _ = tf.clip_by_global_norm(grads, clip_norm=1.0)
# self.learning_rate = tf.train.exponential_decay(
# self.signal2model.learning_rate_val, self.epoch, self.signal2model.count_to_break_max,
# self.signal2model.decay, staircase=True)
# with tf.device('/gpu:1'):
optimizer = tf.train.RMSPropOptimizer(self.learning_rate_gpu)
# optimizer = tf.train.AdamOptimizer(self.learning_rate)
self._optimize_op = optimizer.apply_gradients(grad_var_pairs) | identifier_body |
LibphysGRU.py | )
self.identity = tf.eye(Sd)
return E, U, W, V, b, c, [None, None]
def GRUnn(self, out_prev, x_t):
E, U, W, V, b, c = self.E, self.U, self.W, self.V, self.b, self.c
Hd, Sd, Bd = self.signal2model.hidden_dim, self.signal2model.signal_dim, tf.shape(x_t)[0]
coversion_ones = tf.ones((1, Bd), dtype=tf.float32, name="conversion_matrix")
# s_prev, o_prev, l_prev = out_prev
s_prev, o_prev = out_prev
def GRU(last_input, gru_params):
s_g_prev, u, w, b = gru_params
z = tf.nn.sigmoid(tf.matmul(u[0], last_input) +
tf.matmul(w[0], s_g_prev) +
tf.matmul(tf.reshape(b[0], (Hd, 1)), coversion_ones))
r = tf.nn.sigmoid(tf.matmul(u[1], last_input) + tf.matmul(w[1], s_g_prev) +
tf.matmul(tf.reshape(b[1], (Hd, 1)), coversion_ones))
value = tf.matmul(u[2], last_input) + tf.matmul(w[2], s_g_prev * r) + \
tf.matmul(tf.reshape(b[2], (Hd, 1)), coversion_ones)
s_candidate = tf.nn.tanh(value)
output = tf.add(((tf.ones_like(z) - z) * s_candidate), (z * s_g_prev), name="out_GRU")
return output
# x_e -> (Hd x Mb)
x_e = tf.gather(self.E, x_t, axis=1)
s_t_ = []
s_t_.append(GRU(x_e, [s_prev[0], U[0], W[0], b[0]]))
s_t_.append(GRU(s_t_[0], [s_prev[1], U[1], W[1], b[1]]))
s_t_.append(GRU(s_t_[1], [s_prev[2], U[2], W[2], b[2]]))
s_t = tf.stack(s_t_)
# tf.scan(GRU, (s_prev, self.U, self.W, self.b), initializer=x_e, parallel_iterations=1, name="states")
logits = tf.matmul(self.V, s_t[-1]) + tf.matmul(tf.reshape(self.c, (Sd, 1)), coversion_ones)
o_t = tf.nn.softmax(logits, axis=2)
return [s_t, o_t]#, logits]
def | (self, x_batch):
initial_s = tf.zeros((self.signal2model.n_grus, self.signal2model.hidden_dim, tf.shape(x_batch)[1]), dtype=np.float32)
initial_out = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32)
# initial_l = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32)
# x_batch = (N x Bd) - N (samples); Bd - Batch dimension
# [s, o, l] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out, initial_l], parallel_iterations=1,
# name="network_output")
[_, o] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out], parallel_iterations=1,
name="network_output")
return o
def feed_forward_predict_with_states(self, x_batch):
initial_s = tf.zeros((self.signal2model.n_grus, self.signal2model.hidden_dim, tf.shape(x_batch)[1]),
dtype=np.float32)
initial_out = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32)
# initial_l = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32)
# x_batch = (N x Bd) - N (samples); Bd - Batch dimension
# [s, o, l] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out, initial_l], parallel_iterations=1,
# name="network_output")
[s, o] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out], parallel_iterations=1,
name="network_output")
return [s, o]
def calculate_predictions(self):
# MAP NOT WORKING:
# shape(X)[0] -> Windows
# shape(X)[1] -> Samples
# n_batches = int(signal2model.batch_size / self.signal2model.mini_batch_size)
# N = tf.shape(self.X)[1]
# print(X)
# get the matrices from E with tf.gather(E, X, axis=1, name="X_e")
# transpose these matrices for (batch_size, HD, N)
# reshape to enter map, where each minibatch is entered at the same time (n_batches, mini_batch, HD, N)
# transpose to enter the DNN inside -> (n_batches, N, mini_batch)
return self.feed_forward_predict(tf.transpose(self.X))
def to_one_hot_vector_in_mini_batches(self, matrix):
return self.get_one_hot(matrix)
def get_one_hot(self, columns):
return tf.gather(self.identity, columns)
def calculate_cross_entropy(self):
return None
# logits = tf.transpose(self.logits, perm=[2, 0, 1])
# n_batches = int(self.signal2model.batch_size / self.signal2model.mini_batch_size)
# y = tf.reshape(self.Y, (n_batches, self.signal2model.mini_batch_size, tf.shape(self.Y)[1]))
# self.full_loss = tf.losses.sparse_softmax_cross_entropy_with_logits(labels=y, logits=logits)
# return self.full_loss
def calculate_mse(self):
_y = self.to_one_hot_vector_in_mini_batches(self.Y)
return tf.reduce_mean(tf.subtract(tf.transpose(self.out, perm=[2, 0, 1]), _y) ** 2, axis=2, name="mse")
def calculate_mse_vector_loss(self, x, y):
with tf.variable_scope('vector_loss'):
return tf.reduce_mean(self.calculate_minibatch_mse(x, y), axis=0, name="vector_loss")
def calculate_mse_loss(self):
return tf.reduce_mean(self.calculate_minibatch_mse(), axis=0, name="loss")
@property
def loss_op(self):
""" An Operation that takes one optimization step. """
return self.loss
def init_optimizer(self):
trainables = self.parameters
grads = tf.gradients(self.loss, trainables)
grad_var_pairs = zip(grads, trainables)
# grads, _ = tf.clip_by_global_norm(grads, clip_norm=1.0)
# self.learning_rate = tf.train.exponential_decay(
# self.signal2model.learning_rate_val, self.epoch, self.signal2model.count_to_break_max,
# self.signal2model.decay, staircase=True)
# with tf.device('/gpu:1'):
optimizer = tf.train.RMSPropOptimizer(self.learning_rate_gpu)
# optimizer = tf.train.AdamOptimizer(self.learning_rate)
self._optimize_op = optimizer.apply_gradients(grad_var_pairs)
@property
def optimize_op(self):
""" An Operation that takes one optimization step. """
return self._optimize_op
def train(self, X, Y, signal2model=None):
self.batch_size += np.shape(X)[0]
self.init_time = time.time()
plt.ion()
if signal2model is not None:
self.signal2model = signal2model
plt.ion()
condition_not_met = True
history = []
self.epoch = 0
self.loss_history = []
tf.summary.scalar('loss', self.loss)
merged = tf.summary.merge_all()
train_writer = tf.summary.FileWriter('train',
self.session.graph)
# run_options = tf.RunOptions(report_tensor_allocations_upon_oom=True)
# db_url = 'postgres://belo:passsword@localhost/postgres'
# experiments, steps, model_params = labnotebook.initialize(db_url)
# model_desc = {'loss': 0.}
# experiment = labnotebook.start_experiment(model_desc=model_desc)
tf.global_variables_initializer()
while condition_not_met:
self.epoch += 1
# tic = time.time()
random_indexes = np.random.permutation(self.signal2model.batch_size)
groups = np.reshape(random_indexes,
(int(self.signal2model.batch_size/self.signal2model.mini_batch_size),
self.signal2model.mini_batch_size))
for group in groups:
dictionary = self.shuffle(X, Y, group)
op, group_loss = self.session.run(
[self.optimize_op, self.loss_op],
feed_dict=dictionary)#, options=run_options)
# labnotebook.step_experiment(experiment,
# timestep=str(self.epoch),
# trainacc=0,
# valacc=0,
# trainloss=str(group_loss))
# print("toc: {0} secs".format | feed_forward_predict | identifier_name |
search-table.controller.js | License is distributed on an "AS IS" BASIS, WITHOUT
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
* License for the specific language governing permissions and limitations
* under the License.
*/
(function () {
'use strict';
/**
* @ngdoc controller
* @name SearchTableController
*
* @description
* Controller for the search table.
* Serves as the focal point for table actions.
*/
angular
.module('horizon.dashboard.project.search')
.controller('searchTableController', SearchTableController);
SearchTableController.$inject = [
'$scope',
'$filter',
'$q',
'$timeout',
'searchPluginResourceTypesFilter',
'horizon.framework.conf.resource-type-registry.service',
'horizon.app.core.openstack-service-api.userSession',
'horizon.dashboard.project.search.searchlightFacetUtils',
'horizon.dashboard.project.search.searchlightSearchHelper',
'horizon.dashboard.project.search.settingsService',
'horizon.dashboard.search.search.util.cache.service'
];
function SearchTableController($scope,
$filter,
$q,
$timeout,
searchPluginResourceTypesFilter,
registry,
userSession,
searchlightFacetUtils,
searchlightSearchHelper,
searchSettings,
cache)
{
var ctrl = this;
ctrl.filter = $filter;
ctrl.hits = [];
ctrl.hitsSrc = [];
ctrl.initialized = false;
ctrl.searchFacets = [];
ctrl.excludedTypes = ['OS::Glance::Metadef'];
ctrl.searchSettings = searchSettings;
ctrl.defaultResourceTypes = [];
ctrl.defaultFacets = searchlightFacetUtils.defaultFacets();
ctrl.registry = registry;
ctrl.refresh = searchlightSearchHelper.repeatLastSearchWithLatestSettings;
ctrl.actionResultHandler = actionResultHandler;
ctrl.getSearchlightKey = getSearchlightKey;
ctrl.userSession = {};
var adHocPollInterval = 500;
var adHocPollDuration = 5000;
//ctrl.isNested;
init();
////////////////////////////////
function init() {
ctrl.searchSettings.initScope($scope);
searchlightFacetUtils.initScope($scope);
if (searchlightSearchHelper.lastSearchQueryOptions) {
ctrl.searchFacets = searchlightSearchHelper.lastSearchQueryOptions.searchFacets;
if (searchlightSearchHelper.lastSearchQueryOptions.queryString) {
$timeout(setInput(searchlightSearchHelper.lastSearchQueryOptions.queryString));
}
} else {
ctrl.searchFacets = ctrl.defaultFacets;
}
userSession.get()
.then(function onUserSessionGet(session) {
ctrl.userSession = session;
});
}
function setInput(text) {
return function() {
angular.element('.search-input').val(text);
};
}
/*function isNested (input) {
var result = angular.isArray(input) &&
input.length > 0 &&
angular.isObject(input[0]) &&
Object.keys(input[0]).length > 1;
return result;
}*/
var pluginsUpdatedWatcher = $scope.$on(
ctrl.searchSettings.events.pluginsUpdatedEvent,
pluginsUpdated
);
function pluginsUpdated(event, plugins) {
var pluginToTypesOptions = {
excludedTypes: ctrl.excludedTypes,
flatten: true
};
ctrl.defaultResourceTypes = searchPluginResourceTypesFilter(plugins, pluginToTypesOptions);
ctrl.defaultResourceTypes.forEach(function(type) {
registry.initActions(type, $scope);
});
searchlightFacetUtils.setTypeFacetFromResourceTypes(
ctrl.defaultResourceTypes, ctrl.searchFacets);
searchlightFacetUtils.broadcastFacetsChanged(searchlightSearchHelper.lastSearchQueryOptions);
ctrl.initialized = true;
if (searchlightSearchHelper.lastSearchQueryOptions) {
searchlightSearchHelper.lastSearchQueryOptions.onSearchSuccess = onSearchResult;
searchlightSearchHelper.lastSearchQueryOptions.onSearchError = onSearchResult;
searchlightSearchHelper.repeatLastSearchWithLatestSettings();
} else {
search();
}
}
var fullTextSearchTimeout;
var searchUpdatedWatcher = $scope.$on('serverSearchUpdated', function (event, searchData) {
// Magic search always broadcasts this at startup, so
// we have to not run until we are fully initialized.
if (!ctrl.initialized) {
return;
}
function performSearch() {
fullTextSearchTimeout = null;
search(searchData);
}
if (searchData.queryStringChanged) {
// This keeps the query from being executed too rapidly
// when the user is performing rapid key presses.
if (fullTextSearchTimeout) {
$timeout.cancel(fullTextSearchTimeout);
}
fullTextSearchTimeout = $timeout(
performSearch,
ctrl.searchSettings.settings.fullTextSearch.delayInMS
);
} else if (searchData.magicSearchQueryChanged) {
performSearch();
}
});
var checkFacetsWatcher = $scope.$on('checkFacets', function (event, selectedFacets) {
//Facets are actually DOM elements. This affects the styling.
$timeout(function () {
angular.forEach(selectedFacets, function setIsServerTrue(facet) {
facet.isServer = true;
});
});
});
var searchSettingsUpdatedWatcher = $scope.$on(
ctrl.searchSettings.events.settingsUpdatedEvent,
searchlightSearchHelper.repeatLastSearchWithLatestSettings
);
$scope.$on('$destroy', function cleanupListeners() {
searchlightSearchHelper.stopSearchPolling();
checkFacetsWatcher();
searchUpdatedWatcher();
searchSettingsUpdatedWatcher();
pluginsUpdatedWatcher();
});
function search(queryOptions) {
queryOptions = queryOptions || {};
queryOptions.allFacetDefinitions = ctrl.searchFacets;
queryOptions.searchFacets = ctrl.searchFacets;
queryOptions.defaultResourceTypes = ctrl.defaultResourceTypes;
queryOptions.onSearchSuccess = onSearchResult;
queryOptions.onSearchError = onSearchResult;
return searchlightSearchHelper.search(queryOptions);
}
function onSearchResult(response) {
cache.clean(adHocPollDuration * 3);
ctrl.hitsSrc = response.hits.map(syncWithCache).filter(removeDeletedItems);
ctrl.queryResponse = response;
}
function syncWithCache(searchlight_item) {
return cache.sync(searchlight_item, searchlight_item._id, getSearchlightTimestamp(searchlight_item));
}
function removeDeletedItems(searchlight_item) {
if ( searchlight_item.deleted ) {
return false;
} else {
return true;
}
}
function actionResultHandler(returnValue) {
return $q.when(returnValue, actionSuccessHandler, actionErrorHandler);
}
/*
function repeatUntilChangedResults() {
// For now, all we can do is poll for a period of time.
searchlightSearchHelper.startAdHocPolling(adHocPollInterval, adHocPollDuration);
}
*/
function actionSuccessHandler(result) {
// For now, always poll for 5 seconds after every action. This is not
// needed with default polling enabled.
//repeatUntilChangedResults();
// The action has completed (for whatever "complete" means to that
// action. Notice the view doesn't really need to know the semantics of the
// particular action because the actions return data in a standard form.
// That return includes the id and type of each created, updated, deleted
// and failed item.
//
// This handler is also careful to check the type of each item. This
// is important because actions which create non-images are launched from
// the images page (like create "volume" from image).
var deletedIds, updatedIds, createdIds, failedIds;
if ( result ) {
// Reduce the results to just image ids ignoring other types the action
// may have produced
deletedIds = getIdsOfType(result.deleted, undefined);
updatedIds = getIdsOfType(result.updated, undefined);
createdIds = getIdsOfType(result.created, undefined);
failedIds = getIdsOfType(result.failed, undefined);
addItemsToCache(deletedIds, true);
addItemsToCache(updatedIds);
addItemsToCache(createdIds);
// Handle deleted images
if (deletedIds.length) {
// Do nothing for now
}
// Handle updated and created images
if ( updatedIds.length || createdIds.length ) {
}
// Handle failed images
if ( failedIds ) {
// Do nothing for now
}
} else {
// promise resolved, but no result returned. Because the action didn't
// tell us what happened...reload the displayed items just in case.
}
}
function addItemsToCache(ids, deleted) {
var searchlight_item;
ids.forEach(function addToCache(id) {
var index = ctrl.hitsSrc.findIndex(function findItemWithId(item) {
if (item._source.id === id) {
return item;
}
});
if ( index >= 0 ) {
var searchlight_item = ctrl.hitsSrc[index];
if ( deleted ) |
if ( searchlight_item ) {
searchlight_item.dirty = true;
searchlight_item.deleted = deleted;
cache.add(searchlight_item, searchlight_item._id, getSearchlightTimestamp(searchlight_item));
}
}
});
}
| {
ctrl.hitsSrc.splice(index,1);
} | conditional_block |
search-table.controller.js | the License is distributed on an "AS IS" BASIS, WITHOUT
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
* License for the specific language governing permissions and limitations
* under the License.
*/
(function () {
'use strict';
/**
* @ngdoc controller
* @name SearchTableController
*
* @description
* Controller for the search table.
* Serves as the focal point for table actions.
*/
angular
.module('horizon.dashboard.project.search')
.controller('searchTableController', SearchTableController);
SearchTableController.$inject = [
'$scope',
'$filter',
'$q',
'$timeout',
'searchPluginResourceTypesFilter', | 'horizon.framework.conf.resource-type-registry.service',
'horizon.app.core.openstack-service-api.userSession',
'horizon.dashboard.project.search.searchlightFacetUtils',
'horizon.dashboard.project.search.searchlightSearchHelper',
'horizon.dashboard.project.search.settingsService',
'horizon.dashboard.search.search.util.cache.service'
];
function SearchTableController($scope,
$filter,
$q,
$timeout,
searchPluginResourceTypesFilter,
registry,
userSession,
searchlightFacetUtils,
searchlightSearchHelper,
searchSettings,
cache)
{
var ctrl = this;
ctrl.filter = $filter;
ctrl.hits = [];
ctrl.hitsSrc = [];
ctrl.initialized = false;
ctrl.searchFacets = [];
ctrl.excludedTypes = ['OS::Glance::Metadef'];
ctrl.searchSettings = searchSettings;
ctrl.defaultResourceTypes = [];
ctrl.defaultFacets = searchlightFacetUtils.defaultFacets();
ctrl.registry = registry;
ctrl.refresh = searchlightSearchHelper.repeatLastSearchWithLatestSettings;
ctrl.actionResultHandler = actionResultHandler;
ctrl.getSearchlightKey = getSearchlightKey;
ctrl.userSession = {};
var adHocPollInterval = 500;
var adHocPollDuration = 5000;
//ctrl.isNested;
init();
////////////////////////////////
function init() {
ctrl.searchSettings.initScope($scope);
searchlightFacetUtils.initScope($scope);
if (searchlightSearchHelper.lastSearchQueryOptions) {
ctrl.searchFacets = searchlightSearchHelper.lastSearchQueryOptions.searchFacets;
if (searchlightSearchHelper.lastSearchQueryOptions.queryString) {
$timeout(setInput(searchlightSearchHelper.lastSearchQueryOptions.queryString));
}
} else {
ctrl.searchFacets = ctrl.defaultFacets;
}
userSession.get()
.then(function onUserSessionGet(session) {
ctrl.userSession = session;
});
}
function setInput(text) {
return function() {
angular.element('.search-input').val(text);
};
}
/*function isNested (input) {
var result = angular.isArray(input) &&
input.length > 0 &&
angular.isObject(input[0]) &&
Object.keys(input[0]).length > 1;
return result;
}*/
var pluginsUpdatedWatcher = $scope.$on(
ctrl.searchSettings.events.pluginsUpdatedEvent,
pluginsUpdated
);
function pluginsUpdated(event, plugins) {
var pluginToTypesOptions = {
excludedTypes: ctrl.excludedTypes,
flatten: true
};
ctrl.defaultResourceTypes = searchPluginResourceTypesFilter(plugins, pluginToTypesOptions);
ctrl.defaultResourceTypes.forEach(function(type) {
registry.initActions(type, $scope);
});
searchlightFacetUtils.setTypeFacetFromResourceTypes(
ctrl.defaultResourceTypes, ctrl.searchFacets);
searchlightFacetUtils.broadcastFacetsChanged(searchlightSearchHelper.lastSearchQueryOptions);
ctrl.initialized = true;
if (searchlightSearchHelper.lastSearchQueryOptions) {
searchlightSearchHelper.lastSearchQueryOptions.onSearchSuccess = onSearchResult;
searchlightSearchHelper.lastSearchQueryOptions.onSearchError = onSearchResult;
searchlightSearchHelper.repeatLastSearchWithLatestSettings();
} else {
search();
}
}
var fullTextSearchTimeout;
var searchUpdatedWatcher = $scope.$on('serverSearchUpdated', function (event, searchData) {
// Magic search always broadcasts this at startup, so
// we have to not run until we are fully initialized.
if (!ctrl.initialized) {
return;
}
function performSearch() {
fullTextSearchTimeout = null;
search(searchData);
}
if (searchData.queryStringChanged) {
// This keeps the query from being executed too rapidly
// when the user is performing rapid key presses.
if (fullTextSearchTimeout) {
$timeout.cancel(fullTextSearchTimeout);
}
fullTextSearchTimeout = $timeout(
performSearch,
ctrl.searchSettings.settings.fullTextSearch.delayInMS
);
} else if (searchData.magicSearchQueryChanged) {
performSearch();
}
});
var checkFacetsWatcher = $scope.$on('checkFacets', function (event, selectedFacets) {
//Facets are actually DOM elements. This affects the styling.
$timeout(function () {
angular.forEach(selectedFacets, function setIsServerTrue(facet) {
facet.isServer = true;
});
});
});
var searchSettingsUpdatedWatcher = $scope.$on(
ctrl.searchSettings.events.settingsUpdatedEvent,
searchlightSearchHelper.repeatLastSearchWithLatestSettings
);
$scope.$on('$destroy', function cleanupListeners() {
searchlightSearchHelper.stopSearchPolling();
checkFacetsWatcher();
searchUpdatedWatcher();
searchSettingsUpdatedWatcher();
pluginsUpdatedWatcher();
});
function search(queryOptions) {
queryOptions = queryOptions || {};
queryOptions.allFacetDefinitions = ctrl.searchFacets;
queryOptions.searchFacets = ctrl.searchFacets;
queryOptions.defaultResourceTypes = ctrl.defaultResourceTypes;
queryOptions.onSearchSuccess = onSearchResult;
queryOptions.onSearchError = onSearchResult;
return searchlightSearchHelper.search(queryOptions);
}
function onSearchResult(response) {
cache.clean(adHocPollDuration * 3);
ctrl.hitsSrc = response.hits.map(syncWithCache).filter(removeDeletedItems);
ctrl.queryResponse = response;
}
function syncWithCache(searchlight_item) {
return cache.sync(searchlight_item, searchlight_item._id, getSearchlightTimestamp(searchlight_item));
}
function removeDeletedItems(searchlight_item) {
if ( searchlight_item.deleted ) {
return false;
} else {
return true;
}
}
function actionResultHandler(returnValue) {
return $q.when(returnValue, actionSuccessHandler, actionErrorHandler);
}
/*
function repeatUntilChangedResults() {
// For now, all we can do is poll for a period of time.
searchlightSearchHelper.startAdHocPolling(adHocPollInterval, adHocPollDuration);
}
*/
function actionSuccessHandler(result) {
// For now, always poll for 5 seconds after every action. This is not
// needed with default polling enabled.
//repeatUntilChangedResults();
// The action has completed (for whatever "complete" means to that
// action. Notice the view doesn't really need to know the semantics of the
// particular action because the actions return data in a standard form.
// That return includes the id and type of each created, updated, deleted
// and failed item.
//
// This handler is also careful to check the type of each item. This
// is important because actions which create non-images are launched from
// the images page (like create "volume" from image).
var deletedIds, updatedIds, createdIds, failedIds;
if ( result ) {
// Reduce the results to just image ids ignoring other types the action
// may have produced
deletedIds = getIdsOfType(result.deleted, undefined);
updatedIds = getIdsOfType(result.updated, undefined);
createdIds = getIdsOfType(result.created, undefined);
failedIds = getIdsOfType(result.failed, undefined);
addItemsToCache(deletedIds, true);
addItemsToCache(updatedIds);
addItemsToCache(createdIds);
// Handle deleted images
if (deletedIds.length) {
// Do nothing for now
}
// Handle updated and created images
if ( updatedIds.length || createdIds.length ) {
}
// Handle failed images
if ( failedIds ) {
// Do nothing for now
}
} else {
// promise resolved, but no result returned. Because the action didn't
// tell us what happened...reload the displayed items just in case.
}
}
function addItemsToCache(ids, deleted) {
var searchlight_item;
ids.forEach(function addToCache(id) {
var index = ctrl.hitsSrc.findIndex(function findItemWithId(item) {
if (item._source.id === id) {
return item;
}
});
if ( index >= 0 ) {
var searchlight_item = ctrl.hitsSrc[index];
if ( deleted ) {
ctrl.hitsSrc.splice(index,1);
}
if ( searchlight_item ) {
searchlight_item.dirty = true;
searchlight_item.deleted = deleted;
cache.add(searchlight_item, searchlight_item._id, getSearchlightTimestamp(searchlight_item));
}
}
});
}
| random_line_split |
|
search-table.controller.js | description
* Controller for the search table.
* Serves as the focal point for table actions.
*/
angular
.module('horizon.dashboard.project.search')
.controller('searchTableController', SearchTableController);
SearchTableController.$inject = [
'$scope',
'$filter',
'$q',
'$timeout',
'searchPluginResourceTypesFilter',
'horizon.framework.conf.resource-type-registry.service',
'horizon.app.core.openstack-service-api.userSession',
'horizon.dashboard.project.search.searchlightFacetUtils',
'horizon.dashboard.project.search.searchlightSearchHelper',
'horizon.dashboard.project.search.settingsService',
'horizon.dashboard.search.search.util.cache.service'
];
function SearchTableController($scope,
$filter,
$q,
$timeout,
searchPluginResourceTypesFilter,
registry,
userSession,
searchlightFacetUtils,
searchlightSearchHelper,
searchSettings,
cache)
{
var ctrl = this;
ctrl.filter = $filter;
ctrl.hits = [];
ctrl.hitsSrc = [];
ctrl.initialized = false;
ctrl.searchFacets = [];
ctrl.excludedTypes = ['OS::Glance::Metadef'];
ctrl.searchSettings = searchSettings;
ctrl.defaultResourceTypes = [];
ctrl.defaultFacets = searchlightFacetUtils.defaultFacets();
ctrl.registry = registry;
ctrl.refresh = searchlightSearchHelper.repeatLastSearchWithLatestSettings;
ctrl.actionResultHandler = actionResultHandler;
ctrl.getSearchlightKey = getSearchlightKey;
ctrl.userSession = {};
var adHocPollInterval = 500;
var adHocPollDuration = 5000;
//ctrl.isNested;
init();
////////////////////////////////
function init() {
ctrl.searchSettings.initScope($scope);
searchlightFacetUtils.initScope($scope);
if (searchlightSearchHelper.lastSearchQueryOptions) {
ctrl.searchFacets = searchlightSearchHelper.lastSearchQueryOptions.searchFacets;
if (searchlightSearchHelper.lastSearchQueryOptions.queryString) {
$timeout(setInput(searchlightSearchHelper.lastSearchQueryOptions.queryString));
}
} else {
ctrl.searchFacets = ctrl.defaultFacets;
}
userSession.get()
.then(function onUserSessionGet(session) {
ctrl.userSession = session;
});
}
function setInput(text) {
return function() {
angular.element('.search-input').val(text);
};
}
/*function isNested (input) {
var result = angular.isArray(input) &&
input.length > 0 &&
angular.isObject(input[0]) &&
Object.keys(input[0]).length > 1;
return result;
}*/
var pluginsUpdatedWatcher = $scope.$on(
ctrl.searchSettings.events.pluginsUpdatedEvent,
pluginsUpdated
);
function pluginsUpdated(event, plugins) {
var pluginToTypesOptions = {
excludedTypes: ctrl.excludedTypes,
flatten: true
};
ctrl.defaultResourceTypes = searchPluginResourceTypesFilter(plugins, pluginToTypesOptions);
ctrl.defaultResourceTypes.forEach(function(type) {
registry.initActions(type, $scope);
});
searchlightFacetUtils.setTypeFacetFromResourceTypes(
ctrl.defaultResourceTypes, ctrl.searchFacets);
searchlightFacetUtils.broadcastFacetsChanged(searchlightSearchHelper.lastSearchQueryOptions);
ctrl.initialized = true;
if (searchlightSearchHelper.lastSearchQueryOptions) {
searchlightSearchHelper.lastSearchQueryOptions.onSearchSuccess = onSearchResult;
searchlightSearchHelper.lastSearchQueryOptions.onSearchError = onSearchResult;
searchlightSearchHelper.repeatLastSearchWithLatestSettings();
} else {
search();
}
}
var fullTextSearchTimeout;
var searchUpdatedWatcher = $scope.$on('serverSearchUpdated', function (event, searchData) {
// Magic search always broadcasts this at startup, so
// we have to not run until we are fully initialized.
if (!ctrl.initialized) {
return;
}
function performSearch() {
fullTextSearchTimeout = null;
search(searchData);
}
if (searchData.queryStringChanged) {
// This keeps the query from being executed too rapidly
// when the user is performing rapid key presses.
if (fullTextSearchTimeout) {
$timeout.cancel(fullTextSearchTimeout);
}
fullTextSearchTimeout = $timeout(
performSearch,
ctrl.searchSettings.settings.fullTextSearch.delayInMS
);
} else if (searchData.magicSearchQueryChanged) {
performSearch();
}
});
var checkFacetsWatcher = $scope.$on('checkFacets', function (event, selectedFacets) {
//Facets are actually DOM elements. This affects the styling.
$timeout(function () {
angular.forEach(selectedFacets, function setIsServerTrue(facet) {
facet.isServer = true;
});
});
});
var searchSettingsUpdatedWatcher = $scope.$on(
ctrl.searchSettings.events.settingsUpdatedEvent,
searchlightSearchHelper.repeatLastSearchWithLatestSettings
);
$scope.$on('$destroy', function cleanupListeners() {
searchlightSearchHelper.stopSearchPolling();
checkFacetsWatcher();
searchUpdatedWatcher();
searchSettingsUpdatedWatcher();
pluginsUpdatedWatcher();
});
function search(queryOptions) {
queryOptions = queryOptions || {};
queryOptions.allFacetDefinitions = ctrl.searchFacets;
queryOptions.searchFacets = ctrl.searchFacets;
queryOptions.defaultResourceTypes = ctrl.defaultResourceTypes;
queryOptions.onSearchSuccess = onSearchResult;
queryOptions.onSearchError = onSearchResult;
return searchlightSearchHelper.search(queryOptions);
}
function onSearchResult(response) {
cache.clean(adHocPollDuration * 3);
ctrl.hitsSrc = response.hits.map(syncWithCache).filter(removeDeletedItems);
ctrl.queryResponse = response;
}
function syncWithCache(searchlight_item) {
return cache.sync(searchlight_item, searchlight_item._id, getSearchlightTimestamp(searchlight_item));
}
function removeDeletedItems(searchlight_item) {
if ( searchlight_item.deleted ) {
return false;
} else {
return true;
}
}
function actionResultHandler(returnValue) {
return $q.when(returnValue, actionSuccessHandler, actionErrorHandler);
}
/*
function repeatUntilChangedResults() {
// For now, all we can do is poll for a period of time.
searchlightSearchHelper.startAdHocPolling(adHocPollInterval, adHocPollDuration);
}
*/
function actionSuccessHandler(result) {
// For now, always poll for 5 seconds after every action. This is not
// needed with default polling enabled.
//repeatUntilChangedResults();
// The action has completed (for whatever "complete" means to that
// action. Notice the view doesn't really need to know the semantics of the
// particular action because the actions return data in a standard form.
// That return includes the id and type of each created, updated, deleted
// and failed item.
//
// This handler is also careful to check the type of each item. This
// is important because actions which create non-images are launched from
// the images page (like create "volume" from image).
var deletedIds, updatedIds, createdIds, failedIds;
if ( result ) {
// Reduce the results to just image ids ignoring other types the action
// may have produced
deletedIds = getIdsOfType(result.deleted, undefined);
updatedIds = getIdsOfType(result.updated, undefined);
createdIds = getIdsOfType(result.created, undefined);
failedIds = getIdsOfType(result.failed, undefined);
addItemsToCache(deletedIds, true);
addItemsToCache(updatedIds);
addItemsToCache(createdIds);
// Handle deleted images
if (deletedIds.length) {
// Do nothing for now
}
// Handle updated and created images
if ( updatedIds.length || createdIds.length ) {
}
// Handle failed images
if ( failedIds ) {
// Do nothing for now
}
} else {
// promise resolved, but no result returned. Because the action didn't
// tell us what happened...reload the displayed items just in case.
}
}
function addItemsToCache(ids, deleted) {
var searchlight_item;
ids.forEach(function addToCache(id) {
var index = ctrl.hitsSrc.findIndex(function findItemWithId(item) {
if (item._source.id === id) {
return item;
}
});
if ( index >= 0 ) {
var searchlight_item = ctrl.hitsSrc[index];
if ( deleted ) {
ctrl.hitsSrc.splice(index,1);
}
if ( searchlight_item ) {
searchlight_item.dirty = true;
searchlight_item.deleted = deleted;
cache.add(searchlight_item, searchlight_item._id, getSearchlightTimestamp(searchlight_item));
}
}
});
}
function actionErrorHandler(reason) { // eslint-disable-line no-unused-vars
// Action has failed. Do nothing.
}
function getIdsOfType(items, type) {
var result;
function typeIdReduce(accumulator, item) | {
if (type === undefined || item.type === type) {
accumulator.push(item.id);
}
return accumulator;
} | identifier_body |
|
search-table.controller.js | the License is distributed on an "AS IS" BASIS, WITHOUT
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
* License for the specific language governing permissions and limitations
* under the License.
*/
(function () {
'use strict';
/**
* @ngdoc controller
* @name SearchTableController
*
* @description
* Controller for the search table.
* Serves as the focal point for table actions.
*/
angular
.module('horizon.dashboard.project.search')
.controller('searchTableController', SearchTableController);
SearchTableController.$inject = [
'$scope',
'$filter',
'$q',
'$timeout',
'searchPluginResourceTypesFilter',
'horizon.framework.conf.resource-type-registry.service',
'horizon.app.core.openstack-service-api.userSession',
'horizon.dashboard.project.search.searchlightFacetUtils',
'horizon.dashboard.project.search.searchlightSearchHelper',
'horizon.dashboard.project.search.settingsService',
'horizon.dashboard.search.search.util.cache.service'
];
function SearchTableController($scope,
$filter,
$q,
$timeout,
searchPluginResourceTypesFilter,
registry,
userSession,
searchlightFacetUtils,
searchlightSearchHelper,
searchSettings,
cache)
{
var ctrl = this;
ctrl.filter = $filter;
ctrl.hits = [];
ctrl.hitsSrc = [];
ctrl.initialized = false;
ctrl.searchFacets = [];
ctrl.excludedTypes = ['OS::Glance::Metadef'];
ctrl.searchSettings = searchSettings;
ctrl.defaultResourceTypes = [];
ctrl.defaultFacets = searchlightFacetUtils.defaultFacets();
ctrl.registry = registry;
ctrl.refresh = searchlightSearchHelper.repeatLastSearchWithLatestSettings;
ctrl.actionResultHandler = actionResultHandler;
ctrl.getSearchlightKey = getSearchlightKey;
ctrl.userSession = {};
var adHocPollInterval = 500;
var adHocPollDuration = 5000;
//ctrl.isNested;
init();
////////////////////////////////
function init() {
ctrl.searchSettings.initScope($scope);
searchlightFacetUtils.initScope($scope);
if (searchlightSearchHelper.lastSearchQueryOptions) {
ctrl.searchFacets = searchlightSearchHelper.lastSearchQueryOptions.searchFacets;
if (searchlightSearchHelper.lastSearchQueryOptions.queryString) {
$timeout(setInput(searchlightSearchHelper.lastSearchQueryOptions.queryString));
}
} else {
ctrl.searchFacets = ctrl.defaultFacets;
}
userSession.get()
.then(function onUserSessionGet(session) {
ctrl.userSession = session;
});
}
function setInput(text) {
return function() {
angular.element('.search-input').val(text);
};
}
/*function isNested (input) {
var result = angular.isArray(input) &&
input.length > 0 &&
angular.isObject(input[0]) &&
Object.keys(input[0]).length > 1;
return result;
}*/
var pluginsUpdatedWatcher = $scope.$on(
ctrl.searchSettings.events.pluginsUpdatedEvent,
pluginsUpdated
);
function pluginsUpdated(event, plugins) {
var pluginToTypesOptions = {
excludedTypes: ctrl.excludedTypes,
flatten: true
};
ctrl.defaultResourceTypes = searchPluginResourceTypesFilter(plugins, pluginToTypesOptions);
ctrl.defaultResourceTypes.forEach(function(type) {
registry.initActions(type, $scope);
});
searchlightFacetUtils.setTypeFacetFromResourceTypes(
ctrl.defaultResourceTypes, ctrl.searchFacets);
searchlightFacetUtils.broadcastFacetsChanged(searchlightSearchHelper.lastSearchQueryOptions);
ctrl.initialized = true;
if (searchlightSearchHelper.lastSearchQueryOptions) {
searchlightSearchHelper.lastSearchQueryOptions.onSearchSuccess = onSearchResult;
searchlightSearchHelper.lastSearchQueryOptions.onSearchError = onSearchResult;
searchlightSearchHelper.repeatLastSearchWithLatestSettings();
} else {
search();
}
}
var fullTextSearchTimeout;
var searchUpdatedWatcher = $scope.$on('serverSearchUpdated', function (event, searchData) {
// Magic search always broadcasts this at startup, so
// we have to not run until we are fully initialized.
if (!ctrl.initialized) {
return;
}
function performSearch() {
fullTextSearchTimeout = null;
search(searchData);
}
if (searchData.queryStringChanged) {
// This keeps the query from being executed too rapidly
// when the user is performing rapid key presses.
if (fullTextSearchTimeout) {
$timeout.cancel(fullTextSearchTimeout);
}
fullTextSearchTimeout = $timeout(
performSearch,
ctrl.searchSettings.settings.fullTextSearch.delayInMS
);
} else if (searchData.magicSearchQueryChanged) {
performSearch();
}
});
var checkFacetsWatcher = $scope.$on('checkFacets', function (event, selectedFacets) {
//Facets are actually DOM elements. This affects the styling.
$timeout(function () {
angular.forEach(selectedFacets, function setIsServerTrue(facet) {
facet.isServer = true;
});
});
});
var searchSettingsUpdatedWatcher = $scope.$on(
ctrl.searchSettings.events.settingsUpdatedEvent,
searchlightSearchHelper.repeatLastSearchWithLatestSettings
);
$scope.$on('$destroy', function cleanupListeners() {
searchlightSearchHelper.stopSearchPolling();
checkFacetsWatcher();
searchUpdatedWatcher();
searchSettingsUpdatedWatcher();
pluginsUpdatedWatcher();
});
function search(queryOptions) {
queryOptions = queryOptions || {};
queryOptions.allFacetDefinitions = ctrl.searchFacets;
queryOptions.searchFacets = ctrl.searchFacets;
queryOptions.defaultResourceTypes = ctrl.defaultResourceTypes;
queryOptions.onSearchSuccess = onSearchResult;
queryOptions.onSearchError = onSearchResult;
return searchlightSearchHelper.search(queryOptions);
}
function onSearchResult(response) {
cache.clean(adHocPollDuration * 3);
ctrl.hitsSrc = response.hits.map(syncWithCache).filter(removeDeletedItems);
ctrl.queryResponse = response;
}
function | (searchlight_item) {
return cache.sync(searchlight_item, searchlight_item._id, getSearchlightTimestamp(searchlight_item));
}
function removeDeletedItems(searchlight_item) {
if ( searchlight_item.deleted ) {
return false;
} else {
return true;
}
}
function actionResultHandler(returnValue) {
return $q.when(returnValue, actionSuccessHandler, actionErrorHandler);
}
/*
function repeatUntilChangedResults() {
// For now, all we can do is poll for a period of time.
searchlightSearchHelper.startAdHocPolling(adHocPollInterval, adHocPollDuration);
}
*/
function actionSuccessHandler(result) {
// For now, always poll for 5 seconds after every action. This is not
// needed with default polling enabled.
//repeatUntilChangedResults();
// The action has completed (for whatever "complete" means to that
// action. Notice the view doesn't really need to know the semantics of the
// particular action because the actions return data in a standard form.
// That return includes the id and type of each created, updated, deleted
// and failed item.
//
// This handler is also careful to check the type of each item. This
// is important because actions which create non-images are launched from
// the images page (like create "volume" from image).
var deletedIds, updatedIds, createdIds, failedIds;
if ( result ) {
// Reduce the results to just image ids ignoring other types the action
// may have produced
deletedIds = getIdsOfType(result.deleted, undefined);
updatedIds = getIdsOfType(result.updated, undefined);
createdIds = getIdsOfType(result.created, undefined);
failedIds = getIdsOfType(result.failed, undefined);
addItemsToCache(deletedIds, true);
addItemsToCache(updatedIds);
addItemsToCache(createdIds);
// Handle deleted images
if (deletedIds.length) {
// Do nothing for now
}
// Handle updated and created images
if ( updatedIds.length || createdIds.length ) {
}
// Handle failed images
if ( failedIds ) {
// Do nothing for now
}
} else {
// promise resolved, but no result returned. Because the action didn't
// tell us what happened...reload the displayed items just in case.
}
}
function addItemsToCache(ids, deleted) {
var searchlight_item;
ids.forEach(function addToCache(id) {
var index = ctrl.hitsSrc.findIndex(function findItemWithId(item) {
if (item._source.id === id) {
return item;
}
});
if ( index >= 0 ) {
var searchlight_item = ctrl.hitsSrc[index];
if ( deleted ) {
ctrl.hitsSrc.splice(index,1);
}
if ( searchlight_item ) {
searchlight_item.dirty = true;
searchlight_item.deleted = deleted;
cache.add(searchlight_item, searchlight_item._id, getSearchlightTimestamp(searchlight_item));
}
}
});
}
| syncWithCache | identifier_name |
modal-gallery.component.ts | this.description = description;
this.extUrl = extUrl;
this.selected = selected;
}
}
/**
* Enum `Keyboard` with keys and their relative key codes.
*/
export enum Keyboard {
ESC = 27,
LEFT_ARROW = 37,
RIGHT_ARROW = 39,
UP_ARROW = 38,
DOWN_ARROW = 40
}
/**
* Interface `Description` to change the description, either with a full custom
* description or with a small and simple customization.
*/
export interface Description {
customFullDescription?: string;
imageText?: string;
numberSeparator?: string;
beforeTextDescription?: string;
}
// /**
// * Interface `ButtonsConfig` to show/hide buttons.
// */
// export interface ButtonsConfig {
// download?: boolean;
// extUrl?: boolean;
// close?: boolean;
// }
/**
* Interface `KeyboardConfig` to assign custom keyCodes to ESC, RIGHT and LEFT keyboard's actions.
*/
export interface KeyboardConfig {
esc?: number;
right?: number;
left?: number;
}
/**
* Interface `SlideConfig` to configure sliding features of modal gallery.
*/
export interface SlideConfig {
infinite?: boolean;
}
/**
* Main Component of this library with the modal gallery.
*/
@Component({
selector: 'modal-gallery',
exportAs: 'modalGallery',
styleUrls: ['modal-gallery.scss'],
templateUrl: 'modal-gallery.html'
})
export class AngularModalGalleryComponent implements OnInit, OnDestroy, OnChanges {
/**
* Array or Observable input that represents a list of Images used to show both
* thumbs and the modal gallery.
*/
@Input() modalImages: Observable<Array<Image>> | Array<Image>;
@Input() selectionLimit: number;
/**
* Number to open the modal gallery (passing a value >=0) showing the image with the
* imagePointer's index.
*
* Be careful, because this feature will be probably deprecated/changed in version 4.0.0
*/
@Input() imagePointer: number;
/**
* Boolean required to enable image download with both ctrl+s/cmd+s and download button.
* If you want to show enable button, this is not enough. You have to use also `buttonsConfig`.
*/
@Input() downloadable: boolean = false;
/**
* Description object with the configuration to show image descriptions.
*/
@Input() description: Description;
/**
* Object of type `ButtonsConfig` to show/hide buttons.
* This is used only inside `ngOnInit()` to create `configButtons`
*/
@Input() selectionInfo: SelectionInfo;
@Input() buttonsConfig: ButtonsConfig;
/**
* Object of type `KeyboardConfig` to assign custom keys to ESC, RIGHT and LEFT keyboard's actions.
*/
@Input() keyboardConfig: KeyboardConfig;
/**
* enableCloseOutside's input to enable modal-gallery close's behaviour while clicking
* on the semi-transparent background. Disabled by default.
*/
@Input() enableCloseOutside: boolean = false;
/**
* Object of type `SlideConfig` to configure sliding of modal gallery.
*/
@Input() slideConfig: SlideConfig;
/**
* DEPRECATED
* -----REMOVE THIS IN 4.0.0----- deprecated both showDownloadButton and showExtUrlButton
*/
@Input() showDownloadButton: boolean = false; // deprecated
/**
* DEPRECATED
* -----REMOVE THIS IN 4.0.0----- deprecated both showDownloadButton and showExtUrlButton
*/
@Input() showExtUrlButton: boolean = false; // deprecated
@Output() close: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>();
@Output() show: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>();
@Output() firstImage: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>();
@Output() lastImage: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>();
@Output() hasData: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>();
@Output() selectChanged: EventEmitter<Image> = new EventEmitter<Image>();
selectedImageCount: number = 0;
/**
* Boolean that it is true if the modal gallery is visible
*/
opened: boolean = false;
/**
* Boolean that it is true if an image of the modal gallery is still loading
*/
loading: boolean = false;
/**
* Boolean to open the modal gallery. Closed by default.
*/
showGallery: boolean = false;
/**
* Array of `Image` that represent the model of this library with all images, thumbs and so on.
*/
images: Image[];
/**
* `Image` currently visible.
*/
currentImage: Image;
/**
* Number that represents the index of the current image.
*/
currentImageIndex: number = 0;
/**
* Object of type `ButtonsConfig` used to configure buttons visibility. This is a temporary value
* initialized by the real `buttonsConfig`'s input
*/
configButtons: ButtonsConfig;
/**
* Enum of type `Action` used to pass a click action when you click on the modal image.
* Declared here to be used inside the template.
*/
clickAction: Action = Action.CLICK;
/**
* Boolean that it's true when you are watching the first image (currently visible).
*/
isFirstImage: boolean = false;
/**
* Boolean that it's true when you are watching the last image (currently visible).
*/
isLastImage: boolean = false;
canSelectImage: boolean = false;
/**
* Paging related variables
*/
totalImageCount: number = 0;
pageSize: number = 20;
pageCount: number = 0;
currentPage: number = 0;
/**
* Private SWIPE_ACTION to define all swipe actions used by hammerjs.
*/
private SWIPE_ACTION = {
LEFT: 'swipeleft',
RIGHT: 'swiperight',
UP: 'swipeup',
DOWN: 'swipedown'
};
/**
* When you pass an Observable of `Image`s as `modalImages`, you have to subscribe to that
* Observable. So, to prevent memory leaks, you must store the subscription and call `unsubscribe` in
* OnDestroy.
*/
private subscription: Subscription;
/**
* Listener to catch keyboard's events and call the right method based on the key.
* For instance, pressing esc, this will call `closeGallery(Action.KEYBOARD)` and so on.
* If you passed a valid `keyboardConfig` esc, right and left buttons will be customized based on your data.
* @param e KeyboardEvent caught by the listener.
*/
@HostListener('window:keydown', ['$event'])
onKeyDown(e: KeyboardEvent) {
if (!this.opened) {
return;
}
const esc: number = this.keyboardConfig && this.keyboardConfig.esc ? this.keyboardConfig.esc : Keyboard.ESC;
const right: number = this.keyboardConfig && this.keyboardConfig.right ? this.keyboardConfig.right : Keyboard.RIGHT_ARROW;
const left: number = this.keyboardConfig && this.keyboardConfig.left ? this.keyboardConfig.left : Keyboard.LEFT_ARROW;
switch (e.keyCode) {
case esc:
this.closeGallery(Action.KEYBOARD);
break;
case right:
this.nextImage(Action.KEYBOARD);
break;
case left:
this.prevImage(Action.KEYBOARD);
break;
}
}
/**
* Constructor with the injection of ´KeyboardService´ that initialize some description fields
* based on default values.
*/
constructor(private keyboardService: KeyboardService) {
// if description isn't provided initialize it with a default object
if (!this.description) {
this.description = {
imageText: 'Image ',
numberSeparator: '/',
beforeTextDescription: ' - '
};
}
// if one of the Description fields isn't initialized, provide a default value
this.description.imageText = this.description.imageText || 'Image ';
this.description.numberSeparator = this.description.numberSeparator || '/';
this.description.beforeTextDescription = this.description.beforeTextDescription || ' - ';
}
/**
* This method will initialize the pager when the images are loaded.
*/
initializePager(){
if (this.images.length > 0){
this.totalImageCount = this.images.length;
this.pageCount = this.totalImageCount / this.pageSize;
this.currentPage = 1;
}
}
/**
* Method ´ngOnInit´ to build `configButtons` and to call `initImages()`.
* This is an Angular's lifecycle hook, so its called automatically by Angular itself.
* In particular, it's called only one time!!!
*/
ngOnInit() {
// build configButtons to use it inside upper-buttons
this.configButtons = {
download: this.showDownloadButton || (this.buttonsConfig && this.buttonsConfig.download),
extUrl: this.showExtUrlButton || (this.buttonsConfig && this.buttonsConfig.extUrl),
close: (this.buttonsConfig && this.buttonsConfig.close)
};
// call initImages passing true as parameter, because I want to emit `hasData` event
this.initImages(true);
}
/**
* Method ´ngOnChanges´ to init images preventing errors.
* This | this.width = width;
this.height = height;
this.thumb = thumb; | random_line_split |
|
modal-gallery.component.ts | want to show enable button, this is not enough. You have to use also `buttonsConfig`.
*/
@Input() downloadable: boolean = false;
/**
* Description object with the configuration to show image descriptions.
*/
@Input() description: Description;
/**
* Object of type `ButtonsConfig` to show/hide buttons.
* This is used only inside `ngOnInit()` to create `configButtons`
*/
@Input() selectionInfo: SelectionInfo;
@Input() buttonsConfig: ButtonsConfig;
/**
* Object of type `KeyboardConfig` to assign custom keys to ESC, RIGHT and LEFT keyboard's actions.
*/
@Input() keyboardConfig: KeyboardConfig;
/**
* enableCloseOutside's input to enable modal-gallery close's behaviour while clicking
* on the semi-transparent background. Disabled by default.
*/
@Input() enableCloseOutside: boolean = false;
/**
* Object of type `SlideConfig` to configure sliding of modal gallery.
*/
@Input() slideConfig: SlideConfig;
/**
* DEPRECATED
* -----REMOVE THIS IN 4.0.0----- deprecated both showDownloadButton and showExtUrlButton
*/
@Input() showDownloadButton: boolean = false; // deprecated
/**
* DEPRECATED
* -----REMOVE THIS IN 4.0.0----- deprecated both showDownloadButton and showExtUrlButton
*/
@Input() showExtUrlButton: boolean = false; // deprecated
@Output() close: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>();
@Output() show: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>();
@Output() firstImage: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>();
@Output() lastImage: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>();
@Output() hasData: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>();
@Output() selectChanged: EventEmitter<Image> = new EventEmitter<Image>();
selectedImageCount: number = 0;
/**
* Boolean that it is true if the modal gallery is visible
*/
opened: boolean = false;
/**
* Boolean that it is true if an image of the modal gallery is still loading
*/
loading: boolean = false;
/**
* Boolean to open the modal gallery. Closed by default.
*/
showGallery: boolean = false;
/**
* Array of `Image` that represent the model of this library with all images, thumbs and so on.
*/
images: Image[];
/**
* `Image` currently visible.
*/
currentImage: Image;
/**
* Number that represents the index of the current image.
*/
currentImageIndex: number = 0;
/**
* Object of type `ButtonsConfig` used to configure buttons visibility. This is a temporary value
* initialized by the real `buttonsConfig`'s input
*/
configButtons: ButtonsConfig;
/**
* Enum of type `Action` used to pass a click action when you click on the modal image.
* Declared here to be used inside the template.
*/
clickAction: Action = Action.CLICK;
/**
* Boolean that it's true when you are watching the first image (currently visible).
*/
isFirstImage: boolean = false;
/**
* Boolean that it's true when you are watching the last image (currently visible).
*/
isLastImage: boolean = false;
canSelectImage: boolean = false;
/**
* Paging related variables
*/
totalImageCount: number = 0;
pageSize: number = 20;
pageCount: number = 0;
currentPage: number = 0;
/**
* Private SWIPE_ACTION to define all swipe actions used by hammerjs.
*/
private SWIPE_ACTION = {
LEFT: 'swipeleft',
RIGHT: 'swiperight',
UP: 'swipeup',
DOWN: 'swipedown'
};
/**
* When you pass an Observable of `Image`s as `modalImages`, you have to subscribe to that
* Observable. So, to prevent memory leaks, you must store the subscription and call `unsubscribe` in
* OnDestroy.
*/
private subscription: Subscription;
/**
* Listener to catch keyboard's events and call the right method based on the key.
* For instance, pressing esc, this will call `closeGallery(Action.KEYBOARD)` and so on.
* If you passed a valid `keyboardConfig` esc, right and left buttons will be customized based on your data.
* @param e KeyboardEvent caught by the listener.
*/
@HostListener('window:keydown', ['$event'])
onKeyDown(e: KeyboardEvent) {
if (!this.opened) |
const esc: number = this.keyboardConfig && this.keyboardConfig.esc ? this.keyboardConfig.esc : Keyboard.ESC;
const right: number = this.keyboardConfig && this.keyboardConfig.right ? this.keyboardConfig.right : Keyboard.RIGHT_ARROW;
const left: number = this.keyboardConfig && this.keyboardConfig.left ? this.keyboardConfig.left : Keyboard.LEFT_ARROW;
switch (e.keyCode) {
case esc:
this.closeGallery(Action.KEYBOARD);
break;
case right:
this.nextImage(Action.KEYBOARD);
break;
case left:
this.prevImage(Action.KEYBOARD);
break;
}
}
/**
* Constructor with the injection of ´KeyboardService´ that initialize some description fields
* based on default values.
*/
constructor(private keyboardService: KeyboardService) {
// if description isn't provided initialize it with a default object
if (!this.description) {
this.description = {
imageText: 'Image ',
numberSeparator: '/',
beforeTextDescription: ' - '
};
}
// if one of the Description fields isn't initialized, provide a default value
this.description.imageText = this.description.imageText || 'Image ';
this.description.numberSeparator = this.description.numberSeparator || '/';
this.description.beforeTextDescription = this.description.beforeTextDescription || ' - ';
}
/**
* This method will initialize the pager when the images are loaded.
*/
initializePager(){
if (this.images.length > 0){
this.totalImageCount = this.images.length;
this.pageCount = this.totalImageCount / this.pageSize;
this.currentPage = 1;
}
}
/**
* Method ´ngOnInit´ to build `configButtons` and to call `initImages()`.
* This is an Angular's lifecycle hook, so its called automatically by Angular itself.
* In particular, it's called only one time!!!
*/
ngOnInit() {
// build configButtons to use it inside upper-buttons
this.configButtons = {
download: this.showDownloadButton || (this.buttonsConfig && this.buttonsConfig.download),
extUrl: this.showExtUrlButton || (this.buttonsConfig && this.buttonsConfig.extUrl),
close: (this.buttonsConfig && this.buttonsConfig.close)
};
// call initImages passing true as parameter, because I want to emit `hasData` event
this.initImages(true);
}
/**
* Method ´ngOnChanges´ to init images preventing errors.
* This is an Angular's lifecycle hook, so its called automatically by Angular itself.
* In particular, it's called before `ngOnInit()` and whenever one or more data-bound input properties change.
* @param changes `SimpleChanges` object of current and previous property values provided by Angular.
*/
ngOnChanges(changes: SimpleChanges) {
// to prevent errors when you pass to this library
// the array of images inside a subscribe block, in this way: `...subscribe(val => { this.images = arrayOfImages })`
// As you can see, I'm providing examples in these situations in all official demos
if (this.modalImages) {
// I pass `false` as parameter, because I DON'T want to emit `hasData`
// event (preventing multiple hasData events while initializing)
this.initImages(false);
}
}
getImageCountsToDisplay(){
var selectedImages = this.images.filter(image=>{
return image.selected === true;
});
var selectedImageCount = selectedImages.length;
var tobeselected = this.selectionLimit - selectedImageCount;
this.canSelectImage = tobeselected <= 0 && !this.currentImage.selected;
return "You need to select " + tobeselected + " images."
}
/**
* Method `getDescriptionToDisplay` to get the image description based on input params.
* If you provide a full description this will be the visible description, otherwise,
* it will be built using the `description` object, concatenating its fields.
* @returns String description to display.
*/
getDescriptionToDisplay() {
if (this.description && this.description.customFullDescription) {
return this.description.customFullDescription;
}
// If the current image hasn't a description,
// prevent to write the ' - ' (or this.description.beforeTextDescription)
if (!this.currentImage.description || this.currentImage.description === '') {
return `${this.description.imageText}${this.currentImageIndex + 1}${this.description.numberSeparator}${this.images.length}`;
}
return `${this.description.imageText}${this.currentImageIndex + 1}${this.description.numberSeparator}${this.images.length}${this.description.beforeTextDescription}${this.currentImage.description}`;
}
/**
* Method `swipe | {
return;
} | conditional_block |
modal-gallery.component.ts | want to show enable button, this is not enough. You have to use also `buttonsConfig`.
*/
@Input() downloadable: boolean = false;
/**
* Description object with the configuration to show image descriptions.
*/
@Input() description: Description;
/**
* Object of type `ButtonsConfig` to show/hide buttons.
* This is used only inside `ngOnInit()` to create `configButtons`
*/
@Input() selectionInfo: SelectionInfo;
@Input() buttonsConfig: ButtonsConfig;
/**
* Object of type `KeyboardConfig` to assign custom keys to ESC, RIGHT and LEFT keyboard's actions.
*/
@Input() keyboardConfig: KeyboardConfig;
/**
* enableCloseOutside's input to enable modal-gallery close's behaviour while clicking
* on the semi-transparent background. Disabled by default.
*/
@Input() enableCloseOutside: boolean = false;
/**
* Object of type `SlideConfig` to configure sliding of modal gallery.
*/
@Input() slideConfig: SlideConfig;
/**
* DEPRECATED
* -----REMOVE THIS IN 4.0.0----- deprecated both showDownloadButton and showExtUrlButton
*/
@Input() showDownloadButton: boolean = false; // deprecated
/**
* DEPRECATED
* -----REMOVE THIS IN 4.0.0----- deprecated both showDownloadButton and showExtUrlButton
*/
@Input() showExtUrlButton: boolean = false; // deprecated
@Output() close: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>();
@Output() show: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>();
@Output() firstImage: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>();
@Output() lastImage: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>();
@Output() hasData: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>();
@Output() selectChanged: EventEmitter<Image> = new EventEmitter<Image>();
selectedImageCount: number = 0;
/**
* Boolean that it is true if the modal gallery is visible
*/
opened: boolean = false;
/**
* Boolean that it is true if an image of the modal gallery is still loading
*/
loading: boolean = false;
/**
* Boolean to open the modal gallery. Closed by default.
*/
showGallery: boolean = false;
/**
* Array of `Image` that represent the model of this library with all images, thumbs and so on.
*/
images: Image[];
/**
* `Image` currently visible.
*/
currentImage: Image;
/**
* Number that represents the index of the current image.
*/
currentImageIndex: number = 0;
/**
* Object of type `ButtonsConfig` used to configure buttons visibility. This is a temporary value
* initialized by the real `buttonsConfig`'s input
*/
configButtons: ButtonsConfig;
/**
* Enum of type `Action` used to pass a click action when you click on the modal image.
* Declared here to be used inside the template.
*/
clickAction: Action = Action.CLICK;
/**
* Boolean that it's true when you are watching the first image (currently visible).
*/
isFirstImage: boolean = false;
/**
* Boolean that it's true when you are watching the last image (currently visible).
*/
isLastImage: boolean = false;
canSelectImage: boolean = false;
/**
* Paging related variables
*/
totalImageCount: number = 0;
pageSize: number = 20;
pageCount: number = 0;
currentPage: number = 0;
/**
* Private SWIPE_ACTION to define all swipe actions used by hammerjs.
*/
private SWIPE_ACTION = {
LEFT: 'swipeleft',
RIGHT: 'swiperight',
UP: 'swipeup',
DOWN: 'swipedown'
};
/**
* When you pass an Observable of `Image`s as `modalImages`, you have to subscribe to that
* Observable. So, to prevent memory leaks, you must store the subscription and call `unsubscribe` in
* OnDestroy.
*/
private subscription: Subscription;
/**
* Listener to catch keyboard's events and call the right method based on the key.
* For instance, pressing esc, this will call `closeGallery(Action.KEYBOARD)` and so on.
* If you passed a valid `keyboardConfig` esc, right and left buttons will be customized based on your data.
* @param e KeyboardEvent caught by the listener.
*/
@HostListener('window:keydown', ['$event'])
onKeyDown(e: KeyboardEvent) {
if (!this.opened) {
return;
}
const esc: number = this.keyboardConfig && this.keyboardConfig.esc ? this.keyboardConfig.esc : Keyboard.ESC;
const right: number = this.keyboardConfig && this.keyboardConfig.right ? this.keyboardConfig.right : Keyboard.RIGHT_ARROW;
const left: number = this.keyboardConfig && this.keyboardConfig.left ? this.keyboardConfig.left : Keyboard.LEFT_ARROW;
switch (e.keyCode) {
case esc:
this.closeGallery(Action.KEYBOARD);
break;
case right:
this.nextImage(Action.KEYBOARD);
break;
case left:
this.prevImage(Action.KEYBOARD);
break;
}
}
/**
* Constructor with the injection of ´KeyboardService´ that initialize some description fields
* based on default values.
*/
constructor(private keyboardService: KeyboardService) {
// if description isn't provided initialize it with a default object
if (!this.description) {
this.description = {
imageText: 'Image ',
numberSeparator: '/',
beforeTextDescription: ' - '
};
}
// if one of the Description fields isn't initialized, provide a default value
this.description.imageText = this.description.imageText || 'Image ';
this.description.numberSeparator = this.description.numberSeparator || '/';
this.description.beforeTextDescription = this.description.beforeTextDescription || ' - ';
}
/**
* This method will initialize the pager when the images are loaded.
*/
initializePager(){
if (this.images.length > 0){
this.totalImageCount = this.images.length;
this.pageCount = this.totalImageCount / this.pageSize;
this.currentPage = 1;
}
}
/**
* Method ´ngOnInit´ to build `configButtons` and to call `initImages()`.
* This is an Angular's lifecycle hook, so its called automatically by Angular itself.
* In particular, it's called only one time!!!
*/
ngOnInit() {
// build configButtons to use it inside upper-buttons
this.configButtons = {
download: this.showDownloadButton || (this.buttonsConfig && this.buttonsConfig.download),
extUrl: this.showExtUrlButton || (this.buttonsConfig && this.buttonsConfig.extUrl),
close: (this.buttonsConfig && this.buttonsConfig.close)
};
// call initImages passing true as parameter, because I want to emit `hasData` event
this.initImages(true);
}
/**
* Method ´ngOnChanges´ to init images preventing errors.
* This is an Angular's lifecycle hook, so its called automatically by Angular itself.
* In particular, it's called before `ngOnInit()` and whenever one or more data-bound input properties change.
* @param changes `SimpleChanges` object of current and previous property values provided by Angular.
*/
ngOnChanges(changes: SimpleChanges) {
| tImageCountsToDisplay(){
var selectedImages = this.images.filter(image=>{
return image.selected === true;
});
var selectedImageCount = selectedImages.length;
var tobeselected = this.selectionLimit - selectedImageCount;
this.canSelectImage = tobeselected <= 0 && !this.currentImage.selected;
return "You need to select " + tobeselected + " images."
}
/**
* Method `getDescriptionToDisplay` to get the image description based on input params.
* If you provide a full description this will be the visible description, otherwise,
* it will be built using the `description` object, concatenating its fields.
* @returns String description to display.
*/
getDescriptionToDisplay() {
if (this.description && this.description.customFullDescription) {
return this.description.customFullDescription;
}
// If the current image hasn't a description,
// prevent to write the ' - ' (or this.description.beforeTextDescription)
if (!this.currentImage.description || this.currentImage.description === '') {
return `${this.description.imageText}${this.currentImageIndex + 1}${this.description.numberSeparator}${this.images.length}`;
}
return `${this.description.imageText}${this.currentImageIndex + 1}${this.description.numberSeparator}${this.images.length}${this.description.beforeTextDescription}${this.currentImage.description}`;
}
/**
* Method `swipe | // to prevent errors when you pass to this library
// the array of images inside a subscribe block, in this way: `...subscribe(val => { this.images = arrayOfImages })`
// As you can see, I'm providing examples in these situations in all official demos
if (this.modalImages) {
// I pass `false` as parameter, because I DON'T want to emit `hasData`
// event (preventing multiple hasData events while initializing)
this.initImages(false);
}
}
ge | identifier_body |
modal-gallery.component.ts | of `Image`s as `modalImages`, you have to subscribe to that
* Observable. So, to prevent memory leaks, you must store the subscription and call `unsubscribe` in
* OnDestroy.
*/
private subscription: Subscription;
/**
* Listener to catch keyboard's events and call the right method based on the key.
* For instance, pressing esc, this will call `closeGallery(Action.KEYBOARD)` and so on.
* If you passed a valid `keyboardConfig` esc, right and left buttons will be customized based on your data.
* @param e KeyboardEvent caught by the listener.
*/
@HostListener('window:keydown', ['$event'])
onKeyDown(e: KeyboardEvent) {
if (!this.opened) {
return;
}
const esc: number = this.keyboardConfig && this.keyboardConfig.esc ? this.keyboardConfig.esc : Keyboard.ESC;
const right: number = this.keyboardConfig && this.keyboardConfig.right ? this.keyboardConfig.right : Keyboard.RIGHT_ARROW;
const left: number = this.keyboardConfig && this.keyboardConfig.left ? this.keyboardConfig.left : Keyboard.LEFT_ARROW;
switch (e.keyCode) {
case esc:
this.closeGallery(Action.KEYBOARD);
break;
case right:
this.nextImage(Action.KEYBOARD);
break;
case left:
this.prevImage(Action.KEYBOARD);
break;
}
}
/**
* Constructor with the injection of ´KeyboardService´ that initialize some description fields
* based on default values.
*/
constructor(private keyboardService: KeyboardService) {
// if description isn't provided initialize it with a default object
if (!this.description) {
this.description = {
imageText: 'Image ',
numberSeparator: '/',
beforeTextDescription: ' - '
};
}
// if one of the Description fields isn't initialized, provide a default value
this.description.imageText = this.description.imageText || 'Image ';
this.description.numberSeparator = this.description.numberSeparator || '/';
this.description.beforeTextDescription = this.description.beforeTextDescription || ' - ';
}
/**
* This method will initialize the pager when the images are loaded.
*/
initializePager(){
if (this.images.length > 0){
this.totalImageCount = this.images.length;
this.pageCount = this.totalImageCount / this.pageSize;
this.currentPage = 1;
}
}
/**
* Method ´ngOnInit´ to build `configButtons` and to call `initImages()`.
* This is an Angular's lifecycle hook, so its called automatically by Angular itself.
* In particular, it's called only one time!!!
*/
ngOnInit() {
// build configButtons to use it inside upper-buttons
this.configButtons = {
download: this.showDownloadButton || (this.buttonsConfig && this.buttonsConfig.download),
extUrl: this.showExtUrlButton || (this.buttonsConfig && this.buttonsConfig.extUrl),
close: (this.buttonsConfig && this.buttonsConfig.close)
};
// call initImages passing true as parameter, because I want to emit `hasData` event
this.initImages(true);
}
/**
* Method ´ngOnChanges´ to init images preventing errors.
* This is an Angular's lifecycle hook, so its called automatically by Angular itself.
* In particular, it's called before `ngOnInit()` and whenever one or more data-bound input properties change.
* @param changes `SimpleChanges` object of current and previous property values provided by Angular.
*/
ngOnChanges(changes: SimpleChanges) {
// to prevent errors when you pass to this library
// the array of images inside a subscribe block, in this way: `...subscribe(val => { this.images = arrayOfImages })`
// As you can see, I'm providing examples in these situations in all official demos
if (this.modalImages) {
// I pass `false` as parameter, because I DON'T want to emit `hasData`
// event (preventing multiple hasData events while initializing)
this.initImages(false);
}
}
getImageCountsToDisplay(){
var selectedImages = this.images.filter(image=>{
return image.selected === true;
});
var selectedImageCount = selectedImages.length;
var tobeselected = this.selectionLimit - selectedImageCount;
this.canSelectImage = tobeselected <= 0 && !this.currentImage.selected;
return "You need to select " + tobeselected + " images."
}
/**
* Method `getDescriptionToDisplay` to get the image description based on input params.
* If you provide a full description this will be the visible description, otherwise,
* it will be built using the `description` object, concatenating its fields.
* @returns String description to display.
*/
getDescriptionToDisplay() {
if (this.description && this.description.customFullDescription) {
return this.description.customFullDescription;
}
// If the current image hasn't a description,
// prevent to write the ' - ' (or this.description.beforeTextDescription)
if (!this.currentImage.description || this.currentImage.description === '') {
return `${this.description.imageText}${this.currentImageIndex + 1}${this.description.numberSeparator}${this.images.length}`;
}
return `${this.description.imageText}${this.currentImageIndex + 1}${this.description.numberSeparator}${this.images.length}${this.description.beforeTextDescription}${this.currentImage.description}`;
}
/**
* Method `swipe` used by Hammerjs to support touch gestures.
* @param index Number that represent the current visible index
* @param action String that represent the direction of the swipe action. 'swiperight' by default.
*/
swipe(index: number, action = this.SWIPE_ACTION.RIGHT) {
switch (action) {
case this.SWIPE_ACTION.RIGHT:
this.nextImage(Action.SWIPE);
break;
case this.SWIPE_ACTION.LEFT:
this.prevImage(Action.SWIPE);
break;
// case this.SWIPE_ACTION.UP:
// break;
// case this.SWIPE_ACTION.DOWN:
// break;
}
}
/**
* Method `closeGallery` to close the modal gallery.
* @param action Enum of type `Action` that represents the source
* action that closed the modal gallery. NORMAL by default.
*/
closeGallery(action: Action = Action.NORMAL) {
this.close.emit(new ImageModalEvent(action, true));
this.opened = false;
this.keyboardService.reset();
}
imageSelectionChangedComponent(image: any){
this.selectChanged.emit(image);
}
/**
* Method `prevImage` to go back to the previous image shown into the modal gallery.
* @param action Enum of type `Action` that represents the source
* action that moved back to the previous image. NORMAL by default.
*/
prevImage(action: Action = Action.NORMAL) {
// check if prevImage should be blocked
if (this.isPreventSliding(0)) {
return;
}
this.loading = true;
this.currentImageIndex = this.getPrevIndex(action, this.currentImageIndex);
this.showModalGallery(this.currentImageIndex);
}
/**
* Method `nextImage` to go back to the previous image shown into the modal gallery.
* @param action Enum of type `Action` that represents the source
* action that moved to the next image. NORMAL by default.
*/
nextImage(action: Action = Action.NORMAL) {
// check if nextImage should be blocked
if (this.isPreventSliding(this.images.length - 1)) {
return;
}
this.loading = true;
this.currentImageIndex = this.getNextIndex(action, this.currentImageIndex);
this.showModalGallery(this.currentImageIndex);
}
/**
* Method `onShowModalGallery` called when you click on an image of your gallery.
* The input index is the index of the clicked image thumb.
* @param index Number that represents the index of the clicked image.
*/
onShowModalGallery(index: number) {
this.showModalGallery(index);
}
/**
* Method `showModalGallery` to show the modal gallery displaying the image with
* the index specified as input parameter.
* It will also register a new `keyboardService` to catch keyboard's events to download the current
* image with keyboard's shortcuts. This service, will be removed when modal gallery component will be destroyed.
* @param index Number that represents the index of the image to show.
*/
showModalGallery(index: number) {
this.keyboardService.add((event: KeyboardEvent, combo: string) => {
if (event.preventDefault) {
event.preventDefault();
} else {
// internet explorer
event.returnValue = false;
}
this.downloadImage();
});
// enable/disable 'infinite sliding' based on @Input() slideConfig
this.manageSlideConfig(index);
this.currentImageIndex = index;
this.opened = true;
this.currentImage = this.images[this.currentImageIndex];
this.loading = false;
// emit current visible image index
this.show.emit(new ImageModalEvent(Action.LOAD, this.currentImageIndex + 1));
}
/**
* Method `downloadImage` to download the current visible image, only if `downloadable` is true.
* For IE, this will navigate to the image instead of a direct download as in all modern browsers.
*/
downlo | adImage() {
| identifier_name |
|
lib.rs | (jit: bool) -> Box<dyn TargetIsa + 'static> {
let mut flags_builder = cranelift::codegen::settings::builder();
// `simplejit` requires non-PIC code
if !jit {
// allow creating shared libraries
flags_builder
.enable("is_pic")
.expect("is_pic should be a valid option");
}
// use debug assertions
flags_builder
.enable("enable_verifier")
.expect("enable_verifier should be a valid option");
// don't emit call to __cranelift_probestack
flags_builder
.set("enable_probestack", "false")
.expect("enable_probestack should be a valid option");
let flags = Flags::new(flags_builder);
cranelift::codegen::isa::lookup(TARGET)
.unwrap_or_else(|_| panic!("platform not supported: {}", TARGET))
.finish(flags)
}
pub fn initialize_aot_module(name: String) -> Module<ObjectBackend> {
let builder = ObjectBuilder::new(
get_isa(false),
name,
cranelift_module::default_libcall_names(),
);
Module::new(builder.expect("unsupported binary format or target architecture"))
}
enum Id {
Function(FuncId),
Global(DataId),
Local(StackSlot),
}
struct Compiler<T: Backend> {
module: Module<T>,
debug: bool,
// if false, we last saw a switch
last_saw_loop: bool,
strings: HashMap<Vec<u8>, DataId>,
declarations: HashMap<Symbol, Id>,
loops: Vec<(Block, Block)>,
// switch, default, end
// if default is empty once we get to the end of a switch body,
// we didn't see a default case
switches: Vec<(Switch, Option<Block>, Block)>,
labels: HashMap<InternedStr, Block>,
error_handler: ErrorHandler,
}
impl<B: Backend> Compiler<B> {
fn new(module: Module<B>, debug: bool) -> Compiler<B> {
Compiler {
module,
declarations: HashMap::new(),
loops: Vec::new(),
switches: Vec::new(),
labels: HashMap::new(),
// the initial value doesn't really matter
last_saw_loop: true,
strings: Default::default(),
error_handler: Default::default(),
debug,
}
}
// we have to consider the following cases:
// 1. declaration before definition
// 2. 2nd declaration before definition
// 3. definition
// 4. declaration after definition
// 1. should declare `id` a import unless specified as `static`.
// 3. should always declare `id` as export or local.
// 2. and 4. should be a no-op.
fn declare_func(&mut self, symbol: Symbol, is_definition: bool) -> CompileResult<FuncId> {
use saltwater_parser::get_str;
if !is_definition {
// case 2 and 4
if let Some(Id::Function(func_id)) = self.declarations.get(&symbol) {
return Ok(*func_id);
}
}
let metadata = symbol.get();
let func_type = match &metadata.ctype {
Type::Function(func_type) => func_type,
_ => unreachable!("bug in backend: only functions should be passed to `declare_func`"),
};
let signature = func_type.signature(self.module.isa());
let linkage = match metadata.storage_class {
StorageClass::Auto | StorageClass::Extern if is_definition => Linkage::Export,
StorageClass::Auto | StorageClass::Extern => Linkage::Import,
StorageClass::Static => Linkage::Local,
StorageClass::Register | StorageClass::Typedef => unreachable!(),
};
let func_id = self
.module
.declare_function(get_str!(metadata.id), linkage, &signature)
.unwrap_or_else(|err| panic!("{}", err));
self.declarations.insert(symbol, Id::Function(func_id));
Ok(func_id)
}
/// declare an object on the stack
fn declare_stack(
&mut self,
decl: Declaration,
location: Location,
builder: &mut FunctionBuilder,
) -> CompileResult<()> {
let meta = decl.symbol.get();
if let StorageClass::Typedef = meta.storage_class {
return Ok(());
}
if let Type::Function(_) = &meta.ctype {
self.declare_func(decl.symbol, false)?;
return Ok(());
}
let u64_size = match meta.ctype.sizeof() {
Ok(size) => size,
Err(err) => {
return Err(CompileError::semantic(Locatable {
data: err.into(),
location,
}))
}
};
let kind = StackSlotKind::ExplicitSlot;
let size = match u32::try_from(u64_size) {
Ok(size) => size,
Err(_) => return Err(CompileError::semantic(Locatable {
data: "cannot store items on the stack that are more than 4 GB, it will overflow the stack".into(),
location,
}))
};
let data = StackSlotData {
kind,
size,
offset: None,
};
let stack_slot = builder.create_stack_slot(data);
self.declarations.insert(decl.symbol, Id::Local(stack_slot));
if let Some(init) = decl.init {
self.store_stack(init, stack_slot, builder)?;
}
Ok(())
}
fn store_stack(
&mut self,
init: Initializer,
stack_slot: StackSlot,
builder: &mut FunctionBuilder,
) -> CompileResult<()> {
match init {
Initializer::Scalar(expr) => {
let val = self.compile_expr(*expr, builder)?;
// TODO: replace with `builder.ins().stack_store(val.ir_val, stack_slot, 0);`
// when Cranelift implements stack_store for i8 and i16
let addr = builder.ins().stack_addr(Type::ptr_type(), stack_slot, 0);
builder.ins().store(MemFlags::new(), val.ir_val, addr, 0);
}
Initializer::InitializerList(_) => unimplemented!("aggregate dynamic initialization"),
Initializer::FunctionBody(_) => unreachable!("functions can't be stored on the stack"),
}
Ok(())
}
// TODO: this is grossly inefficient, ask Cranelift devs if
// there's an easier way to make parameters modifiable.
fn store_stack_params(
&mut self,
params: &[Symbol],
func_start: Block,
location: &Location,
builder: &mut FunctionBuilder,
) -> CompileResult<()> {
// Cranelift requires that all block params are declared up front
let ir_vals: Vec<_> = params
.iter()
.map(|param| {
let ir_type = param.get().ctype.as_ir_type();
Ok(builder.append_block_param(func_start, ir_type))
})
.collect::<CompileResult<_>>()?;
for (¶m, ir_val) in params.iter().zip(ir_vals) {
let u64_size = match param.get().ctype.sizeof() {
Err(data) => semantic_err!(data.into(), *location),
Ok(size) => size,
};
let u32_size = match u32::try_from(u64_size) {
Err(_) => semantic_err!(
format!(
"size {} is too large for stack (can only handle 32-bit values)",
u64_size
),
*location
),
Ok(size) => size,
};
let stack_data = StackSlotData {
kind: StackSlotKind::ExplicitSlot,
size: u32_size,
offset: None,
};
let slot = builder.create_stack_slot(stack_data);
// TODO: need to take the address before storing until Cranelift implements
// stores for i8 and i16
// then this can be replaced with `builder.ins().stack_store(ir_val, slot, 0);`
// See https://github.com/CraneStation/cranelift/issues/433
let addr = builder.ins().stack_addr(Type::ptr_type(), slot, 0);
builder.ins().store(MemFlags::new(), ir_val, addr, 0);
self.declarations.insert(param, Id::Local(slot));
}
Ok(())
}
fn compile_func(
&mut self,
symbol: Symbol,
func_type: &FunctionType,
stmts: Vec<Stmt>,
location: Location,
) -> CompileResult<()> {
let func_id = self.declare_func(symbol, true)?;
// TODO: make declare_func should take a `signature` after all?
// This just calculates it twice, it's probably fine
let signature = func_type.signature(self.module.isa());
// external name is meant to be a lookup in a symbol table,
// but we just give it garbage values
let mut func = Function::with_name_signature(ExternalName::user(0, 0), signature);
// this context is just boiler plate
let mut ctx = FunctionBuilderContext::new();
let mut builder = FunctionBuilder::new(&mut func, &mut ctx);
let func_start = builder.create_block();
builder.switch_to_block(func_start);
let should_ret = func_type.should_return();
if | get_isa | identifier_name |
|
lib.rs | : true,
strings: Default::default(),
error_handler: Default::default(),
debug,
}
}
// we have to consider the following cases:
// 1. declaration before definition
// 2. 2nd declaration before definition
// 3. definition
// 4. declaration after definition
// 1. should declare `id` a import unless specified as `static`.
// 3. should always declare `id` as export or local.
// 2. and 4. should be a no-op.
fn declare_func(&mut self, symbol: Symbol, is_definition: bool) -> CompileResult<FuncId> {
use saltwater_parser::get_str;
if !is_definition {
// case 2 and 4
if let Some(Id::Function(func_id)) = self.declarations.get(&symbol) {
return Ok(*func_id);
}
}
let metadata = symbol.get();
let func_type = match &metadata.ctype {
Type::Function(func_type) => func_type,
_ => unreachable!("bug in backend: only functions should be passed to `declare_func`"),
};
let signature = func_type.signature(self.module.isa());
let linkage = match metadata.storage_class {
StorageClass::Auto | StorageClass::Extern if is_definition => Linkage::Export,
StorageClass::Auto | StorageClass::Extern => Linkage::Import,
StorageClass::Static => Linkage::Local,
StorageClass::Register | StorageClass::Typedef => unreachable!(),
};
let func_id = self
.module
.declare_function(get_str!(metadata.id), linkage, &signature)
.unwrap_or_else(|err| panic!("{}", err));
self.declarations.insert(symbol, Id::Function(func_id));
Ok(func_id)
}
/// declare an object on the stack
fn declare_stack(
&mut self,
decl: Declaration,
location: Location,
builder: &mut FunctionBuilder,
) -> CompileResult<()> {
let meta = decl.symbol.get();
if let StorageClass::Typedef = meta.storage_class {
return Ok(());
}
if let Type::Function(_) = &meta.ctype {
self.declare_func(decl.symbol, false)?;
return Ok(());
}
let u64_size = match meta.ctype.sizeof() {
Ok(size) => size,
Err(err) => {
return Err(CompileError::semantic(Locatable {
data: err.into(),
location,
}))
}
};
let kind = StackSlotKind::ExplicitSlot;
let size = match u32::try_from(u64_size) {
Ok(size) => size,
Err(_) => return Err(CompileError::semantic(Locatable {
data: "cannot store items on the stack that are more than 4 GB, it will overflow the stack".into(),
location,
}))
};
let data = StackSlotData {
kind,
size,
offset: None,
};
let stack_slot = builder.create_stack_slot(data);
self.declarations.insert(decl.symbol, Id::Local(stack_slot));
if let Some(init) = decl.init {
self.store_stack(init, stack_slot, builder)?;
}
Ok(())
}
fn store_stack(
&mut self,
init: Initializer,
stack_slot: StackSlot,
builder: &mut FunctionBuilder,
) -> CompileResult<()> |
// TODO: this is grossly inefficient, ask Cranelift devs if
// there's an easier way to make parameters modifiable.
fn store_stack_params(
&mut self,
params: &[Symbol],
func_start: Block,
location: &Location,
builder: &mut FunctionBuilder,
) -> CompileResult<()> {
// Cranelift requires that all block params are declared up front
let ir_vals: Vec<_> = params
.iter()
.map(|param| {
let ir_type = param.get().ctype.as_ir_type();
Ok(builder.append_block_param(func_start, ir_type))
})
.collect::<CompileResult<_>>()?;
for (¶m, ir_val) in params.iter().zip(ir_vals) {
let u64_size = match param.get().ctype.sizeof() {
Err(data) => semantic_err!(data.into(), *location),
Ok(size) => size,
};
let u32_size = match u32::try_from(u64_size) {
Err(_) => semantic_err!(
format!(
"size {} is too large for stack (can only handle 32-bit values)",
u64_size
),
*location
),
Ok(size) => size,
};
let stack_data = StackSlotData {
kind: StackSlotKind::ExplicitSlot,
size: u32_size,
offset: None,
};
let slot = builder.create_stack_slot(stack_data);
// TODO: need to take the address before storing until Cranelift implements
// stores for i8 and i16
// then this can be replaced with `builder.ins().stack_store(ir_val, slot, 0);`
// See https://github.com/CraneStation/cranelift/issues/433
let addr = builder.ins().stack_addr(Type::ptr_type(), slot, 0);
builder.ins().store(MemFlags::new(), ir_val, addr, 0);
self.declarations.insert(param, Id::Local(slot));
}
Ok(())
}
fn compile_func(
&mut self,
symbol: Symbol,
func_type: &FunctionType,
stmts: Vec<Stmt>,
location: Location,
) -> CompileResult<()> {
let func_id = self.declare_func(symbol, true)?;
// TODO: make declare_func should take a `signature` after all?
// This just calculates it twice, it's probably fine
let signature = func_type.signature(self.module.isa());
// external name is meant to be a lookup in a symbol table,
// but we just give it garbage values
let mut func = Function::with_name_signature(ExternalName::user(0, 0), signature);
// this context is just boiler plate
let mut ctx = FunctionBuilderContext::new();
let mut builder = FunctionBuilder::new(&mut func, &mut ctx);
let func_start = builder.create_block();
builder.switch_to_block(func_start);
let should_ret = func_type.should_return();
if func_type.has_params() {
self.store_stack_params(
// TODO: get rid of this clone
&func_type.params,
func_start,
&location,
&mut builder,
)?;
}
self.compile_all(stmts, &mut builder)?;
if !builder.is_filled() {
let id = symbol.get().id;
if id == InternedStr::get_or_intern("main") {
let ir_int = func_type.return_type.as_ir_type();
let zero = [builder.ins().iconst(ir_int, 0)];
builder.ins().return_(&zero);
} else if should_ret {
semantic_err!(
format!(
"expected a return statement before end of function '{}' returning {}",
id, func_type.return_type
),
location
);
} else {
// void function, return nothing
builder.ins().return_(&[]);
}
}
builder.seal_all_blocks();
builder.finalize();
let flags = settings::Flags::new(settings::builder());
if self.debug {
println!("ir: {}", func);
}
if let Err(err) = codegen::verify_function(&func, &flags) {
panic!(
"verification error: {}\nnote: while compiling {}",
err, func
);
}
let mut ctx = codegen::Context::for_function(func);
let mut trap_sink = codegen::binemit::NullTrapSink {};
if let Err(err) = self
.module
.define_function(func_id, &mut ctx, &mut trap_sink)
{
panic!(
"definition error: {}\nnote: while compiling {}",
err, ctx.func
);
}
Ok(())
}
}
pub type Product = <cranelift_object::ObjectBackend as Backend>::Product;
/// Compile and return the declarations and warnings.
pub fn compile<B: Backend>(module: Module<B>, buf: &str, opt: Opt) -> Program<Module<B>> {
use saltwater_parser::{check_semantics, vec_deque};
let debug_asm = opt.debug_asm;
let mut program = check_semantics(buf, opt);
let hir = match program.result {
Ok(hir) => hir,
Err(err) => {
return Program {
result: | {
match init {
Initializer::Scalar(expr) => {
let val = self.compile_expr(*expr, builder)?;
// TODO: replace with `builder.ins().stack_store(val.ir_val, stack_slot, 0);`
// when Cranelift implements stack_store for i8 and i16
let addr = builder.ins().stack_addr(Type::ptr_type(), stack_slot, 0);
builder.ins().store(MemFlags::new(), val.ir_val, addr, 0);
}
Initializer::InitializerList(_) => unimplemented!("aggregate dynamic initialization"),
Initializer::FunctionBody(_) => unreachable!("functions can't be stored on the stack"),
}
Ok(())
} | identifier_body |
lib.rs | ());
}
let u64_size = match meta.ctype.sizeof() {
Ok(size) => size,
Err(err) => {
return Err(CompileError::semantic(Locatable {
data: err.into(),
location,
}))
}
};
let kind = StackSlotKind::ExplicitSlot;
let size = match u32::try_from(u64_size) {
Ok(size) => size,
Err(_) => return Err(CompileError::semantic(Locatable {
data: "cannot store items on the stack that are more than 4 GB, it will overflow the stack".into(),
location,
}))
};
let data = StackSlotData {
kind,
size,
offset: None,
};
let stack_slot = builder.create_stack_slot(data);
self.declarations.insert(decl.symbol, Id::Local(stack_slot));
if let Some(init) = decl.init {
self.store_stack(init, stack_slot, builder)?;
}
Ok(())
}
fn store_stack(
&mut self,
init: Initializer,
stack_slot: StackSlot,
builder: &mut FunctionBuilder,
) -> CompileResult<()> {
match init {
Initializer::Scalar(expr) => {
let val = self.compile_expr(*expr, builder)?;
// TODO: replace with `builder.ins().stack_store(val.ir_val, stack_slot, 0);`
// when Cranelift implements stack_store for i8 and i16
let addr = builder.ins().stack_addr(Type::ptr_type(), stack_slot, 0);
builder.ins().store(MemFlags::new(), val.ir_val, addr, 0);
}
Initializer::InitializerList(_) => unimplemented!("aggregate dynamic initialization"),
Initializer::FunctionBody(_) => unreachable!("functions can't be stored on the stack"),
}
Ok(())
}
// TODO: this is grossly inefficient, ask Cranelift devs if
// there's an easier way to make parameters modifiable.
fn store_stack_params(
&mut self,
params: &[Symbol],
func_start: Block,
location: &Location,
builder: &mut FunctionBuilder,
) -> CompileResult<()> {
// Cranelift requires that all block params are declared up front
let ir_vals: Vec<_> = params
.iter()
.map(|param| {
let ir_type = param.get().ctype.as_ir_type();
Ok(builder.append_block_param(func_start, ir_type))
})
.collect::<CompileResult<_>>()?;
for (¶m, ir_val) in params.iter().zip(ir_vals) {
let u64_size = match param.get().ctype.sizeof() {
Err(data) => semantic_err!(data.into(), *location),
Ok(size) => size,
};
let u32_size = match u32::try_from(u64_size) {
Err(_) => semantic_err!(
format!(
"size {} is too large for stack (can only handle 32-bit values)",
u64_size
),
*location
),
Ok(size) => size,
};
let stack_data = StackSlotData {
kind: StackSlotKind::ExplicitSlot,
size: u32_size,
offset: None,
};
let slot = builder.create_stack_slot(stack_data);
// TODO: need to take the address before storing until Cranelift implements
// stores for i8 and i16
// then this can be replaced with `builder.ins().stack_store(ir_val, slot, 0);`
// See https://github.com/CraneStation/cranelift/issues/433
let addr = builder.ins().stack_addr(Type::ptr_type(), slot, 0);
builder.ins().store(MemFlags::new(), ir_val, addr, 0);
self.declarations.insert(param, Id::Local(slot));
}
Ok(())
}
fn compile_func(
&mut self,
symbol: Symbol,
func_type: &FunctionType,
stmts: Vec<Stmt>,
location: Location,
) -> CompileResult<()> {
let func_id = self.declare_func(symbol, true)?;
// TODO: make declare_func should take a `signature` after all?
// This just calculates it twice, it's probably fine
let signature = func_type.signature(self.module.isa());
// external name is meant to be a lookup in a symbol table,
// but we just give it garbage values
let mut func = Function::with_name_signature(ExternalName::user(0, 0), signature);
// this context is just boiler plate
let mut ctx = FunctionBuilderContext::new();
let mut builder = FunctionBuilder::new(&mut func, &mut ctx);
let func_start = builder.create_block();
builder.switch_to_block(func_start);
let should_ret = func_type.should_return();
if func_type.has_params() {
self.store_stack_params(
// TODO: get rid of this clone
&func_type.params,
func_start,
&location,
&mut builder,
)?;
}
self.compile_all(stmts, &mut builder)?;
if !builder.is_filled() {
let id = symbol.get().id;
if id == InternedStr::get_or_intern("main") {
let ir_int = func_type.return_type.as_ir_type();
let zero = [builder.ins().iconst(ir_int, 0)];
builder.ins().return_(&zero);
} else if should_ret {
semantic_err!(
format!(
"expected a return statement before end of function '{}' returning {}",
id, func_type.return_type
),
location
);
} else {
// void function, return nothing
builder.ins().return_(&[]);
}
}
builder.seal_all_blocks();
builder.finalize();
let flags = settings::Flags::new(settings::builder());
if self.debug {
println!("ir: {}", func);
}
if let Err(err) = codegen::verify_function(&func, &flags) {
panic!(
"verification error: {}\nnote: while compiling {}",
err, func
);
}
let mut ctx = codegen::Context::for_function(func);
let mut trap_sink = codegen::binemit::NullTrapSink {};
if let Err(err) = self
.module
.define_function(func_id, &mut ctx, &mut trap_sink)
{
panic!(
"definition error: {}\nnote: while compiling {}",
err, ctx.func
);
}
Ok(())
}
}
pub type Product = <cranelift_object::ObjectBackend as Backend>::Product;
/// Compile and return the declarations and warnings.
pub fn compile<B: Backend>(module: Module<B>, buf: &str, opt: Opt) -> Program<Module<B>> {
use saltwater_parser::{check_semantics, vec_deque};
let debug_asm = opt.debug_asm;
let mut program = check_semantics(buf, opt);
let hir = match program.result {
Ok(hir) => hir,
Err(err) => {
return Program {
result: Err(err),
warnings: program.warnings,
files: program.files,
}
}
};
// really we'd like to have all errors but that requires a refactor
let mut err = None;
let mut compiler = Compiler::new(module, debug_asm);
for decl in hir {
let meta = decl.data.symbol.get();
if let StorageClass::Typedef = meta.storage_class {
continue;
}
let current = match &meta.ctype {
Type::Function(func_type) => match decl.data.init {
Some(Initializer::FunctionBody(stmts)) => {
compiler.compile_func(decl.data.symbol, &func_type, stmts, decl.location)
}
None => compiler.declare_func(decl.data.symbol, false).map(|_| ()),
_ => unreachable!("functions can only be initialized by a FunctionBody"),
},
Type::Void | Type::Error => unreachable!("parser let an incomplete type through"),
_ => {
if let Some(Initializer::FunctionBody(_)) = &decl.data.init {
unreachable!("only functions should have a function body")
}
compiler.store_static(decl.data.symbol, decl.data.init, decl.location)
}
};
if let Err(e) = current {
err = Some(e);
break;
}
}
let warns = compiler.error_handler.warnings;
let (result, ir_warnings) = if let Some(err) = err {
(Err(err), warns)
} else {
(Ok(compiler.module), warns)
};
program.warnings.extend(ir_warnings);
Program {
result: result.map_err(|errs| vec_deque![errs]),
warnings: program.warnings,
files: program.files,
}
}
pub fn assemble(product: Product, output: &Path) -> Result<(), saltwater_parser::Error> {
use std::fs::File;
use std::io::{self, Write};
let bytes = product.emit().map_err(saltwater_parser::Error::Platform)?;
File::create(output)?
.write_all(&bytes)
.map_err(io::Error::into)
}
pub fn link(obj_file: &Path, output: &Path) -> Result<(), std::io::Error> { | use std::io::{Error, ErrorKind};
use std::process::Command;
| random_line_split |
|
cef.go |
func (d defaultLogger) Panicf(fmt string, args ... interface{}) {
log.Panicf("[cef] " + fmt, args...)
}
// Sandbox is disabled. Including the "cef_sandbox.lib"
// library results in lots of GCC warnings/errors. It is
// compatible only with VS 2010. It would be required to
// build it using GCC. Add -lcef_sandbox to LDFLAGS.
// capi doesn't expose sandbox functions, you need do add
// these before import "C":
// void* cef_sandbox_info_create();
// void cef_sandbox_info_destroy(void* sandbox_info);
var _SandboxInfo unsafe.Pointer
type Settings struct {
CachePath string
LogSeverity int
LogFile string
ResourcesDirPath string
LocalesDirPath string
RemoteDebuggingPort int
PersistSessionCookies bool
IgnoreCertificateErrors int
}
type CefState int
var (
STATE_DEFAULT CefState = 0
STATE_ENABLED CefState = 1
STATE_DISABLED CefState = 2
)
type BrowserSettings struct {
StandardFontFamily string
FixedFontFamily string
SerifFontFamily string
SansSerifFontFamily string
CursiveFontFamily string
FantasyFontFamily string
DefaultFontSize int
DefaultFixedFontSize int
MinimumFontSize int
MinimumLogicalFontSize int
DefaultEncoding string
RemoteFonts CefState
Javascript CefState
JavascriptOpenWindows CefState
JavascriptCloseWindows CefState
JavascriptAccessClipboard CefState
JavascriptDomPaste CefState
CaretBrowsing CefState
Java CefState
Plugins CefState
UniversalAccessFromFileUrls CefState
FileAccessFromFileUrls CefState
WebSecurity CefState
ImageLoading CefState
ImageShrinkStandaloneToFit CefState
TextAreaResize CefState
TabToLinks CefState
LocalStorage CefState
Databases CefState
ApplicationCache CefState
Webgl CefState
BackgroundColor uint32
}
func _InitializeGlobalCStructures() {
_InitializeGlobalCStructuresBase()
//_InitializeGlobalCStructuresApp()
/*
_DisplayHandler = InitializeDisplayHandler()
_DownloadHandler = InitializeDownloadHandler()*/
}
func SetLogger(logger SimpleLogger) {
Logger = logger
}
func ExecuteProcess(programHandle unsafe.Pointer, appHandler AppHandler) int {
Logger.Infof("ExecuteProcess, args=%v", os.Args)
_InitializeGlobalCStructures()
if appHandler.GetAppHandlerT().CStruct == nil {
panic("GetAppHandlerT cannot have a nil CStruct. Call NewAppHandlerT() to create one!")
}
FillMainArgs(_MainArgs, programHandle)
// Sandbox info needs to be passed to both cef_execute_process()
// and cef_initialize().
// OFF: _SandboxInfo = C.cef_sandbox_info_create()
Logger.Infof("MainArgs %X _AppHanlder %X _SandboxInfo %X", _MainArgs, appHandler.GetAppHandlerT().CStruct, _SandboxInfo)
go_AddRef(unsafe.Pointer(_MainArgs))
go_AddRef(unsafe.Pointer(appHandler.GetAppHandlerT().CStruct))
go_AddRef(unsafe.Pointer(_SandboxInfo))
var exitCode C.int = C.cef_execute_process(_MainArgs, appHandler.GetAppHandlerT().CStruct, _SandboxInfo)
if (exitCode >= 0) {
os.Exit(int(exitCode))
}
return int(exitCode)
}
func Initialize(settings Settings, appHandler AppHandler) int {
Logger.Infof("Initialize\n")
if _MainArgs == nil {
// _MainArgs structure is initialized and filled in ExecuteProcess.
// If cef_execute_process is not called, and there is a call
// to cef_initialize, then it would result in creation of infinite
// number of processes. See Issue 1199 in CEF:
// https://code.google.com/p/chromiumembedded/issues/detail?id=1199
Logger.Errorf("ERROR: missing a call to ExecuteProcess\n")
return 0
}
// Initialize cef_settings_t structure.
var cefSettings *C.struct__cef_settings_t
cefSettings = (*C.struct__cef_settings_t)(
C.calloc(1, C.sizeof_struct__cef_settings_t))
cefSettings.size = C.sizeof_struct__cef_settings_t
// cache_path
// ----------
if (settings.CachePath != "") {
Logger.Infof("CachePath=%s\n", settings.CachePath)
}
var cachePath *C.char = C.CString(settings.CachePath)
defer C.free(unsafe.Pointer(cachePath))
C.cef_string_from_utf8(cachePath, C.strlen(cachePath),
C.cefStringCastToCefString16(&cefSettings.cache_path))
// log_severity
// ------------
cefSettings.log_severity =
(C.cef_log_severity_t)(C.int(settings.LogSeverity))
// log_file
// --------
if (settings.LogFile != "") {
Logger.Infof("LogFile=%s\n", settings.LogFile)
}
var logFile *C.char = C.CString(settings.LogFile)
defer C.free(unsafe.Pointer(logFile))
C.cef_string_from_utf8(logFile, C.strlen(logFile),
C.cefStringCastToCefString16(&cefSettings.log_file))
// resources_dir_path
// ------------------
if settings.ResourcesDirPath == "" && runtime.GOOS != "darwin" {
// Setting this path is required for the tests to run fine.
cwd, _ := os.Getwd()
settings.ResourcesDirPath = cwd
}
if (settings.ResourcesDirPath != "") {
Logger.Infof("ResourcesDirPath=%s\n", settings.ResourcesDirPath)
}
var resourcesDirPath *C.char = C.CString(settings.ResourcesDirPath)
defer C.free(unsafe.Pointer(resourcesDirPath))
C.cef_string_from_utf8(resourcesDirPath, C.strlen(resourcesDirPath),
C.cefStringCastToCefString16(&cefSettings.resources_dir_path))
// locales_dir_path
// ----------------
if settings.LocalesDirPath == "" && runtime.GOOS != "darwin" {
// Setting this path is required for the tests to run fine.
cwd, _ := os.Getwd()
settings.LocalesDirPath = cwd + "/locales"
}
if (settings.LocalesDirPath != "") {
Logger.Infof("LocalesDirPath=%s\n", settings.LocalesDirPath)
}
var localesDirPath *C.char = C.CString(settings.LocalesDirPath)
defer C.free(unsafe.Pointer(localesDirPath))
C.cef_string_from_utf8(localesDirPath, C.strlen(localesDirPath),
C.cefStringCastToCefString16(&cefSettings.locales_dir_path))
if settings.PersistSessionCookies {
cefSettings.persist_session_cookies = 1
}
cefSettings.remote_debugging_port = C.int(settings.RemoteDebuggingPort)
cefSettings.ignore_certificate_errors = C.int(settings.IgnoreCertificateErrors)
// no_sandbox
// ----------
cefSettings.no_sandbox = C.int(1)
go_AddRef(unsafe.Pointer(_MainArgs))
go_AddRef(unsafe.Pointer(appHandler.GetAppHandlerT().CStruct))
go_AddRef(unsafe.Pointer(_SandboxInfo))
ret := C.cef_initialize(_MainArgs, cefSettings, appHandler.GetAppHandlerT().CStruct, _SandboxInfo)
return int(ret)
}
func CreateBrowser(hwnd unsafe.Pointer, clientHandler ClientHandler, browserSettings BrowserSettings,
url string) bool {
Logger.Infof("CreateBrowser, url=%s\n", url)
// Initialize cef_window_info_t structure.
var windowInfo *C.cef_window_info_t
windowInfo = (*C.cef_window_info_t)(
C.calloc(1, C.sizeof_cef_window_info_t))
FillWindowInfo(windowInfo, hwnd)
// url
var cefUrl *C.cef_string_t
cefUrl = (*C.cef_string_t)(
C.calloc(1, C.sizeof_cef_string_t))
var charUrl *C.char = C.CString(url)
defer C.free(unsafe.Pointer(charUrl))
C.cef_string_from_utf8(charUrl, C.strlen(charUrl), C.cefStringCastToCefString16(cefUrl))
// Initialize cef_browser_settings_t structure.
cefBrowserSettings := browserSettings.toC()
// Do not create the browser synchronously using the
// cef_browser_host_create_browser_sync() function, as
// it is unreliable. Instead obtain browser object in
// life_span_handler::on_after_created. In that callback
// keep CEF browser objects in a global map (cef window
// handle -> cef browser) and introduce
// a GetBrowserByWindowHandle() function. This | {
log.Printf("[cef] " + fmt, args...)
} | identifier_body |
|
cef.go | LogSeverity int
LogFile string
ResourcesDirPath string
LocalesDirPath string
RemoteDebuggingPort int
PersistSessionCookies bool
IgnoreCertificateErrors int
}
type CefState int
var (
STATE_DEFAULT CefState = 0
STATE_ENABLED CefState = 1
STATE_DISABLED CefState = 2
)
type BrowserSettings struct {
StandardFontFamily string
FixedFontFamily string
SerifFontFamily string
SansSerifFontFamily string
CursiveFontFamily string
FantasyFontFamily string
DefaultFontSize int
DefaultFixedFontSize int
MinimumFontSize int
MinimumLogicalFontSize int
DefaultEncoding string
RemoteFonts CefState
Javascript CefState
JavascriptOpenWindows CefState
JavascriptCloseWindows CefState
JavascriptAccessClipboard CefState
JavascriptDomPaste CefState
CaretBrowsing CefState
Java CefState
Plugins CefState
UniversalAccessFromFileUrls CefState
FileAccessFromFileUrls CefState
WebSecurity CefState
ImageLoading CefState
ImageShrinkStandaloneToFit CefState
TextAreaResize CefState
TabToLinks CefState
LocalStorage CefState
Databases CefState
ApplicationCache CefState
Webgl CefState
BackgroundColor uint32
}
func _InitializeGlobalCStructures() {
_InitializeGlobalCStructuresBase()
//_InitializeGlobalCStructuresApp()
/*
_DisplayHandler = InitializeDisplayHandler()
_DownloadHandler = InitializeDownloadHandler()*/
}
func SetLogger(logger SimpleLogger) {
Logger = logger
}
func ExecuteProcess(programHandle unsafe.Pointer, appHandler AppHandler) int {
Logger.Infof("ExecuteProcess, args=%v", os.Args)
_InitializeGlobalCStructures()
if appHandler.GetAppHandlerT().CStruct == nil {
panic("GetAppHandlerT cannot have a nil CStruct. Call NewAppHandlerT() to create one!")
}
FillMainArgs(_MainArgs, programHandle)
// Sandbox info needs to be passed to both cef_execute_process()
// and cef_initialize().
// OFF: _SandboxInfo = C.cef_sandbox_info_create()
Logger.Infof("MainArgs %X _AppHanlder %X _SandboxInfo %X", _MainArgs, appHandler.GetAppHandlerT().CStruct, _SandboxInfo)
go_AddRef(unsafe.Pointer(_MainArgs))
go_AddRef(unsafe.Pointer(appHandler.GetAppHandlerT().CStruct))
go_AddRef(unsafe.Pointer(_SandboxInfo))
var exitCode C.int = C.cef_execute_process(_MainArgs, appHandler.GetAppHandlerT().CStruct, _SandboxInfo)
if (exitCode >= 0) {
os.Exit(int(exitCode))
}
return int(exitCode)
}
func Initialize(settings Settings, appHandler AppHandler) int {
Logger.Infof("Initialize\n")
if _MainArgs == nil {
// _MainArgs structure is initialized and filled in ExecuteProcess.
// If cef_execute_process is not called, and there is a call
// to cef_initialize, then it would result in creation of infinite
// number of processes. See Issue 1199 in CEF:
// https://code.google.com/p/chromiumembedded/issues/detail?id=1199
Logger.Errorf("ERROR: missing a call to ExecuteProcess\n")
return 0
}
// Initialize cef_settings_t structure.
var cefSettings *C.struct__cef_settings_t
cefSettings = (*C.struct__cef_settings_t)(
C.calloc(1, C.sizeof_struct__cef_settings_t))
cefSettings.size = C.sizeof_struct__cef_settings_t
// cache_path
// ----------
if (settings.CachePath != "") {
Logger.Infof("CachePath=%s\n", settings.CachePath)
}
var cachePath *C.char = C.CString(settings.CachePath)
defer C.free(unsafe.Pointer(cachePath))
C.cef_string_from_utf8(cachePath, C.strlen(cachePath),
C.cefStringCastToCefString16(&cefSettings.cache_path))
// log_severity
// ------------
cefSettings.log_severity =
(C.cef_log_severity_t)(C.int(settings.LogSeverity))
// log_file
// --------
if (settings.LogFile != "") {
Logger.Infof("LogFile=%s\n", settings.LogFile)
}
var logFile *C.char = C.CString(settings.LogFile)
defer C.free(unsafe.Pointer(logFile))
C.cef_string_from_utf8(logFile, C.strlen(logFile),
C.cefStringCastToCefString16(&cefSettings.log_file))
// resources_dir_path
// ------------------
if settings.ResourcesDirPath == "" && runtime.GOOS != "darwin" {
// Setting this path is required for the tests to run fine.
cwd, _ := os.Getwd()
settings.ResourcesDirPath = cwd
}
if (settings.ResourcesDirPath != "") {
Logger.Infof("ResourcesDirPath=%s\n", settings.ResourcesDirPath)
}
var resourcesDirPath *C.char = C.CString(settings.ResourcesDirPath)
defer C.free(unsafe.Pointer(resourcesDirPath))
C.cef_string_from_utf8(resourcesDirPath, C.strlen(resourcesDirPath),
C.cefStringCastToCefString16(&cefSettings.resources_dir_path))
// locales_dir_path
// ----------------
if settings.LocalesDirPath == "" && runtime.GOOS != "darwin" {
// Setting this path is required for the tests to run fine.
cwd, _ := os.Getwd()
settings.LocalesDirPath = cwd + "/locales"
}
if (settings.LocalesDirPath != "") {
Logger.Infof("LocalesDirPath=%s\n", settings.LocalesDirPath)
}
var localesDirPath *C.char = C.CString(settings.LocalesDirPath)
defer C.free(unsafe.Pointer(localesDirPath))
C.cef_string_from_utf8(localesDirPath, C.strlen(localesDirPath),
C.cefStringCastToCefString16(&cefSettings.locales_dir_path))
if settings.PersistSessionCookies {
cefSettings.persist_session_cookies = 1
}
cefSettings.remote_debugging_port = C.int(settings.RemoteDebuggingPort)
cefSettings.ignore_certificate_errors = C.int(settings.IgnoreCertificateErrors)
// no_sandbox
// ----------
cefSettings.no_sandbox = C.int(1)
go_AddRef(unsafe.Pointer(_MainArgs))
go_AddRef(unsafe.Pointer(appHandler.GetAppHandlerT().CStruct))
go_AddRef(unsafe.Pointer(_SandboxInfo))
ret := C.cef_initialize(_MainArgs, cefSettings, appHandler.GetAppHandlerT().CStruct, _SandboxInfo)
return int(ret)
}
func CreateBrowser(hwnd unsafe.Pointer, clientHandler ClientHandler, browserSettings BrowserSettings, | Logger.Infof("CreateBrowser, url=%s\n", url)
// Initialize cef_window_info_t structure.
var windowInfo *C.cef_window_info_t
windowInfo = (*C.cef_window_info_t)(
C.calloc(1, C.sizeof_cef_window_info_t))
FillWindowInfo(windowInfo, hwnd)
// url
var cefUrl *C.cef_string_t
cefUrl = (*C.cef_string_t)(
C.calloc(1, C.sizeof_cef_string_t))
var charUrl *C.char = C.CString(url)
defer C.free(unsafe.Pointer(charUrl))
C.cef_string_from_utf8(charUrl, C.strlen(charUrl), C.cefStringCastToCefString16(cefUrl))
// Initialize cef_browser_settings_t structure.
cefBrowserSettings := browserSettings.toC()
// Do not create the browser synchronously using the
// cef_browser_host_create_browser_sync() function, as
// it is unreliable. Instead obtain browser object in
// life_span_handler::on_after_created. In that callback
// keep CEF browser objects in a global map (cef window
// handle -> cef browser) and introduce
// a GetBrowserByWindowHandle() function. This function
// will first guess the CEF window handle using for example
// WinAPI functions and then search the global map of cef
// browser objects.
go_AddRef(unsafe.Pointer(clientHandler.GetClientHandlerT().CStruct))
result := C.cef_browser_host_create_browser(
windowInfo,
clientHandler.GetClientHandlerT().CStruct,
cefUrl,
cefBrowserSettings,
nil,
)
return result == C.int(1)
}
func RunMessageLoop() {
Logger.Infof("RunMessageLoop\n")
C.cef_run_message_loop()
}
func QuitMessageLoop() {
Logger.Infof("QuitMessageLoop\n")
C.cef_quit_message_loop()
}
func Shutdown() {
Logger.Infof("Shutdown\n")
C.cef_shutdown()
| url string) bool { | random_line_split |
cef.go | (fmt string, args ... interface{}) {
log.Printf("[cef] " + fmt, args...)
}
func (d defaultLogger) Panicf(fmt string, args ... interface{}) {
log.Panicf("[cef] " + fmt, args...)
}
// Sandbox is disabled. Including the "cef_sandbox.lib"
// library results in lots of GCC warnings/errors. It is
// compatible only with VS 2010. It would be required to
// build it using GCC. Add -lcef_sandbox to LDFLAGS.
// capi doesn't expose sandbox functions, you need do add
// these before import "C":
// void* cef_sandbox_info_create();
// void cef_sandbox_info_destroy(void* sandbox_info);
var _SandboxInfo unsafe.Pointer
type Settings struct {
CachePath string
LogSeverity int
LogFile string
ResourcesDirPath string
LocalesDirPath string
RemoteDebuggingPort int
PersistSessionCookies bool
IgnoreCertificateErrors int
}
type CefState int
var (
STATE_DEFAULT CefState = 0
STATE_ENABLED CefState = 1
STATE_DISABLED CefState = 2
)
type BrowserSettings struct {
StandardFontFamily string
FixedFontFamily string
SerifFontFamily string
SansSerifFontFamily string
CursiveFontFamily string
FantasyFontFamily string
DefaultFontSize int
DefaultFixedFontSize int
MinimumFontSize int
MinimumLogicalFontSize int
DefaultEncoding string
RemoteFonts CefState
Javascript CefState
JavascriptOpenWindows CefState
JavascriptCloseWindows CefState
JavascriptAccessClipboard CefState
JavascriptDomPaste CefState
CaretBrowsing CefState
Java CefState
Plugins CefState
UniversalAccessFromFileUrls CefState
FileAccessFromFileUrls CefState
WebSecurity CefState
ImageLoading CefState
ImageShrinkStandaloneToFit CefState
TextAreaResize CefState
TabToLinks CefState
LocalStorage CefState
Databases CefState
ApplicationCache CefState
Webgl CefState
BackgroundColor uint32
}
func _InitializeGlobalCStructures() {
_InitializeGlobalCStructuresBase()
//_InitializeGlobalCStructuresApp()
/*
_DisplayHandler = InitializeDisplayHandler()
_DownloadHandler = InitializeDownloadHandler()*/
}
func SetLogger(logger SimpleLogger) {
Logger = logger
}
func ExecuteProcess(programHandle unsafe.Pointer, appHandler AppHandler) int {
Logger.Infof("ExecuteProcess, args=%v", os.Args)
_InitializeGlobalCStructures()
if appHandler.GetAppHandlerT().CStruct == nil {
panic("GetAppHandlerT cannot have a nil CStruct. Call NewAppHandlerT() to create one!")
}
FillMainArgs(_MainArgs, programHandle)
// Sandbox info needs to be passed to both cef_execute_process()
// and cef_initialize().
// OFF: _SandboxInfo = C.cef_sandbox_info_create()
Logger.Infof("MainArgs %X _AppHanlder %X _SandboxInfo %X", _MainArgs, appHandler.GetAppHandlerT().CStruct, _SandboxInfo)
go_AddRef(unsafe.Pointer(_MainArgs))
go_AddRef(unsafe.Pointer(appHandler.GetAppHandlerT().CStruct))
go_AddRef(unsafe.Pointer(_SandboxInfo))
var exitCode C.int = C.cef_execute_process(_MainArgs, appHandler.GetAppHandlerT().CStruct, _SandboxInfo)
if (exitCode >= 0) {
os.Exit(int(exitCode))
}
return int(exitCode)
}
func Initialize(settings Settings, appHandler AppHandler) int {
Logger.Infof("Initialize\n")
if _MainArgs == nil {
// _MainArgs structure is initialized and filled in ExecuteProcess.
// If cef_execute_process is not called, and there is a call
// to cef_initialize, then it would result in creation of infinite
// number of processes. See Issue 1199 in CEF:
// https://code.google.com/p/chromiumembedded/issues/detail?id=1199
Logger.Errorf("ERROR: missing a call to ExecuteProcess\n")
return 0
}
// Initialize cef_settings_t structure.
var cefSettings *C.struct__cef_settings_t
cefSettings = (*C.struct__cef_settings_t)(
C.calloc(1, C.sizeof_struct__cef_settings_t))
cefSettings.size = C.sizeof_struct__cef_settings_t
// cache_path
// ----------
if (settings.CachePath != "") {
Logger.Infof("CachePath=%s\n", settings.CachePath)
}
var cachePath *C.char = C.CString(settings.CachePath)
defer C.free(unsafe.Pointer(cachePath))
C.cef_string_from_utf8(cachePath, C.strlen(cachePath),
C.cefStringCastToCefString16(&cefSettings.cache_path))
// log_severity
// ------------
cefSettings.log_severity =
(C.cef_log_severity_t)(C.int(settings.LogSeverity))
// log_file
// --------
if (settings.LogFile != "") {
Logger.Infof("LogFile=%s\n", settings.LogFile)
}
var logFile *C.char = C.CString(settings.LogFile)
defer C.free(unsafe.Pointer(logFile))
C.cef_string_from_utf8(logFile, C.strlen(logFile),
C.cefStringCastToCefString16(&cefSettings.log_file))
// resources_dir_path
// ------------------
if settings.ResourcesDirPath == "" && runtime.GOOS != "darwin" {
// Setting this path is required for the tests to run fine.
cwd, _ := os.Getwd()
settings.ResourcesDirPath = cwd
}
if (settings.ResourcesDirPath != "") {
Logger.Infof("ResourcesDirPath=%s\n", settings.ResourcesDirPath)
}
var resourcesDirPath *C.char = C.CString(settings.ResourcesDirPath)
defer C.free(unsafe.Pointer(resourcesDirPath))
C.cef_string_from_utf8(resourcesDirPath, C.strlen(resourcesDirPath),
C.cefStringCastToCefString16(&cefSettings.resources_dir_path))
// locales_dir_path
// ----------------
if settings.LocalesDirPath == "" && runtime.GOOS != "darwin" {
// Setting this path is required for the tests to run fine.
cwd, _ := os.Getwd()
settings.LocalesDirPath = cwd + "/locales"
}
if (settings.LocalesDirPath != "") {
Logger.Infof("LocalesDirPath=%s\n", settings.LocalesDirPath)
}
var localesDirPath *C.char = C.CString(settings.LocalesDirPath)
defer C.free(unsafe.Pointer(localesDirPath))
C.cef_string_from_utf8(localesDirPath, C.strlen(localesDirPath),
C.cefStringCastToCefString16(&cefSettings.locales_dir_path))
if settings.PersistSessionCookies {
cefSettings.persist_session_cookies = 1
}
cefSettings.remote_debugging_port = C.int(settings.RemoteDebuggingPort)
cefSettings.ignore_certificate_errors = C.int(settings.IgnoreCertificateErrors)
// no_sandbox
// ----------
cefSettings.no_sandbox = C.int(1)
go_AddRef(unsafe.Pointer(_MainArgs))
go_AddRef(unsafe.Pointer(appHandler.GetAppHandlerT().CStruct))
go_AddRef(unsafe.Pointer(_SandboxInfo))
ret := C.cef_initialize(_MainArgs, cefSettings, appHandler.GetAppHandlerT().CStruct, _SandboxInfo)
return int(ret)
}
func CreateBrowser(hwnd unsafe.Pointer, clientHandler ClientHandler, browserSettings BrowserSettings,
url string) bool {
Logger.Infof("CreateBrowser, url=%s\n", url)
// Initialize cef_window_info_t structure.
var windowInfo *C.cef_window_info_t
windowInfo = (*C.cef_window_info_t)(
C.calloc(1, C.sizeof_cef_window_info_t))
FillWindowInfo(windowInfo, hwnd)
// url
var cefUrl *C.cef_string_t
cefUrl = (*C.cef_string_t)(
C.calloc(1, C.sizeof_cef_string_t))
var charUrl *C.char = C.CString(url)
defer C.free(unsafe.Pointer(charUrl))
C.cef_string_from_utf8(charUrl, C.strlen(charUrl), C.cefStringCastToCefString16(cefUrl))
// Initialize cef_browser_settings_t structure.
cefBrowserSettings := browserSettings.toC()
// Do not create the browser synchronously using the
// cef_browser_host_create_browser_sync() function, as
// it is unreliable. Instead obtain browser object in
// life_span_handler::on_after_created. In that callback
// keep CEF browser objects in a global map (cef window
// handle -> cef browser) and introduce
// a Get | Errorf | identifier_name |
|
cef.go | efState
JavascriptCloseWindows CefState
JavascriptAccessClipboard CefState
JavascriptDomPaste CefState
CaretBrowsing CefState
Java CefState
Plugins CefState
UniversalAccessFromFileUrls CefState
FileAccessFromFileUrls CefState
WebSecurity CefState
ImageLoading CefState
ImageShrinkStandaloneToFit CefState
TextAreaResize CefState
TabToLinks CefState
LocalStorage CefState
Databases CefState
ApplicationCache CefState
Webgl CefState
BackgroundColor uint32
}
func _InitializeGlobalCStructures() {
_InitializeGlobalCStructuresBase()
//_InitializeGlobalCStructuresApp()
/*
_DisplayHandler = InitializeDisplayHandler()
_DownloadHandler = InitializeDownloadHandler()*/
}
func SetLogger(logger SimpleLogger) {
Logger = logger
}
func ExecuteProcess(programHandle unsafe.Pointer, appHandler AppHandler) int {
Logger.Infof("ExecuteProcess, args=%v", os.Args)
_InitializeGlobalCStructures()
if appHandler.GetAppHandlerT().CStruct == nil {
panic("GetAppHandlerT cannot have a nil CStruct. Call NewAppHandlerT() to create one!")
}
FillMainArgs(_MainArgs, programHandle)
// Sandbox info needs to be passed to both cef_execute_process()
// and cef_initialize().
// OFF: _SandboxInfo = C.cef_sandbox_info_create()
Logger.Infof("MainArgs %X _AppHanlder %X _SandboxInfo %X", _MainArgs, appHandler.GetAppHandlerT().CStruct, _SandboxInfo)
go_AddRef(unsafe.Pointer(_MainArgs))
go_AddRef(unsafe.Pointer(appHandler.GetAppHandlerT().CStruct))
go_AddRef(unsafe.Pointer(_SandboxInfo))
var exitCode C.int = C.cef_execute_process(_MainArgs, appHandler.GetAppHandlerT().CStruct, _SandboxInfo)
if (exitCode >= 0) {
os.Exit(int(exitCode))
}
return int(exitCode)
}
func Initialize(settings Settings, appHandler AppHandler) int {
Logger.Infof("Initialize\n")
if _MainArgs == nil {
// _MainArgs structure is initialized and filled in ExecuteProcess.
// If cef_execute_process is not called, and there is a call
// to cef_initialize, then it would result in creation of infinite
// number of processes. See Issue 1199 in CEF:
// https://code.google.com/p/chromiumembedded/issues/detail?id=1199
Logger.Errorf("ERROR: missing a call to ExecuteProcess\n")
return 0
}
// Initialize cef_settings_t structure.
var cefSettings *C.struct__cef_settings_t
cefSettings = (*C.struct__cef_settings_t)(
C.calloc(1, C.sizeof_struct__cef_settings_t))
cefSettings.size = C.sizeof_struct__cef_settings_t
// cache_path
// ----------
if (settings.CachePath != "") {
Logger.Infof("CachePath=%s\n", settings.CachePath)
}
var cachePath *C.char = C.CString(settings.CachePath)
defer C.free(unsafe.Pointer(cachePath))
C.cef_string_from_utf8(cachePath, C.strlen(cachePath),
C.cefStringCastToCefString16(&cefSettings.cache_path))
// log_severity
// ------------
cefSettings.log_severity =
(C.cef_log_severity_t)(C.int(settings.LogSeverity))
// log_file
// --------
if (settings.LogFile != "") {
Logger.Infof("LogFile=%s\n", settings.LogFile)
}
var logFile *C.char = C.CString(settings.LogFile)
defer C.free(unsafe.Pointer(logFile))
C.cef_string_from_utf8(logFile, C.strlen(logFile),
C.cefStringCastToCefString16(&cefSettings.log_file))
// resources_dir_path
// ------------------
if settings.ResourcesDirPath == "" && runtime.GOOS != "darwin" {
// Setting this path is required for the tests to run fine.
cwd, _ := os.Getwd()
settings.ResourcesDirPath = cwd
}
if (settings.ResourcesDirPath != "") {
Logger.Infof("ResourcesDirPath=%s\n", settings.ResourcesDirPath)
}
var resourcesDirPath *C.char = C.CString(settings.ResourcesDirPath)
defer C.free(unsafe.Pointer(resourcesDirPath))
C.cef_string_from_utf8(resourcesDirPath, C.strlen(resourcesDirPath),
C.cefStringCastToCefString16(&cefSettings.resources_dir_path))
// locales_dir_path
// ----------------
if settings.LocalesDirPath == "" && runtime.GOOS != "darwin" {
// Setting this path is required for the tests to run fine.
cwd, _ := os.Getwd()
settings.LocalesDirPath = cwd + "/locales"
}
if (settings.LocalesDirPath != "") {
Logger.Infof("LocalesDirPath=%s\n", settings.LocalesDirPath)
}
var localesDirPath *C.char = C.CString(settings.LocalesDirPath)
defer C.free(unsafe.Pointer(localesDirPath))
C.cef_string_from_utf8(localesDirPath, C.strlen(localesDirPath),
C.cefStringCastToCefString16(&cefSettings.locales_dir_path))
if settings.PersistSessionCookies {
cefSettings.persist_session_cookies = 1
}
cefSettings.remote_debugging_port = C.int(settings.RemoteDebuggingPort)
cefSettings.ignore_certificate_errors = C.int(settings.IgnoreCertificateErrors)
// no_sandbox
// ----------
cefSettings.no_sandbox = C.int(1)
go_AddRef(unsafe.Pointer(_MainArgs))
go_AddRef(unsafe.Pointer(appHandler.GetAppHandlerT().CStruct))
go_AddRef(unsafe.Pointer(_SandboxInfo))
ret := C.cef_initialize(_MainArgs, cefSettings, appHandler.GetAppHandlerT().CStruct, _SandboxInfo)
return int(ret)
}
func CreateBrowser(hwnd unsafe.Pointer, clientHandler ClientHandler, browserSettings BrowserSettings,
url string) bool {
Logger.Infof("CreateBrowser, url=%s\n", url)
// Initialize cef_window_info_t structure.
var windowInfo *C.cef_window_info_t
windowInfo = (*C.cef_window_info_t)(
C.calloc(1, C.sizeof_cef_window_info_t))
FillWindowInfo(windowInfo, hwnd)
// url
var cefUrl *C.cef_string_t
cefUrl = (*C.cef_string_t)(
C.calloc(1, C.sizeof_cef_string_t))
var charUrl *C.char = C.CString(url)
defer C.free(unsafe.Pointer(charUrl))
C.cef_string_from_utf8(charUrl, C.strlen(charUrl), C.cefStringCastToCefString16(cefUrl))
// Initialize cef_browser_settings_t structure.
cefBrowserSettings := browserSettings.toC()
// Do not create the browser synchronously using the
// cef_browser_host_create_browser_sync() function, as
// it is unreliable. Instead obtain browser object in
// life_span_handler::on_after_created. In that callback
// keep CEF browser objects in a global map (cef window
// handle -> cef browser) and introduce
// a GetBrowserByWindowHandle() function. This function
// will first guess the CEF window handle using for example
// WinAPI functions and then search the global map of cef
// browser objects.
go_AddRef(unsafe.Pointer(clientHandler.GetClientHandlerT().CStruct))
result := C.cef_browser_host_create_browser(
windowInfo,
clientHandler.GetClientHandlerT().CStruct,
cefUrl,
cefBrowserSettings,
nil,
)
return result == C.int(1)
}
func RunMessageLoop() {
Logger.Infof("RunMessageLoop\n")
C.cef_run_message_loop()
}
func QuitMessageLoop() {
Logger.Infof("QuitMessageLoop\n")
C.cef_quit_message_loop()
}
func Shutdown() {
Logger.Infof("Shutdown\n")
C.cef_shutdown()
// OFF: cef_sandbox_info_destroy(_SandboxInfo)
}
func extractCefMultiMap(cefMapPointer C.cef_string_multimap_t) map[string][]string {
numKeys := C.cef_string_multimap_size(cefMapPointer)
goMap := make(map[string][]string)
for i := 0; i < int(numKeys); i++ {
var key *C.cef_string_utf16_t = C.cef_string_userfree_utf16_alloc()
C.cef_string_multimap_key(cefMapPointer, C.int(i), C.cefString16CastToCefString(key))
charKeyUtf8 := C.cefStringToUtf8(C.cefString16CastToCefString(key))
goKey := C.GoString(charKeyUtf8.str)
if _, ok := goMap[goKey]; ok | {
continue
} | conditional_block |
|
romaO.py | 9, 0.76047],
[0.33253, 0.59151, 0.75704],
[0.32893, 0.58412, 0.75351],
[0.32559, 0.57671, 0.74987],
[0.32256, 0.56928, 0.74613],
[0.31978, 0.56186, 0.74228],
[0.31727, 0.55441, 0.7383],
[0.31505, 0.54695, 0.73422],
[0.31311, 0.53948, 0.73002],
[0.31144, 0.53201, 0.72569],
[0.31007, 0.52453, 0.72124],
[0.30897, 0.51704, 0.71667],
[0.30811, 0.50955, 0.71197],
[0.30755, 0.50205, 0.70713],
[0.30726, 0.49456, 0.70216],
[0.30723, 0.48707, 0.69706],
[0.30746, 0.47958, 0.69182],
[0.30795, 0.4721, 0.68643],
[0.3087, 0.46463, 0.6809],
[0.30968, 0.45716, 0.67525],
[0.31088, 0.44973, 0.66944],
[0.31228, 0.44232, 0.6635],
[0.31393, 0.43493, 0.65741],
[0.31578, 0.42758, 0.65118],
[0.3178, 0.42025, 0.64482],
[0.32001, 0.41299, 0.63833],
[0.32238, 0.40577, 0.6317],
[0.32489, 0.39861, 0.62495],
[0.32755, 0.39152, 0.61809],
[0.33035, 0.38448, 0.61111],
[0.33327, 0.37755, 0.60402],
[0.33627, 0.37068, 0.59684],
[0.33939, 0.36392, 0.58955],
[0.34257, 0.35728, 0.58219],
[0.3458, 0.35073, 0.57476],
[0.34912, 0.34428, 0.56727],
[0.35247, 0.33797, 0.55971],
[0.35587, 0.33179, 0.55212],
[0.35927, 0.32574, 0.54448],
[0.36271, 0.31986, 0.53684],
[0.36617, 0.31411, 0.52917],
[0.36961, 0.30852, 0.52148],
[0.37306, 0.30306, 0.51382],
[0.37652, 0.2978, 0.50615],
[0.37994, 0.29269, 0.49854],
[0.38336, 0.28775, 0.49094],
[0.38674, 0.28301, 0.48337],
[0.39011, 0.27842, 0.47586],
[0.39346, 0.27401, 0.4684],
[0.39677, 0.26978, 0.461],
[0.40006, 0.26573, 0.45366],
[0.40333, 0.26185, 0.4464],
[0.40655, 0.25815, 0.43921],
[0.40974, 0.25466, 0.43212],
[0.4129, 0.25132, 0.42509],
[0.41602, 0.24817, 0.41813],
[0.41912, 0.24515, 0.41128],
[0.42218, 0.24235, 0.40451],
[0.42522, 0.23972, 0.39784],
[0.42823, 0.23728, 0.39126],
[0.43121, 0.23498, 0.38475],
[0.43415, 0.23282, 0.37836],
[0.43708, 0.23086, 0.37204],
[0.43998, 0.22907, 0.36583],
[0.44286, 0.22743, 0.3597],
[0.44571, 0.22596, 0.35366],
[0.44855, 0.2246, 0.34773]]
romaO_map = LinearSegmentedColormap.from_list('romaO', cm_data)
# For use of "viscm view"
test_cm = romaO_map
if __name__ == "__main__":
| import matplotlib.pyplot as plt
import numpy as np
try:
from viscm import viscm
viscm(romaO_map)
except ImportError:
print("viscm not found, falling back on simple display")
plt.imshow(np.linspace(0, 100, 256)[None, :], aspect='auto',
cmap=romaO_map)
plt.show() | conditional_block |
|
romaO.py | 2165],
[0.78242, 0.6759, 0.33066],
[0.78669, 0.68481, 0.33988],
[0.79087, 0.69365, 0.34929],
[0.79494, 0.7024, 0.35888],
[0.7989, 0.71106, 0.36867],
[0.80273, 0.71961, 0.37859],
[0.80642, 0.72803, 0.38866],
[0.80996, 0.73631, 0.39885],
[0.81334, 0.74446, 0.40916],
[0.81655, 0.75244, 0.41957],
[0.81956, 0.76025, 0.43004],
[0.82239, 0.76787, 0.44057],
[0.82501, 0.7753, 0.45115],
[0.82742, 0.78252, 0.46174],
[0.8296, 0.78953, 0.47235],
[0.83155, 0.79631, 0.48293],
[0.83326, 0.80287, 0.49349],
[0.83472, 0.80919, 0.50402],
[0.83592, 0.81526, 0.51449],
[0.83686, 0.82109, 0.52487],
[0.83753, 0.82666, 0.53517],
[0.83793, 0.83198, 0.54537],
[0.83805, 0.83703, 0.55546],
[0.83788, 0.84182, 0.56542],
[0.83744, 0.84635, 0.57525],
[0.8367, 0.85061, 0.58493],
[0.83567, 0.85462, 0.59446],
[0.83435, 0.85835, 0.60382],
[0.83274, 0.86183, 0.61301],
[0.83084, 0.86504, 0.62202],
[0.82864, 0.868, 0.63085],
[0.82615, 0.87068, 0.63949],
[0.82337, 0.87312, 0.64792],
[0.8203, 0.87531, 0.65617],
[0.81695, 0.87724, 0.6642],
[0.81331, 0.87892, 0.67203],
[0.80939, 0.88036, 0.67964],
[0.80518, 0.88156, 0.68705],
[0.80071, 0.8825, 0.69424],
[0.79595, 0.88322, 0.70121],
[0.79094, 0.8837, 0.70797],
[0.78566, 0.88395, 0.7145],
[0.78012, 0.88396, 0.72082],
[0.77433, 0.88375, 0.72692],
[0.7683, 0.88331, 0.73279],
[0.76203, 0.88264, 0.73844],
[0.75553, 0.88177, 0.74387],
[0.74879, 0.88066, 0.74908],
[0.74184, 0.87934, 0.75407],
[0.73468, 0.87781, 0.75884],
[0.72731, 0.87607, 0.76339],
[0.71976, 0.87411, 0.76772],
[0.71201, 0.87195, 0.77184],
[0.70408, 0.86958, 0.77573],
[0.69599, 0.86701, 0.77941],
[0.68774, 0.86425, 0.78288],
[0.67934, 0.86127, 0.78614],
[0.67081, 0.85811, 0.78919],
[0.66215, 0.85476, 0.79202],
[0.65336, 0.8512, 0.79465],
[0.64448, 0.84747, 0.79707],
[0.6355, 0.84356, 0.7993],
[0.62645, 0.83947, 0.80131],
[0.61732, 0.83519, 0.80313],
[0.60814, 0.83075, 0.80476],
[0.59891, 0.82614, 0.80619], | [0.58965, 0.82137, 0.80743],
[0.58037, 0.81644, 0.80848],
[0.57108, 0.81135, 0.80935],
[0.56181, 0.80612, 0.81004],
[0.55255, 0.80074, 0.81055], | random_line_split |
|
tags.rs | ;
}
None => unreachable!(),
}
}
let result = if tags.is_empty() {
ctx.psql().get_tags_mapset(mapset_id).await
} else {
let db_result = match action {
Action::Add => ctx.psql().add_tags_mapset(mapset_id, tags).await,
Action::Remove => ctx.psql().remove_tags_mapset(mapset_id, tags).await,
};
match db_result {
Ok(_) => ctx.psql().get_tags_mapset(mapset_id).await,
Err(err) => Err(err),
}
};
// Then show the final tags
match result {
Ok(tags) => {
let content = format!("{OSU_BASE}beatmapsets/{mapset_id} is now tagged as:\n{tags}");
let builder = MessageBuilder::new().content(content);
msg.create_message(&ctx, builder).await?;
}
Err(err) => {
let _ = msg.error(&ctx, GENERAL_ISSUE).await;
return Err(err);
}
}
Ok(())
}
// #[command]
// #[short_desc("Help out tagging backgrounds")]
// #[long_desc(
// "Let me give you mapsets that still need to be tagged.\n\
// React to them properly, then lock it in by reacting with ✅.\n\
// To leave the loop, react with ❌ or just wait 10 minutes.\n\
// Mode can be specified in the first argument, defaults to std.\n\
// **You need to be verified to use this command**, feel free to \
// let Badewanne3 know if you want to help out tagging backgrounds."
// )]
// #[usage("[std / mna]")]
// #[aliases("bgt", "bgtag")]
// #[owner()]
async fn bgtags(ctx: Arc<Context>, data: CommandData) -> Result<()> {
let (msg, mut args) = match data {
CommandData::Message { msg, args, .. } => (msg, args),
CommandData::Interaction { .. } => unreachable!(),
};
// Parse arguments as mode
let mode = match args.next() {
Some(arg) => match arg.cow_to_ascii_lowercase().as_ref() {
"mna" | "mania" | "m" => GameMode::Mania,
"osu" | "std" | "standard" | "o" => GameMode::Osu,
_ => {
let content = "Could not parse first argument as mode. \
Provide either `mna`, or `std`";
return msg.error(&ctx, content).await;
}
},
None => GameMode::Osu,
};
let mut untagged = match ctx.psql().get_all_tags_mapset(mode).await {
Ok(tags) => tags.iter().any(|tag| tag.untagged()),
Err(err) => {
let _ = msg.error(&ctx, GENERAL_ISSUE).await;
return Err(err);
}
};
if !untagged {
let content = "All backgrounds have been tagged, \
here are some random ones you can review again though";
let builder = MessageBuilder::new().content(content);
let _ = msg.create_message(&ctx, builder).await;
}
let mut owner = msg.author.id;
loop {
// Get all mapsets for which tags are missing
let tags_result = if untagged {
ctx.psql().get_all_tags_mapset(mode).await
} else {
ctx.psql()
.get_random_tags_mapset(mode)
.await
.map(|tags| vec![tags])
};
let mapsets = match tags_result {
Ok(mut tags) => {
if untagged {
if tags.iter().any(|tag| tag.untagged()) {
tags.retain(|tag| tag.untagged());
} else {
let content = "All backgrounds have been tagged, \
here are some random ones you can review again though";
let builder = MessageBuilder::new().content(content);
let _ = msg.create_message(&ctx, builder).await;
untagged = false;
tags.truncate(1);
}
}
tags
}
Err(err) => {
let _ = msg.error(&ctx, GENERAL_ISSUE).await;
return Err(err);
}
};
let (mapset_id, img) = get_random_image(mapsets, mode).await;
let content = format!(
"<@{owner}> Which tags should this mapsets get: {OSU_BASE}beatmapsets/{mapset_id}\n\
```\n\
🍋: Easy 🎨: Weeb 😱: Hard name 🗽: English 💯: Tech\n\
🤓: Hard 🍨: Kpop 🪀: Alternate 🌀: Streams ✅: Lock in\n\
🤡: Meme 👨🌾: Farm 🟦: Blue sky 👴: Old ❌: Exit loop\n\
```"
);
let img = Attachment::from_bytes("bg_img.png".to_owned(), img);
// Send response
let response = ctx
.http
.create_message(msg.channel_id)
.content(&content)?
.attachments(&[img])
.unwrap()
.exec()
.await?
.model()
.await?;
let msg_id = response.id;
// Setup collector
let reaction_stream = ctx
.standby
.wait_for_event_stream(move |event: &Event| match event {
Event::ReactionAdd(event) => {
event.message_id == msg_id && event.user_id.get() == OWNER_USER_ID
}
Event::ReactionRemove(event) => {
event.message_id == msg_id && event.user_id.get() == OWNER_USER_ID
}
_ => false,
})
.map(|event| match event {
Event::ReactionAdd(add) => ReactionWrapper::Add(add.0),
Event::ReactionRemove(remove) => ReactionWrapper::Remove(remove.0),
_ => unreachable!(),
})
.timeout(Duration::from_secs(600));
tokio::pin!(reaction_stream);
// Add reactions
let reactions = [
"🍋",
"🤓",
"🤡",
"🎨",
"🍨",
"👨🌾",
"😱",
"🪀",
"🟦",
"🗽",
"🌀",
"👴",
"💯",
"✅",
"❌",
];
for &reaction in reactions.iter() {
let emote = Emote::Custom(reaction);
send_reaction(&*ctx, &response, emote).await?;
}
let mut break_loop = true;
// Run collector
let mut add_tags = MapsetTags::empty();
let mut remove_tags = MapsetTags::empty();
while let Some(Ok(reaction)) = reaction_stream.next().await {
let tag = if let ReactionType::Unicode { ref name } = reaction.as_deref().emoji {
match name.as_str() {
"🍋" => MapsetTags::Easy,
"🤓" => MapsetTags::Hard,
"🤡" => MapsetTags::Meme,
"👴" => MapsetTags::Old,
"😱" => MapsetTags::HardName,
"🟦" => MapsetTags::BlueSky,
"🪀" => MapsetTags::Alternate,
"🗽" => MapsetTags::English,
"👨🌾" => MapsetTags::Farm,
"💯" => MapsetTags::Tech,
"🎨" => MapsetTags::Weeb,
"🌀" => MapsetTags::Streams,
"🍨" => MapsetTags::Kpop,
"✅" => {
owner = reaction.as_deref().user_id;
break_loop = false;
break;
}
"❌" => break,
_ => continue,
}
} else {
continue;
};
match reaction {
ReactionWrapper::Add(_) => {
add_tags.insert(tag);
}
ReactionWrapper::Remove(_) => {
remove_tags.insert(tag);
}
}
}
if !add_tags.is_empty() {
if let Err(err) = ctx.psql().add_tags_mapset(mapset_id, add_tags).await {
warn!(?err, "Failed to add tags");
}
}
if !remove_tags.is_empty() {
if let Err(err) = ctx.psql().remove_tags_mapset(mapset_id, remove_tags).await {
warn!(?err, "Failed to remove tags");
}
}
// Then show the final tags
match ctx.psql().get_tags_mapset(mapset_id).await {
Ok(tags) => {
if !tags.is_empty() {
let content = format!(
"{}beatmapsets/{} is now tagged as:\n{}",
OSU_BASE, mapset_id, tags,
);
let builder = MessageBuilder::new().content(content);
msg.create_message(&ctx, builder).await?;
}
}
Err(err) => {
let _ = msg.error(&ctx, GENERAL_ISSUE).await;
return Err(err);
}
};
if break_loop {
| let builder = MessageBuilder::new().content("Exiting loop :wave:");
msg.create_message( | conditional_block |
|
tags.rs | data: CommandData) -> Result<()> {
let (msg, mut args) = match data {
CommandData::Message { msg, args, .. } => (msg, args),
CommandData::Interaction { .. } => unreachable!(),
};
// Parse mapset id
let mapset_id = match args.next().map(u32::from_str) {
Some(Ok(num)) => num,
Some(Err(_)) => {
let content = "Could not parse mapset id. Be sure to specify it as first argument";
return msg.error(&ctx, content).await;
}
None => {
let content = "Arguments: `[mapset id] [add/a/remove/r] [list of tags]`\n\
Example: `21662 r hard farm streams alternate hardname tech weeb bluesky`\n\
Tags: `farm, streams, alternate, old, meme, hardname, easy, hard, tech, \
weeb, bluesky, english`";
let builder = MessageBuilder::new().content(content);
msg.create_message(&ctx, builder).await?;
return Ok(());
}
};
// Check if there is background for the given mapset id
if ctx.psql().get_tags_mapset(mapset_id).await.is_err() {
let content = "No background entry found with this id";
return msg.error(&ctx, content).await;
}
// Parse action
let action = match args.next().map(Action::from_str) {
Some(Ok(action)) => action,
None | Some(Err(_)) => {
let content = "Could not parse action. \
Be sure to specify `r`, `remove`, `a`, or `add` as second argument";
return msg.error(&ctx, content).await;
}
};
// Parse tags
let mut tags = MapsetTags::empty();
while !args.is_empty() {
match args.next().map(MapsetTags::from_str) {
Some(Ok(tag)) => tags.insert(tag),
Some(Err(tag)) => {
let content = format!(
"Could not parse tag `{tag}`.\n\
Be sure to only give these tags:\n\
`farm, streams, alternate, old, meme, hardname, \
easy, hard, tech, weeb, bluesky, english`"
);
return msg.error(&ctx, content).await;
}
None => unreachable!(),
}
}
let result = if tags.is_empty() {
ctx.psql().get_tags_mapset(mapset_id).await
} else {
let db_result = match action {
Action::Add => ctx.psql().add_tags_mapset(mapset_id, tags).await,
Action::Remove => ctx.psql().remove_tags_mapset(mapset_id, tags).await,
};
match db_result {
Ok(_) => ctx.psql().get_tags_mapset(mapset_id).await,
Err(err) => Err(err),
}
};
// Then show the final tags
match result {
Ok(tags) => {
let content = format!("{OSU_BASE}beatmapsets/{mapset_id} is now tagged as:\n{tags}");
let builder = MessageBuilder::new().content(content);
msg.create_message(&ctx, builder).await?;
}
Err(err) => {
let _ = msg.error(&ctx, GENERAL_ISSUE).await;
return Err(err);
}
}
Ok(())
}
// #[command]
// #[short_desc("Help out tagging backgrounds")]
// #[long_desc(
// "Let me give you mapsets that still need to be tagged.\n\
// React to them properly, then lock it in by reacting with ✅.\n\
// To leave the loop, react with ❌ or just wait 10 minutes.\n\
// Mode can be specified in the first argument, defaults to std.\n\
// **You need to be verified to use this command**, feel free to \
// let Badewanne3 know if you want to help out tagging backgrounds."
// )]
// #[usage("[std / mna]")]
// #[aliases("bgt", "bgtag")]
// #[owner()]
async fn bgtags(ctx: Arc<Context>, data: CommandData) -> Result<()> {
let (msg, mut args) = match data {
CommandData::Message { msg, args, .. } => (msg, args), | let mode = match args.next() {
Some(arg) => match arg.cow_to_ascii_lowercase().as_ref() {
"mna" | "mania" | "m" => GameMode::Mania,
"osu" | "std" | "standard" | "o" => GameMode::Osu,
_ => {
let content = "Could not parse first argument as mode. \
Provide either `mna`, or `std`";
return msg.error(&ctx, content).await;
}
},
None => GameMode::Osu,
};
let mut untagged = match ctx.psql().get_all_tags_mapset(mode).await {
Ok(tags) => tags.iter().any(|tag| tag.untagged()),
Err(err) => {
let _ = msg.error(&ctx, GENERAL_ISSUE).await;
return Err(err);
}
};
if !untagged {
let content = "All backgrounds have been tagged, \
here are some random ones you can review again though";
let builder = MessageBuilder::new().content(content);
let _ = msg.create_message(&ctx, builder).await;
}
let mut owner = msg.author.id;
loop {
// Get all mapsets for which tags are missing
let tags_result = if untagged {
ctx.psql().get_all_tags_mapset(mode).await
} else {
ctx.psql()
.get_random_tags_mapset(mode)
.await
.map(|tags| vec![tags])
};
let mapsets = match tags_result {
Ok(mut tags) => {
if untagged {
if tags.iter().any(|tag| tag.untagged()) {
tags.retain(|tag| tag.untagged());
} else {
let content = "All backgrounds have been tagged, \
here are some random ones you can review again though";
let builder = MessageBuilder::new().content(content);
let _ = msg.create_message(&ctx, builder).await;
untagged = false;
tags.truncate(1);
}
}
tags
}
Err(err) => {
let _ = msg.error(&ctx, GENERAL_ISSUE).await;
return Err(err);
}
};
let (mapset_id, img) = get_random_image(mapsets, mode).await;
let content = format!(
"<@{owner}> Which tags should this mapsets get: {OSU_BASE}beatmapsets/{mapset_id}\n\
```\n\
🍋: Easy 🎨: Weeb 😱: Hard name 🗽: English 💯: Tech\n\
🤓: Hard 🍨: Kpop 🪀: Alternate 🌀: Streams ✅: Lock in\n\
🤡: Meme 👨🌾: Farm 🟦: Blue sky 👴: Old ❌: Exit loop\n\
```"
);
let img = Attachment::from_bytes("bg_img.png".to_owned(), img);
// Send response
let response = ctx
.http
.create_message(msg.channel_id)
.content(&content)?
.attachments(&[img])
.unwrap()
.exec()
.await?
.model()
.await?;
let msg_id = response.id;
// Setup collector
let reaction_stream = ctx
.standby
.wait_for_event_stream(move |event: &Event| match event {
Event::ReactionAdd(event) => {
event.message_id == msg_id && event.user_id.get() == OWNER_USER_ID
}
Event::ReactionRemove(event) => {
event.message_id == msg_id && event.user_id.get() == OWNER_USER_ID
}
_ => false,
})
.map(|event| match event {
Event::ReactionAdd(add) => ReactionWrapper::Add(add.0),
Event::ReactionRemove(remove) => ReactionWrapper::Remove(remove.0),
_ => unreachable!(),
})
.timeout(Duration::from_secs(600));
tokio::pin!(reaction_stream);
// Add reactions
let reactions = [
"🍋",
"🤓",
"🤡",
"🎨",
"🍨",
"👨🌾",
"😱",
"🪀",
"🟦",
"🗽",
"🌀",
"👴",
"💯",
"✅",
"❌",
];
for &reaction in reactions.iter() {
let emote = Emote::Custom(reaction);
send_reaction(&*ctx, &response, emote).await?;
}
let mut break_loop = true;
// Run collector
let mut add_tags = MapsetTags::empty();
let mut remove_tags = MapsetTags::empty();
while let Some(Ok(reaction)) = reaction_stream.next().await {
let | CommandData::Interaction { .. } => unreachable!(),
};
// Parse arguments as mode | random_line_split |
tags.rs | , tags).await,
Action::Remove => ctx.psql().remove_tags_mapset(mapset_id, tags).await,
};
match db_result {
Ok(_) => ctx.psql().get_tags_mapset(mapset_id).await,
Err(err) => Err(err),
}
};
// Then show the final tags
match result {
Ok(tags) => {
let content = format!("{OSU_BASE}beatmapsets/{mapset_id} is now tagged as:\n{tags}");
let builder = MessageBuilder::new().content(content);
msg.create_message(&ctx, builder).await?;
}
Err(err) => {
let _ = msg.error(&ctx, GENERAL_ISSUE).await;
return Err(err);
}
}
Ok(())
}
// #[command]
// #[short_desc("Help out tagging backgrounds")]
// #[long_desc(
// "Let me give you mapsets that still need to be tagged.\n\
// React to them properly, then lock it in by reacting with ✅.\n\
// To leave the loop, react with ❌ or just wait 10 minutes.\n\
// Mode can be specified in the first argument, defaults to std.\n\
// **You need to be verified to use this command**, feel free to \
// let Badewanne3 know if you want to help out tagging backgrounds."
// )]
// #[usage("[std / mna]")]
// #[aliases("bgt", "bgtag")]
// #[owner()]
async fn bgtags(ctx: Arc<Context>, data: CommandData) -> Result<()> {
let (msg, mut args) = match data {
CommandData::Message { msg, args, .. } => (msg, args),
CommandData::Interaction { .. } => unreachable!(),
};
// Parse arguments as mode
let mode = match args.next() {
Some(arg) => match arg.cow_to_ascii_lowercase().as_ref() {
"mna" | "mania" | "m" => GameMode::Mania,
"osu" | "std" | "standard" | "o" => GameMode::Osu,
_ => {
let content = "Could not parse first argument as mode. \
Provide either `mna`, or `std`";
return msg.error(&ctx, content).await;
}
},
None => GameMode::Osu,
};
let mut untagged = match ctx.psql().get_all_tags_mapset(mode).await {
Ok(tags) => tags.iter().any(|tag| tag.untagged()),
Err(err) => {
let _ = msg.error(&ctx, GENERAL_ISSUE).await;
return Err(err);
}
};
if !untagged {
let content = "All backgrounds have been tagged, \
here are some random ones you can review again though";
let builder = MessageBuilder::new().content(content);
let _ = msg.create_message(&ctx, builder).await;
}
let mut owner = msg.author.id;
loop {
// Get all mapsets for which tags are missing
let tags_result = if untagged {
ctx.psql().get_all_tags_mapset(mode).await
} else {
ctx.psql()
.get_random_tags_mapset(mode)
.await
.map(|tags| vec![tags])
};
let mapsets = match tags_result {
Ok(mut tags) => {
if untagged {
if tags.iter().any(|tag| tag.untagged()) {
tags.retain(|tag| tag.untagged());
} else {
let content = "All backgrounds have been tagged, \
here are some random ones you can review again though";
let builder = MessageBuilder::new().content(content);
let _ = msg.create_message(&ctx, builder).await;
untagged = false;
tags.truncate(1);
}
}
tags
}
Err(err) => {
let _ = msg.error(&ctx, GENERAL_ISSUE).await;
return Err(err);
}
};
let (mapset_id, img) = get_random_image(mapsets, mode).await;
let content = format!(
"<@{owner}> Which tags should this mapsets get: {OSU_BASE}beatmapsets/{mapset_id}\n\
```\n\
🍋: Easy 🎨: Weeb 😱: Hard name 🗽: English 💯: Tech\n\
🤓: Hard 🍨: Kpop 🪀: Alternate 🌀: Streams ✅: Lock in\n\
🤡: Meme 👨🌾: Farm 🟦: Blue sky 👴: Old ❌: Exit loop\n\
```"
);
let img = Attachment::from_bytes("bg_img.png".to_owned(), img);
// Send response
let response = ctx
.http
.create_message(msg.channel_id)
.content(&content)?
.attachments(&[img])
.unwrap()
.exec()
.await?
.model()
.await?;
let msg_id = response.id;
// Setup collector
let reaction_stream = ctx
.standby
.wait_for_event_stream(move |event: &Event| match event {
Event::ReactionAdd(event) => {
event.message_id == msg_id && event.user_id.get() == OWNER_USER_ID
}
Event::ReactionRemove(event) => {
event.message_id == msg_id && event.user_id.get() == OWNER_USER_ID
}
_ => false,
})
.map(|event| match event {
Event::ReactionAdd(add) => ReactionWrapper::Add(add.0),
Event::ReactionRemove(remove) => ReactionWrapper::Remove(remove.0),
_ => unreachable!(),
})
.timeout(Duration::from_secs(600));
tokio::pin!(reaction_stream);
// Add reactions
let reactions = [
"🍋",
"🤓",
"🤡",
"🎨",
"🍨",
"👨🌾",
"😱",
"🪀",
"🟦",
"🗽",
"🌀",
"👴",
"💯",
"✅",
"❌",
];
for &reaction in reactions.iter() {
let emote = Emote::Custom(reaction);
send_reaction(&*ctx, &response, emote).await?;
}
let mut break_loop = true;
// Run collector
let mut add_tags = MapsetTags::empty();
let mut remove_tags = MapsetTags::empty();
while let Some(Ok(reaction)) = reaction_stream.next().await {
let tag = if let ReactionType::Unicode { ref name } = reaction.as_deref().emoji {
match name.as_str() {
"🍋" => MapsetTags::Easy,
"🤓" => MapsetTags::Hard,
"🤡" => MapsetTags::Meme,
"👴" => MapsetTags::Old,
"😱" => MapsetTags::HardName,
"🟦" => MapsetTags::BlueSky,
"🪀" => MapsetTags::Alternate,
"🗽" => MapsetTags::English,
"👨🌾" => MapsetTags::Farm,
"💯" => MapsetTags::Tech,
"🎨" => MapsetTags::Weeb,
"🌀" => MapsetTags::Streams,
"🍨" => MapsetTags::Kpop,
"✅" => {
owner = reaction.as_deref().user_id;
break_loop = false;
break;
}
"❌" => break,
_ => continue,
}
} else {
continue;
};
match reaction {
ReactionWrapper::Add(_) => {
add_tags.insert(tag);
}
ReactionWrapper::Remove(_) => {
remove_tags.insert(tag);
}
}
}
if !add_tags.is_empty() {
if let Err(err) = ctx.psql().add_tags_mapset(mapset_id, add_tags).await {
warn!(?err, "Failed to add tags");
}
}
if !remove_tags.is_empty() {
if let Err(err) = ctx.psql().remove_tags_mapset(mapset_id, remove_tags).await {
warn!(?err, "Failed to remove tags");
}
}
// Then show the final tags
match ctx.psql().get_tags_mapset(mapset_id).await {
Ok(tags) => {
if !tags.is_empty() {
let content = format!(
"{}beatmapsets/{} is now tagged as:\n{}",
OSU_BASE, mapset_id, tags,
);
let builder = MessageBuilder::new().content(content);
msg.create_message(&ctx, builder).await?;
}
}
Err(err) => {
let _ = msg.error(&ctx, GENERAL_ISSUE).await;
return Err(err);
}
};
if break_loop {
let builder = MessageBuilder::new().content("Exiting loop :wave:");
msg.create_message(&ctx, builder).await?;
break;
}
}
Ok(())
}
async fn get_random_image(mut mapsets: Vec<MapsetTagWrapper>, mode: GameMode) -> (u32, Vec<u8>) {
let mut path = CONFIG.get().unwrap().paths.backgrounds. | clone();
ma | identifier_name |
|
tags.rs | : CommandData) -> Result<()> {
let (msg, mut args) = match data {
CommandData::Message { msg, args, .. } => (msg, args),
CommandData::Interaction { .. } => unreachable!(),
};
// Parse mapset id
let mapset_id = match args.next().map(u32::from_str) {
Some(Ok(num)) => num,
Some(Err(_)) => {
let content = "Could not parse mapset id. Be sure to specify it as first argument";
return msg.error(&ctx, content).await;
}
None => {
let content = "Arguments: `[mapset id] [add/a/remove/r] [list of tags]`\n\
Example: `21662 r hard farm streams alternate hardname tech weeb bluesky`\n\
Tags: `farm, streams, alternate, old, meme, hardname, easy, hard, tech, \
weeb, bluesky, english`";
let builder = MessageBuilder::new().content(content);
msg.create_message(&ctx, builder).await?;
return Ok(());
}
};
// Check if there is background for the given mapset id
if ctx.psql().get_tags_mapset(mapset_id).await.is_err() {
let content = "No background entry found with this id";
return msg.error(&ctx, content).await;
}
// Parse action
let action = match args.next().map(Action::from_str) {
Some(Ok(action)) => action,
None | Some(Err(_)) => {
let content = "Could not parse action. \
Be sure to specify `r`, `remove`, `a`, or `add` as second argument";
return msg.error(&ctx, content).await;
}
};
// Parse tags
let mut tags = MapsetTags::empty();
while !args.is_empty() {
match args.next().map(MapsetTags::from_str) {
Some(Ok(tag)) => tags.insert(tag),
Some(Err(tag)) => {
let content = format!(
"Could not parse tag `{tag}`.\n\
Be sure to only give these tags:\n\
`farm, streams, alternate, old, meme, hardname, \
easy, hard, tech, weeb, bluesky, english`"
);
return msg.error(&ctx, content).await;
}
None => unreachable!(),
}
}
let result = if tags.is_empty() {
ctx.psql().get_tags_mapset(mapset_id).await
} else {
let db_result = match action {
Action::Add => ctx.psql().add_tags_mapset(mapset_id, tags).await,
Action::Remove => ctx.psql().remove_tags_mapset(mapset_id, tags).await,
};
match db_result {
Ok(_) => ctx.psql().get_tags_mapset(mapset_id).await,
Err(err) => Err(err),
}
};
// Then show the final tags
match result {
Ok(tags) => {
let content = format!("{OSU_BASE}beatmapsets/{mapset_id} is now tagged as:\n{tags}");
let builder = MessageBuilder::new().content(content);
msg.create_message(&ctx, builder).await?;
}
Err(err) => {
let _ = msg.error(&ctx, GENERAL_ISSUE).await;
return Err(err);
}
}
Ok(())
}
// #[command]
// #[short_desc("Help out tagging backgrounds")]
// #[long_desc(
// "Let me give you mapsets that still need to be tagged.\n\
// React to them properly, then lock it in by reacting with ✅.\n\
// To leave the loop, react with ❌ or just wait 10 minutes.\n\
// Mode can be specified in the first argument, defaults to std.\n\
// **You need to be verified to use this command**, feel free to \
// let Badewanne3 know if you want to help out tagging backgrounds."
// )]
// #[usage("[std / mna]")]
// #[aliases("bgt", "bgtag")]
// #[owner()]
async fn bgtags(ctx: Arc<Context>, data: CommandData) -> Result<()> {
| let mut untagged = match ctx.psql().get_all_tags_mapset(mode).await {
Ok(tags) => tags.iter().any(|tag| tag.untagged()),
Err(err) => {
let _ = msg.error(&ctx, GENERAL_ISSUE).await;
return Err(err);
}
};
if !untagged {
let content = "All backgrounds have been tagged, \
here are some random ones you can review again though";
let builder = MessageBuilder::new().content(content);
let _ = msg.create_message(&ctx, builder).await;
}
let mut owner = msg.author.id;
loop {
// Get all mapsets for which tags are missing
let tags_result = if untagged {
ctx.psql().get_all_tags_mapset(mode).await
} else {
ctx.psql()
.get_random_tags_mapset(mode)
.await
.map(|tags| vec![tags])
};
let mapsets = match tags_result {
Ok(mut tags) => {
if untagged {
if tags.iter().any(|tag| tag.untagged()) {
tags.retain(|tag| tag.untagged());
} else {
let content = "All backgrounds have been tagged, \
here are some random ones you can review again though";
let builder = MessageBuilder::new().content(content);
let _ = msg.create_message(&ctx, builder).await;
untagged = false;
tags.truncate(1);
}
}
tags
}
Err(err) => {
let _ = msg.error(&ctx, GENERAL_ISSUE).await;
return Err(err);
}
};
let (mapset_id, img) = get_random_image(mapsets, mode).await;
let content = format!(
"<@{owner}> Which tags should this mapsets get: {OSU_BASE}beatmapsets/{mapset_id}\n\
```\n\
🍋: Easy 🎨: Weeb 😱: Hard name 🗽: English 💯: Tech\n\
🤓: Hard 🍨: Kpop 🪀: Alternate 🌀: Streams ✅: Lock in\n\
🤡: Meme 👨🌾: Farm 🟦: Blue sky 👴: Old ❌: Exit loop\n\
```"
);
let img = Attachment::from_bytes("bg_img.png".to_owned(), img);
// Send response
let response = ctx
.http
.create_message(msg.channel_id)
.content(&content)?
.attachments(&[img])
.unwrap()
.exec()
.await?
.model()
.await?;
let msg_id = response.id;
// Setup collector
let reaction_stream = ctx
.standby
.wait_for_event_stream(move |event: &Event| match event {
Event::ReactionAdd(event) => {
event.message_id == msg_id && event.user_id.get() == OWNER_USER_ID
}
Event::ReactionRemove(event) => {
event.message_id == msg_id && event.user_id.get() == OWNER_USER_ID
}
_ => false,
})
.map(|event| match event {
Event::ReactionAdd(add) => ReactionWrapper::Add(add.0),
Event::ReactionRemove(remove) => ReactionWrapper::Remove(remove.0),
_ => unreachable!(),
})
.timeout(Duration::from_secs(600));
tokio::pin!(reaction_stream);
// Add reactions
let reactions = [
"🍋",
"🤓",
"🤡",
"🎨",
"🍨",
"👨🌾",
"😱",
"🪀",
"🟦",
"🗽",
"🌀",
"👴",
"💯",
"✅",
"❌",
];
for &reaction in reactions.iter() {
let emote = Emote::Custom(reaction);
send_reaction(&*ctx, &response, emote).await?;
}
let mut break_loop = true;
// Run collector
let mut add_tags = MapsetTags::empty();
let mut remove_tags = MapsetTags::empty();
while let Some(Ok(reaction)) = reaction_stream.next().await {
| let (msg, mut args) = match data {
CommandData::Message { msg, args, .. } => (msg, args),
CommandData::Interaction { .. } => unreachable!(),
};
// Parse arguments as mode
let mode = match args.next() {
Some(arg) => match arg.cow_to_ascii_lowercase().as_ref() {
"mna" | "mania" | "m" => GameMode::Mania,
"osu" | "std" | "standard" | "o" => GameMode::Osu,
_ => {
let content = "Could not parse first argument as mode. \
Provide either `mna`, or `std`";
return msg.error(&ctx, content).await;
}
},
None => GameMode::Osu,
};
| identifier_body |
golem_renderer.rs | Float, Dimension::D4)
));
for matrix_counter in 1 ..= shader.description.num_float_matrices {
uniforms.push(Uniform::new(
MATRIX_VARIABLE_NAMES[matrix_counter as usize],
UniformType::Matrix(Dimension::D4)
));
}
for color_counter in 1 ..= shader.description.num_colors {
uniforms.push(Uniform::new(
COLOR_VARIABLE_NAMES[color_counter as usize],
UniformType::Vector(NumberType::Float, Dimension::D4)
));
}
for vector_counter in 1 ..= shader.description.num_float_vectors {
uniforms.push(Uniform::new(
FLOAT_VECTOR_VARIABLE_NAMES[vector_counter as usize],
UniformType::Vector(NumberType::Float, Dimension::D4)
));
}
for vector_counter in 1 ..= shader.description.num_int_vectors {
uniforms.push(Uniform::new(
INT_VECTOR_VARIABLE_NAMES[vector_counter as usize],
UniformType::Vector(NumberType::Int, Dimension::D4)
));
}
for float_counter in 1 ..= shader.description.num_floats {
uniforms.push(Uniform::new(
FLOAT_VARIABLE_NAMES[float_counter as usize],
UniformType::Scalar(NumberType::Float)
));
}
for int_counter in 1 ..= shader.description.num_ints {
uniforms.push(Uniform::new(
INT_VARIABLE_NAMES[int_counter as usize],
UniformType::Scalar(NumberType::Int)
));
}
let shader_description = ShaderDescription {
vertex_input: &[
Attribute::new("vertexInnerPosition", AttributeType::Vector(Dimension::D2))
],
fragment_input: &[
Attribute::new("innerPosition", AttributeType::Vector(Dimension::D2)),
Attribute::new("outerPosition", AttributeType::Vector(Dimension::D2))
],
uniforms: &uniforms,
vertex_shader: "
void main() {
innerPosition = 0.5 * vertexInnerPosition + 0.5;
vec2 bottomLeftBounds = vertexBounds.xy;
vec2 topRightBounds = vertexBounds.zw;
outerPosition = bottomLeftBounds + innerPosition * (topRightBounds - bottomLeftBounds);
gl_Position = vec4(2.0 * outerPosition - vec2(1.0, 1.0), 0.0, 1.0);
}
",
fragment_shader: &shader.description.source_code
};
ShaderProgram::new(golem, shader_description)
}, |shader_program| {
shader_program.set_uniform("vertexBounds", UniformValue::Vector4([min_x, min_y, max_x, max_y]))?;
for matrix_counter in 1 ..= shader.description.num_float_matrices {
let _result = shader_program.set_uniform(
&format!("matrix{}", matrix_counter),
UniformValue::Matrix4(parameters.float_matrices[matrix_counter as usize - 1])
);
}
for color_counter in 1 ..= shader.description.num_colors {
let _result = shader_program.set_uniform(
&format!("color{}", color_counter),
UniformValue::Vector4(parameters.colors[color_counter as usize - 1].to_float_array())
);
}
for vector_counter in 1 ..= shader.description.num_float_vectors {
let _result = shader_program.set_uniform(
&format!("floatVector{}", vector_counter),
UniformValue::Vector4(parameters.float_vectors[vector_counter as usize - 1])
);
}
for vector_counter in 1 ..= shader.description.num_int_vectors {
let _result = shader_program.set_uniform(
&format!("intVector{}", vector_counter),
UniformValue::IVector4(parameters.int_vectors[vector_counter as usize - 1])
);
}
for float_counter in 1 ..= shader.description.num_floats {
let _result = shader_program.set_uniform(
&format!("float{}", float_counter), | );
}
for int_counter in 1 ..= shader.description.num_ints {
let _result = shader_program.set_uniform(
&format!("int{}", int_counter),
UniformValue::Int(parameters.ints[int_counter as usize - 1])
);
}
unsafe {
shader_program.draw(
self.get_quad_vertices(),
self.get_quad_indices(),
0 .. self.get_num_quad_indices(),
GeometryMode::Triangles
)
}
}
).expect("Shader shouldn't fail");
}
/// Gets the golem `Context` of this `Renderer`. Use this context to perform drawing operations
/// that are not covered by the other methods of `Renderer`. Note that using this will damage
/// the portability of the application since this will only work when a Golem renderer is used.
pub fn get_context(&self) -> &Context {
&self.context
}
// This will be handled internally.
pub(super) fn apply_viewport_and_scissor(&self) {
self.get_viewport().set_viewport(&self.context);
self.get_scissor().set_scissor(&self.context);
}
/// Gets a reference to a `VertexBuffer` representing the basic `quad` model (simply the
/// positions [(-1.0, -1.0), (1.0, -1.0), (1.0, 1.0), (-1.0, 1.0)] ).
///
/// This model can be surprisingly useful for `Component`s because this simple model can be
/// quite powerful in combination with the right (fragment) shader: by discarding the right
/// pixels, it is easy to construct other shapes like circles. It is also great for drawing
/// basic images.
///
/// As explained above, it can be useful for many `Component`. It would be a slight waste of
/// resources to let every component create its own quad `VertexBuffer`. To solve this issue,
/// all components in need of the quad model can simply share this one.
pub fn get_quad_vertices(&self) -> &VertexBuffer {
&self.storage.quad_vertices
}
/// Gets a reference to the corresponding `ElementBuffer` of the `VertexBuffer` given by the
/// `get_quad_vertices` method. (These indices are just [(0, 1, 2), (2, 3, 0)].)
pub fn get_quad_indices(&self) -> &ElementBuffer {
&self.storage.quad_indices
}
/// Gets the number of indices in the `ElementBuffer` given by the `get_quad_indices`
/// method, in integers (which is just 6).
pub fn get_num_quad_indices(&self) -> usize {
6
}
/// Checks if the shader with the given *id* has been cached by this `Renderer`. If so, `bind`s
/// that shader and calls the given *use_shader* closure.
///
/// If the shader with the given *id* is **not** found in the cache, the given `create_shader`
/// closure will be called to create this. Then, it will be stored in the cache and its `bind`
/// function will be called. And finally, the given *use_shader* closure will be called.
///
/// ## Motivation
/// Caching shaders can make the implementation of the `render` methods of `Component`s easier
/// while also improving performance: `Component`s often need shader(s) for rendering, and they
/// either need to create it at the start of every call of their `render` method (which is very
/// bad for performance). Or, they could create it lazily during their first `render` call and
/// store it for later (which is annoying to program because it requires adding an extra
/// `Option<ShaderProgram>` field and maintain that). That would be better for performance, but
/// is still suboptimal because every `Component` will need its **own** instance of the
/// shader it need(s), even if many other `Component`s need that exact same shader.
///
/// When `Component`s use this method, they no longer need to worry about storing the shader
/// (because the `Renderer` will take care of that), and it will automatically be shared by all
/// other `Component` that use this method and the same shader **id**.
pub fn use_cached_shader(
&self,
id: &ShaderId,
create_shader: impl FnOnce(&golem::Context) -> Result<ShaderProgram, GolemError>,
use_shader: impl FnOnce(&mut ShaderProgram) -> Result<(), GolemError>,
) -> Result<(), GolemError> {
let mut cache = self.storage.shader_cache.borrow_mut();
cache.use_shader(id, || create_shader(&self.context), use_shader)
}
pub fn load_texture(&self, cpu_texture: &crate::Texture) -> Result<golem::Texture, GolemError> {
let mut gpu_texture = golem::Texture::new(&self.context)?;
let pixel_buffer = cpu_texture.create_pixel_buffer();
gpu_texture.set_image(
Some(&pixel_buffer),
cpu_texture.get_width(),
cpu_texture.get_height(),
ColorFormat::RGBA,
);
gpu_texture.set_wrap_h(TextureWrap::ClampToEdge)?;
gpu_texture.set_wrap_v(TextureWrap::ClampToEdge)?;
gpu_texture.set_magnification(TextureFilter::Linear)?;
gpu_texture.set_minification(TextureFilter::Linear)?;
Ok(gpu_texture)
}
}
pub(super) struct GolemRenderStorage {
// Frequently | UniformValue::Float(parameters.floats[float_counter as usize - 1]) | random_line_split |
golem_renderer.rs | Float, Dimension::D4)
));
for matrix_counter in 1 ..= shader.description.num_float_matrices {
uniforms.push(Uniform::new(
MATRIX_VARIABLE_NAMES[matrix_counter as usize],
UniformType::Matrix(Dimension::D4)
));
}
for color_counter in 1 ..= shader.description.num_colors {
uniforms.push(Uniform::new(
COLOR_VARIABLE_NAMES[color_counter as usize],
UniformType::Vector(NumberType::Float, Dimension::D4)
));
}
for vector_counter in 1 ..= shader.description.num_float_vectors {
uniforms.push(Uniform::new(
FLOAT_VECTOR_VARIABLE_NAMES[vector_counter as usize],
UniformType::Vector(NumberType::Float, Dimension::D4)
));
}
for vector_counter in 1 ..= shader.description.num_int_vectors {
uniforms.push(Uniform::new(
INT_VECTOR_VARIABLE_NAMES[vector_counter as usize],
UniformType::Vector(NumberType::Int, Dimension::D4)
));
}
for float_counter in 1 ..= shader.description.num_floats {
uniforms.push(Uniform::new(
FLOAT_VARIABLE_NAMES[float_counter as usize],
UniformType::Scalar(NumberType::Float)
));
}
for int_counter in 1 ..= shader.description.num_ints {
uniforms.push(Uniform::new(
INT_VARIABLE_NAMES[int_counter as usize],
UniformType::Scalar(NumberType::Int)
));
}
let shader_description = ShaderDescription {
vertex_input: &[
Attribute::new("vertexInnerPosition", AttributeType::Vector(Dimension::D2))
],
fragment_input: &[
Attribute::new("innerPosition", AttributeType::Vector(Dimension::D2)),
Attribute::new("outerPosition", AttributeType::Vector(Dimension::D2))
],
uniforms: &uniforms,
vertex_shader: "
void main() {
innerPosition = 0.5 * vertexInnerPosition + 0.5;
vec2 bottomLeftBounds = vertexBounds.xy;
vec2 topRightBounds = vertexBounds.zw;
outerPosition = bottomLeftBounds + innerPosition * (topRightBounds - bottomLeftBounds);
gl_Position = vec4(2.0 * outerPosition - vec2(1.0, 1.0), 0.0, 1.0);
}
",
fragment_shader: &shader.description.source_code
};
ShaderProgram::new(golem, shader_description)
}, |shader_program| {
shader_program.set_uniform("vertexBounds", UniformValue::Vector4([min_x, min_y, max_x, max_y]))?;
for matrix_counter in 1 ..= shader.description.num_float_matrices {
let _result = shader_program.set_uniform(
&format!("matrix{}", matrix_counter),
UniformValue::Matrix4(parameters.float_matrices[matrix_counter as usize - 1])
);
}
for color_counter in 1 ..= shader.description.num_colors {
let _result = shader_program.set_uniform(
&format!("color{}", color_counter),
UniformValue::Vector4(parameters.colors[color_counter as usize - 1].to_float_array())
);
}
for vector_counter in 1 ..= shader.description.num_float_vectors {
let _result = shader_program.set_uniform(
&format!("floatVector{}", vector_counter),
UniformValue::Vector4(parameters.float_vectors[vector_counter as usize - 1])
);
}
for vector_counter in 1 ..= shader.description.num_int_vectors {
let _result = shader_program.set_uniform(
&format!("intVector{}", vector_counter),
UniformValue::IVector4(parameters.int_vectors[vector_counter as usize - 1])
);
}
for float_counter in 1 ..= shader.description.num_floats {
let _result = shader_program.set_uniform(
&format!("float{}", float_counter),
UniformValue::Float(parameters.floats[float_counter as usize - 1])
);
}
for int_counter in 1 ..= shader.description.num_ints {
let _result = shader_program.set_uniform(
&format!("int{}", int_counter),
UniformValue::Int(parameters.ints[int_counter as usize - 1])
);
}
unsafe {
shader_program.draw(
self.get_quad_vertices(),
self.get_quad_indices(),
0 .. self.get_num_quad_indices(),
GeometryMode::Triangles
)
}
}
).expect("Shader shouldn't fail");
}
/// Gets the golem `Context` of this `Renderer`. Use this context to perform drawing operations
/// that are not covered by the other methods of `Renderer`. Note that using this will damage
/// the portability of the application since this will only work when a Golem renderer is used.
pub fn get_context(&self) -> &Context {
&self.context
}
// This will be handled internally.
pub(super) fn apply_viewport_and_scissor(&self) {
self.get_viewport().set_viewport(&self.context);
self.get_scissor().set_scissor(&self.context);
}
/// Gets a reference to a `VertexBuffer` representing the basic `quad` model (simply the
/// positions [(-1.0, -1.0), (1.0, -1.0), (1.0, 1.0), (-1.0, 1.0)] ).
///
/// This model can be surprisingly useful for `Component`s because this simple model can be
/// quite powerful in combination with the right (fragment) shader: by discarding the right
/// pixels, it is easy to construct other shapes like circles. It is also great for drawing
/// basic images.
///
/// As explained above, it can be useful for many `Component`. It would be a slight waste of
/// resources to let every component create its own quad `VertexBuffer`. To solve this issue,
/// all components in need of the quad model can simply share this one.
pub fn get_quad_vertices(&self) -> &VertexBuffer {
&self.storage.quad_vertices
}
/// Gets a reference to the corresponding `ElementBuffer` of the `VertexBuffer` given by the
/// `get_quad_vertices` method. (These indices are just [(0, 1, 2), (2, 3, 0)].)
pub fn | (&self) -> &ElementBuffer {
&self.storage.quad_indices
}
/// Gets the number of indices in the `ElementBuffer` given by the `get_quad_indices`
/// method, in integers (which is just 6).
pub fn get_num_quad_indices(&self) -> usize {
6
}
/// Checks if the shader with the given *id* has been cached by this `Renderer`. If so, `bind`s
/// that shader and calls the given *use_shader* closure.
///
/// If the shader with the given *id* is **not** found in the cache, the given `create_shader`
/// closure will be called to create this. Then, it will be stored in the cache and its `bind`
/// function will be called. And finally, the given *use_shader* closure will be called.
///
/// ## Motivation
/// Caching shaders can make the implementation of the `render` methods of `Component`s easier
/// while also improving performance: `Component`s often need shader(s) for rendering, and they
/// either need to create it at the start of every call of their `render` method (which is very
/// bad for performance). Or, they could create it lazily during their first `render` call and
/// store it for later (which is annoying to program because it requires adding an extra
/// `Option<ShaderProgram>` field and maintain that). That would be better for performance, but
/// is still suboptimal because every `Component` will need its **own** instance of the
/// shader it need(s), even if many other `Component`s need that exact same shader.
///
/// When `Component`s use this method, they no longer need to worry about storing the shader
/// (because the `Renderer` will take care of that), and it will automatically be shared by all
/// other `Component` that use this method and the same shader **id**.
pub fn use_cached_shader(
&self,
id: &ShaderId,
create_shader: impl FnOnce(&golem::Context) -> Result<ShaderProgram, GolemError>,
use_shader: impl FnOnce(&mut ShaderProgram) -> Result<(), GolemError>,
) -> Result<(), GolemError> {
let mut cache = self.storage.shader_cache.borrow_mut();
cache.use_shader(id, || create_shader(&self.context), use_shader)
}
pub fn load_texture(&self, cpu_texture: &crate::Texture) -> Result<golem::Texture, GolemError> {
let mut gpu_texture = golem::Texture::new(&self.context)?;
let pixel_buffer = cpu_texture.create_pixel_buffer();
gpu_texture.set_image(
Some(&pixel_buffer),
cpu_texture.get_width(),
cpu_texture.get_height(),
ColorFormat::RGBA,
);
gpu_texture.set_wrap_h(TextureWrap::ClampToEdge)?;
gpu_texture.set_wrap_v(TextureWrap::ClampToEdge)?;
gpu_texture.set_magnification(TextureFilter::Linear)?;
gpu_texture.set_minification(TextureFilter::Linear)?;
Ok(gpu_texture)
}
}
pub(super) struct GolemRenderStorage {
// | get_quad_indices | identifier_name |
golem_renderer.rs | ].to_float_array())
);
}
for vector_counter in 1 ..= shader.description.num_float_vectors {
let _result = shader_program.set_uniform(
&format!("floatVector{}", vector_counter),
UniformValue::Vector4(parameters.float_vectors[vector_counter as usize - 1])
);
}
for vector_counter in 1 ..= shader.description.num_int_vectors {
let _result = shader_program.set_uniform(
&format!("intVector{}", vector_counter),
UniformValue::IVector4(parameters.int_vectors[vector_counter as usize - 1])
);
}
for float_counter in 1 ..= shader.description.num_floats {
let _result = shader_program.set_uniform(
&format!("float{}", float_counter),
UniformValue::Float(parameters.floats[float_counter as usize - 1])
);
}
for int_counter in 1 ..= shader.description.num_ints {
let _result = shader_program.set_uniform(
&format!("int{}", int_counter),
UniformValue::Int(parameters.ints[int_counter as usize - 1])
);
}
unsafe {
shader_program.draw(
self.get_quad_vertices(),
self.get_quad_indices(),
0 .. self.get_num_quad_indices(),
GeometryMode::Triangles
)
}
}
).expect("Shader shouldn't fail");
}
/// Gets the golem `Context` of this `Renderer`. Use this context to perform drawing operations
/// that are not covered by the other methods of `Renderer`. Note that using this will damage
/// the portability of the application since this will only work when a Golem renderer is used.
pub fn get_context(&self) -> &Context {
&self.context
}
// This will be handled internally.
pub(super) fn apply_viewport_and_scissor(&self) {
self.get_viewport().set_viewport(&self.context);
self.get_scissor().set_scissor(&self.context);
}
/// Gets a reference to a `VertexBuffer` representing the basic `quad` model (simply the
/// positions [(-1.0, -1.0), (1.0, -1.0), (1.0, 1.0), (-1.0, 1.0)] ).
///
/// This model can be surprisingly useful for `Component`s because this simple model can be
/// quite powerful in combination with the right (fragment) shader: by discarding the right
/// pixels, it is easy to construct other shapes like circles. It is also great for drawing
/// basic images.
///
/// As explained above, it can be useful for many `Component`. It would be a slight waste of
/// resources to let every component create its own quad `VertexBuffer`. To solve this issue,
/// all components in need of the quad model can simply share this one.
pub fn get_quad_vertices(&self) -> &VertexBuffer {
&self.storage.quad_vertices
}
/// Gets a reference to the corresponding `ElementBuffer` of the `VertexBuffer` given by the
/// `get_quad_vertices` method. (These indices are just [(0, 1, 2), (2, 3, 0)].)
pub fn get_quad_indices(&self) -> &ElementBuffer {
&self.storage.quad_indices
}
/// Gets the number of indices in the `ElementBuffer` given by the `get_quad_indices`
/// method, in integers (which is just 6).
pub fn get_num_quad_indices(&self) -> usize {
6
}
/// Checks if the shader with the given *id* has been cached by this `Renderer`. If so, `bind`s
/// that shader and calls the given *use_shader* closure.
///
/// If the shader with the given *id* is **not** found in the cache, the given `create_shader`
/// closure will be called to create this. Then, it will be stored in the cache and its `bind`
/// function will be called. And finally, the given *use_shader* closure will be called.
///
/// ## Motivation
/// Caching shaders can make the implementation of the `render` methods of `Component`s easier
/// while also improving performance: `Component`s often need shader(s) for rendering, and they
/// either need to create it at the start of every call of their `render` method (which is very
/// bad for performance). Or, they could create it lazily during their first `render` call and
/// store it for later (which is annoying to program because it requires adding an extra
/// `Option<ShaderProgram>` field and maintain that). That would be better for performance, but
/// is still suboptimal because every `Component` will need its **own** instance of the
/// shader it need(s), even if many other `Component`s need that exact same shader.
///
/// When `Component`s use this method, they no longer need to worry about storing the shader
/// (because the `Renderer` will take care of that), and it will automatically be shared by all
/// other `Component` that use this method and the same shader **id**.
pub fn use_cached_shader(
&self,
id: &ShaderId,
create_shader: impl FnOnce(&golem::Context) -> Result<ShaderProgram, GolemError>,
use_shader: impl FnOnce(&mut ShaderProgram) -> Result<(), GolemError>,
) -> Result<(), GolemError> {
let mut cache = self.storage.shader_cache.borrow_mut();
cache.use_shader(id, || create_shader(&self.context), use_shader)
}
pub fn load_texture(&self, cpu_texture: &crate::Texture) -> Result<golem::Texture, GolemError> {
let mut gpu_texture = golem::Texture::new(&self.context)?;
let pixel_buffer = cpu_texture.create_pixel_buffer();
gpu_texture.set_image(
Some(&pixel_buffer),
cpu_texture.get_width(),
cpu_texture.get_height(),
ColorFormat::RGBA,
);
gpu_texture.set_wrap_h(TextureWrap::ClampToEdge)?;
gpu_texture.set_wrap_v(TextureWrap::ClampToEdge)?;
gpu_texture.set_magnification(TextureFilter::Linear)?;
gpu_texture.set_minification(TextureFilter::Linear)?;
Ok(gpu_texture)
}
}
pub(super) struct GolemRenderStorage {
// Frequently used and cheap buffers
quad_vertices: VertexBuffer,
quad_indices: ElementBuffer,
shader_cache: RefCell<ShaderCache>,
}
impl GolemRenderStorage {
fn new(context: &Context) -> Result<Self, GolemError> {
let mut quad_vertices = VertexBuffer::new(context)?;
#[rustfmt::skip]
quad_vertices.set_data(&[-1.0, -1.0, 1.0, -1.0, 1.0, 1.0, -1.0, 1.0]);
let mut quad_indices = ElementBuffer::new(context)?;
quad_indices.set_data(&[0, 1, 2, 2, 3, 0]);
// Practice will have to tell whether 200 is good.
let max_cached_shaders = 200;
Ok(Self {
quad_vertices,
quad_indices,
shader_cache: RefCell::new(ShaderCache::new(max_cached_shaders)),
})
}
}
struct ShaderCache {
map: HashMap<ShaderId, CachedShader>,
max_cached_shaders: usize,
current_time: u64,
}
impl ShaderCache {
fn new(max_cached_shaders: usize) -> Self {
assert!(max_cached_shaders > 0);
Self {
map: HashMap::new(),
max_cached_shaders,
current_time: 0,
}
}
fn get_existing(&mut self, id: &ShaderId) -> &mut ShaderProgram {
let cached = self.map.get_mut(id).unwrap();
cached.last_used = self.current_time;
return &mut cached.shader;
}
fn use_shader(
&mut self,
id: &ShaderId,
create_shader: impl FnOnce() -> Result<ShaderProgram, GolemError>,
use_shader: impl FnOnce(&mut ShaderProgram) -> Result<(), GolemError>,
) -> Result<(), GolemError> {
self.current_time += 1;
// If we have the value already, update its last_used and return it
// Unfortunately, we do 2 hash lookups. I tried using only 1, but couldn't convince compiler
let has_already = self.map.contains_key(id);
if has_already {
let shader = self.get_existing(id);
shader.bind();
return use_shader(shader);
}
// If we reach this line, we didn't have the shader yet
let new_length = self.map.len() + 1;
// If we would exceed the maximum number of cached shaders, we remove the least recently used half
if new_length > self.max_cached_shaders | {
let mut last_used_times: Vec<u64> = self
.map
.values()
.map(|cached_shader| cached_shader.last_used)
.collect();
last_used_times.sort();
let median = last_used_times[last_used_times.len() / 2];
// Remove at least half of the cached shaders
self.map
.retain(|_id, cached_shader| cached_shader.last_used > median);
} | conditional_block |
|
TheGiver.py | usersSavedDataObj.keys():
for browserCount in usersSavedDataObj[userName]["previous_browser_counts"]:
if browserCount["name"] == browserName:
return browserCount
def usersBrowserCount(userName, browserName):
if userName in usersObj.keys():
for browserCount in usersObj[userName]["previous_browser_counts"]:
if browserCount["name"] == browserName:
return browserCount
def usersPreviousScore(userName):
return sum(b["score"] for b in usersObj[userName]["previous_browsers"]) | currentLoser = users[0]
currentScore = usersPreviousScore(currentLoser["name"])
for user in users:
userScore = usersPreviousScore(user["name"])
if userScore > currentScore:
currentLoser = user
currentScore = userScore
return currentLoser
# Clean up object
for user in users:
user["score"] = 0.0
user["last_score"] = 0.0
user["loses"] = 0
user["bails"] = 0
user["previous_browsers"] = []
user["previous_browser_counts"] = []
# Load saved user data into users
if user["name"] in usersSavedDataObj.keys():
user["score"] = usersSavedDataObj[user["name"]]["score"]
user["loses"] = usersSavedDataObj[user["name"]]["loses"]
user["previous_browsers"] = usersSavedDataObj[user["name"]]["previous_browsers"]
user["previous_browser_counts"] = usersSavedDataObj[user["name"]]["previous_browser_counts"]
user["bails"] = usersSavedDataObj[user["name"]]["bails"]
for browser in browsers:
browserCount = usersSavedBrowserCount(user["name"], browser["name"])
if browserCount is None:
browserCount = {"name":browser["name"], "count": 0}
user["previous_browser_counts"].append(browserCount)
# Order user by score, highest score more likely to luck out and not get a second browser
orderedUsers = sorted(users, key=lambda k: k["score"])
# reset when needed
if len(sys.argv) > 1:
if sys.argv[1].upper() == "RESET":
for user in users:
user["score"] = 0.0
user["last_score"] = 0.0
user["loses"] = 0
user["bails"] = 0
user["previous_browsers"] = []
user["previous_browser_counts"] = []
for browser in browsers:
browserCount = {"name":browser["name"], "count": 0}
user["previous_browser_counts"].append(browserCount)
# Check Lose Fairness
elif sys.argv[1].upper() == "LOSERS":
print("LOSERS:")
orderedLosers = sorted(users, key=lambda k: k["loses"], reverse=True)
sum = 0
for user in orderedLosers:
sum = sum + user["loses"]
for user in orderedLosers:
perc = 0.0
if sum > perc:
perc = float(user["loses"])/float(sum)
lineLen = int(gridLen*perc)
lineString = ''
for j in range(gridLen):
if j < lineLen:
lineString = lineString + '|'
else:
lineString = lineString + '.'
print(lineString + ' ' + user["name"] + ' (' + str(user["loses"]) + ') : ' + str(int(perc*100)) + '%')
# Swap browser testing for previous results
elif sys.argv[1].upper() == "SWAP":
print("SWAP:")
print('\n'.join('[' + str(i) + '] ' + users[i]["name"] + ' (' + str(users[i]["score"]) + ')' for i in range(len(users))))
indexA = int(raw_input('Lucky SOB\'s index: '))
indexB = int(raw_input('Unlucky SOB\' index: '))
if indexA < len(users) and indexB < len(users):
loserUser = previousRoundLoser()
userA = users[indexA]
userB = users[indexB]
browsersA = userA["previous_browsers"]
browsersB = userB["previous_browsers"]
print('')
print(userA["name"] + ' can swap the following browsers:')
print('\n'.join('[' + str(i) + '] ' + browsersA[i]["name"] + ' (' + str(browsersA[i]["score"]) + ')' for i in range(len(browsersA))))
indexC = int(raw_input('Browser index: '))
if (indexC < len(browsersA)):
browserC = browsersA[indexC]
confirm = raw_input('Take ' + browserC["name"] +
' from ' + userA["name"] +
' and give it to ' + userB["name"] + ' (y/n)? ')
print('')
if confirm is 'y':
browsersA.pop(indexC)
browsersB.append(browserC)
# update saved scores
userA["score"] = userA["score"] - browserC["score"]
userB["score"] = userB["score"] + browserC["score"]
# update tested browser counts
browserCountA = usersBrowserCount(userA["name"], browserC["name"])
browserCountA["count"] = browserCountA["count"] - 1
browserCountB = usersBrowserCount(userB["name"], browserC["name"])
browserCountB["count"] = browserCountB["count"] + 1
# update last round's user if needed
if usersPreviousScore(userB["name"]) > usersPreviousScore(loserUser["name"]):
print('Previous Loser: ' + str(usersPreviousScore(loserUser["name"])) + ' ' + loserUser["name"])
print('New Loser: ' + str(usersPreviousScore(userB["name"])) + ' ' + userB["name"])
print('')
loserUser["loses"] = loserUser["loses"] - 1
userB["loses"] = userB["loses"] + 1
print(userA["name"] + '\'s browsers:')
if (len(browsersA) > 0):
print('\n'.join('[' + str(i) + '] ' + browsersA[i]["name"] for i in range(len(browsersA))))
print('')
print(userB["name"] + '\'s browsers:')
if (len(browsersB) > 0):
print('\n'.join('[' + str(i) + '] ' + browsersB[i]["name"] for i in range(len(browsersB))))
# Setup for SAVE
writeSave = True
else:
print('Invalid Browser Index!')
else:
print('Invalid User Index!')
# Check randomness
elif len(sys.argv[1]) > 0:
for user in orderedUsers:
if sys.argv[1].upper() == user["name"].upper():
print(sys.argv[1].upper() + ' CHECK:')
browserCounts = []
sum = 0
for browserCount in user["previous_browser_counts"]:
browserCounts.append(browserCount)
sum = sum + browserCount["count"]
browserCounts = sorted(browserCounts, key=lambda k: k["count"], reverse=True)
for browserCount in browserCounts:
perc = 0.0
if sum > perc:
perc = float(browserCount["count"])/float(sum)
lineLen = int(gridLen*perc)
lineString = ''
for j in range(gridLen):
if j < lineLen:
lineString = lineString + '|'
else:
lineString = lineString + '.'
print(lineString + ' ' + browserCount["name"] + ': ' + str(int(perc*100)) + '%')
# Do work
else:
# init previous browser lists
for user in users:
user["previous_browsers"] = []
# assign random browsers to users
userIndex = 0
usersBrowsers = {}
remainingBrowsers = list(browsers)
random.shuffle(remainingBrowsers)
while len(remainingBrowsers) > 0:
user = orderedUsers[userIndex%len(orderedUsers)]
browser = remainingBrowsers.pop(random.randrange(100)%len(remainingBrowsers))
user["previous_browsers"].append(browser)
userIndex = userIndex + 1
# Identify just_awful double Jeopardy
zeroJeopardyUsers = []
doubleJeopardyUsers = []
for user in orderedUsers:
ieCount = 0
for browser in user["previous_browsers"]:
if browser["just_awful"]:
ieCount = ieCount + 1
if ieCount == 0:
zeroJeopardyUsers.append(user)
elif ieCount == 2:
doubleJeopardyUsers.append(user)
# Resolve just_awful double Jeopardy
for i in range(min(len(zeroJeopardyUsers), len(doubleJeopardyUsers))):
tempBrowser = zeroJeopardyUsers[i]["previous_browsers"][0]
zeroJeopardyUsers[i]["previous_browsers"][0] = doubleJeopardyUsers[i]["previous_browsers"][0]
|
def previousRoundLoser(): | random_line_split |
TheGiver.py | currentLoser
# Clean up object
for user in users:
user["score"] = 0.0
user["last_score"] = 0.0
user["loses"] = 0
user["bails"] = 0
user["previous_browsers"] = []
user["previous_browser_counts"] = []
# Load saved user data into users
if user["name"] in usersSavedDataObj.keys():
user["score"] = usersSavedDataObj[user["name"]]["score"]
user["loses"] = usersSavedDataObj[user["name"]]["loses"]
user["previous_browsers"] = usersSavedDataObj[user["name"]]["previous_browsers"]
user["previous_browser_counts"] = usersSavedDataObj[user["name"]]["previous_browser_counts"]
user["bails"] = usersSavedDataObj[user["name"]]["bails"]
for browser in browsers:
browserCount = usersSavedBrowserCount(user["name"], browser["name"])
if browserCount is None:
browserCount = {"name":browser["name"], "count": 0}
user["previous_browser_counts"].append(browserCount)
# Order user by score, highest score more likely to luck out and not get a second browser
orderedUsers = sorted(users, key=lambda k: k["score"])
# reset when needed
if len(sys.argv) > 1:
if sys.argv[1].upper() == "RESET":
for user in users:
user["score"] = 0.0
user["last_score"] = 0.0
user["loses"] = 0
user["bails"] = 0
user["previous_browsers"] = []
user["previous_browser_counts"] = []
for browser in browsers:
browserCount = {"name":browser["name"], "count": 0}
user["previous_browser_counts"].append(browserCount)
# Check Lose Fairness
elif sys.argv[1].upper() == "LOSERS":
print("LOSERS:")
orderedLosers = sorted(users, key=lambda k: k["loses"], reverse=True)
sum = 0
for user in orderedLosers:
sum = sum + user["loses"]
for user in orderedLosers:
perc = 0.0
if sum > perc:
perc = float(user["loses"])/float(sum)
lineLen = int(gridLen*perc)
lineString = ''
for j in range(gridLen):
if j < lineLen:
lineString = lineString + '|'
else:
lineString = lineString + '.'
print(lineString + ' ' + user["name"] + ' (' + str(user["loses"]) + ') : ' + str(int(perc*100)) + '%')
# Swap browser testing for previous results
elif sys.argv[1].upper() == "SWAP":
print("SWAP:")
print('\n'.join('[' + str(i) + '] ' + users[i]["name"] + ' (' + str(users[i]["score"]) + ')' for i in range(len(users))))
indexA = int(raw_input('Lucky SOB\'s index: '))
indexB = int(raw_input('Unlucky SOB\' index: '))
if indexA < len(users) and indexB < len(users):
loserUser = previousRoundLoser()
userA = users[indexA]
userB = users[indexB]
browsersA = userA["previous_browsers"]
browsersB = userB["previous_browsers"]
print('')
print(userA["name"] + ' can swap the following browsers:')
print('\n'.join('[' + str(i) + '] ' + browsersA[i]["name"] + ' (' + str(browsersA[i]["score"]) + ')' for i in range(len(browsersA))))
indexC = int(raw_input('Browser index: '))
if (indexC < len(browsersA)):
browserC = browsersA[indexC]
confirm = raw_input('Take ' + browserC["name"] +
' from ' + userA["name"] +
' and give it to ' + userB["name"] + ' (y/n)? ')
print('')
if confirm is 'y':
browsersA.pop(indexC)
browsersB.append(browserC)
# update saved scores
userA["score"] = userA["score"] - browserC["score"]
userB["score"] = userB["score"] + browserC["score"]
# update tested browser counts
browserCountA = usersBrowserCount(userA["name"], browserC["name"])
browserCountA["count"] = browserCountA["count"] - 1
browserCountB = usersBrowserCount(userB["name"], browserC["name"])
browserCountB["count"] = browserCountB["count"] + 1
# update last round's user if needed
if usersPreviousScore(userB["name"]) > usersPreviousScore(loserUser["name"]):
print('Previous Loser: ' + str(usersPreviousScore(loserUser["name"])) + ' ' + loserUser["name"])
print('New Loser: ' + str(usersPreviousScore(userB["name"])) + ' ' + userB["name"])
print('')
loserUser["loses"] = loserUser["loses"] - 1
userB["loses"] = userB["loses"] + 1
print(userA["name"] + '\'s browsers:')
if (len(browsersA) > 0):
print('\n'.join('[' + str(i) + '] ' + browsersA[i]["name"] for i in range(len(browsersA))))
print('')
print(userB["name"] + '\'s browsers:')
if (len(browsersB) > 0):
print('\n'.join('[' + str(i) + '] ' + browsersB[i]["name"] for i in range(len(browsersB))))
# Setup for SAVE
writeSave = True
else:
print('Invalid Browser Index!')
else:
print('Invalid User Index!')
# Check randomness
elif len(sys.argv[1]) > 0:
for user in orderedUsers:
if sys.argv[1].upper() == user["name"].upper():
print(sys.argv[1].upper() + ' CHECK:')
browserCounts = []
sum = 0
for browserCount in user["previous_browser_counts"]:
browserCounts.append(browserCount)
sum = sum + browserCount["count"]
browserCounts = sorted(browserCounts, key=lambda k: k["count"], reverse=True)
for browserCount in browserCounts:
perc = 0.0
if sum > perc:
perc = float(browserCount["count"])/float(sum)
lineLen = int(gridLen*perc)
lineString = ''
for j in range(gridLen):
if j < lineLen:
lineString = lineString + '|'
else:
lineString = lineString + '.'
print(lineString + ' ' + browserCount["name"] + ': ' + str(int(perc*100)) + '%')
# Do work
else:
# init previous browser lists
for user in users:
user["previous_browsers"] = []
# assign random browsers to users
userIndex = 0
usersBrowsers = {}
remainingBrowsers = list(browsers)
random.shuffle(remainingBrowsers)
while len(remainingBrowsers) > 0:
user = orderedUsers[userIndex%len(orderedUsers)]
browser = remainingBrowsers.pop(random.randrange(100)%len(remainingBrowsers))
user["previous_browsers"].append(browser)
userIndex = userIndex + 1
# Identify just_awful double Jeopardy
zeroJeopardyUsers = []
doubleJeopardyUsers = []
for user in orderedUsers:
ieCount = 0
for browser in user["previous_browsers"]:
if browser["just_awful"]:
ieCount = ieCount + 1
if ieCount == 0:
zeroJeopardyUsers.append(user)
elif ieCount == 2:
doubleJeopardyUsers.append(user)
# Resolve just_awful double Jeopardy
for i in range(min(len(zeroJeopardyUsers), len(doubleJeopardyUsers))):
tempBrowser = zeroJeopardyUsers[i]["previous_browsers"][0]
zeroJeopardyUsers[i]["previous_browsers"][0] = doubleJeopardyUsers[i]["previous_browsers"][0]
doubleJeopardyUsers[i]["previous_browsers"][0] = tempBrowser
# print results and clean up user objects
thisLoser = ''
thisLosingScore = 0
biggestLoser = ''
biggestLosingScore = 0
scoreLinesForPrint = {}
for user in orderedUsers:
scoreThisRound = 0
usersBrowsersString = ''
for browser in user["previous_browsers"]:
| scoreThisRound = scoreThisRound + browser["score"]
usersBrowsersString = usersBrowsersString + '[' + browser["name"] + '] '
# update the number of times user has tested this browser
browserCount = usersBrowserCount(user["name"], browser["name"])
browserCount["count"] = browserCount["count"] + 1 | conditional_block |
|
TheGiver.py | usersSavedDataObj.keys():
for browserCount in usersSavedDataObj[userName]["previous_browser_counts"]:
if browserCount["name"] == browserName:
return browserCount
def usersBrowserCount(userName, browserName):
if userName in usersObj.keys():
for browserCount in usersObj[userName]["previous_browser_counts"]:
if browserCount["name"] == browserName:
return browserCount
def usersPreviousScore(userName):
return sum(b["score"] for b in usersObj[userName]["previous_browsers"])
def | ():
currentLoser = users[0]
currentScore = usersPreviousScore(currentLoser["name"])
for user in users:
userScore = usersPreviousScore(user["name"])
if userScore > currentScore:
currentLoser = user
currentScore = userScore
return currentLoser
# Clean up object
for user in users:
user["score"] = 0.0
user["last_score"] = 0.0
user["loses"] = 0
user["bails"] = 0
user["previous_browsers"] = []
user["previous_browser_counts"] = []
# Load saved user data into users
if user["name"] in usersSavedDataObj.keys():
user["score"] = usersSavedDataObj[user["name"]]["score"]
user["loses"] = usersSavedDataObj[user["name"]]["loses"]
user["previous_browsers"] = usersSavedDataObj[user["name"]]["previous_browsers"]
user["previous_browser_counts"] = usersSavedDataObj[user["name"]]["previous_browser_counts"]
user["bails"] = usersSavedDataObj[user["name"]]["bails"]
for browser in browsers:
browserCount = usersSavedBrowserCount(user["name"], browser["name"])
if browserCount is None:
browserCount = {"name":browser["name"], "count": 0}
user["previous_browser_counts"].append(browserCount)
# Order user by score, highest score more likely to luck out and not get a second browser
orderedUsers = sorted(users, key=lambda k: k["score"])
# reset when needed
if len(sys.argv) > 1:
if sys.argv[1].upper() == "RESET":
for user in users:
user["score"] = 0.0
user["last_score"] = 0.0
user["loses"] = 0
user["bails"] = 0
user["previous_browsers"] = []
user["previous_browser_counts"] = []
for browser in browsers:
browserCount = {"name":browser["name"], "count": 0}
user["previous_browser_counts"].append(browserCount)
# Check Lose Fairness
elif sys.argv[1].upper() == "LOSERS":
print("LOSERS:")
orderedLosers = sorted(users, key=lambda k: k["loses"], reverse=True)
sum = 0
for user in orderedLosers:
sum = sum + user["loses"]
for user in orderedLosers:
perc = 0.0
if sum > perc:
perc = float(user["loses"])/float(sum)
lineLen = int(gridLen*perc)
lineString = ''
for j in range(gridLen):
if j < lineLen:
lineString = lineString + '|'
else:
lineString = lineString + '.'
print(lineString + ' ' + user["name"] + ' (' + str(user["loses"]) + ') : ' + str(int(perc*100)) + '%')
# Swap browser testing for previous results
elif sys.argv[1].upper() == "SWAP":
print("SWAP:")
print('\n'.join('[' + str(i) + '] ' + users[i]["name"] + ' (' + str(users[i]["score"]) + ')' for i in range(len(users))))
indexA = int(raw_input('Lucky SOB\'s index: '))
indexB = int(raw_input('Unlucky SOB\' index: '))
if indexA < len(users) and indexB < len(users):
loserUser = previousRoundLoser()
userA = users[indexA]
userB = users[indexB]
browsersA = userA["previous_browsers"]
browsersB = userB["previous_browsers"]
print('')
print(userA["name"] + ' can swap the following browsers:')
print('\n'.join('[' + str(i) + '] ' + browsersA[i]["name"] + ' (' + str(browsersA[i]["score"]) + ')' for i in range(len(browsersA))))
indexC = int(raw_input('Browser index: '))
if (indexC < len(browsersA)):
browserC = browsersA[indexC]
confirm = raw_input('Take ' + browserC["name"] +
' from ' + userA["name"] +
' and give it to ' + userB["name"] + ' (y/n)? ')
print('')
if confirm is 'y':
browsersA.pop(indexC)
browsersB.append(browserC)
# update saved scores
userA["score"] = userA["score"] - browserC["score"]
userB["score"] = userB["score"] + browserC["score"]
# update tested browser counts
browserCountA = usersBrowserCount(userA["name"], browserC["name"])
browserCountA["count"] = browserCountA["count"] - 1
browserCountB = usersBrowserCount(userB["name"], browserC["name"])
browserCountB["count"] = browserCountB["count"] + 1
# update last round's user if needed
if usersPreviousScore(userB["name"]) > usersPreviousScore(loserUser["name"]):
print('Previous Loser: ' + str(usersPreviousScore(loserUser["name"])) + ' ' + loserUser["name"])
print('New Loser: ' + str(usersPreviousScore(userB["name"])) + ' ' + userB["name"])
print('')
loserUser["loses"] = loserUser["loses"] - 1
userB["loses"] = userB["loses"] + 1
print(userA["name"] + '\'s browsers:')
if (len(browsersA) > 0):
print('\n'.join('[' + str(i) + '] ' + browsersA[i]["name"] for i in range(len(browsersA))))
print('')
print(userB["name"] + '\'s browsers:')
if (len(browsersB) > 0):
print('\n'.join('[' + str(i) + '] ' + browsersB[i]["name"] for i in range(len(browsersB))))
# Setup for SAVE
writeSave = True
else:
print('Invalid Browser Index!')
else:
print('Invalid User Index!')
# Check randomness
elif len(sys.argv[1]) > 0:
for user in orderedUsers:
if sys.argv[1].upper() == user["name"].upper():
print(sys.argv[1].upper() + ' CHECK:')
browserCounts = []
sum = 0
for browserCount in user["previous_browser_counts"]:
browserCounts.append(browserCount)
sum = sum + browserCount["count"]
browserCounts = sorted(browserCounts, key=lambda k: k["count"], reverse=True)
for browserCount in browserCounts:
perc = 0.0
if sum > perc:
perc = float(browserCount["count"])/float(sum)
lineLen = int(gridLen*perc)
lineString = ''
for j in range(gridLen):
if j < lineLen:
lineString = lineString + '|'
else:
lineString = lineString + '.'
print(lineString + ' ' + browserCount["name"] + ': ' + str(int(perc*100)) + '%')
# Do work
else:
# init previous browser lists
for user in users:
user["previous_browsers"] = []
# assign random browsers to users
userIndex = 0
usersBrowsers = {}
remainingBrowsers = list(browsers)
random.shuffle(remainingBrowsers)
while len(remainingBrowsers) > 0:
user = orderedUsers[userIndex%len(orderedUsers)]
browser = remainingBrowsers.pop(random.randrange(100)%len(remainingBrowsers))
user["previous_browsers"].append(browser)
userIndex = userIndex + 1
# Identify just_awful double Jeopardy
zeroJeopardyUsers = []
doubleJeopardyUsers = []
for user in orderedUsers:
ieCount = 0
for browser in user["previous_browsers"]:
if browser["just_awful"]:
ieCount = ieCount + 1
if ieCount == 0:
zeroJeopardyUsers.append(user)
elif ieCount == 2:
doubleJeopardyUsers.append(user)
# Resolve just_awful double Jeopardy
for i in range(min(len(zeroJeopardyUsers), len(doubleJeopardyUsers))):
tempBrowser = zeroJeopardyUsers[i]["previous_browsers"][0]
zeroJeopardyUsers[i]["previous_browsers"][0] = doubleJeopardyUsers[i]["previous_browsers"][0]
| previousRoundLoser | identifier_name |
TheGiver.py | usersSavedDataObj.keys():
for browserCount in usersSavedDataObj[userName]["previous_browser_counts"]:
if browserCount["name"] == browserName:
return browserCount
def usersBrowserCount(userName, browserName):
if userName in usersObj.keys():
for browserCount in usersObj[userName]["previous_browser_counts"]:
if browserCount["name"] == browserName:
return browserCount
def usersPreviousScore(userName):
|
def previousRoundLoser():
currentLoser = users[0]
currentScore = usersPreviousScore(currentLoser["name"])
for user in users:
userScore = usersPreviousScore(user["name"])
if userScore > currentScore:
currentLoser = user
currentScore = userScore
return currentLoser
# Clean up object
for user in users:
user["score"] = 0.0
user["last_score"] = 0.0
user["loses"] = 0
user["bails"] = 0
user["previous_browsers"] = []
user["previous_browser_counts"] = []
# Load saved user data into users
if user["name"] in usersSavedDataObj.keys():
user["score"] = usersSavedDataObj[user["name"]]["score"]
user["loses"] = usersSavedDataObj[user["name"]]["loses"]
user["previous_browsers"] = usersSavedDataObj[user["name"]]["previous_browsers"]
user["previous_browser_counts"] = usersSavedDataObj[user["name"]]["previous_browser_counts"]
user["bails"] = usersSavedDataObj[user["name"]]["bails"]
for browser in browsers:
browserCount = usersSavedBrowserCount(user["name"], browser["name"])
if browserCount is None:
browserCount = {"name":browser["name"], "count": 0}
user["previous_browser_counts"].append(browserCount)
# Order user by score, highest score more likely to luck out and not get a second browser
orderedUsers = sorted(users, key=lambda k: k["score"])
# reset when needed
if len(sys.argv) > 1:
if sys.argv[1].upper() == "RESET":
for user in users:
user["score"] = 0.0
user["last_score"] = 0.0
user["loses"] = 0
user["bails"] = 0
user["previous_browsers"] = []
user["previous_browser_counts"] = []
for browser in browsers:
browserCount = {"name":browser["name"], "count": 0}
user["previous_browser_counts"].append(browserCount)
# Check Lose Fairness
elif sys.argv[1].upper() == "LOSERS":
print("LOSERS:")
orderedLosers = sorted(users, key=lambda k: k["loses"], reverse=True)
sum = 0
for user in orderedLosers:
sum = sum + user["loses"]
for user in orderedLosers:
perc = 0.0
if sum > perc:
perc = float(user["loses"])/float(sum)
lineLen = int(gridLen*perc)
lineString = ''
for j in range(gridLen):
if j < lineLen:
lineString = lineString + '|'
else:
lineString = lineString + '.'
print(lineString + ' ' + user["name"] + ' (' + str(user["loses"]) + ') : ' + str(int(perc*100)) + '%')
# Swap browser testing for previous results
elif sys.argv[1].upper() == "SWAP":
print("SWAP:")
print('\n'.join('[' + str(i) + '] ' + users[i]["name"] + ' (' + str(users[i]["score"]) + ')' for i in range(len(users))))
indexA = int(raw_input('Lucky SOB\'s index: '))
indexB = int(raw_input('Unlucky SOB\' index: '))
if indexA < len(users) and indexB < len(users):
loserUser = previousRoundLoser()
userA = users[indexA]
userB = users[indexB]
browsersA = userA["previous_browsers"]
browsersB = userB["previous_browsers"]
print('')
print(userA["name"] + ' can swap the following browsers:')
print('\n'.join('[' + str(i) + '] ' + browsersA[i]["name"] + ' (' + str(browsersA[i]["score"]) + ')' for i in range(len(browsersA))))
indexC = int(raw_input('Browser index: '))
if (indexC < len(browsersA)):
browserC = browsersA[indexC]
confirm = raw_input('Take ' + browserC["name"] +
' from ' + userA["name"] +
' and give it to ' + userB["name"] + ' (y/n)? ')
print('')
if confirm is 'y':
browsersA.pop(indexC)
browsersB.append(browserC)
# update saved scores
userA["score"] = userA["score"] - browserC["score"]
userB["score"] = userB["score"] + browserC["score"]
# update tested browser counts
browserCountA = usersBrowserCount(userA["name"], browserC["name"])
browserCountA["count"] = browserCountA["count"] - 1
browserCountB = usersBrowserCount(userB["name"], browserC["name"])
browserCountB["count"] = browserCountB["count"] + 1
# update last round's user if needed
if usersPreviousScore(userB["name"]) > usersPreviousScore(loserUser["name"]):
print('Previous Loser: ' + str(usersPreviousScore(loserUser["name"])) + ' ' + loserUser["name"])
print('New Loser: ' + str(usersPreviousScore(userB["name"])) + ' ' + userB["name"])
print('')
loserUser["loses"] = loserUser["loses"] - 1
userB["loses"] = userB["loses"] + 1
print(userA["name"] + '\'s browsers:')
if (len(browsersA) > 0):
print('\n'.join('[' + str(i) + '] ' + browsersA[i]["name"] for i in range(len(browsersA))))
print('')
print(userB["name"] + '\'s browsers:')
if (len(browsersB) > 0):
print('\n'.join('[' + str(i) + '] ' + browsersB[i]["name"] for i in range(len(browsersB))))
# Setup for SAVE
writeSave = True
else:
print('Invalid Browser Index!')
else:
print('Invalid User Index!')
# Check randomness
elif len(sys.argv[1]) > 0:
for user in orderedUsers:
if sys.argv[1].upper() == user["name"].upper():
print(sys.argv[1].upper() + ' CHECK:')
browserCounts = []
sum = 0
for browserCount in user["previous_browser_counts"]:
browserCounts.append(browserCount)
sum = sum + browserCount["count"]
browserCounts = sorted(browserCounts, key=lambda k: k["count"], reverse=True)
for browserCount in browserCounts:
perc = 0.0
if sum > perc:
perc = float(browserCount["count"])/float(sum)
lineLen = int(gridLen*perc)
lineString = ''
for j in range(gridLen):
if j < lineLen:
lineString = lineString + '|'
else:
lineString = lineString + '.'
print(lineString + ' ' + browserCount["name"] + ': ' + str(int(perc*100)) + '%')
# Do work
else:
# init previous browser lists
for user in users:
user["previous_browsers"] = []
# assign random browsers to users
userIndex = 0
usersBrowsers = {}
remainingBrowsers = list(browsers)
random.shuffle(remainingBrowsers)
while len(remainingBrowsers) > 0:
user = orderedUsers[userIndex%len(orderedUsers)]
browser = remainingBrowsers.pop(random.randrange(100)%len(remainingBrowsers))
user["previous_browsers"].append(browser)
userIndex = userIndex + 1
# Identify just_awful double Jeopardy
zeroJeopardyUsers = []
doubleJeopardyUsers = []
for user in orderedUsers:
ieCount = 0
for browser in user["previous_browsers"]:
if browser["just_awful"]:
ieCount = ieCount + 1
if ieCount == 0:
zeroJeopardyUsers.append(user)
elif ieCount == 2:
doubleJeopardyUsers.append(user)
# Resolve just_awful double Jeopardy
for i in range(min(len(zeroJeopardyUsers), len(doubleJeopardyUsers))):
tempBrowser = zeroJeopardyUsers[i]["previous_browsers"][0]
zeroJeopardyUsers[i]["previous_browsers"][0] = doubleJeopardyUsers[i]["previous_browsers"][0 | return sum(b["score"] for b in usersObj[userName]["previous_browsers"]) | identifier_body |
repr.rs | _profile(&profile);
if let Ok(static_credentials) = try_static {
break BaseProvider::AccessKey(static_credentials);
}
}
let next_profile = match chain_provider(&profile) {
// this provider wasn't a chain provider, reload it as a base provider
None => {
break base_provider(profile).map_err(|err| {
ProfileFileError::InvalidCredentialSource {
profile: profile.name().into(),
message: format!("could not load source profile: {}", err).into(),
}
})?;
}
Some(result) => {
let (chain_profile, next) = result?;
chain.push(chain_profile);
next
}
};
match next_profile {
NextProfile::SelfReference => {
// self referential profile, don't go through the loop because it will error
// on the infinite loop check. Instead, reload this profile as a base profile
// and exit.
break base_provider(profile)?;
}
NextProfile::Named(name) => source_profile_name = name,
}
};
chain.reverse();
Ok(ProfileChain { base, chain })
}
mod role {
pub const ROLE_ARN: &str = "role_arn";
pub const EXTERNAL_ID: &str = "external_id";
pub const SESSION_NAME: &str = "role_session_name";
pub const CREDENTIAL_SOURCE: &str = "credential_source";
pub const SOURCE_PROFILE: &str = "source_profile";
}
mod web_identity_token {
pub const TOKEN_FILE: &str = "web_identity_token_file";
}
mod static_credentials {
pub const AWS_ACCESS_KEY_ID: &str = "aws_access_key_id";
pub const AWS_SECRET_ACCESS_KEY: &str = "aws_secret_access_key";
pub const AWS_SESSION_TOKEN: &str = "aws_session_token";
}
const PROVIDER_NAME: &str = "ProfileFile";
fn base_provider(profile: &Profile) -> Result<BaseProvider, ProfileFileError> {
// the profile must define either a `CredentialsSource` or a concrete set of access keys
match profile.get(role::CREDENTIAL_SOURCE) {
Some(source) => Ok(BaseProvider::NamedSource(source)),
None => web_identity_token_from_profile(profile)
.unwrap_or_else(|| Ok(BaseProvider::AccessKey(static_creds_from_profile(profile)?))),
}
}
enum NextProfile<'a> {
SelfReference,
Named(&'a str),
}
fn chain_provider(profile: &Profile) -> Option<Result<(RoleArn, NextProfile), ProfileFileError>> {
let role_provider = role_arn_from_profile(&profile)?;
let (source_profile, credential_source) = (
profile.get(role::SOURCE_PROFILE),
profile.get(role::CREDENTIAL_SOURCE),
);
let profile = match (source_profile, credential_source) {
(Some(_), Some(_)) => Err(ProfileFileError::InvalidCredentialSource {
profile: profile.name().to_string(),
message: "profile contained both source_profile and credential_source. \
Only one or the other can be defined"
.into(),
}),
(None, None) => Err(ProfileFileError::InvalidCredentialSource {
profile: profile.name().to_string(),
message:
"profile must contain `source_profile` or `credential_source` but neither were defined"
.into(),
}),
(Some(source_profile), None) if source_profile == profile.name() => {
Ok((role_provider, NextProfile::SelfReference))
}
(Some(source_profile), None) => Ok((role_provider, NextProfile::Named(source_profile))),
// we want to loop back into this profile and pick up the credential source
(None, Some(_credential_source)) => Ok((role_provider, NextProfile::SelfReference)),
};
Some(profile)
}
fn role_arn_from_profile(profile: &Profile) -> Option<RoleArn> {
// Web Identity Tokens are root providers, not chained roles
if profile.get(web_identity_token::TOKEN_FILE).is_some() {
return None;
}
let role_arn = profile.get(role::ROLE_ARN)?;
let session_name = profile.get(role::SESSION_NAME);
let external_id = profile.get(role::EXTERNAL_ID);
Some(RoleArn {
role_arn,
external_id,
session_name,
})
}
fn web_identity_token_from_profile(
profile: &Profile,
) -> Option<Result<BaseProvider, ProfileFileError>> {
let session_name = profile.get(role::SESSION_NAME);
match (
profile.get(role::ROLE_ARN),
profile.get(web_identity_token::TOKEN_FILE),
) {
(Some(role_arn), Some(token_file)) => Some(Ok(BaseProvider::WebIdentityTokenRole {
role_arn,
web_identity_token_file: token_file,
session_name,
})),
(None, None) => None,
(Some(_role_arn), None) => None,
(None, Some(_token_file)) => Some(Err(ProfileFileError::InvalidCredentialSource {
profile: profile.name().to_string(),
message: "`web_identity_token_file` was specified but `role_arn` was missing".into(),
})),
}
}
/// Load static credentials from a profile
///
/// Example:
/// ```ini
/// [profile B]
/// aws_access_key_id = abc123
/// aws_secret_access_key = def456
/// ```
fn static_creds_from_profile(profile: &Profile) -> Result<Credentials, ProfileFileError> {
use static_credentials::*;
let access_key = profile.get(AWS_ACCESS_KEY_ID);
let secret_key = profile.get(AWS_SECRET_ACCESS_KEY);
let session_token = profile.get(AWS_SESSION_TOKEN);
if let (None, None, None) = (access_key, secret_key, session_token) {
return Err(ProfileFileError::ProfileDidNotContainCredentials {
profile: profile.name().to_string(),
});
}
let access_key = access_key.ok_or_else(|| ProfileFileError::InvalidCredentialSource {
profile: profile.name().to_string(),
message: "profile missing aws_access_key_id".into(),
})?;
let secret_key = secret_key.ok_or_else(|| ProfileFileError::InvalidCredentialSource {
profile: profile.name().to_string(),
message: "profile missing aws_secret_access_key".into(),
})?;
Ok(Credentials::new(
access_key,
secret_key,
session_token.map(|s| s.to_string()),
None,
PROVIDER_NAME,
))
}
#[cfg(test)]
mod tests {
use crate::profile::credentials::repr::{resolve_chain, BaseProvider, ProfileChain};
use crate::profile::ProfileSet;
use serde::Deserialize;
use std::collections::HashMap;
use std::error::Error;
use std::fs;
#[test]
fn run_test_cases() -> Result<(), Box<dyn Error>> {
let test_cases: Vec<TestCase> =
serde_json::from_str(&fs::read_to_string("./test-data/assume-role-tests.json")?)?;
for test_case in test_cases {
print!("checking: {}...", test_case.docs);
check(test_case);
println!("ok")
}
Ok(())
}
fn check(test_case: TestCase) {
let source = ProfileSet::new(test_case.input.profile, test_case.input.selected_profile);
let actual = resolve_chain(&source, None);
let expected = test_case.output;
match (expected, actual) {
(TestOutput::Error(s), Err(e)) => assert!(
format!("{}", e).contains(&s),
"expected {} to contain `{}`",
e,
s
),
(TestOutput::ProfileChain(expected), Ok(actual)) => {
assert_eq!(to_test_output(actual), expected)
}
(expected, actual) => panic!(
"error/success mismatch. Expected:\n {:?}\nActual:\n {:?}",
&expected, actual
),
}
}
#[derive(Deserialize)]
struct TestCase {
docs: String,
input: TestInput,
output: TestOutput,
}
#[derive(Deserialize)]
struct TestInput {
profile: HashMap<String, HashMap<String, String>>,
selected_profile: String,
}
fn to_test_output(profile_chain: ProfileChain) -> Vec<Provider> {
let mut output = vec![];
match profile_chain.base {
BaseProvider::NamedSource(name) => output.push(Provider::NamedSource(name.into())),
BaseProvider::AccessKey(creds) => output.push(Provider::AccessKey {
access_key_id: creds.access_key_id().into(),
secret_access_key: creds.secret_access_key().into(),
session_token: creds.session_token().map(|tok| tok.to_string()),
}),
BaseProvider::WebIdentityTokenRole {
role_arn,
web_identity_token_file,
session_name,
} => output.push(Provider::WebIdentityToken {
role_arn: role_arn.into(),
web_identity_token_file: web_identity_token_file.into(),
role_session_name: session_name.map(|sess| sess.to_string()),
}),
};
for role in profile_chain.chain {
output.push(Provider::AssumeRole {
role_arn: role.role_arn.into(),
external_id: role.external_id.map(ToString::to_string),
role_session_name: role.session_name.map(ToString::to_string),
})
}
output
}
#[derive(Deserialize, Debug, PartialEq, Eq)]
enum TestOutput {
ProfileChain(Vec<Provider>),
Error(String),
}
#[derive(Deserialize, Debug, Eq, PartialEq)]
enum | Provider | identifier_name |
|
repr.rs | BaseProvider<'a> {
/// A profile that specifies a named credential source
/// Eg: `credential_source = Ec2InstanceMetadata`
///
/// The following profile produces two separate `ProfileProvider` rows:
/// 1. `BaseProvider::NamedSource("Ec2InstanceMetadata")`
/// 2. `RoleArn { role_arn: "...", ... }
/// ```ini
/// [profile assume-role]
/// role_arn = arn:aws:iam::123456789:role/MyRole
/// credential_source = Ec2InstanceMetadata
/// ```
NamedSource(&'a str),
/// A profile with explicitly configured access keys
///
/// Example
/// ```ini
/// [profile C]
/// aws_access_key_id = abc123
/// aws_secret_access_key = def456
/// ```
AccessKey(Credentials),
WebIdentityTokenRole {
role_arn: &'a str,
web_identity_token_file: &'a str,
session_name: Option<&'a str>,
}, // TODO: add SSO support
/*
/// An SSO Provider
Sso {
sso_account_id: &'a str,
sso_region: &'a str,
sso_role_name: &'a str,
sso_start_url: &'a str,
},
*/
}
/// A profile that specifies a role to assume
///
/// A RoleArn can only be created from either a profile with `source_profile`
/// or one with `credential_source`.
#[derive(Debug)]
pub struct RoleArn<'a> {
/// Role to assume
pub role_arn: &'a str,
/// external_id parameter to pass to the assume role provider
pub external_id: Option<&'a str>,
/// session name parameter to pass to the assume role provider
pub session_name: Option<&'a str>,
}
/// Resolve a ProfileChain from a ProfileSet or return an error
pub fn resolve_chain<'a>(
profile_set: &'a ProfileSet,
profile_override: Option<&str>,
) -> Result<ProfileChain<'a>, ProfileFileError> {
if profile_set.is_empty() {
return Err(ProfileFileError::NoProfilesDefined);
}
let mut source_profile_name =
profile_override.unwrap_or_else(|| profile_set.selected_profile());
let mut visited_profiles = vec![];
let mut chain = vec![];
let base = loop {
let profile = profile_set.get_profile(source_profile_name).ok_or(
ProfileFileError::MissingProfile {
profile: source_profile_name.into(),
message: format!(
"could not find source profile {} referenced from {}",
source_profile_name,
visited_profiles.last().unwrap_or(&"the root profile")
)
.into(),
},
)?;
if visited_profiles.contains(&source_profile_name) {
return Err(ProfileFileError::CredentialLoop {
profiles: visited_profiles
.into_iter()
.map(|s| s.to_string())
.collect(),
next: source_profile_name.to_string(),
});
}
visited_profiles.push(&source_profile_name);
// After the first item in the chain, we will prioritize static credentials if they exist
if visited_profiles.len() > 1 {
let try_static = static_creds_from_profile(&profile);
if let Ok(static_credentials) = try_static {
break BaseProvider::AccessKey(static_credentials);
}
}
let next_profile = match chain_provider(&profile) {
// this provider wasn't a chain provider, reload it as a base provider
None => {
break base_provider(profile).map_err(|err| {
ProfileFileError::InvalidCredentialSource {
profile: profile.name().into(),
message: format!("could not load source profile: {}", err).into(),
}
})?;
}
Some(result) => {
let (chain_profile, next) = result?;
chain.push(chain_profile);
next
}
};
match next_profile {
NextProfile::SelfReference => {
// self referential profile, don't go through the loop because it will error
// on the infinite loop check. Instead, reload this profile as a base profile
// and exit.
break base_provider(profile)?;
}
NextProfile::Named(name) => source_profile_name = name,
}
};
chain.reverse();
Ok(ProfileChain { base, chain })
}
mod role {
pub const ROLE_ARN: &str = "role_arn";
pub const EXTERNAL_ID: &str = "external_id";
pub const SESSION_NAME: &str = "role_session_name";
pub const CREDENTIAL_SOURCE: &str = "credential_source";
pub const SOURCE_PROFILE: &str = "source_profile";
}
mod web_identity_token {
pub const TOKEN_FILE: &str = "web_identity_token_file";
}
mod static_credentials {
pub const AWS_ACCESS_KEY_ID: &str = "aws_access_key_id";
pub const AWS_SECRET_ACCESS_KEY: &str = "aws_secret_access_key";
pub const AWS_SESSION_TOKEN: &str = "aws_session_token";
}
const PROVIDER_NAME: &str = "ProfileFile";
fn base_provider(profile: &Profile) -> Result<BaseProvider, ProfileFileError> {
// the profile must define either a `CredentialsSource` or a concrete set of access keys
match profile.get(role::CREDENTIAL_SOURCE) {
Some(source) => Ok(BaseProvider::NamedSource(source)),
None => web_identity_token_from_profile(profile)
.unwrap_or_else(|| Ok(BaseProvider::AccessKey(static_creds_from_profile(profile)?))),
}
}
enum NextProfile<'a> {
SelfReference,
Named(&'a str),
}
fn chain_provider(profile: &Profile) -> Option<Result<(RoleArn, NextProfile), ProfileFileError>> {
let role_provider = role_arn_from_profile(&profile)?;
let (source_profile, credential_source) = (
profile.get(role::SOURCE_PROFILE),
profile.get(role::CREDENTIAL_SOURCE),
);
let profile = match (source_profile, credential_source) {
(Some(_), Some(_)) => Err(ProfileFileError::InvalidCredentialSource {
profile: profile.name().to_string(),
message: "profile contained both source_profile and credential_source. \
Only one or the other can be defined"
.into(),
}),
(None, None) => Err(ProfileFileError::InvalidCredentialSource {
profile: profile.name().to_string(),
message:
"profile must contain `source_profile` or `credential_source` but neither were defined"
.into(),
}),
(Some(source_profile), None) if source_profile == profile.name() => {
Ok((role_provider, NextProfile::SelfReference))
}
(Some(source_profile), None) => Ok((role_provider, NextProfile::Named(source_profile))),
// we want to loop back into this profile and pick up the credential source
(None, Some(_credential_source)) => Ok((role_provider, NextProfile::SelfReference)),
};
Some(profile)
}
fn role_arn_from_profile(profile: &Profile) -> Option<RoleArn> {
// Web Identity Tokens are root providers, not chained roles
if profile.get(web_identity_token::TOKEN_FILE).is_some() {
return None;
}
let role_arn = profile.get(role::ROLE_ARN)?;
let session_name = profile.get(role::SESSION_NAME);
let external_id = profile.get(role::EXTERNAL_ID);
Some(RoleArn {
role_arn,
external_id,
session_name,
})
}
fn web_identity_token_from_profile(
profile: &Profile,
) -> Option<Result<BaseProvider, ProfileFileError>> |
/// Load static credentials from a profile
///
/// Example:
/// ```ini
/// [profile B]
/// aws_access_key_id = abc123
/// aws_secret_access_key = def456
/// ```
fn static_creds_from_profile(profile: &Profile) -> Result<Credentials, ProfileFileError> {
use static_credentials::*;
let access_key = profile.get(AWS_ACCESS_KEY_ID);
let secret_key = profile.get(AWS_SECRET_ACCESS_KEY);
let session_token = profile.get(AWS_SESSION_TOKEN);
if let (None, None, None) = (access_key, secret_key, session_token) {
return Err(ProfileFileError::ProfileDidNotContainCredentials {
profile: profile.name().to_string(),
});
}
let access_key = access_key.ok_or_else(|| ProfileFileError::InvalidCredentialSource {
profile: profile.name().to_string(),
message: "profile missing aws_access_key_id".into(),
})?;
let secret_key = secret_key.ok_or_else(|| ProfileFileError::InvalidCredentialSource {
profile: profile.name().to_string(),
message: "profile missing aws_secret | {
let session_name = profile.get(role::SESSION_NAME);
match (
profile.get(role::ROLE_ARN),
profile.get(web_identity_token::TOKEN_FILE),
) {
(Some(role_arn), Some(token_file)) => Some(Ok(BaseProvider::WebIdentityTokenRole {
role_arn,
web_identity_token_file: token_file,
session_name,
})),
(None, None) => None,
(Some(_role_arn), None) => None,
(None, Some(_token_file)) => Some(Err(ProfileFileError::InvalidCredentialSource {
profile: profile.name().to_string(),
message: "`web_identity_token_file` was specified but `role_arn` was missing".into(),
})),
}
} | identifier_body |
repr.rs | }
pub fn chain(&self) -> &[RoleArn<'a>] {
&self.chain.as_slice()
}
}
/// A base member of the profile chain
///
/// Base providers do not require input credentials to provide their own credentials,
/// eg. IMDS, ECS, Environment variables
#[derive(Debug, Clone)]
#[non_exhaustive]
pub enum BaseProvider<'a> {
/// A profile that specifies a named credential source
/// Eg: `credential_source = Ec2InstanceMetadata`
///
/// The following profile produces two separate `ProfileProvider` rows:
/// 1. `BaseProvider::NamedSource("Ec2InstanceMetadata")`
/// 2. `RoleArn { role_arn: "...", ... }
/// ```ini
/// [profile assume-role]
/// role_arn = arn:aws:iam::123456789:role/MyRole
/// credential_source = Ec2InstanceMetadata
/// ```
NamedSource(&'a str),
/// A profile with explicitly configured access keys
///
/// Example
/// ```ini
/// [profile C]
/// aws_access_key_id = abc123
/// aws_secret_access_key = def456
/// ```
AccessKey(Credentials),
WebIdentityTokenRole {
role_arn: &'a str,
web_identity_token_file: &'a str,
session_name: Option<&'a str>,
}, // TODO: add SSO support
/*
/// An SSO Provider
Sso {
sso_account_id: &'a str,
sso_region: &'a str,
sso_role_name: &'a str,
sso_start_url: &'a str,
},
*/
}
/// A profile that specifies a role to assume
///
/// A RoleArn can only be created from either a profile with `source_profile`
/// or one with `credential_source`.
#[derive(Debug)]
pub struct RoleArn<'a> {
/// Role to assume
pub role_arn: &'a str,
/// external_id parameter to pass to the assume role provider
pub external_id: Option<&'a str>,
/// session name parameter to pass to the assume role provider
pub session_name: Option<&'a str>,
}
/// Resolve a ProfileChain from a ProfileSet or return an error
pub fn resolve_chain<'a>(
profile_set: &'a ProfileSet,
profile_override: Option<&str>,
) -> Result<ProfileChain<'a>, ProfileFileError> {
if profile_set.is_empty() {
return Err(ProfileFileError::NoProfilesDefined);
}
let mut source_profile_name =
profile_override.unwrap_or_else(|| profile_set.selected_profile());
let mut visited_profiles = vec![];
let mut chain = vec![];
let base = loop {
let profile = profile_set.get_profile(source_profile_name).ok_or(
ProfileFileError::MissingProfile {
profile: source_profile_name.into(),
message: format!(
"could not find source profile {} referenced from {}",
source_profile_name,
visited_profiles.last().unwrap_or(&"the root profile")
)
.into(),
},
)?;
if visited_profiles.contains(&source_profile_name) {
return Err(ProfileFileError::CredentialLoop {
profiles: visited_profiles
.into_iter()
.map(|s| s.to_string())
.collect(),
next: source_profile_name.to_string(),
});
}
visited_profiles.push(&source_profile_name);
// After the first item in the chain, we will prioritize static credentials if they exist
if visited_profiles.len() > 1 {
let try_static = static_creds_from_profile(&profile);
if let Ok(static_credentials) = try_static {
break BaseProvider::AccessKey(static_credentials);
}
}
let next_profile = match chain_provider(&profile) {
// this provider wasn't a chain provider, reload it as a base provider
None => {
break base_provider(profile).map_err(|err| {
ProfileFileError::InvalidCredentialSource {
profile: profile.name().into(),
message: format!("could not load source profile: {}", err).into(),
}
})?;
}
Some(result) => {
let (chain_profile, next) = result?;
chain.push(chain_profile);
next
}
};
match next_profile {
NextProfile::SelfReference => {
// self referential profile, don't go through the loop because it will error
// on the infinite loop check. Instead, reload this profile as a base profile
// and exit.
break base_provider(profile)?;
}
NextProfile::Named(name) => source_profile_name = name,
}
};
chain.reverse();
Ok(ProfileChain { base, chain })
}
mod role {
pub const ROLE_ARN: &str = "role_arn";
pub const EXTERNAL_ID: &str = "external_id";
pub const SESSION_NAME: &str = "role_session_name";
pub const CREDENTIAL_SOURCE: &str = "credential_source";
pub const SOURCE_PROFILE: &str = "source_profile";
}
mod web_identity_token {
pub const TOKEN_FILE: &str = "web_identity_token_file";
}
mod static_credentials {
pub const AWS_ACCESS_KEY_ID: &str = "aws_access_key_id";
pub const AWS_SECRET_ACCESS_KEY: &str = "aws_secret_access_key";
pub const AWS_SESSION_TOKEN: &str = "aws_session_token";
}
const PROVIDER_NAME: &str = "ProfileFile";
fn base_provider(profile: &Profile) -> Result<BaseProvider, ProfileFileError> {
// the profile must define either a `CredentialsSource` or a concrete set of access keys
match profile.get(role::CREDENTIAL_SOURCE) {
Some(source) => Ok(BaseProvider::NamedSource(source)),
None => web_identity_token_from_profile(profile)
.unwrap_or_else(|| Ok(BaseProvider::AccessKey(static_creds_from_profile(profile)?))),
}
}
enum NextProfile<'a> {
SelfReference,
Named(&'a str),
}
fn chain_provider(profile: &Profile) -> Option<Result<(RoleArn, NextProfile), ProfileFileError>> {
let role_provider = role_arn_from_profile(&profile)?;
let (source_profile, credential_source) = (
profile.get(role::SOURCE_PROFILE),
profile.get(role::CREDENTIAL_SOURCE),
);
let profile = match (source_profile, credential_source) {
(Some(_), Some(_)) => Err(ProfileFileError::InvalidCredentialSource {
profile: profile.name().to_string(),
message: "profile contained both source_profile and credential_source. \
Only one or the other can be defined"
.into(),
}),
(None, None) => Err(ProfileFileError::InvalidCredentialSource {
profile: profile.name().to_string(),
message:
"profile must contain `source_profile` or `credential_source` but neither were defined"
.into(),
}),
(Some(source_profile), None) if source_profile == profile.name() => {
Ok((role_provider, NextProfile::SelfReference))
}
(Some(source_profile), None) => Ok((role_provider, NextProfile::Named(source_profile))),
// we want to loop back into this profile and pick up the credential source
(None, Some(_credential_source)) => Ok((role_provider, NextProfile::SelfReference)),
};
Some(profile)
}
fn role_arn_from_profile(profile: &Profile) -> Option<RoleArn> {
// Web Identity Tokens are root providers, not chained roles
if profile.get(web_identity_token::TOKEN_FILE).is_some() {
return None;
}
let role_arn = profile.get(role::ROLE_ARN)?;
let session_name = profile.get(role::SESSION_NAME);
let external_id = profile.get(role::EXTERNAL_ID);
Some(RoleArn {
role_arn,
external_id,
session_name,
})
}
fn web_identity_token_from_profile(
profile: &Profile,
) -> Option<Result<BaseProvider, ProfileFileError>> {
let session_name = profile.get(role::SESSION_NAME);
match (
profile.get(role::ROLE_ARN),
profile.get(web_identity_token::TOKEN_FILE),
) {
(Some(role_arn), Some(token_file)) => Some(Ok(BaseProvider::WebIdentityTokenRole {
role_arn,
web_identity_token_file: token_file,
session_name,
})),
(None, None) => None,
(Some(_role_arn), None) => None,
(None, Some(_token_file)) => Some(Err(ProfileFileError::InvalidCredentialSource {
profile: profile.name().to_string(),
message: "`web_identity_token_file` was specified but `role_arn` was missing".into(),
})),
}
}
/// Load static credentials from a profile
///
/// Example:
/// ```ini
/// [profile B]
/// aws_access_key_id = abc123
/// aws_secret_access_key = def456
/// ```
fn static_creds_from_profile(profile: &Profile) -> Result<Credentials, ProfileFileError> {
use static_credentials::*;
let access_key = profile.get(AWS_ACCESS_KEY_ID);
let secret_key = profile.get(AWS_SECRET_ACCESS_KEY);
let session_token = profile.get(AWS_SESSION_TOKEN);
if let (None, None, None) = (access_key, secret_key, session_token) {
return Err(ProfileFileError::ProfileDidNotContainCredentials {
profile: profile.name().to_string(),
});
}
let access_key = access | random_line_split |
||
neural_network.py | hat = logistic(self.W @ Z).T
return Z, Yhat
def loss(self, X, V, W, Y):
Z, Yhat = self.forward(X, V, W)
log_Yhat = log(Yhat)
log_Yhat_inv = log(1 - Yhat)
log_Yhat[Y == 0] = 0
log_Yhat_inv[Y == 1] = 0
if not (np.isfinite(log_Yhat).all() and
np.isfinite(log_Yhat_inv).all()):
stderr.write('parameters incompatible with data '
'(log() arising in loss calculation).\n')
stderr.flush()
log_Yhat[~np.isfinite(log_Yhat)] = log(EPSILON)
log_Yhat_inv[~np.isfinite(log_Yhat_inv)] = log(EPSILON)
return -(Y * log_Yhat + (1 - Y) * log_Yhat_inv).sum()
def fit(self, X, y):
"""
\grad_{W_k} L = \partiald{L}{\yhat_k} \grad_{W_k} \yhat_k
\partiald{L}{\yhat_k} = \frac{y_k - \yhat_k}{\yhat_k (1 - \yhat_k)}
\grad_{W_k} \yhat_k = z \yhat_k (1 - \yhat_k)
\grad_z L = \sum_k \partiald{L}{\yhat_k} \grad_z \yhat_k
\grad_z \yhat_k = W_k \yhat_k (1 - \yhat_k)
\grad_{V_h} L = \partiald{L}{z_h} \grad_{V_h} z_h
\grad_{V_h} z_h = x(1 - z_h^2)
"""
assert self.stop_factor or self.n_iterations is not None
X, Y = self.prepare_data(X, y)
H = self.H
K = self.K
n, d = X.shape
# X has extra offset dimension containing all 1s
# The hidden layer z also has a unit whose value is always 1
d -= 1
if self.V is None:
self.V = random_normal(0, 0.1, (H + 1, d + 1))
if self.W is None:
self.W = random_normal(0, 0.1, (K, H + 1))
V, W = self.V, self.W
sample_indices = list(range(n))
shuffle(sample_indices)
if self.parallel:
pool = Pool()
starmap = pool.starmap
else:
starmap = itertools.starmap
it = -self.batch_size
while True:
it += self.batch_size
if it >= self.n_iterations:
break
if it % 10000 == 0:
print('%6d/%-6d %.3f' % (it, self.n_iterations, self.loss(X, V, W, Y)))
def args(it):
i = sample_indices[it % n]
return (X[[i], :], V, W, Y[i, :])
gradients = starmap(
self.gradient,
map(args, range(it, it + self.batch_size))
)
grad__L__V, grad__L__W = [
reduce(np.add, grads) / len(grads)
for grads in zip(*gradients)
]
W -= self.learning_rate * grad__L__W
V -= self.learning_rate * grad__L__V
return self
def gradient(self, X, V, W, Y):
"""
Compute gradient of loss with respect to V and W.
"""
one, d_plus_one = X.shape
K, H_plus_one = W.shape
d = d_plus_one - 1
H = H_plus_one - 1
Z, Yhat = self.forward(X, V, W)
assert one == 1
x = X
y = Y
z = Z.ravel()
yhat = Yhat.ravel()
# Update W
# grad__L__yhat = (yhat - y) / np.clip(yhat * (1 - yhat), EPSILON, inf)
# grad__L__z[:] = 0.0
# for k in range(K):
# grad__yhat_k__W_k = z * yhat[k] * (1 - yhat[k])
# # Last element corresponds to constant offset 1 appended to z
# # vector; it does not change / has no derivative.
# grad__yhat_k__z = W[k, :-1] * yhat[k] * (1 - yhat[k])
# grad__L__z += grad__L__yhat[k] * grad__yhat_k__z
# W[k, :] -= self.learning_rate * grad__L__yhat[k] * grad__yhat_k__W_k
grad__L__z = (W.T * (yhat - y)).sum(axis=1)
zz = z.reshape((1, H + 1)).repeat(K, 0)
grad__L__W = diag(yhat - y) @ zz
# Update V
# for h in range(H):
# grad__z_h__V_h = x * (1 - z[h] ** 2)
# grad__L__V_h = grad__L__z[h] * grad__z_h__V_h
# V[h, :] -= self.learning_rate * grad__L__V_h
xx = x.reshape((1, d + 1)).repeat(H + 1, 0)
grad__L__V = diag((1 - z ** 2) * grad__L__z) @ xx
return grad__L__V, grad__L__W
def estimate_grad__z_h__V_h(self, h, x, V, grad):
eps = EPSILON_FINITE_DIFFERENCE
def d(j):
eps_vec = np.zeros_like(V)
eps_vec[h, j] = eps
z_plus = tanh((V + eps_vec) @ x)
z_minus = tanh((V - eps_vec) @ x)
z_plus[-1] = 1
z_minus[-1] = 1
return (z_plus[h] - z_minus[h]) / (2 * eps)
return self._do_finite_difference_estimate(
d,
V[h, :],
'grad__z[%d]__V[%d,:]' % (h, h),
grad,
)
def estimate_grad__yhat_k__z(self, k, z, W, y, grad):
eps = EPSILON_FINITE_DIFFERENCE
def d(h):
eps_vec = np.zeros_like(z)
eps_vec[h] = eps
yhat_plus = logistic(W @ (z + eps_vec))
yhat_minus = logistic(W @ (z - eps_vec))
return (yhat_plus[k] - yhat_minus[k]) / (2 * eps)
return self._do_finite_difference_estimate(
d,
z,
'grad__yhat[%d]__z' % k,
grad,
)
def estimate_grad__yhat_k__W_k(self, k, z, W, y, grad):
eps = EPSILON_FINITE_DIFFERENCE
def d(h):
eps_vec = np.zeros_like(W)
eps_vec[k, h] = eps
yhat_plus = logistic((W + eps_vec) @ z)
yhat_minus = logistic((W - eps_vec) @ z)
return (yhat_plus[k] - yhat_minus[k]) / (2 * eps)
return self._do_finite_difference_estimate(
d,
W[k, :],
'grad__yhat[%d]__W[%d,:]' % (k, k),
grad,
)
def estimate_grad__L__yhat(self, yhat, y, grad):
eps = EPSILON_FINITE_DIFFERENCE
def d(k):
eps_vec = np.zeros_like(yhat)
eps_vec[k] = eps
L_plus = self.loss(yhat + eps_vec, y)
L_minus = self.loss(yhat - eps_vec, y)
return (L_plus - L_minus) / (2 * eps)
return self._do_finite_difference_estimate(
d,
yhat,
'grad__L__yhat',
grad,
)
def estimate_grad__L__z(self, z, W, y, grad):
| eps = EPSILON_FINITE_DIFFERENCE
def d(h):
eps_vec = np.zeros_like(z)
eps_vec[h] = eps
yhat_plus = logistic(W @ (z + eps_vec))
yhat_minus = logistic(W @ (z - eps_vec))
L_plus = self.loss(yhat_plus, y)
L_minus = self.loss(yhat_minus, y)
return (L_plus - L_minus) / (2 * eps)
return self._do_finite_difference_estimate(
d,
z,
'grad__L__z',
grad,
) | identifier_body |
|
neural_network.py | x K). We use stochastic gradient
descent, i.e. compute and update gradients for a single input row at a
time, so in backpropagation we work with x (d x 1) and y (K x 1).
| Input | x | d x 1 |
| First weight matrix | V | H x d |
| Hidden layer | Z = tanh(Vx) | H x 1 |
| Second weight matrix | W | K x H |
| Output | yhat | K x 1 |
| Loss | L | scalar |
The loss function is the cross-entropy
-sum_k { y_k log(yhat_k) + (1 - y_k) log(1 - yhat_k) }
"""
def __init__(self,
n_hidden_units,
learning_rate,
n_iterations=None,
batch_size=1,
parallel=False,
stop_factor=None,
stop_window_size=None,
outfile=None):
if stop_window_size:
assert DEBUG
self.H = n_hidden_units
self.K = None # Determined empirically as distinct training labels
self.V = None
self.W = None
self.learning_rate = learning_rate
self.n_iterations = n_iterations
self.batch_size = batch_size
self.parallel = parallel
self.stop_factor = stop_factor
self.stop_window_size = stop_window_size
self.outfile = outfile
def predict(self, X):
X = self.prepare_data(X)
Z, Yhat = self.forward(X, self.V, self.W)
return Yhat
def forward(self, X, V, W):
Z = tanh(V @ X.T)
Z[-1, :] = 1 # The last row of V is unused; z[-1] must always be 1, just as x[-1].
Yhat = logistic(self.W @ Z).T
return Z, Yhat
def loss(self, X, V, W, Y):
Z, Yhat = self.forward(X, V, W)
log_Yhat = log(Yhat)
log_Yhat_inv = log(1 - Yhat)
log_Yhat[Y == 0] = 0
log_Yhat_inv[Y == 1] = 0
if not (np.isfinite(log_Yhat).all() and
np.isfinite(log_Yhat_inv).all()):
stderr.write('parameters incompatible with data '
'(log() arising in loss calculation).\n')
stderr.flush()
log_Yhat[~np.isfinite(log_Yhat)] = log(EPSILON)
log_Yhat_inv[~np.isfinite(log_Yhat_inv)] = log(EPSILON)
return -(Y * log_Yhat + (1 - Y) * log_Yhat_inv).sum()
def fit(self, X, y):
"""
\grad_{W_k} L = \partiald{L}{\yhat_k} \grad_{W_k} \yhat_k
\partiald{L}{\yhat_k} = \frac{y_k - \yhat_k}{\yhat_k (1 - \yhat_k)}
\grad_{W_k} \yhat_k = z \yhat_k (1 - \yhat_k)
\grad_z L = \sum_k \partiald{L}{\yhat_k} \grad_z \yhat_k
\grad_z \yhat_k = W_k \yhat_k (1 - \yhat_k)
\grad_{V_h} L = \partiald{L}{z_h} \grad_{V_h} z_h
\grad_{V_h} z_h = x(1 - z_h^2)
"""
assert self.stop_factor or self.n_iterations is not None
X, Y = self.prepare_data(X, y)
H = self.H
K = self.K
n, d = X.shape
# X has extra offset dimension containing all 1s
# The hidden layer z also has a unit whose value is always 1
d -= 1
if self.V is None:
self.V = random_normal(0, 0.1, (H + 1, d + 1))
if self.W is None:
self.W = random_normal(0, 0.1, (K, H + 1))
V, W = self.V, self.W
sample_indices = list(range(n))
shuffle(sample_indices)
if self.parallel:
pool = Pool()
starmap = pool.starmap
else:
starmap = itertools.starmap
it = -self.batch_size
while True:
it += self.batch_size
if it >= self.n_iterations:
break
if it % 10000 == 0:
print('%6d/%-6d %.3f' % (it, self.n_iterations, self.loss(X, V, W, Y)))
def args(it):
i = sample_indices[it % n]
return (X[[i], :], V, W, Y[i, :])
gradients = starmap(
self.gradient,
map(args, range(it, it + self.batch_size))
)
grad__L__V, grad__L__W = [
reduce(np.add, grads) / len(grads)
for grads in zip(*gradients)
]
W -= self.learning_rate * grad__L__W
V -= self.learning_rate * grad__L__V
return self
def gradient(self, X, V, W, Y):
"""
Compute gradient of loss with respect to V and W.
"""
one, d_plus_one = X.shape
K, H_plus_one = W.shape
d = d_plus_one - 1
H = H_plus_one - 1
Z, Yhat = self.forward(X, V, W)
assert one == 1
x = X
y = Y
z = Z.ravel()
yhat = Yhat.ravel()
# Update W
# grad__L__yhat = (yhat - y) / np.clip(yhat * (1 - yhat), EPSILON, inf)
# grad__L__z[:] = 0.0
# for k in range(K):
# grad__yhat_k__W_k = z * yhat[k] * (1 - yhat[k])
# # Last element corresponds to constant offset 1 appended to z
# # vector; it does not change / has no derivative.
# grad__yhat_k__z = W[k, :-1] * yhat[k] * (1 - yhat[k])
# grad__L__z += grad__L__yhat[k] * grad__yhat_k__z
# W[k, :] -= self.learning_rate * grad__L__yhat[k] * grad__yhat_k__W_k
grad__L__z = (W.T * (yhat - y)).sum(axis=1)
zz = z.reshape((1, H + 1)).repeat(K, 0)
grad__L__W = diag(yhat - y) @ zz
# Update V
# for h in range(H): | # grad__z_h__V_h = x * (1 - z[h] ** 2)
# grad__L__V_h = grad__L__z[h] * grad__z_h__V_h
# V[h, :] -= self.learning_rate * grad__L__V_h
xx = x.reshape((1, d + 1)).repeat(H + 1, 0)
grad__L__V = diag((1 - z ** 2) * grad__L__z) @ xx
return grad__L__V, grad__L__W
def estimate_grad__z_h__V_h(self, h, x, V, grad):
eps = EPSILON_FINITE_DIFFERENCE
def d(j):
eps_vec = np.zeros_like(V)
eps_vec[h, j] = eps
z_plus = tanh((V + eps_vec) @ x)
z_minus = tanh((V - eps_vec) @ x)
z_plus[-1] = 1
z_minus[-1] = 1
return (z_plus[h] - z_minus[h]) / (2 * eps)
return self._do_finite_difference_estimate(
d,
V[h, :],
'grad__z[%d]__V[%d,:]' % (h, h),
grad,
)
def estimate_grad__yhat_k__z(self, k, z, W, y, grad):
eps = EPSILON_FINITE_DIFFERENCE
def d(h):
eps_vec = np.zeros_like(z)
eps_vec[h] = eps
yhat_plus = logistic(W @ (z + eps_vec))
yhat_minus = logistic(W @ (z - eps_vec))
return (yhat_plus[k] - yhat_minus[k]) / (2 * eps)
return self._do_finite_difference_estimate(
d,
z,
'grad__yhat[%d]__z' % k,
grad,
)
def | random_line_split |
|
neural_network.py | y)
H = self.H
K = self.K
n, d = X.shape
# X has extra offset dimension containing all 1s
# The hidden layer z also has a unit whose value is always 1
d -= 1
if self.V is None:
self.V = random_normal(0, 0.1, (H + 1, d + 1))
if self.W is None:
self.W = random_normal(0, 0.1, (K, H + 1))
V, W = self.V, self.W
sample_indices = list(range(n))
shuffle(sample_indices)
if self.parallel:
pool = Pool()
starmap = pool.starmap
else:
starmap = itertools.starmap
it = -self.batch_size
while True:
it += self.batch_size
if it >= self.n_iterations:
break
if it % 10000 == 0:
print('%6d/%-6d %.3f' % (it, self.n_iterations, self.loss(X, V, W, Y)))
def args(it):
i = sample_indices[it % n]
return (X[[i], :], V, W, Y[i, :])
gradients = starmap(
self.gradient,
map(args, range(it, it + self.batch_size))
)
grad__L__V, grad__L__W = [
reduce(np.add, grads) / len(grads)
for grads in zip(*gradients)
]
W -= self.learning_rate * grad__L__W
V -= self.learning_rate * grad__L__V
return self
def gradient(self, X, V, W, Y):
"""
Compute gradient of loss with respect to V and W.
"""
one, d_plus_one = X.shape
K, H_plus_one = W.shape
d = d_plus_one - 1
H = H_plus_one - 1
Z, Yhat = self.forward(X, V, W)
assert one == 1
x = X
y = Y
z = Z.ravel()
yhat = Yhat.ravel()
# Update W
# grad__L__yhat = (yhat - y) / np.clip(yhat * (1 - yhat), EPSILON, inf)
# grad__L__z[:] = 0.0
# for k in range(K):
# grad__yhat_k__W_k = z * yhat[k] * (1 - yhat[k])
# # Last element corresponds to constant offset 1 appended to z
# # vector; it does not change / has no derivative.
# grad__yhat_k__z = W[k, :-1] * yhat[k] * (1 - yhat[k])
# grad__L__z += grad__L__yhat[k] * grad__yhat_k__z
# W[k, :] -= self.learning_rate * grad__L__yhat[k] * grad__yhat_k__W_k
grad__L__z = (W.T * (yhat - y)).sum(axis=1)
zz = z.reshape((1, H + 1)).repeat(K, 0)
grad__L__W = diag(yhat - y) @ zz
# Update V
# for h in range(H):
# grad__z_h__V_h = x * (1 - z[h] ** 2)
# grad__L__V_h = grad__L__z[h] * grad__z_h__V_h
# V[h, :] -= self.learning_rate * grad__L__V_h
xx = x.reshape((1, d + 1)).repeat(H + 1, 0)
grad__L__V = diag((1 - z ** 2) * grad__L__z) @ xx
return grad__L__V, grad__L__W
def estimate_grad__z_h__V_h(self, h, x, V, grad):
eps = EPSILON_FINITE_DIFFERENCE
def d(j):
eps_vec = np.zeros_like(V)
eps_vec[h, j] = eps
z_plus = tanh((V + eps_vec) @ x)
z_minus = tanh((V - eps_vec) @ x)
z_plus[-1] = 1
z_minus[-1] = 1
return (z_plus[h] - z_minus[h]) / (2 * eps)
return self._do_finite_difference_estimate(
d,
V[h, :],
'grad__z[%d]__V[%d,:]' % (h, h),
grad,
)
def estimate_grad__yhat_k__z(self, k, z, W, y, grad):
eps = EPSILON_FINITE_DIFFERENCE
def d(h):
eps_vec = np.zeros_like(z)
eps_vec[h] = eps
yhat_plus = logistic(W @ (z + eps_vec))
yhat_minus = logistic(W @ (z - eps_vec))
return (yhat_plus[k] - yhat_minus[k]) / (2 * eps)
return self._do_finite_difference_estimate(
d,
z,
'grad__yhat[%d]__z' % k,
grad,
)
def estimate_grad__yhat_k__W_k(self, k, z, W, y, grad):
eps = EPSILON_FINITE_DIFFERENCE
def d(h):
eps_vec = np.zeros_like(W)
eps_vec[k, h] = eps
yhat_plus = logistic((W + eps_vec) @ z)
yhat_minus = logistic((W - eps_vec) @ z)
return (yhat_plus[k] - yhat_minus[k]) / (2 * eps)
return self._do_finite_difference_estimate(
d,
W[k, :],
'grad__yhat[%d]__W[%d,:]' % (k, k),
grad,
)
def estimate_grad__L__yhat(self, yhat, y, grad):
eps = EPSILON_FINITE_DIFFERENCE
def d(k):
eps_vec = np.zeros_like(yhat)
eps_vec[k] = eps
L_plus = self.loss(yhat + eps_vec, y)
L_minus = self.loss(yhat - eps_vec, y)
return (L_plus - L_minus) / (2 * eps)
return self._do_finite_difference_estimate(
d,
yhat,
'grad__L__yhat',
grad,
)
def estimate_grad__L__z(self, z, W, y, grad):
eps = EPSILON_FINITE_DIFFERENCE
def d(h):
eps_vec = np.zeros_like(z)
eps_vec[h] = eps
yhat_plus = logistic(W @ (z + eps_vec))
yhat_minus = logistic(W @ (z - eps_vec))
L_plus = self.loss(yhat_plus, y)
L_minus = self.loss(yhat_minus, y)
return (L_plus - L_minus) / (2 * eps)
return self._do_finite_difference_estimate(
d,
z,
'grad__L__z',
grad,
)
def estimate_grad__L__V_h(self, h, x, V, W, y, grad):
eps = EPSILON_FINITE_DIFFERENCE
def d(j):
eps_vec = np.zeros_like(V)
eps_vec[h, j] = eps
z_plus = tanh((V + eps_vec) @ x)
z_minus = tanh((V - eps_vec) @ x)
z_plus[-1] = 1
z_minus[-1] = 1
yhat_plus = logistic(W @ z_plus)
yhat_minus = logistic(W @ z_minus)
L_plus = self.loss(yhat_plus, y)
L_minus = self.loss(yhat_minus, y)
return (L_plus - L_minus) / (2 * eps)
return self._do_finite_difference_estimate(
d,
V[h, :],
'grad__L__V_h',
grad,
)
@staticmethod
def _do_finite_difference_estimate(d, wrt, label, grad):
grad__n = np.array(list(map(d, range(len(wrt)))))
if DEBUG:
col = get_colour(re.subn(r'\d+', '%d', label))
print(col('%s = %s' % (label, grad__n)))
print(col(', '.join('%.9f' % g for g in describe(grad__n - grad).minmax)))
return grad__n
def prepare_data(self, X, y=None):
n, d = X.shape
X = np.hstack([X, np.ones((n, 1))])
if y is None:
return X
nY, = y.shape
assert nY == n
K = len(set(y))
# Demand that labels are integers 1...max(y)
if not np.issubdtype(y.dtype, np.int):
| y_int = np.floor(y).astype(np.int)
assert (y_int == y).all()
y = y_int | conditional_block |
|
neural_network.py | (self, yhat):
"""
Map values in output layer to classification predictions.
"""
raise NotImplementedError
class SingleLayerTanhLogisticNeuralNetwork(NeuralNetwork):
"""
A classification neural net with one hidden layer.
The hidden layer uses the tanh activation function.
The output layer uses the logistic activation function.
Model:
The input data are X (n x d) and Y (n x K). We use stochastic gradient
descent, i.e. compute and update gradients for a single input row at a
time, so in backpropagation we work with x (d x 1) and y (K x 1).
| Input | x | d x 1 |
| First weight matrix | V | H x d |
| Hidden layer | Z = tanh(Vx) | H x 1 |
| Second weight matrix | W | K x H |
| Output | yhat | K x 1 |
| Loss | L | scalar |
The loss function is the cross-entropy
-sum_k { y_k log(yhat_k) + (1 - y_k) log(1 - yhat_k) }
"""
def __init__(self,
n_hidden_units,
learning_rate,
n_iterations=None,
batch_size=1,
parallel=False,
stop_factor=None,
stop_window_size=None,
outfile=None):
if stop_window_size:
assert DEBUG
self.H = n_hidden_units
self.K = None # Determined empirically as distinct training labels
self.V = None
self.W = None
self.learning_rate = learning_rate
self.n_iterations = n_iterations
self.batch_size = batch_size
self.parallel = parallel
self.stop_factor = stop_factor
self.stop_window_size = stop_window_size
self.outfile = outfile
def predict(self, X):
X = self.prepare_data(X)
Z, Yhat = self.forward(X, self.V, self.W)
return Yhat
def forward(self, X, V, W):
Z = tanh(V @ X.T)
Z[-1, :] = 1 # The last row of V is unused; z[-1] must always be 1, just as x[-1].
Yhat = logistic(self.W @ Z).T
return Z, Yhat
def loss(self, X, V, W, Y):
Z, Yhat = self.forward(X, V, W)
log_Yhat = log(Yhat)
log_Yhat_inv = log(1 - Yhat)
log_Yhat[Y == 0] = 0
log_Yhat_inv[Y == 1] = 0
if not (np.isfinite(log_Yhat).all() and
np.isfinite(log_Yhat_inv).all()):
stderr.write('parameters incompatible with data '
'(log() arising in loss calculation).\n')
stderr.flush()
log_Yhat[~np.isfinite(log_Yhat)] = log(EPSILON)
log_Yhat_inv[~np.isfinite(log_Yhat_inv)] = log(EPSILON)
return -(Y * log_Yhat + (1 - Y) * log_Yhat_inv).sum()
def fit(self, X, y):
"""
\grad_{W_k} L = \partiald{L}{\yhat_k} \grad_{W_k} \yhat_k
\partiald{L}{\yhat_k} = \frac{y_k - \yhat_k}{\yhat_k (1 - \yhat_k)}
\grad_{W_k} \yhat_k = z \yhat_k (1 - \yhat_k)
\grad_z L = \sum_k \partiald{L}{\yhat_k} \grad_z \yhat_k
\grad_z \yhat_k = W_k \yhat_k (1 - \yhat_k)
\grad_{V_h} L = \partiald{L}{z_h} \grad_{V_h} z_h
\grad_{V_h} z_h = x(1 - z_h^2)
"""
assert self.stop_factor or self.n_iterations is not None
X, Y = self.prepare_data(X, y)
H = self.H
K = self.K
n, d = X.shape
# X has extra offset dimension containing all 1s
# The hidden layer z also has a unit whose value is always 1
d -= 1
if self.V is None:
self.V = random_normal(0, 0.1, (H + 1, d + 1))
if self.W is None:
self.W = random_normal(0, 0.1, (K, H + 1))
V, W = self.V, self.W
sample_indices = list(range(n))
shuffle(sample_indices)
if self.parallel:
pool = Pool()
starmap = pool.starmap
else:
starmap = itertools.starmap
it = -self.batch_size
while True:
it += self.batch_size
if it >= self.n_iterations:
break
if it % 10000 == 0:
print('%6d/%-6d %.3f' % (it, self.n_iterations, self.loss(X, V, W, Y)))
def args(it):
i = sample_indices[it % n]
return (X[[i], :], V, W, Y[i, :])
gradients = starmap(
self.gradient,
map(args, range(it, it + self.batch_size))
)
grad__L__V, grad__L__W = [
reduce(np.add, grads) / len(grads)
for grads in zip(*gradients)
]
W -= self.learning_rate * grad__L__W
V -= self.learning_rate * grad__L__V
return self
def gradient(self, X, V, W, Y):
"""
Compute gradient of loss with respect to V and W.
"""
one, d_plus_one = X.shape
K, H_plus_one = W.shape
d = d_plus_one - 1
H = H_plus_one - 1
Z, Yhat = self.forward(X, V, W)
assert one == 1
x = X
y = Y
z = Z.ravel()
yhat = Yhat.ravel()
# Update W
# grad__L__yhat = (yhat - y) / np.clip(yhat * (1 - yhat), EPSILON, inf)
# grad__L__z[:] = 0.0
# for k in range(K):
# grad__yhat_k__W_k = z * yhat[k] * (1 - yhat[k])
# # Last element corresponds to constant offset 1 appended to z
# # vector; it does not change / has no derivative.
# grad__yhat_k__z = W[k, :-1] * yhat[k] * (1 - yhat[k])
# grad__L__z += grad__L__yhat[k] * grad__yhat_k__z
# W[k, :] -= self.learning_rate * grad__L__yhat[k] * grad__yhat_k__W_k
grad__L__z = (W.T * (yhat - y)).sum(axis=1)
zz = z.reshape((1, H + 1)).repeat(K, 0)
grad__L__W = diag(yhat - y) @ zz
# Update V
# for h in range(H):
# grad__z_h__V_h = x * (1 - z[h] ** 2)
# grad__L__V_h = grad__L__z[h] * grad__z_h__V_h
# V[h, :] -= self.learning_rate * grad__L__V_h
xx = x.reshape((1, d + 1)).repeat(H + 1, 0)
grad__L__V = diag((1 - z ** 2) * grad__L__z) @ xx
return grad__L__V, grad__L__W
def estimate_grad__z_h__V_h(self, h, x, V, grad):
eps = EPSILON_FINITE_DIFFERENCE
def d(j):
eps_vec = np.zeros_like(V)
eps_vec[h, j] = eps
z_plus = tanh((V + eps_vec) @ x)
z_minus = tanh((V - eps_vec) @ x)
z_plus[-1] = 1
z_minus[-1] = 1
return (z_plus[h] - z_minus[h]) / (2 * eps)
return self._do_finite_difference_estimate(
d,
V[h, :],
'grad__z[%d]__V[%d,:]' % (h, h),
grad,
)
def estimate_grad__yhat_k__z(self, k, z, W, y, grad):
eps = EPSILON_FINITE_DIFFERENCE
def d(h):
eps_vec = np.zeros_like(z)
eps_vec | prediction_fn | identifier_name |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.