file_name
large_stringlengths
4
140
prefix
large_stringlengths
0
12.1k
suffix
large_stringlengths
0
12k
middle
large_stringlengths
0
7.51k
fim_type
large_stringclasses
4 values
script.js
send too many ajax requests */ function addRowToTable(action, name, instructions)
'src', 'picture.php?' + $.param({'medicineName': name}) ); newElement.children('td').eq(0).append(pictureElement); pictureElement.ready(function () { pictureElement.height(pictureElement.width()); }); } else if (action === "update") { row = 'table tbody tr[name=' + htmlName + '] td:nth-child(2)'; $(row).html(instructions); } else if (action === "remove") { row = 'table tbody tr[name=' + htmlName + ']'; console.log(row); $(row).remove(); } else { console.error(action); } } function extractName(medicineName) { 'use strict'; return medicineName.substr(medicineName.indexOf('_') + 1); } $(document).ready(function () { 'use strict'; var request, qrcode, width, nameElement, dataElement, text; //jQuery('#qrcode').qrcode("this plugin is great"); //new QRCode(document.getElementById("qrcode"), "http://jindo.dev.naver.com/collie"); width = $('button#add_bttn').width(); console.log(width); // turn #qrcode div into QR code qrcode = new QRCode("qrcode", { text: " ", width: 170, height: 170, colorDark : "#000000", colorLight : "#ffffff", correctLevel : QRCode.CorrectLevel.H }); nameElement = $("textarea[name='medicineName']"); dataElement = $("textarea[name='medicineData']"); nameElement.on('change keydown paste input', function () { // whenever name field (nameElement) content changes, set QR code value arccordingly. console.log("NCHAN"); var name, isEmpty, googleElement; name = $.trim($(this).val()); isEmpty = true; //console.log(name); qrcode.makeCode(name); //console.log(name); $('table#medicineTable').find('tr').each(function () { var rowName = extractName($(this).attr('name')); //console.log(rowName); if (rowName == 'google') { } else if (rowName.indexOf(name) !== -1) { $(this).css('display', 'table-row'); isEmpty = false; } else { $(this).css('display', 'none'); } }); if (isEmpty == true) { $("tr[name='google']").css("display", "table-row"); googleElement = $("tr[name='google'] td:nth-child(2)"); googleElement.empty(); googleElement.append("Internal search has no results. Google search "); googleElement.append($('<b/>', {"id": "google_query"}).text(name)); googleElement.append(' instead?'); //console.log("EMPTY"); } else { $("tr[name='google']").css("display", "none"); } }); /* setTimeout( function () {nameElement.trigger("input");}, 0 ); */ $('div#searchTable').height($('div#formDiv').height()); //Ajax call to getAll.php to get all medicine entries in database request = $.ajax({ url: 'getAll.php', type: 'post', success: function (data, textStatus, xhr) { var name, instructions, medicines, htmlName, row, k; console.log("RAWDATA", data); medicines = JSON.parse(data); //add medicine names + instructions to table for (k = 0; k < medicines.length; k += 1) { row = medicines[k]; console.log(row); console.log(row.name, row.instructions); addRowToTable('insert', row.name, row.instructions); } //console.log(medicines); nameElement.trigger("input") }, error: function (xhr, desc, err) { console.log(xhr); console.log("Details: " + desc + "\nError:" + err); } }); $('table#medicineTable').on('click', 'tr', function () { console.log($(this).attr('name')); var medicineName, medicineData, pictureElement, query; if ($(this).hasClass('clicked')) { return; } if ($(this).attr('id') === 'google_search_tr') { query = $("b#google_query").text(); window.location.replace("externalSearch.php?q=" + query); return; } medicineName = $(this).attr('name'); medicineData = $(this).children('td').eq(1).text(); medicineName = extractName(medicineName); $('table#medicineTable *').removeClass('clicked'); $(this).addClass('clicked'); qrcode.makeCode(medicineName); $("textarea[name='medicineName']").val(medicineName); $("textarea[name='medicineData']").val(medicineData); }); // Bind to the submit event of our form console.log("potato", $("p#removeBttn")); $("p#removeBttn").click(function (event) { var form, name, instructions, inputs; console.log('hiya'); event.preventDefault(); if (request) {request.abort(); } form = $(this); inputs = form.find("input, select, button, textarea"); inputs.prop("disabled", true); console.log(inputs); // Let's select and cache all the fields name = $.trim(nameElement.val()); instructions = $.trim(dataElement.val()); if (name === '') { alert("name is empty!"); return; } request = $.ajax({ url: 'remove.php', type: 'post', data: {'name': name, 'instructions': instructions}, success: function (data, textStatus, xhr) { var name, instructions, action, result, htmlName, row; console.log("RAWDATA", data); result = JSON.parse(data); /* result[0] is action, is "update" if php script updated database and "insert" if php script inserted new medicine */ action = result[0]; name = result[1]; instructions = result[2]; console.log(action); if (action === 'remove') { //console.log("REMOVING"); addRowToTable('remove', name, instructions); } console.log(result); inputs.prop("disabled", false); }, error: function (xhr, desc, err) { console.log(xhr); console.log("Details: " + desc + "\nError:" + err); inputs.prop("disabled", false); } }); }); $("form#editForm").submit('click', function (event) { var form, name, instructions, inputs; console.log('hiya'); // Prevent default posting of form - put here to work in case of errors event.preventDefault(); // Abort any pending request if (request) {request.abort(); } // setup some local variables form = $(this); //name = $("form textarea[name='medicineName']").val(); //instructions = $("form textarea[name='medicineName']").val(); //console.log("NAME", name, instructions); // Let's select and cache all the fields name = $.trim(nameElement.val()); instructions = $.trim(dataElement.val()); if (name === '' || instructions === '') { alert("name or instructions is empty!"); return; } inputs = form.find("input, select, button, textarea"); // Serialize the data in the form //serializedData = form.serialize(); console.log(inputs); // Let's disable the inputs for the duration of the Ajax request. // Note: we disable elements AFTER the form data has been serialized. // Disabled form elements will not be serialized. inputs.prop("disabled", true); // Fire off the request to /form.php //console.log("SERIAL", serializedData, inputs); request = $.ajax({ url: 'docterEnter.php', type: 'post', data: {'name': name, 'instructions': instructions}, success: function (data, textStatus, xhr) { var name, instructions, action, result, htmlName, row; console.log("RAWDATA", data); result = JSON.parse(data); /* result[0] is action, value is "update" if php script updated database and "insert" if php script inserted new medicine into database
{ 'use strict'; var htmlName, newElement, pictureElement, medicineName, row; htmlName = JSON.stringify("medicine_" + name); console.log(name, instructions); if (action === "insert") { newElement = $("<tr name=" + htmlName + ">").append( $('<td/>').text(name), $('<td/>').text(instructions) ); $('table tbody').append(newElement); console.log("APPENDED"); pictureElement = $( "<iframe id='picture' scrolling='no'/>", {'class': "images"} ); pictureElement.attr(
identifier_body
script.js
send too many ajax requests */ function
(action, name, instructions) { 'use strict'; var htmlName, newElement, pictureElement, medicineName, row; htmlName = JSON.stringify("medicine_" + name); console.log(name, instructions); if (action === "insert") { newElement = $("<tr name=" + htmlName + ">").append( $('<td/>').text(name), $('<td/>').text(instructions) ); $('table tbody').append(newElement); console.log("APPENDED"); pictureElement = $( "<iframe id='picture' scrolling='no'/>", {'class': "images"} ); pictureElement.attr( 'src', 'picture.php?' + $.param({'medicineName': name}) ); newElement.children('td').eq(0).append(pictureElement); pictureElement.ready(function () { pictureElement.height(pictureElement.width()); }); } else if (action === "update") { row = 'table tbody tr[name=' + htmlName + '] td:nth-child(2)'; $(row).html(instructions); } else if (action === "remove") { row = 'table tbody tr[name=' + htmlName + ']'; console.log(row); $(row).remove(); } else { console.error(action); } } function extractName(medicineName) { 'use strict'; return medicineName.substr(medicineName.indexOf('_') + 1); } $(document).ready(function () { 'use strict'; var request, qrcode, width, nameElement, dataElement, text; //jQuery('#qrcode').qrcode("this plugin is great"); //new QRCode(document.getElementById("qrcode"), "http://jindo.dev.naver.com/collie"); width = $('button#add_bttn').width(); console.log(width); // turn #qrcode div into QR code qrcode = new QRCode("qrcode", { text: " ", width: 170, height: 170, colorDark : "#000000", colorLight : "#ffffff", correctLevel : QRCode.CorrectLevel.H }); nameElement = $("textarea[name='medicineName']"); dataElement = $("textarea[name='medicineData']"); nameElement.on('change keydown paste input', function () { // whenever name field (nameElement) content changes, set QR code value arccordingly. console.log("NCHAN"); var name, isEmpty, googleElement; name = $.trim($(this).val()); isEmpty = true; //console.log(name); qrcode.makeCode(name); //console.log(name); $('table#medicineTable').find('tr').each(function () { var rowName = extractName($(this).attr('name')); //console.log(rowName); if (rowName == 'google') { } else if (rowName.indexOf(name) !== -1) { $(this).css('display', 'table-row'); isEmpty = false; } else { $(this).css('display', 'none'); } }); if (isEmpty == true) { $("tr[name='google']").css("display", "table-row"); googleElement = $("tr[name='google'] td:nth-child(2)"); googleElement.empty(); googleElement.append("Internal search has no results. Google search "); googleElement.append($('<b/>', {"id": "google_query"}).text(name)); googleElement.append(' instead?'); //console.log("EMPTY"); } else { $("tr[name='google']").css("display", "none"); } }); /* setTimeout( function () {nameElement.trigger("input");}, 0 ); */ $('div#searchTable').height($('div#formDiv').height()); //Ajax call to getAll.php to get all medicine entries in database request = $.ajax({ url: 'getAll.php', type: 'post', success: function (data, textStatus, xhr) { var name, instructions, medicines, htmlName, row, k; console.log("RAWDATA", data); medicines = JSON.parse(data); //add medicine names + instructions to table for (k = 0; k < medicines.length; k += 1) { row = medicines[k]; console.log(row); console.log(row.name, row.instructions); addRowToTable('insert', row.name, row.instructions); } //console.log(medicines); nameElement.trigger("input") }, error: function (xhr, desc, err) { console.log(xhr); console.log("Details: " + desc + "\nError:" + err); } }); $('table#medicineTable').on('click', 'tr', function () { console.log($(this).attr('name')); var medicineName, medicineData, pictureElement, query; if ($(this).hasClass('clicked')) { return; } if ($(this).attr('id') === 'google_search_tr') { query = $("b#google_query").text(); window.location.replace("externalSearch.php?q=" + query); return; } medicineName = $(this).attr('name'); medicineData = $(this).children('td').eq(1).text(); medicineName = extractName(medicineName); $('table#medicineTable *').removeClass('clicked'); $(this).addClass('clicked'); qrcode.makeCode(medicineName); $("textarea[name='medicineName']").val(medicineName); $("textarea[name='medicineData']").val(medicineData); }); // Bind to the submit event of our form console.log("potato", $("p#removeBttn")); $("p#removeBttn").click(function (event) { var form, name, instructions, inputs; console.log('hiya'); event.preventDefault(); if (request) {request.abort(); } form = $(this); inputs = form.find("input, select, button, textarea"); inputs.prop("disabled", true); console.log(inputs); // Let's select and cache all the fields name = $.trim(nameElement.val()); instructions = $.trim(dataElement.val()); if (name === '') { alert("name is empty!"); return; } request = $.ajax({ url: 'remove.php', type: 'post', data: {'name': name, 'instructions': instructions}, success: function (data, textStatus, xhr) { var name, instructions, action, result, htmlName, row; console.log("RAWDATA", data); result = JSON.parse(data); /* result[0] is action, is "update" if php script updated database and "insert" if php script inserted new medicine */ action = result[0]; name = result[1]; instructions = result[2]; console.log(action); if (action === 'remove') { //console.log("REMOVING"); addRowToTable('remove', name, instructions); } console.log(result); inputs.prop("disabled", false); }, error: function (xhr, desc, err) { console.log(xhr); console.log("Details: " + desc + "\nError:" + err); inputs.prop("disabled", false); } }); }); $("form#editForm").submit('click', function (event) { var form, name, instructions, inputs; console.log('hiya'); // Prevent default posting of form - put here to work in case of errors event.preventDefault(); // Abort any pending request if (request) {request.abort(); } // setup some local variables form = $(this); //name = $("form textarea[name='medicineName']").val(); //instructions = $("form textarea[name='medicineName']").val(); //console.log("NAME", name, instructions); // Let's select and cache all the fields name = $.trim(nameElement.val()); instructions = $.trim(dataElement.val()); if (name === '' || instructions === '') { alert("name or instructions is empty!"); return; } inputs = form.find("input, select, button, textarea"); // Serialize the data in the form //serializedData = form.serialize(); console.log(inputs); // Let's disable the inputs for the duration of the Ajax request. // Note: we disable elements AFTER the form data has been serialized. // Disabled form elements will not be serialized. inputs.prop("disabled", true); // Fire off the request to /form.php //console.log("SERIAL", serializedData, inputs); request = $.ajax({ url: 'docterEnter.php', type: 'post', data: {'name': name, 'instructions': instructions}, success: function (data, textStatus, xhr) { var name, instructions, action, result, htmlName, row; console.log("RAWDATA", data); result = JSON.parse(data); /* result[0] is action, value is "update" if php script updated database and "insert" if php script inserted new medicine into database
addRowToTable
identifier_name
script.js
send too many ajax requests */ function addRowToTable(action, name, instructions) { 'use strict'; var htmlName, newElement, pictureElement, medicineName, row; htmlName = JSON.stringify("medicine_" + name); console.log(name, instructions); if (action === "insert") { newElement = $("<tr name=" + htmlName + ">").append( $('<td/>').text(name), $('<td/>').text(instructions) ); $('table tbody').append(newElement); console.log("APPENDED"); pictureElement = $( "<iframe id='picture' scrolling='no'/>", {'class': "images"} ); pictureElement.attr( 'src', 'picture.php?' + $.param({'medicineName': name}) ); newElement.children('td').eq(0).append(pictureElement); pictureElement.ready(function () { pictureElement.height(pictureElement.width()); }); } else if (action === "update") { row = 'table tbody tr[name=' + htmlName + '] td:nth-child(2)'; $(row).html(instructions); } else if (action === "remove") { row = 'table tbody tr[name=' + htmlName + ']'; console.log(row); $(row).remove(); } else { console.error(action); } } function extractName(medicineName) { 'use strict'; return medicineName.substr(medicineName.indexOf('_') + 1); } $(document).ready(function () { 'use strict'; var request, qrcode, width, nameElement, dataElement, text; //jQuery('#qrcode').qrcode("this plugin is great"); //new QRCode(document.getElementById("qrcode"), "http://jindo.dev.naver.com/collie"); width = $('button#add_bttn').width(); console.log(width); // turn #qrcode div into QR code qrcode = new QRCode("qrcode", { text: " ", width: 170, height: 170, colorDark : "#000000", colorLight : "#ffffff", correctLevel : QRCode.CorrectLevel.H }); nameElement = $("textarea[name='medicineName']"); dataElement = $("textarea[name='medicineData']"); nameElement.on('change keydown paste input', function () { // whenever name field (nameElement) content changes, set QR code value arccordingly. console.log("NCHAN"); var name, isEmpty, googleElement; name = $.trim($(this).val()); isEmpty = true; //console.log(name); qrcode.makeCode(name); //console.log(name); $('table#medicineTable').find('tr').each(function () { var rowName = extractName($(this).attr('name')); //console.log(rowName); if (rowName == 'google') { } else if (rowName.indexOf(name) !== -1) { $(this).css('display', 'table-row'); isEmpty = false; } else { $(this).css('display', 'none'); } }); if (isEmpty == true) { $("tr[name='google']").css("display", "table-row"); googleElement = $("tr[name='google'] td:nth-child(2)"); googleElement.empty(); googleElement.append("Internal search has no results. Google search "); googleElement.append($('<b/>', {"id": "google_query"}).text(name)); googleElement.append(' instead?'); //console.log("EMPTY"); } else
}); /* setTimeout( function () {nameElement.trigger("input");}, 0 ); */ $('div#searchTable').height($('div#formDiv').height()); //Ajax call to getAll.php to get all medicine entries in database request = $.ajax({ url: 'getAll.php', type: 'post', success: function (data, textStatus, xhr) { var name, instructions, medicines, htmlName, row, k; console.log("RAWDATA", data); medicines = JSON.parse(data); //add medicine names + instructions to table for (k = 0; k < medicines.length; k += 1) { row = medicines[k]; console.log(row); console.log(row.name, row.instructions); addRowToTable('insert', row.name, row.instructions); } //console.log(medicines); nameElement.trigger("input") }, error: function (xhr, desc, err) { console.log(xhr); console.log("Details: " + desc + "\nError:" + err); } }); $('table#medicineTable').on('click', 'tr', function () { console.log($(this).attr('name')); var medicineName, medicineData, pictureElement, query; if ($(this).hasClass('clicked')) { return; } if ($(this).attr('id') === 'google_search_tr') { query = $("b#google_query").text(); window.location.replace("externalSearch.php?q=" + query); return; } medicineName = $(this).attr('name'); medicineData = $(this).children('td').eq(1).text(); medicineName = extractName(medicineName); $('table#medicineTable *').removeClass('clicked'); $(this).addClass('clicked'); qrcode.makeCode(medicineName); $("textarea[name='medicineName']").val(medicineName); $("textarea[name='medicineData']").val(medicineData); }); // Bind to the submit event of our form console.log("potato", $("p#removeBttn")); $("p#removeBttn").click(function (event) { var form, name, instructions, inputs; console.log('hiya'); event.preventDefault(); if (request) {request.abort(); } form = $(this); inputs = form.find("input, select, button, textarea"); inputs.prop("disabled", true); console.log(inputs); // Let's select and cache all the fields name = $.trim(nameElement.val()); instructions = $.trim(dataElement.val()); if (name === '') { alert("name is empty!"); return; } request = $.ajax({ url: 'remove.php', type: 'post', data: {'name': name, 'instructions': instructions}, success: function (data, textStatus, xhr) { var name, instructions, action, result, htmlName, row; console.log("RAWDATA", data); result = JSON.parse(data); /* result[0] is action, is "update" if php script updated database and "insert" if php script inserted new medicine */ action = result[0]; name = result[1]; instructions = result[2]; console.log(action); if (action === 'remove') { //console.log("REMOVING"); addRowToTable('remove', name, instructions); } console.log(result); inputs.prop("disabled", false); }, error: function (xhr, desc, err) { console.log(xhr); console.log("Details: " + desc + "\nError:" + err); inputs.prop("disabled", false); } }); }); $("form#editForm").submit('click', function (event) { var form, name, instructions, inputs; console.log('hiya'); // Prevent default posting of form - put here to work in case of errors event.preventDefault(); // Abort any pending request if (request) {request.abort(); } // setup some local variables form = $(this); //name = $("form textarea[name='medicineName']").val(); //instructions = $("form textarea[name='medicineName']").val(); //console.log("NAME", name, instructions); // Let's select and cache all the fields name = $.trim(nameElement.val()); instructions = $.trim(dataElement.val()); if (name === '' || instructions === '') { alert("name or instructions is empty!"); return; } inputs = form.find("input, select, button, textarea"); // Serialize the data in the form //serializedData = form.serialize(); console.log(inputs); // Let's disable the inputs for the duration of the Ajax request. // Note: we disable elements AFTER the form data has been serialized. // Disabled form elements will not be serialized. inputs.prop("disabled", true); // Fire off the request to /form.php //console.log("SERIAL", serializedData, inputs); request = $.ajax({ url: 'docterEnter.php', type: 'post', data: {'name': name, 'instructions': instructions}, success: function (data, textStatus, xhr) { var name, instructions, action, result, htmlName, row; console.log("RAWDATA", data); result = JSON.parse(data); /* result[0] is action, value is "update" if php script updated database and "insert" if php script inserted new medicine into database
{ $("tr[name='google']").css("display", "none"); }
conditional_block
script.js
't send too many ajax requests */ function addRowToTable(action, name, instructions) { 'use strict'; var htmlName, newElement, pictureElement, medicineName, row; htmlName = JSON.stringify("medicine_" + name); console.log(name, instructions); if (action === "insert") { newElement = $("<tr name=" + htmlName + ">").append( $('<td/>').text(name), $('<td/>').text(instructions) ); $('table tbody').append(newElement); console.log("APPENDED"); pictureElement = $( "<iframe id='picture' scrolling='no'/>", {'class': "images"} ); pictureElement.attr( 'src', 'picture.php?' + $.param({'medicineName': name}) ); newElement.children('td').eq(0).append(pictureElement); pictureElement.ready(function () { pictureElement.height(pictureElement.width()); }); } else if (action === "update") { row = 'table tbody tr[name=' + htmlName + '] td:nth-child(2)'; $(row).html(instructions); } else if (action === "remove") { row = 'table tbody tr[name=' + htmlName + ']'; console.log(row); $(row).remove(); } else { console.error(action); } } function extractName(medicineName) { 'use strict'; return medicineName.substr(medicineName.indexOf('_') + 1); } $(document).ready(function () { 'use strict'; var request, qrcode, width, nameElement, dataElement, text; //jQuery('#qrcode').qrcode("this plugin is great"); //new QRCode(document.getElementById("qrcode"), "http://jindo.dev.naver.com/collie"); width = $('button#add_bttn').width(); console.log(width); // turn #qrcode div into QR code qrcode = new QRCode("qrcode", { text: " ", width: 170, height: 170, colorDark : "#000000", colorLight : "#ffffff", correctLevel : QRCode.CorrectLevel.H }); nameElement = $("textarea[name='medicineName']"); dataElement = $("textarea[name='medicineData']"); nameElement.on('change keydown paste input', function () { // whenever name field (nameElement) content changes, set QR code value arccordingly. console.log("NCHAN"); var name, isEmpty, googleElement; name = $.trim($(this).val()); isEmpty = true; //console.log(name); qrcode.makeCode(name); //console.log(name); $('table#medicineTable').find('tr').each(function () { var rowName = extractName($(this).attr('name')); //console.log(rowName); if (rowName == 'google') { } else if (rowName.indexOf(name) !== -1) { $(this).css('display', 'table-row'); isEmpty = false; } else { $(this).css('display', 'none'); } }); if (isEmpty == true) { $("tr[name='google']").css("display", "table-row"); googleElement = $("tr[name='google'] td:nth-child(2)"); googleElement.empty(); googleElement.append("Internal search has no results. Google search "); googleElement.append($('<b/>', {"id": "google_query"}).text(name)); googleElement.append(' instead?'); //console.log("EMPTY"); } else { $("tr[name='google']").css("display", "none"); } }); /* setTimeout( function () {nameElement.trigger("input");}, 0 ); */ $('div#searchTable').height($('div#formDiv').height()); //Ajax call to getAll.php to get all medicine entries in database request = $.ajax({ url: 'getAll.php', type: 'post', success: function (data, textStatus, xhr) { var name, instructions, medicines, htmlName, row, k; console.log("RAWDATA", data); medicines = JSON.parse(data); //add medicine names + instructions to table for (k = 0; k < medicines.length; k += 1) { row = medicines[k]; console.log(row); console.log(row.name, row.instructions); addRowToTable('insert', row.name, row.instructions); } //console.log(medicines); nameElement.trigger("input") }, error: function (xhr, desc, err) { console.log(xhr); console.log("Details: " + desc + "\nError:" + err); } }); $('table#medicineTable').on('click', 'tr', function () { console.log($(this).attr('name')); var medicineName, medicineData, pictureElement, query; if ($(this).hasClass('clicked')) { return; } if ($(this).attr('id') === 'google_search_tr') { query = $("b#google_query").text(); window.location.replace("externalSearch.php?q=" + query); return; } medicineName = $(this).attr('name'); medicineData = $(this).children('td').eq(1).text(); medicineName = extractName(medicineName); $('table#medicineTable *').removeClass('clicked'); $(this).addClass('clicked'); qrcode.makeCode(medicineName); $("textarea[name='medicineName']").val(medicineName); $("textarea[name='medicineData']").val(medicineData); }); // Bind to the submit event of our form console.log("potato", $("p#removeBttn")); $("p#removeBttn").click(function (event) { var form, name, instructions, inputs; console.log('hiya'); event.preventDefault(); if (request) {request.abort(); } form = $(this); inputs = form.find("input, select, button, textarea"); inputs.prop("disabled", true); console.log(inputs); // Let's select and cache all the fields name = $.trim(nameElement.val()); instructions = $.trim(dataElement.val()); if (name === '') { alert("name is empty!"); return; } request = $.ajax({ url: 'remove.php', type: 'post', data: {'name': name, 'instructions': instructions}, success: function (data, textStatus, xhr) { var name, instructions, action, result, htmlName, row; console.log("RAWDATA", data); result = JSON.parse(data); /* result[0] is action, is "update" if php script updated database and "insert" if php script inserted new medicine */ action = result[0]; name = result[1]; instructions = result[2]; console.log(action); if (action === 'remove') { //console.log("REMOVING"); addRowToTable('remove', name, instructions); } console.log(result); inputs.prop("disabled", false); }, error: function (xhr, desc, err) { console.log(xhr); console.log("Details: " + desc + "\nError:" + err); inputs.prop("disabled", false); } }); }); $("form#editForm").submit('click', function (event) { var form, name, instructions, inputs; console.log('hiya'); // Prevent default posting of form - put here to work in case of errors event.preventDefault(); // Abort any pending request if (request) {request.abort(); } // setup some local variables form = $(this); //name = $("form textarea[name='medicineName']").val(); //instructions = $("form textarea[name='medicineName']").val(); //console.log("NAME", name, instructions); // Let's select and cache all the fields name = $.trim(nameElement.val()); instructions = $.trim(dataElement.val()); if (name === '' || instructions === '') { alert("name or instructions is empty!"); return; }
// Serialize the data in the form //serializedData = form.serialize(); console.log(inputs); // Let's disable the inputs for the duration of the Ajax request. // Note: we disable elements AFTER the form data has been serialized. // Disabled form elements will not be serialized. inputs.prop("disabled", true); // Fire off the request to /form.php //console.log("SERIAL", serializedData, inputs); request = $.ajax({ url: 'docterEnter.php', type: 'post', data: {'name': name, 'instructions': instructions}, success: function (data, textStatus, xhr) { var name, instructions, action, result, htmlName, row; console.log("RAWDATA", data); result = JSON.parse(data); /* result[0] is action, value is "update" if php script updated database and "insert" if php script inserted new medicine into database
inputs = form.find("input, select, button, textarea");
random_line_split
fourier.py
the paper "Rethinking Attention with Performers" https://arxiv.org/pdf/2009.14794.pdf and the traditional random Fourier features that approximate the RBF kernel. """ from math import sqrt, log from typing import Optional from einops import repeat import torch from .base import Kernel def orthogonal_random_matrix_( num_rows: int, num_columns: int, scaling: float = 0, device: Optional[torch.device] = None ):
else: raise ValueError(f"Invalid scaling {scaling}") return torch.diag(multiplier) @ final_matrix def orthogonal_matrix_chunk( cols: int, device: torch.device = None ) -> torch.Tensor: unstructured_block = torch.randn((cols, cols), device=device) q, _ = torch.qr(unstructured_block.cpu(), some=True) q = q.to(device) return q.t() class RandomFourierFeatures(Kernel): """Random Fourier Features for the RBF kernel according to [1]. [1]: "Weighted Sums of Random Kitchen Sinks: Replacing minimization with randomization in learning" by A. Rahimi and Benjamin Recht. Arguments --------- hidden_size: int, The input query dimensions in order to sample the noise matrix softmax_temp: float, The temerature for the Gaussian kernel approximation exp(-t * |x-y|^2) (default: 1/sqrt(query_dimensions)) orthogonal: bool, When True the random matrix is initialized for orthogonal random features to reduce the approximation variance (default: False) """ def __init__( self, head_size: int, kernel_size: Optional[int] = None, softmax_temp: Optional[float] = None, orthogonal: bool = False ): super(RandomFourierFeatures, self).__init__(head_size) assert kernel_size % 2 == 0, "kernel size not divisible by 2" self.kernel_size = kernel_size self.orthogonal = orthogonal self.softmax_temp = ( 1/sqrt(head_size) if softmax_temp is None else softmax_temp ) # Make a buffer for storing the sampled omega self.register_buffer( "omega", torch.zeros(head_size, self.kernel_size//2) ) def new_kernel(self): if self.orthogonal: orthogonal_random_matrix_(self.omega) else: self.omega.normal_() def forward( self, x: torch.Tensor, **kwargs ) -> torch.Tensor: x = x * sqrt(self.softmax_temp) u = x.unsqueeze(-2).matmul(self.omega).squeeze(-2) phi = torch.cat([torch.cos(u), torch.sin(u)], dim=-1) return phi * sqrt(2/self.kernel_size) class SmoothedRandomFourierFeatures(RandomFourierFeatures): """Simply add a constant value to the dot product in order to avoid possible numerical instabilities when the feature map is slightly negative. Implements K(x, y) = exp(-|x-y|^2) + s. Arguments --------- query_dimensions: int, The input query dimensions in order to sample the noise matrix n_dims: int, The size of the feature map (should be divisible by 2) (default: query_dimensions) softmax_temp: float, The temerature for the Gaussian kernel approximation exp(-t * |x-y|^2) (default: 1/sqrt(query_dimensions)) orthogonal: bool, When True the random matrix is initialized for orthogonal random features to reduce the approximation variance (default: False) smoothing: float, The smoothing parameter to add to the dot product. """ def __init__(self, query_dimensions, n_dims=None, softmax_temp=None, orthogonal=False, smoothing=1.0): super(SmoothedRandomFourierFeatures, self).__init__( query_dimensions, n_dims=query_dimensions-1 if n_dims is None else n_dims-1, softmax_temp=softmax_temp, orthogonal=orthogonal, ) self.smoothing = smoothing def forward(self, x): y = super().forward(x) smoothing = torch.full( y.shape[:-1] + (1,), self.smoothing, dtype=y.dtype, device=y.device ) return torch.cat([y, smoothing], dim=-1) class SoftmaxKernel(Kernel): """Positive orthogonal random features that approximate the softmax kernel. Basically implementation of Lemma 1 from "Rethinking Attention with Performers". Arguments --------- head_size: int, The input query dimensions in order to sample the noise matrix kernel_size: int, The size of the feature map (should be divisible by 2) (default: query_dimensions) softmax_temp: float, The temerature for the softmax approximation (default: 1/sqrt(query_dimensions)) orthogonal: bool, If set to true then the random matrix should be orthogonal which results in lower approximation variance (default: True) stabilize: bool, If set to True subtract the max norm from the exponentials to make sure that there are no infinities. It is equivalent to a robust implementation of softmax where the max is subtracted before the exponentiation. (default: False) """ def __init__( self, head_size: int, kernel_size: Optional[int] = None, ortho_scaling: Optional[float] = 0, causal: bool = False, orthogonal: bool = True, eps: float = 1e-6 ): super(SoftmaxKernel, self).__init__(head_size) kernel_size = int(self.head_size * log(self.head_size)) if kernel_size is None else kernel_size self.kernel_size = kernel_size self.ortho_scaling = ortho_scaling self.causal = causal self.orthogonal = orthogonal self.eps = eps self.register_buffer( "omegas", self.new_kernel() ) if self.causal: raise NotImplementedError("linear causal attention not yet implemented") def new_kernel( self, device: Optional[torch.device] = "cpu" ): return orthogonal_random_matrix_( self.kernel_size, self.head_size, scaling=self.ortho_scaling, device=device ) def forward( self, x: torch.Tensor, is_query: bool, normalize_data: bool = True, ) -> torch.Tensor: b, h, *_ = x.shape if normalize_data: x_norm = 1. / (x.shape[-1] ** 0.25) else: x_norm = 1. ratio = 1. / (self.omegas.shape[0] ** 0.5) projection_matrix = repeat(self.omegas, 'j d -> b h j d', b=b, h=h) data_dash = torch.einsum('...id,...jd->...ij', (x_norm * x), projection_matrix) diag_x = torch.sum(x ** 2, dim=-1) diag_x = ((diag_x / 2.0) * (x_norm ** 2)).unsqueeze(dim=-1) if is_query: data_dash = ratio * ( torch.exp(data_dash - diag_x - torch.max(data_dash, dim=-1, keepdim=True).values) + self.eps ) else: data_dash = ratio * ( torch.exp(data_dash - diag_x - torch.max(data_dash)) + self.eps ) return data_dash class GeneralizedRandomFeatures(RandomFourierFeatures): """Implements the generalized random Fourier features from Performers. It computes φ(χ) = [f(ω_1 χ), f(ω_2 χ), ..., f(ω_n χ)] where f(.) is the passed in `kernel_fn`. Arguments --------- query_dimensions: int, The input query dimensions in order to sample the noise matrix n_dims: int, The size of the feature map (default: query_dimensions) softmax_temp: float, A normalizer for the dot products that is multiplied to the input features before the feature map application (default: 1.0) orthogonal: bool, If set to true then the random matrix should be orthogonal which results in lower approximation variance (default: True) kernel_fn: callable, defines the f used for the feature map. (default: relu) """ def __init__(self, query_dimensions, n_dims=None, softmax_temp=
num_full_blocks = int(num_rows / num_columns) block_list = [] for _ in range(num_full_blocks): q = orthogonal_matrix_chunk(num_columns, device) block_list.append(q) remaining_rows = num_rows - (num_full_blocks * num_columns) if remaining_rows > 0: q = orthogonal_matrix_chunk(num_columns, device) block_list.append(q[:remaining_rows]) final_matrix = torch.cat(block_list) if scaling == 0: multiplier = torch.randn((num_rows, num_columns), device=device)\ .norm(dim=1) elif scaling == 1: multiplier = sqrt((float(num_columns))) * torch.ones((num_rows,), device=device)
identifier_body
fourier.py
paper "Rethinking Attention with Performers" https://arxiv.org/pdf/2009.14794.pdf and the traditional random Fourier features that approximate the RBF kernel. """ from math import sqrt, log from typing import Optional from einops import repeat import torch from .base import Kernel def orthogonal_random_matrix_( num_rows: int, num_columns: int, scaling: float = 0, device: Optional[torch.device] = None ): num_full_blocks = int(num_rows / num_columns) block_list = [] for _ in range(num_full_blocks): q = orthogonal_matrix_chunk(num_columns, device) block_list.append(q) remaining_rows = num_rows - (num_full_blocks * num_columns) if remaining_rows > 0: q = orthogonal_matrix_chunk(num_columns, device) block_list.append(q[:remaining_rows]) final_matrix = torch.cat(block_list) if scaling == 0: multiplier = torch.randn((num_rows, num_columns), device=device)\ .norm(dim=1) elif scaling == 1:
else: raise ValueError(f"Invalid scaling {scaling}") return torch.diag(multiplier) @ final_matrix def orthogonal_matrix_chunk( cols: int, device: torch.device = None ) -> torch.Tensor: unstructured_block = torch.randn((cols, cols), device=device) q, _ = torch.qr(unstructured_block.cpu(), some=True) q = q.to(device) return q.t() class RandomFourierFeatures(Kernel): """Random Fourier Features for the RBF kernel according to [1]. [1]: "Weighted Sums of Random Kitchen Sinks: Replacing minimization with randomization in learning" by A. Rahimi and Benjamin Recht. Arguments --------- hidden_size: int, The input query dimensions in order to sample the noise matrix softmax_temp: float, The temerature for the Gaussian kernel approximation exp(-t * |x-y|^2) (default: 1/sqrt(query_dimensions)) orthogonal: bool, When True the random matrix is initialized for orthogonal random features to reduce the approximation variance (default: False) """ def __init__( self, head_size: int, kernel_size: Optional[int] = None, softmax_temp: Optional[float] = None, orthogonal: bool = False ): super(RandomFourierFeatures, self).__init__(head_size) assert kernel_size % 2 == 0, "kernel size not divisible by 2" self.kernel_size = kernel_size self.orthogonal = orthogonal self.softmax_temp = ( 1/sqrt(head_size) if softmax_temp is None else softmax_temp ) # Make a buffer for storing the sampled omega self.register_buffer( "omega", torch.zeros(head_size, self.kernel_size//2) ) def new_kernel(self): if self.orthogonal: orthogonal_random_matrix_(self.omega) else: self.omega.normal_() def forward( self, x: torch.Tensor, **kwargs ) -> torch.Tensor: x = x * sqrt(self.softmax_temp) u = x.unsqueeze(-2).matmul(self.omega).squeeze(-2) phi = torch.cat([torch.cos(u), torch.sin(u)], dim=-1) return phi * sqrt(2/self.kernel_size) class SmoothedRandomFourierFeatures(RandomFourierFeatures): """Simply add a constant value to the dot product in order to avoid possible numerical instabilities when the feature map is slightly negative. Implements K(x, y) = exp(-|x-y|^2) + s. Arguments --------- query_dimensions: int, The input query dimensions in order to sample the noise matrix n_dims: int, The size of the feature map (should be divisible by 2) (default: query_dimensions) softmax_temp: float, The temerature for the Gaussian kernel approximation exp(-t * |x-y|^2) (default: 1/sqrt(query_dimensions)) orthogonal: bool, When True the random matrix is initialized for orthogonal random features to reduce the approximation variance (default: False) smoothing: float, The smoothing parameter to add to the dot product. """ def __init__(self, query_dimensions, n_dims=None, softmax_temp=None, orthogonal=False, smoothing=1.0): super(SmoothedRandomFourierFeatures, self).__init__( query_dimensions, n_dims=query_dimensions-1 if n_dims is None else n_dims-1, softmax_temp=softmax_temp, orthogonal=orthogonal, ) self.smoothing = smoothing def forward(self, x): y = super().forward(x) smoothing = torch.full( y.shape[:-1] + (1,), self.smoothing, dtype=y.dtype, device=y.device ) return torch.cat([y, smoothing], dim=-1) class SoftmaxKernel(Kernel): """Positive orthogonal random features that approximate the softmax kernel. Basically implementation of Lemma 1 from "Rethinking Attention with Performers". Arguments --------- head_size: int, The input query dimensions in order to sample the noise matrix kernel_size: int, The size of the feature map (should be divisible by 2) (default: query_dimensions) softmax_temp: float, The temerature for the softmax approximation (default: 1/sqrt(query_dimensions)) orthogonal: bool, If set to true then the random matrix should be orthogonal which results in lower approximation variance (default: True) stabilize: bool, If set to True subtract the max norm from the exponentials to make sure that there are no infinities. It is equivalent to a robust implementation of softmax where the max is subtracted before the exponentiation. (default: False) """ def __init__( self, head_size: int, kernel_size: Optional[int] = None, ortho_scaling: Optional[float] = 0, causal: bool = False, orthogonal: bool = True, eps: float = 1e-6 ): super(SoftmaxKernel, self).__init__(head_size) kernel_size = int(self.head_size * log(self.head_size)) if kernel_size is None else kernel_size self.kernel_size = kernel_size self.ortho_scaling = ortho_scaling self.causal = causal self.orthogonal = orthogonal self.eps = eps self.register_buffer( "omegas", self.new_kernel() ) if self.causal: raise NotImplementedError("linear causal attention not yet implemented") def new_kernel( self, device: Optional[torch.device] = "cpu" ): return orthogonal_random_matrix_( self.kernel_size, self.head_size, scaling=self.ortho_scaling, device=device ) def forward( self, x: torch.Tensor, is_query: bool, normalize_data: bool = True, ) -> torch.Tensor: b, h, *_ = x.shape if normalize_data: x_norm = 1. / (x.shape[-1] ** 0.25) else: x_norm = 1. ratio = 1. / (self.omegas.shape[0] ** 0.5) projection_matrix = repeat(self.omegas, 'j d -> b h j d', b=b, h=h) data_dash = torch.einsum('...id,...jd->...ij', (x_norm * x), projection_matrix) diag_x = torch.sum(x ** 2, dim=-1) diag_x = ((diag_x / 2.0) * (x_norm ** 2)).unsqueeze(dim=-1) if is_query: data_dash = ratio * ( torch.exp(data_dash - diag_x - torch.max(data_dash, dim=-1, keepdim=True).values) + self.eps ) else: data_dash = ratio * ( torch.exp(data_dash - diag_x - torch.max(data_dash)) + self.eps ) return data_dash class GeneralizedRandomFeatures(RandomFourierFeatures): """Implements the generalized random Fourier features from Performers. It computes φ(χ) = [f(ω_1 χ), f(ω_2 χ), ..., f(ω_n χ)] where f(.) is the passed in `kernel_fn`. Arguments --------- query_dimensions: int, The input query dimensions in order to sample the noise matrix n_dims: int, The size of the feature map (default: query_dimensions) softmax_temp: float, A normalizer for the dot products that is multiplied to the input features before the feature map application (default: 1.0) orthogonal: bool, If set to true then the random matrix should be orthogonal which results in lower approximation variance (default: True) kernel_fn: callable, defines the f used for the feature map. (default: relu) """ def __init__(self, query_dimensions, n_dims=None, softmax_temp
multiplier = sqrt((float(num_columns))) * torch.ones((num_rows,), device=device)
conditional_block
fourier.py
the paper "Rethinking Attention with Performers" https://arxiv.org/pdf/2009.14794.pdf and the traditional random Fourier features that approximate the RBF kernel. """ from math import sqrt, log from typing import Optional from einops import repeat import torch from .base import Kernel def orthogonal_random_matrix_( num_rows: int, num_columns: int, scaling: float = 0, device: Optional[torch.device] = None ): num_full_blocks = int(num_rows / num_columns) block_list = [] for _ in range(num_full_blocks): q = orthogonal_matrix_chunk(num_columns, device) block_list.append(q) remaining_rows = num_rows - (num_full_blocks * num_columns) if remaining_rows > 0: q = orthogonal_matrix_chunk(num_columns, device) block_list.append(q[:remaining_rows]) final_matrix = torch.cat(block_list) if scaling == 0: multiplier = torch.randn((num_rows, num_columns), device=device)\ .norm(dim=1) elif scaling == 1: multiplier = sqrt((float(num_columns))) * torch.ones((num_rows,), device=device) else: raise ValueError(f"Invalid scaling {scaling}") return torch.diag(multiplier) @ final_matrix def orthogonal_matrix_chunk( cols: int, device: torch.device = None ) -> torch.Tensor: unstructured_block = torch.randn((cols, cols), device=device) q, _ = torch.qr(unstructured_block.cpu(), some=True) q = q.to(device) return q.t() class RandomFourierFeatures(Kernel): """Random Fourier Features for the RBF kernel according to [1]. [1]: "Weighted Sums of Random Kitchen Sinks: Replacing minimization with randomization in learning" by A. Rahimi and Benjamin Recht. Arguments --------- hidden_size: int, The input query dimensions in order to sample the noise matrix softmax_temp: float, The temerature for the Gaussian kernel approximation exp(-t * |x-y|^2) (default: 1/sqrt(query_dimensions)) orthogonal: bool, When True the random matrix is initialized for orthogonal random features to reduce the approximation variance (default: False) """ def __init__( self, head_size: int, kernel_size: Optional[int] = None, softmax_temp: Optional[float] = None, orthogonal: bool = False ): super(RandomFourierFeatures, self).__init__(head_size) assert kernel_size % 2 == 0, "kernel size not divisible by 2" self.kernel_size = kernel_size self.orthogonal = orthogonal self.softmax_temp = ( 1/sqrt(head_size) if softmax_temp is None else softmax_temp ) # Make a buffer for storing the sampled omega self.register_buffer( "omega", torch.zeros(head_size, self.kernel_size//2) ) def new_kernel(self): if self.orthogonal: orthogonal_random_matrix_(self.omega) else: self.omega.normal_() def forward( self, x: torch.Tensor, **kwargs ) -> torch.Tensor: x = x * sqrt(self.softmax_temp) u = x.unsqueeze(-2).matmul(self.omega).squeeze(-2) phi = torch.cat([torch.cos(u), torch.sin(u)], dim=-1) return phi * sqrt(2/self.kernel_size) class SmoothedRandomFourierFeatures(RandomFourierFeatures): """Simply add a constant value to the dot product in order to avoid possible numerical instabilities when the feature map is slightly negative. Implements K(x, y) = exp(-|x-y|^2) + s. Arguments --------- query_dimensions: int, The input query dimensions in order to sample the noise matrix n_dims: int, The size of the feature map (should be divisible by 2) (default: query_dimensions) softmax_temp: float, The temerature for the Gaussian kernel approximation exp(-t * |x-y|^2) (default: 1/sqrt(query_dimensions)) orthogonal: bool, When True the random matrix is initialized for orthogonal random features to reduce the approximation variance (default: False) smoothing: float, The smoothing parameter to add to the dot product. """ def __init__(self, query_dimensions, n_dims=None, softmax_temp=None, orthogonal=False, smoothing=1.0): super(SmoothedRandomFourierFeatures, self).__init__( query_dimensions, n_dims=query_dimensions-1 if n_dims is None else n_dims-1, softmax_temp=softmax_temp, orthogonal=orthogonal, ) self.smoothing = smoothing def forward(self, x): y = super().forward(x) smoothing = torch.full( y.shape[:-1] + (1,), self.smoothing, dtype=y.dtype, device=y.device ) return torch.cat([y, smoothing], dim=-1) class SoftmaxKernel(Kernel): """Positive orthogonal random features that approximate the softmax kernel. Basically implementation of Lemma 1 from "Rethinking Attention with Performers". Arguments --------- head_size: int, The input query dimensions in order to sample the noise matrix kernel_size: int, The size of the feature map (should be divisible by 2) (default: query_dimensions) softmax_temp: float, The temerature for the softmax approximation (default: 1/sqrt(query_dimensions)) orthogonal: bool, If set to true then the random matrix should be orthogonal which results in lower approximation variance (default: True) stabilize: bool, If set to True subtract the max norm from the exponentials to make sure that there are no infinities. It is equivalent to a robust implementation of softmax where the max is subtracted before the exponentiation. (default: False) """ def __init__( self, head_size: int, kernel_size: Optional[int] = None, ortho_scaling: Optional[float] = 0, causal: bool = False, orthogonal: bool = True, eps: float = 1e-6 ): super(SoftmaxKernel, self).__init__(head_size) kernel_size = int(self.head_size * log(self.head_size)) if kernel_size is None else kernel_size self.kernel_size = kernel_size self.ortho_scaling = ortho_scaling self.causal = causal self.orthogonal = orthogonal self.eps = eps self.register_buffer( "omegas", self.new_kernel() ) if self.causal: raise NotImplementedError("linear causal attention not yet implemented") def new_kernel( self, device: Optional[torch.device] = "cpu" ): return orthogonal_random_matrix_( self.kernel_size, self.head_size, scaling=self.ortho_scaling, device=device ) def forward( self, x: torch.Tensor, is_query: bool, normalize_data: bool = True, ) -> torch.Tensor: b, h, *_ = x.shape if normalize_data: x_norm = 1. / (x.shape[-1] ** 0.25) else: x_norm = 1. ratio = 1. / (self.omegas.shape[0] ** 0.5) projection_matrix = repeat(self.omegas, 'j d -> b h j d', b=b, h=h) data_dash = torch.einsum('...id,...jd->...ij', (x_norm * x), projection_matrix) diag_x = torch.sum(x ** 2, dim=-1) diag_x = ((diag_x / 2.0) * (x_norm ** 2)).unsqueeze(dim=-1) if is_query: data_dash = ratio * ( torch.exp(data_dash - diag_x - torch.max(data_dash, dim=-1, keepdim=True).values) + self.eps ) else: data_dash = ratio * ( torch.exp(data_dash - diag_x - torch.max(data_dash)) + self.eps ) return data_dash class GeneralizedRandomFeatures(RandomFourierFeatures): """Implements the generalized random Fourier features from Performers. It computes φ(χ) = [f(ω_1 χ), f(ω_2 χ), ..., f(ω_n χ)] where f(.) is the
query_dimensions: int, The input query dimensions in order to sample the noise matrix n_dims: int, The size of the feature map (default: query_dimensions) softmax_temp: float, A normalizer for the dot products that is multiplied to the input features before the feature map application (default: 1.0) orthogonal: bool, If set to true then the random matrix should be orthogonal which results in lower approximation variance (default: True) kernel_fn: callable, defines the f used for the feature map. (default: relu) """ def __init__(self, query_dimensions, n_dims=None, softmax_temp=1.
passed in `kernel_fn`. Arguments ---------
random_line_split
fourier.py
paper "Rethinking Attention with Performers" https://arxiv.org/pdf/2009.14794.pdf and the traditional random Fourier features that approximate the RBF kernel. """ from math import sqrt, log from typing import Optional from einops import repeat import torch from .base import Kernel def orthogonal_random_matrix_( num_rows: int, num_columns: int, scaling: float = 0, device: Optional[torch.device] = None ): num_full_blocks = int(num_rows / num_columns) block_list = [] for _ in range(num_full_blocks): q = orthogonal_matrix_chunk(num_columns, device) block_list.append(q) remaining_rows = num_rows - (num_full_blocks * num_columns) if remaining_rows > 0: q = orthogonal_matrix_chunk(num_columns, device) block_list.append(q[:remaining_rows]) final_matrix = torch.cat(block_list) if scaling == 0: multiplier = torch.randn((num_rows, num_columns), device=device)\ .norm(dim=1) elif scaling == 1: multiplier = sqrt((float(num_columns))) * torch.ones((num_rows,), device=device) else: raise ValueError(f"Invalid scaling {scaling}") return torch.diag(multiplier) @ final_matrix def orthogonal_matrix_chunk( cols: int, device: torch.device = None ) -> torch.Tensor: unstructured_block = torch.randn((cols, cols), device=device) q, _ = torch.qr(unstructured_block.cpu(), some=True) q = q.to(device) return q.t() class RandomFourierFeatures(Kernel): """Random Fourier Features for the RBF kernel according to [1]. [1]: "Weighted Sums of Random Kitchen Sinks: Replacing minimization with randomization in learning" by A. Rahimi and Benjamin Recht. Arguments --------- hidden_size: int, The input query dimensions in order to sample the noise matrix softmax_temp: float, The temerature for the Gaussian kernel approximation exp(-t * |x-y|^2) (default: 1/sqrt(query_dimensions)) orthogonal: bool, When True the random matrix is initialized for orthogonal random features to reduce the approximation variance (default: False) """ def __init__( self, head_size: int, kernel_size: Optional[int] = None, softmax_temp: Optional[float] = None, orthogonal: bool = False ): super(RandomFourierFeatures, self).__init__(head_size) assert kernel_size % 2 == 0, "kernel size not divisible by 2" self.kernel_size = kernel_size self.orthogonal = orthogonal self.softmax_temp = ( 1/sqrt(head_size) if softmax_temp is None else softmax_temp ) # Make a buffer for storing the sampled omega self.register_buffer( "omega", torch.zeros(head_size, self.kernel_size//2) ) def new_kernel(self): if self.orthogonal: orthogonal_random_matrix_(self.omega) else: self.omega.normal_() def forward( self, x: torch.Tensor, **kwargs ) -> torch.Tensor: x = x * sqrt(self.softmax_temp) u = x.unsqueeze(-2).matmul(self.omega).squeeze(-2) phi = torch.cat([torch.cos(u), torch.sin(u)], dim=-1) return phi * sqrt(2/self.kernel_size) class SmoothedRandomFourierFeatures(RandomFourierFeatures): """Simply add a constant value to the dot product in order to avoid possible numerical instabilities when the feature map is slightly negative. Implements K(x, y) = exp(-|x-y|^2) + s. Arguments --------- query_dimensions: int, The input query dimensions in order to sample the noise matrix n_dims: int, The size of the feature map (should be divisible by 2) (default: query_dimensions) softmax_temp: float, The temerature for the Gaussian kernel approximation exp(-t * |x-y|^2) (default: 1/sqrt(query_dimensions)) orthogonal: bool, When True the random matrix is initialized for orthogonal random features to reduce the approximation variance (default: False) smoothing: float, The smoothing parameter to add to the dot product. """ def __init__(self, query_dimensions, n_dims=None, softmax_temp=None, orthogonal=False, smoothing=1.0): super(SmoothedRandomFourierFeatures, self).__init__( query_dimensions, n_dims=query_dimensions-1 if n_dims is None else n_dims-1, softmax_temp=softmax_temp, orthogonal=orthogonal, ) self.smoothing = smoothing def forward(self, x): y = super().forward(x) smoothing = torch.full( y.shape[:-1] + (1,), self.smoothing, dtype=y.dtype, device=y.device ) return torch.cat([y, smoothing], dim=-1) class SoftmaxKernel(Kernel): """Positive orthogonal random features that approximate the softmax kernel. Basically implementation of Lemma 1 from "Rethinking Attention with Performers". Arguments --------- head_size: int, The input query dimensions in order to sample the noise matrix kernel_size: int, The size of the feature map (should be divisible by 2) (default: query_dimensions) softmax_temp: float, The temerature for the softmax approximation (default: 1/sqrt(query_dimensions)) orthogonal: bool, If set to true then the random matrix should be orthogonal which results in lower approximation variance (default: True) stabilize: bool, If set to True subtract the max norm from the exponentials to make sure that there are no infinities. It is equivalent to a robust implementation of softmax where the max is subtracted before the exponentiation. (default: False) """ def __init__( self, head_size: int, kernel_size: Optional[int] = None, ortho_scaling: Optional[float] = 0, causal: bool = False, orthogonal: bool = True, eps: float = 1e-6 ): super(SoftmaxKernel, self).__init__(head_size) kernel_size = int(self.head_size * log(self.head_size)) if kernel_size is None else kernel_size self.kernel_size = kernel_size self.ortho_scaling = ortho_scaling self.causal = causal self.orthogonal = orthogonal self.eps = eps self.register_buffer( "omegas", self.new_kernel() ) if self.causal: raise NotImplementedError("linear causal attention not yet implemented") def
( self, device: Optional[torch.device] = "cpu" ): return orthogonal_random_matrix_( self.kernel_size, self.head_size, scaling=self.ortho_scaling, device=device ) def forward( self, x: torch.Tensor, is_query: bool, normalize_data: bool = True, ) -> torch.Tensor: b, h, *_ = x.shape if normalize_data: x_norm = 1. / (x.shape[-1] ** 0.25) else: x_norm = 1. ratio = 1. / (self.omegas.shape[0] ** 0.5) projection_matrix = repeat(self.omegas, 'j d -> b h j d', b=b, h=h) data_dash = torch.einsum('...id,...jd->...ij', (x_norm * x), projection_matrix) diag_x = torch.sum(x ** 2, dim=-1) diag_x = ((diag_x / 2.0) * (x_norm ** 2)).unsqueeze(dim=-1) if is_query: data_dash = ratio * ( torch.exp(data_dash - diag_x - torch.max(data_dash, dim=-1, keepdim=True).values) + self.eps ) else: data_dash = ratio * ( torch.exp(data_dash - diag_x - torch.max(data_dash)) + self.eps ) return data_dash class GeneralizedRandomFeatures(RandomFourierFeatures): """Implements the generalized random Fourier features from Performers. It computes φ(χ) = [f(ω_1 χ), f(ω_2 χ), ..., f(ω_n χ)] where f(.) is the passed in `kernel_fn`. Arguments --------- query_dimensions: int, The input query dimensions in order to sample the noise matrix n_dims: int, The size of the feature map (default: query_dimensions) softmax_temp: float, A normalizer for the dot products that is multiplied to the input features before the feature map application (default: 1.0) orthogonal: bool, If set to true then the random matrix should be orthogonal which results in lower approximation variance (default: True) kernel_fn: callable, defines the f used for the feature map. (default: relu) """ def __init__(self, query_dimensions, n_dims=None, softmax_temp=
new_kernel
identifier_name
differentiation.go
5. Traverse the graph from output towards input. On each visit, perform the symbolic differentiation // // For most cases, Grad() should be used instead of Backpropagate(), as Grad() performs several checks which would be the general use case, before calling Backpropagate() func Backpropagate(outputs, gradOutputs, wrt Nodes) (retVal Nodes, err error) { symdiffLogf("BACKPROP START") symdiffLogf("Outputs: %d", outputs) symdiffLogf("gradOutputs: %d", gradOutputs) symdiffLogf("WRT: %d", wrt) enterLogScope() defer leaveLogScope() g := outputs[0].g // this entire section about removing foreveralone nodes need a rethink symdiffLogf("removing foreveralone nodes") enterLogScope() for i := 0; i < len(g.AllNodes()); i++ { n := g.AllNodes()[i] fr := g.From(n.ID()).Len() to := g.To(n.ID()).Len() if fr == 0 && to == 0 && !n.isConstant() && !n.isInput() { g.RemoveNode(n) symdiffLogf("removed %v(%p); %x; %s", n, n, n.ID(), n.Name()) } } leaveLogScope() var sortedNodes Nodes if sortedNodes, err = Sort(g); err != nil { return nil, errors.Wrap(err, sortFail) } symdiffLogf("sorted nodes: %v", sortedNodes) symdiffLogf("sorted nodes: %d", sortedNodes) var affectsOutput NodeSet var affectedByOutput NodeSet if affectsOutput, err = forwardDiffAnalysis(outputs, sortedNodes); err != nil { return nil, errors.Wrap(err, "Failed during forward differentiation analysis") } if affectedByOutput, err = backwardDiffAnalysis(wrt, sortedNodes); err != nil { return nil, errors.Wrap(err, "Failed during forward differentiation analysis") } symdiffLogf("affects output: %v", affectsOutput) symdiffLogf("affected by output : %v", affectedByOutput) wrtSet := wrt.mapSet() badWRTs := wrtSet.Difference(affectsOutput) if len(badWRTs) > 0 { return nil, SymDiffError{nodes: badWRTs.ToSlice(), err: errors.Errorf("Non Differentiable WRTs: %v", badWRTs)} } outputSet := outputs.mapSet() badOutputs := outputSet.Difference(affectedByOutput) if len(badOutputs) > 0 { symdiffLogf("badOutputs: %#v", badOutputs) return nil, SymDiffError{nodes: badOutputs.ToSlice(), err: errors.Errorf("Non-Differentable Outputs: %v", badOutputs)} } // map a node to a list of gradient terms // these gradient terms will be summed up when we visit the node // when iterating through the nondes in reverse topological order nodeGradMap := make(map[*Node]Nodes) for i, n := range outputs { symdiffLogf("Adding outputs for %x", n.ID()) nodeGradMap[n] = Nodes{gradOutputs[i]} } // "active" nodes are the ones that are differentially influenced by the inputs // and also differentiably influence the outputs. These are the nodes where we need to call the // "pullback" function to backpropagate derivatives activeNodes := affectsOutput.Intersect(affectedByOutput) symdiffLogf("Active: %v", activeNodes) symdiffLogf("Sorted: %d", sortedNodes) symdiffLogf("nodeGradMap: %+#d", FmtNodeMap(nodeGradMap)) enterLogScope() for _, node := range sortedNodes { if _, ok := activeNodes[node]; !ok { symdiffLogf("skipping %x", node.ID()) continue } if node.deriv != nil { symdiffLogf("skipping %x - previously differentiated", node.ID()) nodeGradMap[node] = append(nodeGradMap[node], node.deriv) continue } symdiffLogf("Working on %x %v", node.ID(), node) enterLogScope() // Check if there is any grads coming into this node if len(nodeGradMap[node]) < 1 { leaveLogScope() return nil, SymDiffError{ single: node, gradMap: nodeGradMap, err: errors.New("No gradients found for node"), } } // once we've reached a node, we already backpropagated from its dependents // so we sum up the gradients symdiffLogf("nodeGradMap[%x]: %d", node.ID(), nodeGradMap[node]) if len(nodeGradMap[node]) > 1 { var n *Node symdiffLogf("reduce adding") if n, err = ReduceAdd(nodeGradMap[node], WithGroupName(gradClust)); err != nil { leaveLogScope() return nil, SymDiffError{ single: node, nodes: nodeGradMap[node], gradMap: nodeGradMap, err: errors.Wrap(err, "ReduceAdd failed during differentiation"), } } symdiffLogf("reduced to... %x", n.ID()) // node.derives = append(node.derives, n) n.derivOf = append(n.derivOf, node) node.deriv = n nodeGradMap[node] = Nodes{n} // } } else if len(nodeGradMap[node]) == 1 { deriv := nodeGradMap[node][0] deriv.derivOf = append(deriv.derivOf, node) node.deriv = deriv } gradNode := nodeGradMap[node][0] if !node.isInput() { symdiffLogf("differentiating %x (%v)", node.ID(), node.op) enterLogScope() var op SDOp var childrenGrads Nodes var ok bool if op, ok = node.op.(SDOp); !ok { return nil, SymDiffError{ single: node, err: errors.New("Not a SymDifOp"), } } symdiffLogf("op: %v || optype: %v || node: %v || Children: %#Y || Grad: %v", node.op, node.op.Type(), node.t, node.children, gradNode) if childrenGrads, err = op.SymDiff(node.children, node, gradNode); err != nil { leaveLogScope() return nil, SymDiffError{ single: node, grad: gradNode, gradMap: nodeGradMap, err: errors.Wrapf(err, ".SymDiff() failed"), } } symdiffLogf("Derived(%d): %P", len(childrenGrads), childrenGrads) leaveLogScope() diffs := node.diffWRT() for i, child := range node.children { symdiffLogf("child is %v, i: %v", child, i) differentiable := diffs[i] childGrad := childrenGrads[i] if differentiable { childGrad.setGroup(gradClust) if grads, ok := nodeGradMap[child]; ok { grads = append(grads, childGrad) nodeGradMap[child] = grads } else { nodeGradMap[child] = Nodes{childGrad} } } else { symdiffLogf("Child %x is non differentiable", child.ID()) if childGrad != nil { childGrad.setGroup(strayClust) } } } } else { symdiffLogf("iz input") symdiffLogf("%d ", nodeGradMap[node]) } leaveLogScope() } leaveLogScope() // only we already summed up the gradients for the input nodes, so just take // 0th element for _, n := range wrt { symdiffLogf("nodeGradMap wrt: %d", nodeGradMap[n]) retVal = append(retVal, nodeGradMap[n][0]) } return } // SetDerivOf is used to hack around the fundamental limitations of Gorgonia. // // Specifically it is used to set a node as the derivative of another node, // used in the cuDNN version of batch norm. // // The cuDNN BatchNorm operation produces the derivatives for the scale and bias as a side effect // of calculating the derivative of the input. Because Gorgonia's Ops are modelled as pure functions (and no tuples) // this causes a bit of trouble. With the clever use of scratch space ops multireturn can be simulated. // But this causes derivatives to not be set correctly. func SetDerivOf(deriv, of *Node)
{ deriv.derivOf = append(deriv.derivOf, of) of.deriv = deriv }
identifier_body
differentiation.go
list of Nodes that are affected by differentiating output. // Given a list of WRTs, we want to find a list of nodes that will be affected when backpropagating. func backwardDiffAnalysis(wrt, sortedNodes Nodes) (retVal NodeSet, err error) { symdiffLogf("Backwards analysis") enterLogScope() defer leaveLogScope() if !wrt.AllSameGraph() { return nil, errors.New("The supplied output Nodes are not the same graph") } diffSet := wrt.mapSet() symdiffLogf("wrt:%d diffset: %d", len(wrt), len(diffSet)) symdiffLogf("%v", diffSet) symdiffLogf("sorted: %d", sortedNodes) enterLogScope() for i := len(sortedNodes) - 1; i >= 0; i-- { n := sortedNodes[i] symdiffLogf("working on %v. Has %d children", n, len(n.children)) var op SDOp var ok bool var diffs []bool if op, ok = n.op.(SDOp); ok { diffs = op.DiffWRT(len(n.children)) } symdiffLogf("differentiable WRT: %v", diffs) enterLogScope() symdiffLogf("Children: %v", n.children) if len(diffs) == 0 { // check if this makes nodes unreachable. If it does, then error out if n.isStmt { symdiffLogf("Statement nodes are Non differentiable!") leaveLogScope() continue } else if n.isInput() { symdiffLogf("Input nodes are Non differentiable") leaveLogScope() continue } else if len(n.children) == 0 { symdiffLogf("Leaf nodes have no children") leaveLogScope() continue } g := n.g for _, child := range n.children { parents := graph.NodesOf(g.To(child.ID())) if len(parents) == 1 && len(child.children) > 0 { leaveLogScope() return nil, errors.Errorf("Being unable to differentiate %v would leave a portion of the graph unreachable. Unable to continue", n) } } symdiffLogf("SKIPPING... Non differentiable!") leaveLogScope() continue } inner: for j, child := range n.children { d := diffs[j] if diffSet.Contains(child) && d { symdiffLogf("Adding %x to differentiable set", child.ID()) diffSet.Add(n) break inner } } leaveLogScope() } leaveLogScope() return diffSet, nil } // Backpropagate backpropagates errors by performing reverse-mode symbolic differentiation, starting from the outputs, and working its way towads the inputs. // // This is the rough algorithm: // 1. Filter out nodes that are unreachable // 2. Forwards analysis, where a list of nodes affecting the output is added to consideration // 3. Backwards analysis, where a list of nodes affected by differentiating the output are added to the consideration // 4. If there is a difference in both sets, it will cause an error (both sets should be the same) // 5. Traverse the graph from output towards input. On each visit, perform the symbolic differentiation // // For most cases, Grad() should be used instead of Backpropagate(), as Grad() performs several checks which would be the general use case, before calling Backpropagate() func Backpropagate(outputs, gradOutputs, wrt Nodes) (retVal Nodes, err error) { symdiffLogf("BACKPROP START") symdiffLogf("Outputs: %d", outputs) symdiffLogf("gradOutputs: %d", gradOutputs) symdiffLogf("WRT: %d", wrt) enterLogScope() defer leaveLogScope() g := outputs[0].g // this entire section about removing foreveralone nodes need a rethink symdiffLogf("removing foreveralone nodes") enterLogScope() for i := 0; i < len(g.AllNodes()); i++ { n := g.AllNodes()[i] fr := g.From(n.ID()).Len() to := g.To(n.ID()).Len() if fr == 0 && to == 0 && !n.isConstant() && !n.isInput() { g.RemoveNode(n) symdiffLogf("removed %v(%p); %x; %s", n, n, n.ID(), n.Name()) } } leaveLogScope() var sortedNodes Nodes if sortedNodes, err = Sort(g); err != nil { return nil, errors.Wrap(err, sortFail) } symdiffLogf("sorted nodes: %v", sortedNodes) symdiffLogf("sorted nodes: %d", sortedNodes) var affectsOutput NodeSet var affectedByOutput NodeSet if affectsOutput, err = forwardDiffAnalysis(outputs, sortedNodes); err != nil { return nil, errors.Wrap(err, "Failed during forward differentiation analysis") } if affectedByOutput, err = backwardDiffAnalysis(wrt, sortedNodes); err != nil { return nil, errors.Wrap(err, "Failed during forward differentiation analysis") } symdiffLogf("affects output: %v", affectsOutput) symdiffLogf("affected by output : %v", affectedByOutput) wrtSet := wrt.mapSet() badWRTs := wrtSet.Difference(affectsOutput) if len(badWRTs) > 0 { return nil, SymDiffError{nodes: badWRTs.ToSlice(), err: errors.Errorf("Non Differentiable WRTs: %v", badWRTs)} } outputSet := outputs.mapSet() badOutputs := outputSet.Difference(affectedByOutput) if len(badOutputs) > 0 { symdiffLogf("badOutputs: %#v", badOutputs) return nil, SymDiffError{nodes: badOutputs.ToSlice(), err: errors.Errorf("Non-Differentable Outputs: %v", badOutputs)} } // map a node to a list of gradient terms // these gradient terms will be summed up when we visit the node // when iterating through the nondes in reverse topological order nodeGradMap := make(map[*Node]Nodes) for i, n := range outputs { symdiffLogf("Adding outputs for %x", n.ID()) nodeGradMap[n] = Nodes{gradOutputs[i]} } // "active" nodes are the ones that are differentially influenced by the inputs // and also differentiably influence the outputs. These are the nodes where we need to call the // "pullback" function to backpropagate derivatives activeNodes := affectsOutput.Intersect(affectedByOutput) symdiffLogf("Active: %v", activeNodes) symdiffLogf("Sorted: %d", sortedNodes) symdiffLogf("nodeGradMap: %+#d", FmtNodeMap(nodeGradMap)) enterLogScope() for _, node := range sortedNodes { if _, ok := activeNodes[node]; !ok { symdiffLogf("skipping %x", node.ID()) continue } if node.deriv != nil { symdiffLogf("skipping %x - previously differentiated", node.ID()) nodeGradMap[node] = append(nodeGradMap[node], node.deriv) continue } symdiffLogf("Working on %x %v", node.ID(), node) enterLogScope() // Check if there is any grads coming into this node if len(nodeGradMap[node]) < 1 { leaveLogScope() return nil, SymDiffError{ single: node, gradMap: nodeGradMap, err: errors.New("No gradients found for node"), } } // once we've reached a node, we already backpropagated from its dependents // so we sum up the gradients symdiffLogf("nodeGradMap[%x]: %d", node.ID(), nodeGradMap[node]) if len(nodeGradMap[node]) > 1 { var n *Node symdiffLogf("reduce adding") if n, err = ReduceAdd(nodeGradMap[node], WithGroupName(gradClust)); err != nil
symdiffLogf("reduced to... %x", n.ID()) // node.derives = append(node.derives, n) n.derivOf = append(n.derivOf, node) node.deriv = n nodeGradMap[node] = Nodes{n} // } } else if len(nodeGradMap[node]) == 1 { deriv := nodeGradMap[node][0] deriv.deriv
{ leaveLogScope() return nil, SymDiffError{ single: node, nodes: nodeGradMap[node], gradMap: nodeGradMap, err: errors.Wrap(err, "ReduceAdd failed during differentiation"), } }
conditional_block
differentiation.go
list of Nodes that are affected by differentiating output. // Given a list of WRTs, we want to find a list of nodes that will be affected when backpropagating. func backwardDiffAnalysis(wrt, sortedNodes Nodes) (retVal NodeSet, err error) { symdiffLogf("Backwards analysis") enterLogScope() defer leaveLogScope() if !wrt.AllSameGraph() { return nil, errors.New("The supplied output Nodes are not the same graph") } diffSet := wrt.mapSet() symdiffLogf("wrt:%d diffset: %d", len(wrt), len(diffSet)) symdiffLogf("%v", diffSet) symdiffLogf("sorted: %d", sortedNodes) enterLogScope() for i := len(sortedNodes) - 1; i >= 0; i-- { n := sortedNodes[i] symdiffLogf("working on %v. Has %d children", n, len(n.children)) var op SDOp var ok bool var diffs []bool if op, ok = n.op.(SDOp); ok { diffs = op.DiffWRT(len(n.children)) } symdiffLogf("differentiable WRT: %v", diffs) enterLogScope() symdiffLogf("Children: %v", n.children) if len(diffs) == 0 { // check if this makes nodes unreachable. If it does, then error out if n.isStmt { symdiffLogf("Statement nodes are Non differentiable!") leaveLogScope() continue } else if n.isInput() { symdiffLogf("Input nodes are Non differentiable") leaveLogScope() continue } else if len(n.children) == 0 { symdiffLogf("Leaf nodes have no children") leaveLogScope() continue } g := n.g for _, child := range n.children { parents := graph.NodesOf(g.To(child.ID())) if len(parents) == 1 && len(child.children) > 0 { leaveLogScope() return nil, errors.Errorf("Being unable to differentiate %v would leave a portion of the graph unreachable. Unable to continue", n) } } symdiffLogf("SKIPPING... Non differentiable!") leaveLogScope() continue } inner: for j, child := range n.children { d := diffs[j] if diffSet.Contains(child) && d { symdiffLogf("Adding %x to differentiable set", child.ID()) diffSet.Add(n) break inner } } leaveLogScope() } leaveLogScope() return diffSet, nil } // Backpropagate backpropagates errors by performing reverse-mode symbolic differentiation, starting from the outputs, and working its way towads the inputs. // // This is the rough algorithm: // 1. Filter out nodes that are unreachable // 2. Forwards analysis, where a list of nodes affecting the output is added to consideration // 3. Backwards analysis, where a list of nodes affected by differentiating the output are added to the consideration // 4. If there is a difference in both sets, it will cause an error (both sets should be the same) // 5. Traverse the graph from output towards input. On each visit, perform the symbolic differentiation // // For most cases, Grad() should be used instead of Backpropagate(), as Grad() performs several checks which would be the general use case, before calling Backpropagate() func Backpropagate(outputs, gradOutputs, wrt Nodes) (retVal Nodes, err error) { symdiffLogf("BACKPROP START") symdiffLogf("Outputs: %d", outputs) symdiffLogf("gradOutputs: %d", gradOutputs) symdiffLogf("WRT: %d", wrt) enterLogScope() defer leaveLogScope() g := outputs[0].g // this entire section about removing foreveralone nodes need a rethink symdiffLogf("removing foreveralone nodes") enterLogScope() for i := 0; i < len(g.AllNodes()); i++ { n := g.AllNodes()[i] fr := g.From(n.ID()).Len() to := g.To(n.ID()).Len() if fr == 0 && to == 0 && !n.isConstant() && !n.isInput() { g.RemoveNode(n) symdiffLogf("removed %v(%p); %x; %s", n, n, n.ID(), n.Name()) } } leaveLogScope() var sortedNodes Nodes if sortedNodes, err = Sort(g); err != nil { return nil, errors.Wrap(err, sortFail) } symdiffLogf("sorted nodes: %v", sortedNodes) symdiffLogf("sorted nodes: %d", sortedNodes) var affectsOutput NodeSet var affectedByOutput NodeSet if affectsOutput, err = forwardDiffAnalysis(outputs, sortedNodes); err != nil { return nil, errors.Wrap(err, "Failed during forward differentiation analysis") } if affectedByOutput, err = backwardDiffAnalysis(wrt, sortedNodes); err != nil { return nil, errors.Wrap(err, "Failed during forward differentiation analysis") } symdiffLogf("affects output: %v", affectsOutput) symdiffLogf("affected by output : %v", affectedByOutput) wrtSet := wrt.mapSet() badWRTs := wrtSet.Difference(affectsOutput) if len(badWRTs) > 0 { return nil, SymDiffError{nodes: badWRTs.ToSlice(), err: errors.Errorf("Non Differentiable WRTs: %v", badWRTs)} } outputSet := outputs.mapSet() badOutputs := outputSet.Difference(affectedByOutput) if len(badOutputs) > 0 { symdiffLogf("badOutputs: %#v", badOutputs) return nil, SymDiffError{nodes: badOutputs.ToSlice(), err: errors.Errorf("Non-Differentable Outputs: %v", badOutputs)} } // map a node to a list of gradient terms // these gradient terms will be summed up when we visit the node // when iterating through the nondes in reverse topological order nodeGradMap := make(map[*Node]Nodes) for i, n := range outputs { symdiffLogf("Adding outputs for %x", n.ID()) nodeGradMap[n] = Nodes{gradOutputs[i]} } // "active" nodes are the ones that are differentially influenced by the inputs // and also differentiably influence the outputs. These are the nodes where we need to call the // "pullback" function to backpropagate derivatives activeNodes := affectsOutput.Intersect(affectedByOutput) symdiffLogf("Active: %v", activeNodes) symdiffLogf("Sorted: %d", sortedNodes) symdiffLogf("nodeGradMap: %+#d", FmtNodeMap(nodeGradMap)) enterLogScope() for _, node := range sortedNodes { if _, ok := activeNodes[node]; !ok { symdiffLogf("skipping %x", node.ID()) continue } if node.deriv != nil { symdiffLogf("skipping %x - previously differentiated", node.ID()) nodeGradMap[node] = append(nodeGradMap[node], node.deriv) continue } symdiffLogf("Working on %x %v", node.ID(), node) enterLogScope() // Check if there is any grads coming into this node if len(nodeGradMap[node]) < 1 { leaveLogScope() return nil, SymDiffError{ single: node, gradMap: nodeGradMap, err: errors.New("No gradients found for node"), } } // once we've reached a node, we already backpropagated from its dependents // so we sum up the gradients symdiffLogf("nodeGradMap[%x]: %d", node.ID(), nodeGradMap[node]) if len(nodeGradMap[node]) > 1 { var n *Node symdiffLogf("reduce adding") if n, err = ReduceAdd(nodeGradMap[node], WithGroupName(gradClust)); err != nil {
single: node, nodes: nodeGradMap[node], gradMap: nodeGradMap, err: errors.Wrap(err, "ReduceAdd failed during differentiation"), } } symdiffLogf("reduced to... %x", n.ID()) // node.derives = append(node.derives, n) n.derivOf = append(n.derivOf, node) node.deriv = n nodeGradMap[node] = Nodes{n} // } } else if len(nodeGradMap[node]) == 1 { deriv := nodeGradMap[node][0] deriv.derivOf =
leaveLogScope() return nil, SymDiffError{
random_line_split
differentiation.go
(outputs, sortedNodes Nodes) (retVal NodeSet, err error) { symdiffLogf("Forward analysis. Already sorted?") enterLogScope() defer leaveLogScope() if !outputs.AllSameGraph() { return nil, errors.New("The supplied output Nodes are not the same graph") } diffSet := outputs.mapSet() symdiffLogf("Diff Set: %v", diffSet) symdiffLogf("%d", sortedNodes) for _, n := range sortedNodes { if diffSet.Contains(n) && !n.isInput() { diffs := n.diffWRT() for j, child := range n.children { d := diffs[j] if d { symdiffLogf("Adding %x to differentiable set", child.ID()) diffSet.Add(child) } } } } return diffSet, nil } // backwardDiffAnalysis returns a list of Nodes that are affected by differentiating output. // Given a list of WRTs, we want to find a list of nodes that will be affected when backpropagating. func backwardDiffAnalysis(wrt, sortedNodes Nodes) (retVal NodeSet, err error) { symdiffLogf("Backwards analysis") enterLogScope() defer leaveLogScope() if !wrt.AllSameGraph() { return nil, errors.New("The supplied output Nodes are not the same graph") } diffSet := wrt.mapSet() symdiffLogf("wrt:%d diffset: %d", len(wrt), len(diffSet)) symdiffLogf("%v", diffSet) symdiffLogf("sorted: %d", sortedNodes) enterLogScope() for i := len(sortedNodes) - 1; i >= 0; i-- { n := sortedNodes[i] symdiffLogf("working on %v. Has %d children", n, len(n.children)) var op SDOp var ok bool var diffs []bool if op, ok = n.op.(SDOp); ok { diffs = op.DiffWRT(len(n.children)) } symdiffLogf("differentiable WRT: %v", diffs) enterLogScope() symdiffLogf("Children: %v", n.children) if len(diffs) == 0 { // check if this makes nodes unreachable. If it does, then error out if n.isStmt { symdiffLogf("Statement nodes are Non differentiable!") leaveLogScope() continue } else if n.isInput() { symdiffLogf("Input nodes are Non differentiable") leaveLogScope() continue } else if len(n.children) == 0 { symdiffLogf("Leaf nodes have no children") leaveLogScope() continue } g := n.g for _, child := range n.children { parents := graph.NodesOf(g.To(child.ID())) if len(parents) == 1 && len(child.children) > 0 { leaveLogScope() return nil, errors.Errorf("Being unable to differentiate %v would leave a portion of the graph unreachable. Unable to continue", n) } } symdiffLogf("SKIPPING... Non differentiable!") leaveLogScope() continue } inner: for j, child := range n.children { d := diffs[j] if diffSet.Contains(child) && d { symdiffLogf("Adding %x to differentiable set", child.ID()) diffSet.Add(n) break inner } } leaveLogScope() } leaveLogScope() return diffSet, nil } // Backpropagate backpropagates errors by performing reverse-mode symbolic differentiation, starting from the outputs, and working its way towads the inputs. // // This is the rough algorithm: // 1. Filter out nodes that are unreachable // 2. Forwards analysis, where a list of nodes affecting the output is added to consideration // 3. Backwards analysis, where a list of nodes affected by differentiating the output are added to the consideration // 4. If there is a difference in both sets, it will cause an error (both sets should be the same) // 5. Traverse the graph from output towards input. On each visit, perform the symbolic differentiation // // For most cases, Grad() should be used instead of Backpropagate(), as Grad() performs several checks which would be the general use case, before calling Backpropagate() func Backpropagate(outputs, gradOutputs, wrt Nodes) (retVal Nodes, err error) { symdiffLogf("BACKPROP START") symdiffLogf("Outputs: %d", outputs) symdiffLogf("gradOutputs: %d", gradOutputs) symdiffLogf("WRT: %d", wrt) enterLogScope() defer leaveLogScope() g := outputs[0].g // this entire section about removing foreveralone nodes need a rethink symdiffLogf("removing foreveralone nodes") enterLogScope() for i := 0; i < len(g.AllNodes()); i++ { n := g.AllNodes()[i] fr := g.From(n.ID()).Len() to := g.To(n.ID()).Len() if fr == 0 && to == 0 && !n.isConstant() && !n.isInput() { g.RemoveNode(n) symdiffLogf("removed %v(%p); %x; %s", n, n, n.ID(), n.Name()) } } leaveLogScope() var sortedNodes Nodes if sortedNodes, err = Sort(g); err != nil { return nil, errors.Wrap(err, sortFail) } symdiffLogf("sorted nodes: %v", sortedNodes) symdiffLogf("sorted nodes: %d", sortedNodes) var affectsOutput NodeSet var affectedByOutput NodeSet if affectsOutput, err = forwardDiffAnalysis(outputs, sortedNodes); err != nil { return nil, errors.Wrap(err, "Failed during forward differentiation analysis") } if affectedByOutput, err = backwardDiffAnalysis(wrt, sortedNodes); err != nil { return nil, errors.Wrap(err, "Failed during forward differentiation analysis") } symdiffLogf("affects output: %v", affectsOutput) symdiffLogf("affected by output : %v", affectedByOutput) wrtSet := wrt.mapSet() badWRTs := wrtSet.Difference(affectsOutput) if len(badWRTs) > 0 { return nil, SymDiffError{nodes: badWRTs.ToSlice(), err: errors.Errorf("Non Differentiable WRTs: %v", badWRTs)} } outputSet := outputs.mapSet() badOutputs := outputSet.Difference(affectedByOutput) if len(badOutputs) > 0 { symdiffLogf("badOutputs: %#v", badOutputs) return nil, SymDiffError{nodes: badOutputs.ToSlice(), err: errors.Errorf("Non-Differentable Outputs: %v", badOutputs)} } // map a node to a list of gradient terms // these gradient terms will be summed up when we visit the node // when iterating through the nondes in reverse topological order nodeGradMap := make(map[*Node]Nodes) for i, n := range outputs { symdiffLogf("Adding outputs for %x", n.ID()) nodeGradMap[n] = Nodes{gradOutputs[i]} } // "active" nodes are the ones that are differentially influenced by the inputs // and also differentiably influence the outputs. These are the nodes where we need to call the // "pullback" function to backpropagate derivatives activeNodes := affectsOutput.Intersect(affectedByOutput) symdiffLogf("Active: %v", activeNodes) symdiffLogf("Sorted: %d", sortedNodes) symdiffLogf("nodeGradMap: %+#d", FmtNodeMap(nodeGradMap)) enterLogScope() for _, node := range sortedNodes { if _, ok := activeNodes[node]; !ok { symdiffLogf("skipping %x", node.ID()) continue } if node.deriv != nil { symdiffLogf("skipping %x - previously differentiated", node.ID()) nodeGradMap[node] = append(nodeGradMap[node], node.deriv) continue } symdiffLogf("Working on %x %v", node.ID(), node) enterLogScope() // Check if there is any grads coming into this node if len(nodeGradMap[node]) < 1 { leaveLogScope() return nil, SymDiffError{ single: node, gradMap: nodeGradMap, err: errors.New("No gradients found for node"), } } // once we've reached a node, we already backpropagated from its dependents // so we sum up the gradients symdiffLogf("nodeGradMap[%x]: %d", node.ID(), nodeGradMap[node]) if len(nodeGradMap[node]) >
forwardDiffAnalysis
identifier_name
TypedAction.ts
.g. to a Redux `Store.dispatch`. * * The true benefit of TypedActions come on the Reducer-side. See * the TypedReducer class for more on creating a TypedAction-savvy Reducer for Redux. * * Conforms to Flux Standard Action recommendations. * * @see TypedActionDef#create */ export interface TypedAction<T, E extends string = string> { /** * The type string of the action, used to uniquely identify the Action with its Definition. * * The TypeScript typing of this value is refined to the actual string given to `TypedAction.define()`. */ type: E; /** * The payload associated with the action, whose shape is dictated by the Definition * that generated this action. As allowed by the payload restrictions chosen by the consumer, * this payload could possibly be `undefined` or `null`. * * N.B. A NoPayloadDefinition doesn't actually define this key for generated actions, so such * actions are usually just `{type: "..."}`. */ payload: T; /** * Optional metadata assigned to this action, which has no restrictions. * Interesting usages of metadata: * * 1. To add a timestamp for when the action was first created. * 1. To correlate a set of actions behind a single user event (such as a clickId). * 1. To track progression of an async task, such as a {loading => success} or {loading => error} set. * 1. To identify which actions are being triggered by a continually running job. * */ meta?: any; } export namespace TypedAction { /** * **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()` * over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35 * * Options to TypedAction.define(). * * @deprecated */ export interface DefineOptions<T> { /** * A function used to validate the (runtime) correctness of payloads attached to a Definition's * actions. This can be useful to track down a noncompliant _source_ of actions, * as otherwise debugging where incorrect actions are coming from on the Reducer side can be challenging. * * Not run in production. */ validate?: (payload: T) => boolean; } /** * **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()` * over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35 * * One of the core functions of Redoodle, `TypedAction.define` creates a Definition * to manage all Redux actions of a specific type string, such as `"myapp::set_foo_value"`. * * Each Definition also associates a payload type `T` for all of its matching actions. * For example, the `"myapp::set_foo_value"` Action can associate a required payload shape * `{foo: string, value: number}`, which means that all actions in the application * with type `"myapp::set_foo_value"` *must* have payloads with a `foo` and a `value`. * * The syntax for invoking the function is slightly awkward, in favor of more predictable type inferencing. * An example invocation is below; note the extra `()` after the payload type declaration in `<{}>`s. * * * ``` * export const SetFooValueAction = TypedAction.define("myapp::set_foo_value")<{ * foo: string; * value: number; * }>(); * ``` * * * All Definitions for a Redux-enabled application MUST have unique strings. * * @deprecated */ export function define<E extends string>( type: E, ): <T>(options?: DefineOptions<T>) => Definition<E, T> { return <T>(options?: DefineOptions<T>) => { if ( process.env.NODE_ENV !== "production" && options !== undefined && options.validate !== undefined ) { return createDefinitionWithValidator<E, T>(type, options.validate); } else { return createDefinition<E, T>(type); } }; } /** * **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()` * over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35 * * Similar to TypedAction.define, creates a NoPayloadDefinition for the given Action type * string, like `"example::clear_foo"`. In practice, actions without payloads are * usually of the "clear" or "invalidate" variety. * * The syntax for invoking the function is slightly awkward, in favor of more predictable type inferencing. * An example invocation is below; note the extra pair of `()`, for consistency with its sibling `define` * function and for better future-compatibility of options. * * * ``` * export const SetFooValueAction = TypedAction.defineWithoutPayload("myapp::set_foo_value")(); * ``` * * * All Definitions for a Redux-enabled application MUST have unique strings. * * @deprecated */ export function defineWithoutPayload<E extends string>( type: E, ): () => NoPayloadDefinition<E> { return () => { return createNoPayloadDefinition<E>(type); }; } /** * A central type of Redoodle, the TypedAction.Definition manages all Redux Actions * of a specific type string, such as `"myapp::set_foo_value"`. * * - Definitions should be used to create Actions. * - Definitions can be used to identify an Action, based on its own `type`. * * All Definitions for a Redux-enabled application MUST have unique strings. */ export interface Definition<E extends string, T> { /** * Creates an Action of this type with the given payload. * Functionally equivalent to the explicit Definition.create(). */ (payload: T): { type: E; payload: T }; /** * The Type of a TypedAction refers to the physical `{type}` string * given to matching Actions. This TypedActionString is branded * with the payload type as well for e.g. TypedReducer type inferencing. */ TYPE: TypedActionString<T, E>; /** * Hidden field used for some workflows that need to extract the payload type back out of * a TypedAction definition. For example, `const payload: typeof MyAction.__PAYLOAD = { ... };` * can be used to define a payload conforming to MyAction. * * This value should only be used for constructing Types in TypeScript. It never holds a real value. * Future versions of Redoodle may throw when attempting accessing this value at runtime * to catch accidental misuse. */ __PAYLOAD: T; /** * Creates an Action of this type with the given payload. */ create(payload: T): { type: E; payload: T }; /** * Creates an Action of this type with the given payload and meta. */ createWithMeta<M>(payload: T, meta: M): { type: E; payload: T; meta: M }; /** * Checks whether the given Action matches this Definition, based on its own `type`. * If so, we can safely narrow the Action's payload type based on this Definition. * * While this function can be used for action identification while Reducing, * TypedReducers provide much stronger utilities when working with TypedActions. */ is(action: Action): action is TypedAction<T, E>; } /** * @deprecated use top level PayloadOf */ export type PayloadOf< D extends Definition<any, any> > = D["TYPE"]["__type__"]["withPayload"]; /** * A TypedAction.NoPayloadDefinition manages all Redux actions of a specific type string, * such as `"myapp::clear_foo"`. Unlike the sibling TypedAction.Definition, * actions matching this Definition are associated with no payload data. In practice, * actions without payloads are usually of the "clear" or "invalidate" variety.
* All Definitions for a Redux-enabled application MUST have unique strings. */ export interface NoPayloadDefinition<E extends string> { /** * Creates an Action of this type (and no payload). * Functionally equivalent to the explicit NoPayloadDefinition.create(). */ (): { type: E; payload: never }; /** * The Type of a TypedAction refers to the physical `{type}` string * given to matching Actions. This TypedActionString is branded * with the payload type as well for e.g. TypedReducer type inferencing. */ TYPE: TypedActionString<never, E>; /** * Creates an Action of this type (and no payload). */ create(): { type: E; payload: never }; /** * Creates an Action of this type with the given meta (and no payload). */ createWithMeta<M>(meta:
* * - Definitions should be used to create Actions. * - Definitions can be used to identify an Action, based on its own `type`. *
random_line_split
TypedAction.ts
**DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()` * over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35 * * Options to TypedAction.define(). * * @deprecated */ export interface DefineOptions<T> { /** * A function used to validate the (runtime) correctness of payloads attached to a Definition's * actions. This can be useful to track down a noncompliant _source_ of actions, * as otherwise debugging where incorrect actions are coming from on the Reducer side can be challenging. * * Not run in production. */ validate?: (payload: T) => boolean; } /** * **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()` * over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35 * * One of the core functions of Redoodle, `TypedAction.define` creates a Definition * to manage all Redux actions of a specific type string, such as `"myapp::set_foo_value"`. * * Each Definition also associates a payload type `T` for all of its matching actions. * For example, the `"myapp::set_foo_value"` Action can associate a required payload shape * `{foo: string, value: number}`, which means that all actions in the application * with type `"myapp::set_foo_value"` *must* have payloads with a `foo` and a `value`. * * The syntax for invoking the function is slightly awkward, in favor of more predictable type inferencing. * An example invocation is below; note the extra `()` after the payload type declaration in `<{}>`s. * * * ``` * export const SetFooValueAction = TypedAction.define("myapp::set_foo_value")<{ * foo: string; * value: number; * }>(); * ``` * * * All Definitions for a Redux-enabled application MUST have unique strings. * * @deprecated */ export function define<E extends string>( type: E, ): <T>(options?: DefineOptions<T>) => Definition<E, T> { return <T>(options?: DefineOptions<T>) => { if ( process.env.NODE_ENV !== "production" && options !== undefined && options.validate !== undefined ) { return createDefinitionWithValidator<E, T>(type, options.validate); } else { return createDefinition<E, T>(type); } }; } /** * **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()` * over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35 * * Similar to TypedAction.define, creates a NoPayloadDefinition for the given Action type * string, like `"example::clear_foo"`. In practice, actions without payloads are * usually of the "clear" or "invalidate" variety. * * The syntax for invoking the function is slightly awkward, in favor of more predictable type inferencing. * An example invocation is below; note the extra pair of `()`, for consistency with its sibling `define` * function and for better future-compatibility of options. * * * ``` * export const SetFooValueAction = TypedAction.defineWithoutPayload("myapp::set_foo_value")(); * ``` * * * All Definitions for a Redux-enabled application MUST have unique strings. * * @deprecated */ export function defineWithoutPayload<E extends string>( type: E, ): () => NoPayloadDefinition<E> { return () => { return createNoPayloadDefinition<E>(type); }; } /** * A central type of Redoodle, the TypedAction.Definition manages all Redux Actions * of a specific type string, such as `"myapp::set_foo_value"`. * * - Definitions should be used to create Actions. * - Definitions can be used to identify an Action, based on its own `type`. * * All Definitions for a Redux-enabled application MUST have unique strings. */ export interface Definition<E extends string, T> { /** * Creates an Action of this type with the given payload. * Functionally equivalent to the explicit Definition.create(). */ (payload: T): { type: E; payload: T }; /** * The Type of a TypedAction refers to the physical `{type}` string * given to matching Actions. This TypedActionString is branded * with the payload type as well for e.g. TypedReducer type inferencing. */ TYPE: TypedActionString<T, E>; /** * Hidden field used for some workflows that need to extract the payload type back out of * a TypedAction definition. For example, `const payload: typeof MyAction.__PAYLOAD = { ... };` * can be used to define a payload conforming to MyAction. * * This value should only be used for constructing Types in TypeScript. It never holds a real value. * Future versions of Redoodle may throw when attempting accessing this value at runtime * to catch accidental misuse. */ __PAYLOAD: T; /** * Creates an Action of this type with the given payload. */ create(payload: T): { type: E; payload: T }; /** * Creates an Action of this type with the given payload and meta. */ createWithMeta<M>(payload: T, meta: M): { type: E; payload: T; meta: M }; /** * Checks whether the given Action matches this Definition, based on its own `type`. * If so, we can safely narrow the Action's payload type based on this Definition. * * While this function can be used for action identification while Reducing, * TypedReducers provide much stronger utilities when working with TypedActions. */ is(action: Action): action is TypedAction<T, E>; } /** * @deprecated use top level PayloadOf */ export type PayloadOf< D extends Definition<any, any> > = D["TYPE"]["__type__"]["withPayload"]; /** * A TypedAction.NoPayloadDefinition manages all Redux actions of a specific type string, * such as `"myapp::clear_foo"`. Unlike the sibling TypedAction.Definition, * actions matching this Definition are associated with no payload data. In practice, * actions without payloads are usually of the "clear" or "invalidate" variety. * * - Definitions should be used to create Actions. * - Definitions can be used to identify an Action, based on its own `type`. * * All Definitions for a Redux-enabled application MUST have unique strings. */ export interface NoPayloadDefinition<E extends string> { /** * Creates an Action of this type (and no payload). * Functionally equivalent to the explicit NoPayloadDefinition.create(). */ (): { type: E; payload: never }; /** * The Type of a TypedAction refers to the physical `{type}` string * given to matching Actions. This TypedActionString is branded * with the payload type as well for e.g. TypedReducer type inferencing. */ TYPE: TypedActionString<never, E>; /** * Creates an Action of this type (and no payload). */ create(): { type: E; payload: never }; /** * Creates an Action of this type with the given meta (and no payload). */ createWithMeta<M>(meta: M): { type: E; payload: never; meta: M }; /** * Checks whether the given Action matches this Definition, based on its own `type`. * If so, we can safely narrow the Action's payload type based on this Definition. * * While this function can be used for action identification while Reducing, * TypedReducers provide much stronger utilities when working with TypedActions. */ is(action: Action): action is TypedAction<never, E>; } function createDefinition<E extends string, T>(type: E): Definition<E, T> { const create = (payload: T): { type: E; payload: T } => { return { type, payload }; }; const createWithMeta = <M>( payload: T, meta: M, ): { type: E; payload: T; meta: M } => { return { type, payload, meta }; }; const is = (action: Action): action is TypedAction<T, E> => { return action.type === type; }; const def = create as Definition<E, T>; def.create = create; def.createWithMeta = createWithMeta; def.is = is; def.TYPE = type as TypedActionString<T, E>; return def; } function createDefinitionWithValidator<E extends string, T>( type: E, validate: (payload: T) => boolean, ): Definition<E, T> { const create = (payload: T): { type: E; payload: T } => { if (!validate(payload))
{ throw new Error(`'${type}' validation failed`); }
conditional_block
TypedAction.ts
/** * **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()` * over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35 * * One of the core functions of Redoodle, `TypedAction.define` creates a Definition * to manage all Redux actions of a specific type string, such as `"myapp::set_foo_value"`. * * Each Definition also associates a payload type `T` for all of its matching actions. * For example, the `"myapp::set_foo_value"` Action can associate a required payload shape * `{foo: string, value: number}`, which means that all actions in the application * with type `"myapp::set_foo_value"` *must* have payloads with a `foo` and a `value`. * * The syntax for invoking the function is slightly awkward, in favor of more predictable type inferencing. * An example invocation is below; note the extra `()` after the payload type declaration in `<{}>`s. * * * ``` * export const SetFooValueAction = TypedAction.define("myapp::set_foo_value")<{ * foo: string; * value: number; * }>(); * ``` * * * All Definitions for a Redux-enabled application MUST have unique strings. * * @deprecated */ export function define<E extends string>( type: E, ): <T>(options?: DefineOptions<T>) => Definition<E, T> { return <T>(options?: DefineOptions<T>) => { if ( process.env.NODE_ENV !== "production" && options !== undefined && options.validate !== undefined ) { return createDefinitionWithValidator<E, T>(type, options.validate); } else { return createDefinition<E, T>(type); } }; } /** * **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()` * over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35 * * Similar to TypedAction.define, creates a NoPayloadDefinition for the given Action type * string, like `"example::clear_foo"`. In practice, actions without payloads are * usually of the "clear" or "invalidate" variety. * * The syntax for invoking the function is slightly awkward, in favor of more predictable type inferencing. * An example invocation is below; note the extra pair of `()`, for consistency with its sibling `define` * function and for better future-compatibility of options. * * * ``` * export const SetFooValueAction = TypedAction.defineWithoutPayload("myapp::set_foo_value")(); * ``` * * * All Definitions for a Redux-enabled application MUST have unique strings. * * @deprecated */ export function defineWithoutPayload<E extends string>( type: E, ): () => NoPayloadDefinition<E> { return () => { return createNoPayloadDefinition<E>(type); }; } /** * A central type of Redoodle, the TypedAction.Definition manages all Redux Actions * of a specific type string, such as `"myapp::set_foo_value"`. * * - Definitions should be used to create Actions. * - Definitions can be used to identify an Action, based on its own `type`. * * All Definitions for a Redux-enabled application MUST have unique strings. */ export interface Definition<E extends string, T> { /** * Creates an Action of this type with the given payload. * Functionally equivalent to the explicit Definition.create(). */ (payload: T): { type: E; payload: T }; /** * The Type of a TypedAction refers to the physical `{type}` string * given to matching Actions. This TypedActionString is branded * with the payload type as well for e.g. TypedReducer type inferencing. */ TYPE: TypedActionString<T, E>; /** * Hidden field used for some workflows that need to extract the payload type back out of * a TypedAction definition. For example, `const payload: typeof MyAction.__PAYLOAD = { ... };` * can be used to define a payload conforming to MyAction. * * This value should only be used for constructing Types in TypeScript. It never holds a real value. * Future versions of Redoodle may throw when attempting accessing this value at runtime * to catch accidental misuse. */ __PAYLOAD: T; /** * Creates an Action of this type with the given payload. */ create(payload: T): { type: E; payload: T }; /** * Creates an Action of this type with the given payload and meta. */ createWithMeta<M>(payload: T, meta: M): { type: E; payload: T; meta: M }; /** * Checks whether the given Action matches this Definition, based on its own `type`. * If so, we can safely narrow the Action's payload type based on this Definition. * * While this function can be used for action identification while Reducing, * TypedReducers provide much stronger utilities when working with TypedActions. */ is(action: Action): action is TypedAction<T, E>; } /** * @deprecated use top level PayloadOf */ export type PayloadOf< D extends Definition<any, any> > = D["TYPE"]["__type__"]["withPayload"]; /** * A TypedAction.NoPayloadDefinition manages all Redux actions of a specific type string, * such as `"myapp::clear_foo"`. Unlike the sibling TypedAction.Definition, * actions matching this Definition are associated with no payload data. In practice, * actions without payloads are usually of the "clear" or "invalidate" variety. * * - Definitions should be used to create Actions. * - Definitions can be used to identify an Action, based on its own `type`. * * All Definitions for a Redux-enabled application MUST have unique strings. */ export interface NoPayloadDefinition<E extends string> { /** * Creates an Action of this type (and no payload). * Functionally equivalent to the explicit NoPayloadDefinition.create(). */ (): { type: E; payload: never }; /** * The Type of a TypedAction refers to the physical `{type}` string * given to matching Actions. This TypedActionString is branded * with the payload type as well for e.g. TypedReducer type inferencing. */ TYPE: TypedActionString<never, E>; /** * Creates an Action of this type (and no payload). */ create(): { type: E; payload: never }; /** * Creates an Action of this type with the given meta (and no payload). */ createWithMeta<M>(meta: M): { type: E; payload: never; meta: M }; /** * Checks whether the given Action matches this Definition, based on its own `type`. * If so, we can safely narrow the Action's payload type based on this Definition. * * While this function can be used for action identification while Reducing, * TypedReducers provide much stronger utilities when working with TypedActions. */ is(action: Action): action is TypedAction<never, E>; } function createDefinition<E extends string, T>(type: E): Definition<E, T> { const create = (payload: T): { type: E; payload: T } => { return { type, payload }; }; const createWithMeta = <M>( payload: T, meta: M, ): { type: E; payload: T; meta: M } => { return { type, payload, meta }; }; const is = (action: Action): action is TypedAction<T, E> => { return action.type === type; }; const def = create as Definition<E, T>; def.create = create; def.createWithMeta = createWithMeta; def.is = is; def.TYPE = type as TypedActionString<T, E>; return def; } function createDefinitionWithValidator<E extends string, T>( type: E, validate: (payload: T) => boolean, ): Definition<E, T> { const create = (payload: T): { type: E; payload: T } => { if (!validate(payload)) { throw new Error(`'${type}' validation failed`); } return { type, payload }; }; const createWithMeta = <M>( payload: T, meta: M, ): { type: E; payload: T; meta: M } => { if (!validate(payload)) { throw new Error(`'${type}' validation failed`); } return { type, payload, meta }; }; const is = (action: Action): action is TypedAction<T, E> => { return action.type === type; }; const def = create as Definition<E, T>; def.create = create; def.createWithMeta = createWithMeta; def.is = is; def.TYPE = type as TypedActionString<T, E>; return def; } function
createNoPayloadDefinition
identifier_name
TypedAction.ts
. to a Redux `Store.dispatch`. * * The true benefit of TypedActions come on the Reducer-side. See * the TypedReducer class for more on creating a TypedAction-savvy Reducer for Redux. * * Conforms to Flux Standard Action recommendations. * * @see TypedActionDef#create */ export interface TypedAction<T, E extends string = string> { /** * The type string of the action, used to uniquely identify the Action with its Definition. * * The TypeScript typing of this value is refined to the actual string given to `TypedAction.define()`. */ type: E; /** * The payload associated with the action, whose shape is dictated by the Definition * that generated this action. As allowed by the payload restrictions chosen by the consumer, * this payload could possibly be `undefined` or `null`. * * N.B. A NoPayloadDefinition doesn't actually define this key for generated actions, so such * actions are usually just `{type: "..."}`. */ payload: T; /** * Optional metadata assigned to this action, which has no restrictions. * Interesting usages of metadata: * * 1. To add a timestamp for when the action was first created. * 1. To correlate a set of actions behind a single user event (such as a clickId). * 1. To track progression of an async task, such as a {loading => success} or {loading => error} set. * 1. To identify which actions are being triggered by a continually running job. * */ meta?: any; } export namespace TypedAction { /** * **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()` * over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35 * * Options to TypedAction.define(). * * @deprecated */ export interface DefineOptions<T> { /** * A function used to validate the (runtime) correctness of payloads attached to a Definition's * actions. This can be useful to track down a noncompliant _source_ of actions, * as otherwise debugging where incorrect actions are coming from on the Reducer side can be challenging. * * Not run in production. */ validate?: (payload: T) => boolean; } /** * **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()` * over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35 * * One of the core functions of Redoodle, `TypedAction.define` creates a Definition * to manage all Redux actions of a specific type string, such as `"myapp::set_foo_value"`. * * Each Definition also associates a payload type `T` for all of its matching actions. * For example, the `"myapp::set_foo_value"` Action can associate a required payload shape * `{foo: string, value: number}`, which means that all actions in the application * with type `"myapp::set_foo_value"` *must* have payloads with a `foo` and a `value`. * * The syntax for invoking the function is slightly awkward, in favor of more predictable type inferencing. * An example invocation is below; note the extra `()` after the payload type declaration in `<{}>`s. * * * ``` * export const SetFooValueAction = TypedAction.define("myapp::set_foo_value")<{ * foo: string; * value: number; * }>(); * ``` * * * All Definitions for a Redux-enabled application MUST have unique strings. * * @deprecated */ export function define<E extends string>( type: E, ): <T>(options?: DefineOptions<T>) => Definition<E, T>
/** * **DEPRECATED**: As of Redoodle 2.5.0, consumers should prefer `defineAction()` * over than `TypedAction.define()`. See https://github.com/palantir/redoodle/issues/35 * * Similar to TypedAction.define, creates a NoPayloadDefinition for the given Action type * string, like `"example::clear_foo"`. In practice, actions without payloads are * usually of the "clear" or "invalidate" variety. * * The syntax for invoking the function is slightly awkward, in favor of more predictable type inferencing. * An example invocation is below; note the extra pair of `()`, for consistency with its sibling `define` * function and for better future-compatibility of options. * * * ``` * export const SetFooValueAction = TypedAction.defineWithoutPayload("myapp::set_foo_value")(); * ``` * * * All Definitions for a Redux-enabled application MUST have unique strings. * * @deprecated */ export function defineWithoutPayload<E extends string>( type: E, ): () => NoPayloadDefinition<E> { return () => { return createNoPayloadDefinition<E>(type); }; } /** * A central type of Redoodle, the TypedAction.Definition manages all Redux Actions * of a specific type string, such as `"myapp::set_foo_value"`. * * - Definitions should be used to create Actions. * - Definitions can be used to identify an Action, based on its own `type`. * * All Definitions for a Redux-enabled application MUST have unique strings. */ export interface Definition<E extends string, T> { /** * Creates an Action of this type with the given payload. * Functionally equivalent to the explicit Definition.create(). */ (payload: T): { type: E; payload: T }; /** * The Type of a TypedAction refers to the physical `{type}` string * given to matching Actions. This TypedActionString is branded * with the payload type as well for e.g. TypedReducer type inferencing. */ TYPE: TypedActionString<T, E>; /** * Hidden field used for some workflows that need to extract the payload type back out of * a TypedAction definition. For example, `const payload: typeof MyAction.__PAYLOAD = { ... };` * can be used to define a payload conforming to MyAction. * * This value should only be used for constructing Types in TypeScript. It never holds a real value. * Future versions of Redoodle may throw when attempting accessing this value at runtime * to catch accidental misuse. */ __PAYLOAD: T; /** * Creates an Action of this type with the given payload. */ create(payload: T): { type: E; payload: T }; /** * Creates an Action of this type with the given payload and meta. */ createWithMeta<M>(payload: T, meta: M): { type: E; payload: T; meta: M }; /** * Checks whether the given Action matches this Definition, based on its own `type`. * If so, we can safely narrow the Action's payload type based on this Definition. * * While this function can be used for action identification while Reducing, * TypedReducers provide much stronger utilities when working with TypedActions. */ is(action: Action): action is TypedAction<T, E>; } /** * @deprecated use top level PayloadOf */ export type PayloadOf< D extends Definition<any, any> > = D["TYPE"]["__type__"]["withPayload"]; /** * A TypedAction.NoPayloadDefinition manages all Redux actions of a specific type string, * such as `"myapp::clear_foo"`. Unlike the sibling TypedAction.Definition, * actions matching this Definition are associated with no payload data. In practice, * actions without payloads are usually of the "clear" or "invalidate" variety. * * - Definitions should be used to create Actions. * - Definitions can be used to identify an Action, based on its own `type`. * * All Definitions for a Redux-enabled application MUST have unique strings. */ export interface NoPayloadDefinition<E extends string> { /** * Creates an Action of this type (and no payload). * Functionally equivalent to the explicit NoPayloadDefinition.create(). */ (): { type: E; payload: never }; /** * The Type of a TypedAction refers to the physical `{type}` string * given to matching Actions. This TypedActionString is branded * with the payload type as well for e.g. TypedReducer type inferencing. */ TYPE: TypedActionString<never, E>; /** * Creates an Action of this type (and no payload). */ create(): { type: E; payload: never }; /** * Creates an Action of this type with the given meta (and no payload). */ createWithMeta<M>(meta
{ return <T>(options?: DefineOptions<T>) => { if ( process.env.NODE_ENV !== "production" && options !== undefined && options.validate !== undefined ) { return createDefinitionWithValidator<E, T>(type, options.validate); } else { return createDefinition<E, T>(type); } }; }
identifier_body
time-sentiment.component.ts
renderedChart = false; notDataWarn = false; appliedFilter = false; chartHeight = 300; chartRange1; chartRange2; chartRangeFilter1; chartRangeFilter2; constructor(private chartService: ChartService, private _element: ElementRef) { } ngOnInit() { this.sentimentLineChart = dc.lineChart('#sentimentChartLine'); this.chartService.GetData().subscribe((data) => { this.data = data; }); // Crossfilter this.chartService.getCrossfilter().subscribe((filter) => { this.cfilter = filter; this.setDimension(); if (this.data && this.data.length > 0) { this.sentGroups = this.getSentGroups(); if (this.sentGroups[0]) { this.notDataWarn = false; this.appliedFilter = false; this.renderChart(); } else { this.notDataWarn = true; } } }); // Collapsible view this.chartService.GetChartMode().subscribe(mode => { if (this.data && this.data.length > 0) { if (mode && mode === 'small') { this.chartHeight = 85; this.renderChart(); } else if (mode && mode === 'big') { this.chartHeight = 300; this.renderChart(); } } }); // Gets the range through the chart service from the mainVis Component this.chartService.getChartRange().subscribe((range) => { if (range.chart === null) { if (this.data && range.range) { (this.diff_months(range.range[0], range.range[1]) < 2) ? this.notDataWarn = true : this.notDataWarn = false; this.chartRangeFilter1 = range.range[0]; this.chartRangeFilter2 = range.range[1]; this.sentimentLineChart .x(d3.scaleTime().domain([this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2])) .y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2)])) .round(d3.timeMonth); this.appliedFilter = true; this.sentimentLineChart.redraw(); } else { if (!dc.chartRegistry.list().some((c) => c.hasFilter())) { this.notDataWarn = false; this.sentimentLineChart .x(d3.scaleTime().domain([this.chartRange1, this.chartRange2])) .y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)])); this.appliedFilter = false; } } } }); this.renderedChart = false; } // Buttons and Front-End //////////////////////////////////////////////////////////////////////////////////////////// // sets the tooltip on mouseover setTooltipInfo(event: MouseEvent, tooltip: HTMLSpanElement) { tooltip.style.position = 'fixed'; tooltip.style.top = (event.clientY) + 'px'; tooltip.style.left = (event.clientX - tooltip.offsetWidth - 5) + 'px'; } // sets the crossfilter dimension setDimension() { this.dimension = this.cfilter.dimension((d: any) => { const splitted = d.publishedAt.split('-'); return new Date(splitted[0] + '-' + splitted[1]); }); } // Time-based Stacked Chart ///////////////////////////////////////////////////////////////////////////////////////// // returns a crossfilter-group for each sentiment x private getSentGroups(): { group: CrossFilter.Group<{}, Date, any>, sent: string}[] { if (this.data && this.data.length < 0) { return; } const groups: { group: CrossFilter.Group<{}, Date, any>, sent: string}[] = []; // group by sentiment const nested = d3.nest() .key((d: any) => { if (d.analysis) { if (d.analysis.mainSentiment === 'mixed') { return 'Mix'; } else if (d.analysis.mainSentiment === 'positive') { return 'Pos'; } else if (d.analysis.mainSentiment === 'neutral') { return 'Neu'; } else if (d.analysis.mainSentiment === 'negative') { return 'Neg'; } else if (d.analysis.mainSentiment === 'na') { return 'N/A'; } } }) .entries(this.data); nested.forEach((sentiment) => { const g = this.dimension.group().reduceSum((d: any) => { if (d.analysis) { let mainsentiment = ''; if (d.analysis.mainSentiment === 'mixed') { mainsentiment = 'Mix'; } else if (d.analysis.mainSentiment === 'positive') { mainsentiment = 'Pos'; } else if (d.analysis.mainSentiment === 'neutral') { mainsentiment = 'Neu'; } else if (d.analysis.mainSentiment === 'negative') { mainsentiment = 'Neg'; } else if (d.analysis.mainSentiment === 'na') { mainsentiment = 'N/A'; } return mainsentiment === sentiment.key; } }); groups.push({group: g, sent: sentiment.key }); }); return groups; } getStartDate(previousDate) { const date = new Date(previousDate); date.setDate(1); return date; } // Renders line chart (aggregation) renderChart () { this.chartRange1 = d3.min(this.data, (d: any) => new Date(d.publishedAt)); this.chartRange1 = this.getStartDate(this.chartRange1); this.chartRange2 = d3.max(this.data, (d: any) => new Date(d.publishedAt)); const sentGroupsOrdered = this.reorderGroups(); const chartColors = this.defineChartColors(); let firstItem = 0; while (!sentGroupsOrdered[firstItem] && firstItem < 5) {firstItem++; } const group1 = sentGroupsOrdered[firstItem]; this.sentimentLineChart .renderArea(true) .width(900) .height(this.chartHeight) .ordinalColors(chartColors) .useViewBoxResizing(true) .dimension(this.dimension) .x(d3.scaleTime().domain([this.chartRange1, this.chartRange2])) .y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)])) .yAxisLabel('Comments') .interpolate('monotone') .legend(dc.legend().x(850).y(0).itemHeight(9).gap(5)) .brushOn(false) .group(group1.group, group1.sent) .valueAccessor(function (d) { return d.value; }) .xAxis().ticks(7); let maxSent = 0; if (sentGroupsOrdered.length > 1) { sentGroupsOrdered.forEach((group) => { if (group.group === group1.group || maxSent === 4) { return; } // stacks the groups this.sentimentLineChart .stack(group.group, group.sent, function (d) { return d.value; }); maxSent++; }); } // When filter is applied before refreshing the chart if (this.appliedFilter) { this.sentimentLineChart.x(d3.scaleTime().domain([this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2])); } // Brush: get range and send it to the other charts on brush-filtering this.sentimentLineChart.on('filtered', (chart, filter) => { if (filter) { this.sentimentLineChart.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(filter[0], filter[1])])); } else { this.sentimentLineChart.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)])); } this.chartService.setChartRange({range: filter, chart: chart}); }); // Adapt chart for smaller view (this.chartHeight < 300) ? this.sentimentLineChart.yAxis().ticks(2) : this.sentimentLineChart.yAxis().ticks(10); (this.chartHeight < 300) ? this.sentimentLineChart.xAxisLabel('') : this.sentimentLineChart.xAxisLabel('Date'); this.sentimentLineChart.xAxis().tickFormat(d3.timeFormat('%b %Y')); // month this.sentimentLineChart.render(); } // Adaptable color scale defineChartColors() { const colorArray = []; const sentGroupsOrdered = this.reorderGroups(); sentGroupsOrdered.forEach((g) => { if (g.sent === 'Pos') { colorArray.push('#4daf4a'); } else if (g.sent === 'Neu') { colorArray.push('#666666'); } else if (g.sent === 'Neg') { colorArray.push('#ff7f00'); } else if (g.sent === 'Mix') { colorArray.push('#984ea3'); } else if (g.sent === 'N/A')
{ colorArray.push('#DDDDDD'); }
conditional_block
time-sentiment.component.ts
styleUrls: ['./time-sentiment.component.scss'] }) export class TimeSentimentComponent implements OnInit { aggrView = true; compView = false; data: any[]; cfilter: CrossFilter.CrossFilter<{}>; dimension: CrossFilter.Dimension<{}, Date>; sentGroups: { group: CrossFilter.Group<{}, Date, any>, sent: string}[]; sentimentLineChart: dc.LineChart; renderedChart = false; notDataWarn = false; appliedFilter = false; chartHeight = 300; chartRange1; chartRange2; chartRangeFilter1; chartRangeFilter2; constructor(private chartService: ChartService, private _element: ElementRef) { } ngOnInit() { this.sentimentLineChart = dc.lineChart('#sentimentChartLine'); this.chartService.GetData().subscribe((data) => { this.data = data; }); // Crossfilter this.chartService.getCrossfilter().subscribe((filter) => { this.cfilter = filter; this.setDimension(); if (this.data && this.data.length > 0) { this.sentGroups = this.getSentGroups(); if (this.sentGroups[0]) { this.notDataWarn = false; this.appliedFilter = false; this.renderChart(); } else { this.notDataWarn = true; } } }); // Collapsible view this.chartService.GetChartMode().subscribe(mode => { if (this.data && this.data.length > 0) { if (mode && mode === 'small') { this.chartHeight = 85; this.renderChart(); } else if (mode && mode === 'big') { this.chartHeight = 300; this.renderChart(); } } }); // Gets the range through the chart service from the mainVis Component this.chartService.getChartRange().subscribe((range) => { if (range.chart === null) { if (this.data && range.range) { (this.diff_months(range.range[0], range.range[1]) < 2) ? this.notDataWarn = true : this.notDataWarn = false; this.chartRangeFilter1 = range.range[0]; this.chartRangeFilter2 = range.range[1]; this.sentimentLineChart .x(d3.scaleTime().domain([this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2])) .y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2)])) .round(d3.timeMonth); this.appliedFilter = true; this.sentimentLineChart.redraw(); } else { if (!dc.chartRegistry.list().some((c) => c.hasFilter())) { this.notDataWarn = false; this.sentimentLineChart .x(d3.scaleTime().domain([this.chartRange1, this.chartRange2])) .y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)])); this.appliedFilter = false; } } } }); this.renderedChart = false; } // Buttons and Front-End //////////////////////////////////////////////////////////////////////////////////////////// // sets the tooltip on mouseover setTooltipInfo(event: MouseEvent, tooltip: HTMLSpanElement) { tooltip.style.position = 'fixed'; tooltip.style.top = (event.clientY) + 'px'; tooltip.style.left = (event.clientX - tooltip.offsetWidth - 5) + 'px'; } // sets the crossfilter dimension setDimension() { this.dimension = this.cfilter.dimension((d: any) => { const splitted = d.publishedAt.split('-'); return new Date(splitted[0] + '-' + splitted[1]); }); } // Time-based Stacked Chart ///////////////////////////////////////////////////////////////////////////////////////// // returns a crossfilter-group for each sentiment x private getSentGroups(): { group: CrossFilter.Group<{}, Date, any>, sent: string}[] { if (this.data && this.data.length < 0) { return; } const groups: { group: CrossFilter.Group<{}, Date, any>, sent: string}[] = []; // group by sentiment const nested = d3.nest() .key((d: any) => { if (d.analysis) { if (d.analysis.mainSentiment === 'mixed') { return 'Mix'; } else if (d.analysis.mainSentiment === 'positive') { return 'Pos'; } else if (d.analysis.mainSentiment === 'neutral') { return 'Neu'; } else if (d.analysis.mainSentiment === 'negative') { return 'Neg'; } else if (d.analysis.mainSentiment === 'na') { return 'N/A'; } } }) .entries(this.data); nested.forEach((sentiment) => { const g = this.dimension.group().reduceSum((d: any) => { if (d.analysis) { let mainsentiment = ''; if (d.analysis.mainSentiment === 'mixed') { mainsentiment = 'Mix'; } else if (d.analysis.mainSentiment === 'positive') { mainsentiment = 'Pos'; } else if (d.analysis.mainSentiment === 'neutral') { mainsentiment = 'Neu'; } else if (d.analysis.mainSentiment === 'negative') { mainsentiment = 'Neg'; } else if (d.analysis.mainSentiment === 'na') { mainsentiment = 'N/A'; } return mainsentiment === sentiment.key; } }); groups.push({group: g, sent: sentiment.key }); }); return groups; } getStartDate(previousDate) { const date = new Date(previousDate); date.setDate(1); return date; } // Renders line chart (aggregation) renderChart () { this.chartRange1 = d3.min(this.data, (d: any) => new Date(d.publishedAt)); this.chartRange1 = this.getStartDate(this.chartRange1); this.chartRange2 = d3.max(this.data, (d: any) => new Date(d.publishedAt)); const sentGroupsOrdered = this.reorderGroups(); const chartColors = this.defineChartColors(); let firstItem = 0; while (!sentGroupsOrdered[firstItem] && firstItem < 5) {firstItem++; } const group1 = sentGroupsOrdered[firstItem]; this.sentimentLineChart .renderArea(true) .width(900) .height(this.chartHeight) .ordinalColors(chartColors) .useViewBoxResizing(true) .dimension(this.dimension) .x(d3.scaleTime().domain([this.chartRange1, this.chartRange2])) .y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)])) .yAxisLabel('Comments') .interpolate('monotone') .legend(dc.legend().x(850).y(0).itemHeight(9).gap(5)) .brushOn(false) .group(group1.group, group1.sent) .valueAccessor(function (d) { return d.value; }) .xAxis().ticks(7); let maxSent = 0; if (sentGroupsOrdered.length > 1) { sentGroupsOrdered.forEach((group) => { if (group.group === group1.group || maxSent === 4) { return; } // stacks the groups this.sentimentLineChart .stack(group.group, group.sent, function (d) { return d.value; }); maxSent++; }); } // When filter is applied before refreshing the chart if (this.appliedFilter) { this.sentimentLineChart.x(d3.scaleTime().domain([this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2])); } // Brush: get range and send it to the other charts on brush-filtering this.sentimentLineChart.on('filtered', (chart, filter) => { if (filter) { this.sentimentLineChart.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(filter[0], filter[1])])); } else { this.sentimentLineChart.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)])); } this.chartService.setChartRange({range: filter, chart: chart}); }); // Adapt chart for smaller view (this.chartHeight < 300) ? this.sentimentLineChart.yAxis().ticks(2) : this.sentimentLineChart.yAxis().ticks(10); (this.chartHeight < 300) ? this.sentimentLineChart.xAxisLabel('') : this.sentimentLineChart.xAxisLabel('Date'); this.sentimentLineChart.xAxis().tickFormat(d3.timeFormat('%b %Y')); // month this.sentimentLineChart.render(); } // Adaptable color scale defineChartColors() { const colorArray = []; const sentGroupsOrdered = this.reorderGroups(); sentGroupsOrdered.forEach((g) => { if (g.sent === 'Pos') { colorArray.push('#4daf4a'); } else if (
random_line_split
time-sentiment.component.ts
sentimentLineChart: dc.LineChart; renderedChart = false; notDataWarn = false; appliedFilter = false; chartHeight = 300; chartRange1; chartRange2; chartRangeFilter1; chartRangeFilter2; constructor(private chartService: ChartService, private _element: ElementRef) { } ngOnInit()
// Collapsible view this.chartService.GetChartMode().subscribe(mode => { if (this.data && this.data.length > 0) { if (mode && mode === 'small') { this.chartHeight = 85; this.renderChart(); } else if (mode && mode === 'big') { this.chartHeight = 300; this.renderChart(); } } }); // Gets the range through the chart service from the mainVis Component this.chartService.getChartRange().subscribe((range) => { if (range.chart === null) { if (this.data && range.range) { (this.diff_months(range.range[0], range.range[1]) < 2) ? this.notDataWarn = true : this.notDataWarn = false; this.chartRangeFilter1 = range.range[0]; this.chartRangeFilter2 = range.range[1]; this.sentimentLineChart .x(d3.scaleTime().domain([this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2])) .y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2)])) .round(d3.timeMonth); this.appliedFilter = true; this.sentimentLineChart.redraw(); } else { if (!dc.chartRegistry.list().some((c) => c.hasFilter())) { this.notDataWarn = false; this.sentimentLineChart .x(d3.scaleTime().domain([this.chartRange1, this.chartRange2])) .y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)])); this.appliedFilter = false; } } } }); this.renderedChart = false; } // Buttons and Front-End //////////////////////////////////////////////////////////////////////////////////////////// // sets the tooltip on mouseover setTooltipInfo(event: MouseEvent, tooltip: HTMLSpanElement) { tooltip.style.position = 'fixed'; tooltip.style.top = (event.clientY) + 'px'; tooltip.style.left = (event.clientX - tooltip.offsetWidth - 5) + 'px'; } // sets the crossfilter dimension setDimension() { this.dimension = this.cfilter.dimension((d: any) => { const splitted = d.publishedAt.split('-'); return new Date(splitted[0] + '-' + splitted[1]); }); } // Time-based Stacked Chart ///////////////////////////////////////////////////////////////////////////////////////// // returns a crossfilter-group for each sentiment x private getSentGroups(): { group: CrossFilter.Group<{}, Date, any>, sent: string}[] { if (this.data && this.data.length < 0) { return; } const groups: { group: CrossFilter.Group<{}, Date, any>, sent: string}[] = []; // group by sentiment const nested = d3.nest() .key((d: any) => { if (d.analysis) { if (d.analysis.mainSentiment === 'mixed') { return 'Mix'; } else if (d.analysis.mainSentiment === 'positive') { return 'Pos'; } else if (d.analysis.mainSentiment === 'neutral') { return 'Neu'; } else if (d.analysis.mainSentiment === 'negative') { return 'Neg'; } else if (d.analysis.mainSentiment === 'na') { return 'N/A'; } } }) .entries(this.data); nested.forEach((sentiment) => { const g = this.dimension.group().reduceSum((d: any) => { if (d.analysis) { let mainsentiment = ''; if (d.analysis.mainSentiment === 'mixed') { mainsentiment = 'Mix'; } else if (d.analysis.mainSentiment === 'positive') { mainsentiment = 'Pos'; } else if (d.analysis.mainSentiment === 'neutral') { mainsentiment = 'Neu'; } else if (d.analysis.mainSentiment === 'negative') { mainsentiment = 'Neg'; } else if (d.analysis.mainSentiment === 'na') { mainsentiment = 'N/A'; } return mainsentiment === sentiment.key; } }); groups.push({group: g, sent: sentiment.key }); }); return groups; } getStartDate(previousDate) { const date = new Date(previousDate); date.setDate(1); return date; } // Renders line chart (aggregation) renderChart () { this.chartRange1 = d3.min(this.data, (d: any) => new Date(d.publishedAt)); this.chartRange1 = this.getStartDate(this.chartRange1); this.chartRange2 = d3.max(this.data, (d: any) => new Date(d.publishedAt)); const sentGroupsOrdered = this.reorderGroups(); const chartColors = this.defineChartColors(); let firstItem = 0; while (!sentGroupsOrdered[firstItem] && firstItem < 5) {firstItem++; } const group1 = sentGroupsOrdered[firstItem]; this.sentimentLineChart .renderArea(true) .width(900) .height(this.chartHeight) .ordinalColors(chartColors) .useViewBoxResizing(true) .dimension(this.dimension) .x(d3.scaleTime().domain([this.chartRange1, this.chartRange2])) .y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)])) .yAxisLabel('Comments') .interpolate('monotone') .legend(dc.legend().x(850).y(0).itemHeight(9).gap(5)) .brushOn(false) .group(group1.group, group1.sent) .valueAccessor(function (d) { return d.value; }) .xAxis().ticks(7); let maxSent = 0; if (sentGroupsOrdered.length > 1) { sentGroupsOrdered.forEach((group) => { if (group.group === group1.group || maxSent === 4) { return; } // stacks the groups this.sentimentLineChart .stack(group.group, group.sent, function (d) { return d.value; }); maxSent++; }); } // When filter is applied before refreshing the chart if (this.appliedFilter) { this.sentimentLineChart.x(d3.scaleTime().domain([this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2])); } // Brush: get range and send it to the other charts on brush-filtering this.sentimentLineChart.on('filtered', (chart, filter) => { if (filter) { this.sentimentLineChart.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(filter[0], filter[1])])); } else { this.sentimentLineChart.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)])); } this.chartService.setChartRange({range: filter, chart: chart}); }); // Adapt chart for smaller view (this.chartHeight < 300) ? this.sentimentLineChart.yAxis().ticks(2) : this.sentimentLineChart.yAxis().ticks(10); (this.chartHeight < 300) ? this.sentimentLineChart.xAxisLabel('') : this.sentimentLineChart.xAxisLabel('Date'); this.sentimentLineChart.xAxis().tickFormat(d3.timeFormat('%b %Y')); // month this.sentimentLineChart.render(); } // Adaptable color scale defineChartColors() { const colorArray = []; const sentGroupsOrdered = this.reorderGroups(); sentGroupsOrdered.forEach((g) => { if (g.sent === 'Pos') { colorArray.push('#4daf4a'); } else if (g.sent === 'Neu') { colorArray.push('#666666'); } else if (g.sent === 'Neg') { colorArray.push('#ff7f00'); } else if (g.sent === 'Mix') { colorArray.push('#984ea3'); } else if (g.sent === 'N/A') {
{ this.sentimentLineChart = dc.lineChart('#sentimentChartLine'); this.chartService.GetData().subscribe((data) => { this.data = data; }); // Crossfilter this.chartService.getCrossfilter().subscribe((filter) => { this.cfilter = filter; this.setDimension(); if (this.data && this.data.length > 0) { this.sentGroups = this.getSentGroups(); if (this.sentGroups[0]) { this.notDataWarn = false; this.appliedFilter = false; this.renderChart(); } else { this.notDataWarn = true; } } });
identifier_body
time-sentiment.component.ts
implements OnInit { aggrView = true; compView = false; data: any[]; cfilter: CrossFilter.CrossFilter<{}>; dimension: CrossFilter.Dimension<{}, Date>; sentGroups: { group: CrossFilter.Group<{}, Date, any>, sent: string}[]; sentimentLineChart: dc.LineChart; renderedChart = false; notDataWarn = false; appliedFilter = false; chartHeight = 300; chartRange1; chartRange2; chartRangeFilter1; chartRangeFilter2; constructor(private chartService: ChartService, private _element: ElementRef) { } ngOnInit() { this.sentimentLineChart = dc.lineChart('#sentimentChartLine'); this.chartService.GetData().subscribe((data) => { this.data = data; }); // Crossfilter this.chartService.getCrossfilter().subscribe((filter) => { this.cfilter = filter; this.setDimension(); if (this.data && this.data.length > 0) { this.sentGroups = this.getSentGroups(); if (this.sentGroups[0]) { this.notDataWarn = false; this.appliedFilter = false; this.renderChart(); } else { this.notDataWarn = true; } } }); // Collapsible view this.chartService.GetChartMode().subscribe(mode => { if (this.data && this.data.length > 0) { if (mode && mode === 'small') { this.chartHeight = 85; this.renderChart(); } else if (mode && mode === 'big') { this.chartHeight = 300; this.renderChart(); } } }); // Gets the range through the chart service from the mainVis Component this.chartService.getChartRange().subscribe((range) => { if (range.chart === null) { if (this.data && range.range) { (this.diff_months(range.range[0], range.range[1]) < 2) ? this.notDataWarn = true : this.notDataWarn = false; this.chartRangeFilter1 = range.range[0]; this.chartRangeFilter2 = range.range[1]; this.sentimentLineChart .x(d3.scaleTime().domain([this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2])) .y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2)])) .round(d3.timeMonth); this.appliedFilter = true; this.sentimentLineChart.redraw(); } else { if (!dc.chartRegistry.list().some((c) => c.hasFilter())) { this.notDataWarn = false; this.sentimentLineChart .x(d3.scaleTime().domain([this.chartRange1, this.chartRange2])) .y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)])); this.appliedFilter = false; } } } }); this.renderedChart = false; } // Buttons and Front-End //////////////////////////////////////////////////////////////////////////////////////////// // sets the tooltip on mouseover setTooltipInfo(event: MouseEvent, tooltip: HTMLSpanElement) { tooltip.style.position = 'fixed'; tooltip.style.top = (event.clientY) + 'px'; tooltip.style.left = (event.clientX - tooltip.offsetWidth - 5) + 'px'; } // sets the crossfilter dimension setDimension() { this.dimension = this.cfilter.dimension((d: any) => { const splitted = d.publishedAt.split('-'); return new Date(splitted[0] + '-' + splitted[1]); }); } // Time-based Stacked Chart ///////////////////////////////////////////////////////////////////////////////////////// // returns a crossfilter-group for each sentiment x private getSentGroups(): { group: CrossFilter.Group<{}, Date, any>, sent: string}[] { if (this.data && this.data.length < 0) { return; } const groups: { group: CrossFilter.Group<{}, Date, any>, sent: string}[] = []; // group by sentiment const nested = d3.nest() .key((d: any) => { if (d.analysis) { if (d.analysis.mainSentiment === 'mixed') { return 'Mix'; } else if (d.analysis.mainSentiment === 'positive') { return 'Pos'; } else if (d.analysis.mainSentiment === 'neutral') { return 'Neu'; } else if (d.analysis.mainSentiment === 'negative') { return 'Neg'; } else if (d.analysis.mainSentiment === 'na') { return 'N/A'; } } }) .entries(this.data); nested.forEach((sentiment) => { const g = this.dimension.group().reduceSum((d: any) => { if (d.analysis) { let mainsentiment = ''; if (d.analysis.mainSentiment === 'mixed') { mainsentiment = 'Mix'; } else if (d.analysis.mainSentiment === 'positive') { mainsentiment = 'Pos'; } else if (d.analysis.mainSentiment === 'neutral') { mainsentiment = 'Neu'; } else if (d.analysis.mainSentiment === 'negative') { mainsentiment = 'Neg'; } else if (d.analysis.mainSentiment === 'na') { mainsentiment = 'N/A'; } return mainsentiment === sentiment.key; } }); groups.push({group: g, sent: sentiment.key }); }); return groups; } getStartDate(previousDate) { const date = new Date(previousDate); date.setDate(1); return date; } // Renders line chart (aggregation) renderChart () { this.chartRange1 = d3.min(this.data, (d: any) => new Date(d.publishedAt)); this.chartRange1 = this.getStartDate(this.chartRange1); this.chartRange2 = d3.max(this.data, (d: any) => new Date(d.publishedAt)); const sentGroupsOrdered = this.reorderGroups(); const chartColors = this.defineChartColors(); let firstItem = 0; while (!sentGroupsOrdered[firstItem] && firstItem < 5) {firstItem++; } const group1 = sentGroupsOrdered[firstItem]; this.sentimentLineChart .renderArea(true) .width(900) .height(this.chartHeight) .ordinalColors(chartColors) .useViewBoxResizing(true) .dimension(this.dimension) .x(d3.scaleTime().domain([this.chartRange1, this.chartRange2])) .y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)])) .yAxisLabel('Comments') .interpolate('monotone') .legend(dc.legend().x(850).y(0).itemHeight(9).gap(5)) .brushOn(false) .group(group1.group, group1.sent) .valueAccessor(function (d) { return d.value; }) .xAxis().ticks(7); let maxSent = 0; if (sentGroupsOrdered.length > 1) { sentGroupsOrdered.forEach((group) => { if (group.group === group1.group || maxSent === 4) { return; } // stacks the groups this.sentimentLineChart .stack(group.group, group.sent, function (d) { return d.value; }); maxSent++; }); } // When filter is applied before refreshing the chart if (this.appliedFilter) { this.sentimentLineChart.x(d3.scaleTime().domain([this.getStartDate(this.chartRangeFilter1), this.chartRangeFilter2])); } // Brush: get range and send it to the other charts on brush-filtering this.sentimentLineChart.on('filtered', (chart, filter) => { if (filter) { this.sentimentLineChart.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(filter[0], filter[1])])); } else { this.sentimentLineChart.y(d3.scaleLinear().domain([0, this.getMaxGroupValue(this.chartRange1, this.chartRange2)])); } this.chartService.setChartRange({range: filter, chart: chart}); }); // Adapt chart for smaller view (this.chartHeight < 300) ? this.sentimentLineChart.yAxis().ticks(2) : this.sentimentLineChart.yAxis().ticks(10); (this.chartHeight < 300) ? this.sentimentLineChart.xAxisLabel('') : this.sentimentLineChart.xAxisLabel('Date'); this.sentimentLineChart.xAxis().tickFormat(d3.timeFormat('%b %Y')); // month this.sentimentLineChart.render(); } // Adaptable color scale defineChartColors() { const colorArray = []; const sentGroupsOrdered = this.reorderGroups(); sentGroupsOrdered.forEach((g) => { if (g.sent === 'Pos') { colorArray.push('#4daf4a'); } else if (g.sent === 'Neu') { colorArray.push('#6
TimeSentimentComponent
identifier_name
text.rs
/// as "&quot;". /// pub(crate) fn escape(input: &str) -> String { let mut result = String::with_capacity(input.len()); for c in input.chars() { match c { XML_ESC_AMP_CHAR => result.push_str(&to_entity(XML_ESC_AMP_CHAR)), XML_ESC_APOS_CHAR => result.push_str(&to_entity(XML_ESC_APOS_CHAR)), XML_ESC_GT_CHAR => result.push_str(&to_entity(XML_ESC_GT_CHAR)), XML_ESC_LT_CHAR => result.push_str(&to_entity(XML_ESC_LT_CHAR)), XML_ESC_QUOT_CHAR => result.push_str(&to_entity(XML_ESC_QUOT_CHAR)), o => result.push(o), } } result } pub(crate) fn to_entity(c: char) -> String { format!( "{}{}{}", XML_NUMBERED_ENTITYREF_START, c as u16, XML_ENTITYREF_END ) } #[allow(dead_code)] pub(crate) fn to_entity_hex(c: char) -> String { format!( "{}{:X}{}", XML_HEX_NUMBERED_ENTITYREF_START, c as u16, XML_ENTITYREF_END ) } fn char_from_entity(entity: &str) -> String { assert!(entity.starts_with("&#")); assert!(entity.ends_with(';')); let code_point = if &entity[2..3] == "x" { let code_point = &entity[3..entity.len() - 1]; u32::from_str_radix(code_point, 16).unwrap() } else { let code_point = &entity[2..entity.len() - 1]; u32::from_str_radix(code_point, 10).unwrap() }; let character = char::try_from(code_point).unwrap(); character.to_string() } /// /// From [XML 1.0 §2.2](https://www.w3.org/TR/REC-xml/#charsets) /// /// Definition: A parsed entity contains **text**, a sequence of characters, which may represent /// markup or character data. Definition: A **character** is an atomic unit of text as specified by /// ISO/IEC 10646:2000. Legal characters are tab, carriage return, line feed, and the legal /// characters of Unicode and ISO/IEC 10646. The versions of these standards cited in A.1 Normative /// References were current at the time this document was prepared. New characters may be added to /// these standards by amendments or new editions. Consequently, XML processors must accept any /// character in the range specified for `Char`. /// /// ```ebnf /// Char ::= #x9 | #xA | #xD | [#x20-#xD7FF] | [#xE000-#xFFFD] | [#x10000-#x10FFFF] /// /* any Unicode character, excluding the surrogate blocks, FFFE, and FFFF. */ /// ``` /// /// Document authors are encouraged to avoid "compatibility characters", as defined in section 2.3 /// of Unicode. The characters defined in the following ranges are also discouraged. They are either /// control characters or permanently undefined Unicode characters: /// /// ```text /// [#x7F-#x84], [#x86-#x9F], [#xFDD0-#xFDEF], /// [#x1FFFE-#x1FFFF], [#x2FFFE-#x2FFFF], [#x3FFFE-#x3FFFF], /// [#x4FFFE-#x4FFFF], [#x5FFFE-#x5FFFF], [#x6FFFE-#x6FFFF], /// [#x7FFFE-#x7FFFF], [#x8FFFE-#x8FFFF], [#x9FFFE-#x9FFFF], /// [#xAFFFE-#xAFFFF], [#xBFFFE-#xBFFFF], [#xCFFFE-#xCFFFF], /// [#xDFFFE-#xDFFFF], [#xEFFFE-#xEFFFF], [#xFFFFE-#xFFFFF], /// [#x10FFFE-#x10FFFF]. /// ``` /// #[allow(dead_code)] pub(crate) fn is_xml_10_char(c: char) -> bool { c == '\u{0009}' || c == '\u{000A}' || c == '\u{000D}' || (c >= '\u{0020}' && c <= '\u{D7FF}') || (c >= '\u{E000}' && c <= '\u{FFFD}') || (c >= '\u{10000}' && c <= '\u{10FFF}') } #[allow(dead_code)] pub(crate) fn is_xml_10_restricted_char(c: char) -> bool { c == XML_ESC_AMP_CHAR || c == XML_ESC_APOS_CHAR || c == XML_ESC_GT_CHAR || c == XML_ESC_LT_CHAR || c == XML_ESC_QUOT_CHAR } /// /// From [XML 11 §2.2](https://www.w3.org/TR/xml11/#charsets) /// /// ```ebnf /// Char ::= [#x1-#xD7FF] | [#xE000-#xFFFD] | [#x10000-#x10FFFF] /// /* any Unicode character, excluding the surrogate blocks, FFFE, and FFFF. */ /// ``` /// #[allow(dead_code)] pub(crate) fn is_xml_11_char(c: char) -> bool { // // below ranges are always valid for XML 1.1 documents // from https://en.wikipedia.org/wiki/XML#Valid_characters // (c >= '\u{0001}' && c <= '\u{D7FF}') || (c >= '\u{E000}' && c <= '\u{FFFD}') || (c >= '\u{10000}' && c <= '\u{10FFF}') } /// /// From [XML 11 §2.2](https://www.w3.org/TR/xml11/#charsets) /// /// ```ebnf /// RestrictedChar ::= #x1-#x8] | [#xB-#xC] | [#xE-#x1F] | [#x7F-#x84] | [#x86-#x9F] /// ``` /// #[allow(dead_code)] pub(crate) fn is_xml_11_restricted_char(c: char) -> bool { // // below ranges are always valid for XML 1.1 documents // from https://en.wikipedia.org/wiki/XML#Valid_characters // (c >= '\u{01}' && c <= '\u{08}') || (c >= '\u{0B}' && c <= '\u{0C}') || (c >= '\u{0E}' && c <= '\u{1F}') || (c >= '\u{7F}' && c <= '\u{84}') || (c >= '\u{86}' && c <= '\u{9F}') } /// /// S (white space) consists of one or more space (#x20) characters, carriage returns, line feeds, /// or tabs. /// /// ```ebnf /// S ::= (#x20 | #x9 | #xD | #xA)+ /// ``` /// /// The presence of #xD in the above production is maintained purely for backward compatibility /// with the First Edition. As explained in 2.11 End-of-Line Handling, all #xD characters literally /// present in an XML document are either removed or replaced by #xA characters before any other /// processing is done. The only way to get a #xD character to match this production is to use a /// character reference in an entity value literal. /// #[allow(dead_code)] pub(crate) fn is_xml_space(c: char) -> bool { c == '\u{09}' || c == '\u{0A}' || c == '\u{0D}' || c == '\u{20}' } /// /// ```ebnf /// NameStartChar ::= ":" | [A-Z] | "_" | [a-z] | [#xC0-#xD6] | [#xD8-#xF6] | [#xF8-#x2FF] | /// [#x370-#x37D] | [#x37F-#x1FFF] | [#x200C-#x200D] | [#x2070-#x218F] | /// [#x2C00-#x2FEF] | [#x3001-#xD7FF] | [#xF900-#xFDCF] | [#xFDF0-#xFFFD] | /// [#x10000-#xEFFFF] /// ``` /// #[allow(dead_code)] pub(crate) fn is_xml_name_start_char(c: char) -> bool { c == ':' || (c >= 'A' && c <= 'Z')
|| c == '_' || (c >= 'a' && c <= 'z') || (c >= '\u{C0}' && c <= '\u{D6}')
random_line_split
text.rs
else if let Some(a_match) = capture.name("char") { let replacement = char_from_entity(a_match.as_str()); (a_match.start(), a_match.end(), replacement) } else if let Some(a_match) = capture.name("char_hex") { let replacement = char_from_entity(a_match.as_str()); (a_match.start(), a_match.end(), replacement) } else if let Some(a_match) = capture.name("ws") { (a_match.start(), a_match.end(), "\u{20}".to_string()) } else { panic!("unexpected result"); }; step_2.push_str(&step_1[last_end..start]); step_2.push_str(&replacement); last_end = end; } if last_end < value.len() { step_2.push_str(&step_1[last_end..]); } step_2 }; if is_cdata { step_3 } else { step_3.trim_matches(' ').to_string() } } /// /// From XML 1.1 §2.11 [End-of-Line Handling](https://www.w3.org/TR/xml11/#sec-line-ends): /// /// XML parsed entities are often stored in computer files which, for editing convenience, are /// organized into lines. These lines are typically separated by some combination of the characters /// CARRIAGE RETURN `(#xD`) and LINE FEED (`#xA`). /// /// To simplify the tasks of applications, the XML processor must behave as if it normalized all line /// breaks in external parsed entities (including the document entity) on input, before parsing, by /// translating all of the following to a single `#xA` character: /// /// * the two-character sequence `#xD` `#xA` /// * the two-character sequence `#xD` `#x85` /// * the single character `#x85` /// * the single character `#x2028` /// * any `#xD` character that is not immediately followed by `#xA` or `#x85`. /// /// The characters `#x85` and `#x2028` cannot be reliably recognized and translated until an entity's /// encoding declaration (if present) has been read. Therefore, it is a fatal error to use them /// within the XML declaration or text declaration. /// pub(crate) fn normalize_end_of_lines(value: &str) -> String { if value.is_empty() { value.to_string() } else { let line_ends = regex::Regex::new(r"\u{0D}[\u{0A}\u{85}]?|\u{85}|\u{2028}").unwrap(); line_ends.replace_all(value, "\u{0A}").to_string() } } /// /// Escape character data according to XML 1.1 /// [§2.4 Character Data and Markup](https://www.w3.org/TR/xml11/#dt-chardata). This is the /// do-everything version, not attempting to separate the rules defined below by node type. /// /// # Specification /// /// Text consists of intermingled character data and markup. [Definition: **Markup** takes the form /// of start-tags, end-tags, empty-element tags, entity references, character references, comments, /// CDATA section delimiters, document type declarations, processing instructions, XML declarations, /// text declarations, and any white space that is at the top level of the document entity (that is, /// outside the document element and not inside any other markup).] /// /// [Definition: All text that is not markup constitutes the **character data** of the document]. /// /// The ampersand character (&) and the left angle bracket (<) must not appear in their literal /// form, except when used as markup delimiters, or within a comment, a processing instruction, or /// a CDATA section. If they are needed elsewhere, they must be escaped using either numeric /// character references or the strings "&amp;" and "&lt;" respectively. The right angle bracket /// (>) may be represented using the string "&gt;", and must, for compatibility, be escaped using /// either "&gt;" or a character reference when it appears in the string "]]>" in content, when that /// string is not marking the end of a CDATA section. /// /// In the content of elements, character data is any string of characters which does not contain /// the start-delimiter of any markup or the CDATA-section-close delimiter, "]]>". In a CDATA /// section, character data is any string of characters not including the CDATA-section-close /// delimiter. /// /// To allow attribute values to contain both single and double quotes, the apostrophe or /// single-quote character (') may be represented as "&apos;", and the double-quote character (") /// as "&quot;". /// pub(crate) fn escape(input: &str) -> String { let mut result = String::with_capacity(input.len()); for c in input.chars() { match c { XML_ESC_AMP_CHAR => result.push_str(&to_entity(XML_ESC_AMP_CHAR)), XML_ESC_APOS_CHAR => result.push_str(&to_entity(XML_ESC_APOS_CHAR)), XML_ESC_GT_CHAR => result.push_str(&to_entity(XML_ESC_GT_CHAR)), XML_ESC_LT_CHAR => result.push_str(&to_entity(XML_ESC_LT_CHAR)), XML_ESC_QUOT_CHAR => result.push_str(&to_entity(XML_ESC_QUOT_CHAR)), o => result.push(o), } } result } pub(crate) fn to_entity(c: char) -> String { format!( "{}{}{}", XML_NUMBERED_ENTITYREF_START, c as u16, XML_ENTITYREF_END ) } #[allow(dead_code)] pub(crate) fn to_entity_hex(c: char) -> String { format!( "{}{:X}{}", XML_HEX_NUMBERED_ENTITYREF_START, c as u16, XML_ENTITYREF_END ) } fn char_from_entity(entity: &str) -> String { assert!(entity.starts_with("&#")); assert!(entity.ends_with(';')); let code_point = if &entity[2..3] == "x" { let code_point = &entity[3..entity.len() - 1]; u32::from_str_radix(code_point, 16).unwrap() } else { let code_point = &entity[2..entity.len() - 1]; u32::from_str_radix(code_point, 10).unwrap() }; let character = char::try_from(code_point).unwrap(); character.to_string() } /// /// From [XML 1.0 §2.2](https://www.w3.org/TR/REC-xml/#charsets) /// /// Definition: A parsed entity contains **text**, a sequence of characters, which may represent /// markup or character data. Definition: A **character** is an atomic unit of text as specified by /// ISO/IEC 10646:2000. Legal characters are tab, carriage return, line feed, and the legal /// characters of Unicode and ISO/IEC 10646. The versions of these standards cited in A.1 Normative /// References were current at the time this document was prepared. New characters may be added to /// these standards by amendments or new editions. Consequently, XML processors must accept any /// character in the range specified for `Char`. /// /// ```ebnf /// Char ::= #x9 | #xA | #xD | [#x20-#xD7FF] | [#xE000-#xFFFD] | [#x10000-#x10FFFF] /// /* any Unicode character, excluding the surrogate blocks, FFFE, and FFFF. */ /// ``` /// /// Document authors are encouraged to avoid "compatibility characters", as defined in section 2.3 /// of Unicode. The characters defined in the following ranges are also discouraged. They are either /// control characters or permanently undefined Unicode characters: /// /// ```text /// [#x7F-#x84], [#x86-#x9F], [#xFDD0-#xFDEF], /// [#x1FFFE-#x1FFFF], [#x2FFFE-#x2FFFF], [#x3FFFE-#x3FFFF], /// [#x4FFFE-#x4FFFF], [#x5FFFE-#x5FFFF], [#x6FFFE-#x6FFFF], /// [#x7FFFE-#x7FFFF], [#x8FFFE-#x8FFFF], [#x9FFFE-#x9FFFF], /// [#xAFFFE-#xAFFFF], [#xBFFFE-#xBFFFF], [#xCFFFE-#xCFFFF], /// [#xDFFFE-#xDFFFF], [#xEFFFE-#xEFFFF], [#xFFFFE-#xFFFFF], /// [#x10FFFE-#x10FFFF]. /// ``` /// #[allow(dead_code)] pub(crate) fn is_xml_10
// // TODO: this does not yet deal with entity references. // let replacement = match resolver.resolve(a_match.as_str()) { None => panic!("unknown entity reference {}", a_match.as_str()), Some(replacement) => { normalize_attribute_value(&replacement, resolver, is_cdata) } }; (a_match.start(), a_match.end(), replacement) }
conditional_block
text.rs
xFFFFE-#xFFFFF], /// [#x10FFFE-#x10FFFF]. /// ``` /// #[allow(dead_code)] pub(crate) fn is_xml_10_char(c: char) -> bool { c == '\u{0009}' || c == '\u{000A}' || c == '\u{000D}' || (c >= '\u{0020}' && c <= '\u{D7FF}') || (c >= '\u{E000}' && c <= '\u{FFFD}') || (c >= '\u{10000}' && c <= '\u{10FFF}') } #[allow(dead_code)] pub(crate) fn is_xml_10_restricted_char(c: char) -> bool { c == XML_ESC_AMP_CHAR || c == XML_ESC_APOS_CHAR || c == XML_ESC_GT_CHAR || c == XML_ESC_LT_CHAR || c == XML_ESC_QUOT_CHAR } /// /// From [XML 11 §2.2](https://www.w3.org/TR/xml11/#charsets) /// /// ```ebnf /// Char ::= [#x1-#xD7FF] | [#xE000-#xFFFD] | [#x10000-#x10FFFF] /// /* any Unicode character, excluding the surrogate blocks, FFFE, and FFFF. */ /// ``` /// #[allow(dead_code)] pub(crate) fn is_xml_11_char(c: char) -> bool { // // below ranges are always valid for XML 1.1 documents // from https://en.wikipedia.org/wiki/XML#Valid_characters // (c >= '\u{0001}' && c <= '\u{D7FF}') || (c >= '\u{E000}' && c <= '\u{FFFD}') || (c >= '\u{10000}' && c <= '\u{10FFF}') } /// /// From [XML 11 §2.2](https://www.w3.org/TR/xml11/#charsets) /// /// ```ebnf /// RestrictedChar ::= #x1-#x8] | [#xB-#xC] | [#xE-#x1F] | [#x7F-#x84] | [#x86-#x9F] /// ``` /// #[allow(dead_code)] pub(crate) fn is_xml_11_restricted_char(c: char) -> bool { // // below ranges are always valid for XML 1.1 documents // from https://en.wikipedia.org/wiki/XML#Valid_characters // (c >= '\u{01}' && c <= '\u{08}') || (c >= '\u{0B}' && c <= '\u{0C}') || (c >= '\u{0E}' && c <= '\u{1F}') || (c >= '\u{7F}' && c <= '\u{84}') || (c >= '\u{86}' && c <= '\u{9F}') } /// /// S (white space) consists of one or more space (#x20) characters, carriage returns, line feeds, /// or tabs. /// /// ```ebnf /// S ::= (#x20 | #x9 | #xD | #xA)+ /// ``` /// /// The presence of #xD in the above production is maintained purely for backward compatibility /// with the First Edition. As explained in 2.11 End-of-Line Handling, all #xD characters literally /// present in an XML document are either removed or replaced by #xA characters before any other /// processing is done. The only way to get a #xD character to match this production is to use a /// character reference in an entity value literal. /// #[allow(dead_code)] pub(crate) fn is_xml_space(c: char) -> bool { c == '\u{09}' || c == '\u{0A}' || c == '\u{0D}' || c == '\u{20}' } /// /// ```ebnf /// NameStartChar ::= ":" | [A-Z] | "_" | [a-z] | [#xC0-#xD6] | [#xD8-#xF6] | [#xF8-#x2FF] | /// [#x370-#x37D] | [#x37F-#x1FFF] | [#x200C-#x200D] | [#x2070-#x218F] | /// [#x2C00-#x2FEF] | [#x3001-#xD7FF] | [#xF900-#xFDCF] | [#xFDF0-#xFFFD] | /// [#x10000-#xEFFFF] /// ``` /// #[allow(dead_code)] pub(crate) fn is_xml_name_start_char(c: char) -> bool { c == ':' || (c >= 'A' && c <= 'Z') || c == '_' || (c >= 'a' && c <= 'z') || (c >= '\u{C0}' && c <= '\u{D6}') || (c >= '\u{D8}' && c <= '\u{F6}') || (c >= '\u{0F8}' && c <= '\u{2FF}') || (c >= '\u{370}' && c <= '\u{37D}') || (c >= '\u{037F}' && c <= '\u{1FFF}') || (c >= '\u{200C}' && c <= '\u{200D}') || (c >= '\u{2070}' && c <= '\u{218F}') || (c >= '\u{2C00}' && c <= '\u{2FEF}') || (c >= '\u{3001}' && c <= '\u{D7FF}') || (c >= '\u{F900}' && c <= '\u{FDCF}') || (c >= '\u{FDF0}' && c <= '\u{FFFD}') || (c >= '\u{10000}' && c <= '\u{EFFFF}') } /// /// ```ebnf /// NameChar ::= NameStartChar | "-" | "." | [0-9] | #xB7 | /// [#x0300-#x036F] | [#x203F-#x2040] /// ``` /// pub(crate) fn is_xml_name_char(c: char) -> bool { is_xml_name_start_char(c) || c == '-' || c == '.' || (c >= '0' && c <= '9') || c == '\u{B7}' || (c >= '\u{0300}' && c <= '\u{036F}') || (c >= '\u{203F}' && c <= '\u{2040}') } /// /// ```ebnf /// Name ::= NameStartChar (NameChar)* /// ``` /// pub(crate) fn is_xml_name(s: &str) -> bool { !s.is_empty() && s.starts_with(is_xml_name_start_char) && s[1..].chars().all(is_xml_name_char) } /// /// ```ebnf /// Names ::= Name (#x20 Name)* /// ``` /// #[allow(dead_code)] pub(crate) fn is_xml_names(s: &str) -> bool { !s.is_empty() && s.split(' ').all(is_xml_name) } /// /// ```ebnf /// Nmtoken ::= (NameChar)+ /// ``` /// #[allow(dead_code)] pub(crate) fn is_xml_nmtoken(s: &str) -> bool { !s.is_empty() && s.chars().all(is_xml_name_char) } /// /// ```ebnf /// Nmtokens ::= Nmtoken (#x20 Nmtoken)* /// ``` /// #[allow(dead_code)] pub(crate) fn is_xml_nmtokens(s: &str) -> bool { !s.is_empty() && s.split(' ').all(is_xml_nmtoken) } // ------------------------------------------------------------------------------------------------ // Implementations // ------------------------------------------------------------------------------------------------ impl Default for SpaceHandling { fn default() -> Self { SpaceHandling::Default } } // ------------------------------------------------------------------------------------------------ impl Display for SpaceHandling { fn fmt(&self, f: &mut Formatter<'_>) -> FmtResult { write!( f, "{}{}{}=\"{}\"", XML_NS_ATTRIBUTE, XML_NS_SEPARATOR, XML_NS_ATTR_SPACE, match self { SpaceHandling::Default => XML_NS_ATTR_SPACE_DEFAULT, SpaceHandling::Preserve => XML_NS_ATTR_SPACE_PRESERVE, } ) } } // ------------------------------------------------------------------------------------------------ impl FromStr for SpaceHandling { type Err = (); fn from_str(s: &str) -> Result<Self, Self::Err> {
if s == XML_NS_ATTR_SPACE_DEFAULT { Ok(SpaceHandling::Default) } else if s == XML_NS_ATTR_SPACE_PRESERVE { Ok(SpaceHandling::Preserve) } else { Err(()) } } } //
identifier_body
text.rs
{ c == '\u{09}' || c == '\u{0A}' || c == '\u{0D}' || c == '\u{20}' } /// /// ```ebnf /// NameStartChar ::= ":" | [A-Z] | "_" | [a-z] | [#xC0-#xD6] | [#xD8-#xF6] | [#xF8-#x2FF] | /// [#x370-#x37D] | [#x37F-#x1FFF] | [#x200C-#x200D] | [#x2070-#x218F] | /// [#x2C00-#x2FEF] | [#x3001-#xD7FF] | [#xF900-#xFDCF] | [#xFDF0-#xFFFD] | /// [#x10000-#xEFFFF] /// ``` /// #[allow(dead_code)] pub(crate) fn is_xml_name_start_char(c: char) -> bool { c == ':' || (c >= 'A' && c <= 'Z') || c == '_' || (c >= 'a' && c <= 'z') || (c >= '\u{C0}' && c <= '\u{D6}') || (c >= '\u{D8}' && c <= '\u{F6}') || (c >= '\u{0F8}' && c <= '\u{2FF}') || (c >= '\u{370}' && c <= '\u{37D}') || (c >= '\u{037F}' && c <= '\u{1FFF}') || (c >= '\u{200C}' && c <= '\u{200D}') || (c >= '\u{2070}' && c <= '\u{218F}') || (c >= '\u{2C00}' && c <= '\u{2FEF}') || (c >= '\u{3001}' && c <= '\u{D7FF}') || (c >= '\u{F900}' && c <= '\u{FDCF}') || (c >= '\u{FDF0}' && c <= '\u{FFFD}') || (c >= '\u{10000}' && c <= '\u{EFFFF}') } /// /// ```ebnf /// NameChar ::= NameStartChar | "-" | "." | [0-9] | #xB7 | /// [#x0300-#x036F] | [#x203F-#x2040] /// ``` /// pub(crate) fn is_xml_name_char(c: char) -> bool { is_xml_name_start_char(c) || c == '-' || c == '.' || (c >= '0' && c <= '9') || c == '\u{B7}' || (c >= '\u{0300}' && c <= '\u{036F}') || (c >= '\u{203F}' && c <= '\u{2040}') } /// /// ```ebnf /// Name ::= NameStartChar (NameChar)* /// ``` /// pub(crate) fn is_xml_name(s: &str) -> bool { !s.is_empty() && s.starts_with(is_xml_name_start_char) && s[1..].chars().all(is_xml_name_char) } /// /// ```ebnf /// Names ::= Name (#x20 Name)* /// ``` /// #[allow(dead_code)] pub(crate) fn is_xml_names(s: &str) -> bool { !s.is_empty() && s.split(' ').all(is_xml_name) } /// /// ```ebnf /// Nmtoken ::= (NameChar)+ /// ``` /// #[allow(dead_code)] pub(crate) fn is_xml_nmtoken(s: &str) -> bool { !s.is_empty() && s.chars().all(is_xml_name_char) } /// /// ```ebnf /// Nmtokens ::= Nmtoken (#x20 Nmtoken)* /// ``` /// #[allow(dead_code)] pub(crate) fn is_xml_nmtokens(s: &str) -> bool { !s.is_empty() && s.split(' ').all(is_xml_nmtoken) } // ------------------------------------------------------------------------------------------------ // Implementations // ------------------------------------------------------------------------------------------------ impl Default for SpaceHandling { fn default() -> Self { SpaceHandling::Default } } // ------------------------------------------------------------------------------------------------ impl Display for SpaceHandling { fn fmt(&self, f: &mut Formatter<'_>) -> FmtResult { write!( f, "{}{}{}=\"{}\"", XML_NS_ATTRIBUTE, XML_NS_SEPARATOR, XML_NS_ATTR_SPACE, match self { SpaceHandling::Default => XML_NS_ATTR_SPACE_DEFAULT, SpaceHandling::Preserve => XML_NS_ATTR_SPACE_PRESERVE, } ) } } // ------------------------------------------------------------------------------------------------ impl FromStr for SpaceHandling { type Err = (); fn from_str(s: &str) -> Result<Self, Self::Err> { if s == XML_NS_ATTR_SPACE_DEFAULT { Ok(SpaceHandling::Default) } else if s == XML_NS_ATTR_SPACE_PRESERVE { Ok(SpaceHandling::Preserve) } else { Err(()) } } } // ------------------------------------------------------------------------------------------------ // Unit Tests // ------------------------------------------------------------------------------------------------ #[cfg(test)] mod tests { use super::*; use std::borrow::Borrow; use std::collections::HashMap; #[test] fn test_space_handling_default() { let sh = SpaceHandling::default(); assert_eq!(sh, SpaceHandling::Default); } #[test] fn test_space_handling_display() { assert_eq!( format!("{}", SpaceHandling::Default), format!( "{}{}{}=\"{}\"", XML_NS_ATTRIBUTE, XML_NS_SEPARATOR, XML_NS_ATTR_SPACE, XML_NS_ATTR_SPACE_DEFAULT ) ); assert_eq!( format!("{}", SpaceHandling::Preserve), format!( "{}{}{}=\"{}\"", XML_NS_ATTRIBUTE, XML_NS_SEPARATOR, XML_NS_ATTR_SPACE, XML_NS_ATTR_SPACE_PRESERVE ) ); } #[test] fn test_space_handling_from_str() { assert_eq!( SpaceHandling::from_str(XML_NS_ATTR_SPACE_DEFAULT).unwrap(), SpaceHandling::Default ); assert_eq!( SpaceHandling::from_str(XML_NS_ATTR_SPACE_PRESERVE).unwrap(), SpaceHandling::Preserve ); assert!(SpaceHandling::from_str("").is_err()); assert!(SpaceHandling::from_str("other").is_err()); } #[test] fn test_end_of_line_handling() { let input = "one\u{0D}two\u{0D}\u{0A}\u{0A}three\u{0A}\u{0D}\u{85}four\u{85}five\u{2028}"; let output = normalize_end_of_lines(&input.to_string()); assert_eq!( output, "one\u{0A}two\u{0A}\u{0A}three\u{0A}\u{0A}four\u{0A}five\u{0A}".to_string() ) } struct NoneEntityResolver {} impl EntityResolver for NoneEntityResolver { fn resolve(&self, name: &str) -> Option<String> { let result: Option<String> = None; println!("EntityResolver::resolve({:?}) -> {:?}", name, result); result } } pub(crate) fn none_entity_resolver() -> Box<dyn EntityResolver> { let resolver = NoneEntityResolver {}; Box::new(resolver) } #[test] fn test_normalize_avalue_trim() { let resolver = none_entity_resolver(); let resolver = resolver.borrow(); assert_eq!( normalize_attribute_value(" abc ", resolver, true), " abc " ); assert_eq!(normalize_attribute_value(" abc ", resolver, false), "abc"); } struct TestResolver { entity_map: HashMap<String, String>, } impl EntityResolver for TestResolver { fn resolve(&self, entity: &str) -> Option<String> { self.entity_map.get(entity).cloned() } } impl TestResolver { pub(crate) fn new() -> Self { let mut new_self = Self { entity_map: Default::default(), }; let _safe_to_ignore = new_self .entity_map .insert("&pound;".to_string(), "£".to_string()); let _safe_to_ignore = new_self .entity_map .insert("&yen;".to_string(), "¥".to_string()); let _safe_to_ignore = new_self .entity_map .insert("&euro;".to_string(), "€".to_string()); let _safe_to_ignore = new_self.entity_map.insert( "&currency;".to_string(), "$, &pound;, &euro;, and &yen;".to_string(), ); new_self } } fn test_resolver() -> Box<dyn EntityResolver> { let resolver = TestResolver::new(); Box::new(resolver) } #[test] fn test_norma
lize_avalue_entity_resolver() {
identifier_name
ionic-native-map.ts
"); let userID = localStorage.getItem("userID"); let URL = globalVars.UserAddress(userID); // console.log(locationExists); let data = { alias: this.locationAlias, address: this.address, lat: this.lat, long: this.lng } if(this.validate()){ // let locationExists: boolean = false; // this.addressResponse.forEach(address => { // locationExists = locationExists || (address.alias == this.locationAlias); // console.log(address.alias, this.locationAlias); // console.log(address.alias == this.locationAlias); // }); // console.log('location Exists: ', locationExists); // if(!locationExists){ this.authService.patchCall(URL, data) .subscribe(res => { if (res.status == 200) { console.log(res['_body']); } }); }else{ this.map.setClickable(false); let result = this.alertCntrl.openAlertDialog('Location exits', 'Please enter a location.'); // result.then(value => { // if(value){ // this.map.setClickable(true); // } // }) } // }else{ // // this.alertCntrl.openAlertDialog('Error', 'Location already Exists.') // } } openAdditionalNoteDialog(myEvent) { this.map.setClickable(false); this.isModalVisible = this.isModalVisible ? false : true; setTimeout(() => { this.isModalVisible = this.isModalVisible ? false : true; }, 200); let popover = this.popoverCtrl.create(AdditionalNote, {}, { showBackdrop: true }); popover.present({ ev: myEvent }); popover.onDidDismiss(data => { if(data){ this.map.setClickable(true); console.log(data); this.additionalInfoText = data + "\n"; localStorage.setItem("additionalInfoText", this.additionalInfoText); } }) } additionButtonClicked(myEvent) { this.addition = this.addition ? false : true; console.log("additionButtonClicked"); this.openAdditionalNoteDialog(myEvent); } locationClickedBool; locationClicked(location) { console.log("You have clicked on: ", location); this.locationClickedBool = false; this.hide = true; if(!!location){ this.inputFieldValue = ''; if(!!location.name){ this.locationClickedBool = true; console.log(location); this.inputFieldValue = location.name || ''; localStorage.setItem("Location", JSON.stringify(location)); this.lat = location.geometry.location.lat; this.lng = location.geometry.location.lng; this.address = location.formatted_address; this.locationAlias = location.name; }else{ console.log('Here'); this.locationClickedBool = true; this.inputFieldValue = location.alias || ''; localStorage.setItem("Location", JSON.stringify(location)); this.lat = location.lat; this.lng = location.long; this.address = location.address; this.locationAlias = location.alias; }; setTimeout(() => { this.available_locations = []}, 200); }else{ console.log('Here'); this.locationClickedBool = true; this.inputFieldValue = location.alias || ''; localStorage.setItem("Location", JSON.stringify(location)); this.lat = location.lat; this.lng = location.long; this.address = location.address; this.locationAlias = location.alias; }; setTimeout(() => { this.available_locations = []}, 200); //gMap = new google.maps.Map(document.getElementById('map')); // this.postion = new google.maps.LatLng(this.lat, this.lng); // this.map.setCenter(this.postion); // this.addMarkerMoveCamera(this.map, new LatLng(this.lat, this.lng)); // this.map.center = new google.maps.LatLng(this.lat, this.lng); this.addMarker(this.map, new LatLng(this.lat, this.lng)); this.moveCamera(this.map, new LatLng(this.lat, this.lng)); } validate():boolean { return (this.lat != null && this.lng != null && this.address != null) ? true :false; } startNextScreen() { console.log("Next clicked!"); let valid:boolean = this.validate(); console.log(valid); if(valid === true && this.locationClickedBool == true) { console.log(this.preGenData); this.navCtrl.push(LaundryItems, { preGenData: this.preGenData, pickupDetails: { location: { lat: this.lat, lng: this.lng, address: this.address } }, }); } else{ this.map.setClickable(false); this.alertCntrl.openAlertDialog("What's missing?","No location selected."); } } loadMap(){ let element: HTMLElement = ViewChild('map'); let mapOptions = { "featureType": "all", "elementType": "geometry", styles: [ { elementType: 'geometry', stylers: [{ color: '#15151b' }] }, { elementType: 'labels.text.stroke', stylers: [{ color: '#242f3e' }] }, { elementType: 'labels.text.fill', stylers: [{ color: '#746855' }] }, { featureType: 'administrative', elementType: 'labels', stylers: [{ visibility: 'off' }] }, { featureType: 'poi', elementType: 'labels', stylers: [{ visibility: 'off' }] }, { featureType: 'poi.park', elementType: 'geometry', stylers: [{ visibility: 'off' }] }, { featureType: 'poi.park', elementType: 'labels.text.fill', stylers: [{ visibility: 'off' }] }, { featureType: 'road', elementType: 'geometry', stylers: [{ color: '#000000' }] } // #38414e , { featureType: 'road', elementType: 'geometry.stroke', stylers: [{ color: '#000000' }]//212a37 }, { featureType: 'road', elementType: 'labels.text.fill', stylers: [{ color: '#ffffff' }]//9ca5b3 }, { featureType: 'road.highway', elementType: 'geometry', stylers: [{ color: '#000000' }]//746855 }, { featureType: 'road.highway', elementType: 'geometry.stroke', stylers: [{ color: '#1f2835' }] }, { featureType: 'road.highway', elementType: 'labels.text.fill', stylers: [{ color: '#f3d19c' }] }, { featureType: 'transit', elementType: 'all', stylers: [{ visibility: 'off' }] }, { featureType: 'transit.station', elementType: 'labels.text.fill', stylers: [{ visibility: 'off' }] }, { featureType: 'water', elementType: 'geometry', stylers: [{ color: '#17263c' }] }, { featureType: 'water', elementType: 'labels.text.fill', stylers: [{ visibility: 'off' }] }, { featureType: 'water', elementType: 'labels.text.stroke', stylers: [{ visibility: 'off' }] } ] // mapTypeId: google.maps.MapTypeId.ROADMAP1 }; let map: GoogleMap = this.googleMaps.create(element); map = new GoogleMap('map'); this.map = map; // listen to MAP_READY event // You must wait for this event to fire before adding something to the map or modifying it in anyway map.one(GoogleMapsEvent.MAP_READY).then( () => { console.log('Map is ready!'); // Now you can add elements to the map like the marker map.setOptions(mapOptions); map.setMyLocationEnabled(true); //map.setBackgroundColor('black'); map.setPadding(0, 80, 150, 0); this.latLng = this.getLocation(map); map.setCompassEnabled(false); }); } getLocation(map: GoogleMap) { let latLng: string; map.getMyLocation().then( location => { latLng = location.latLng.lat + ',' + location.latLng.lng; console.log("165", JSON.stringify(location.latLng)); console.log(485, ":", latLng); this.newLocation = new LatLng(location.latLng.lat, location.latLng.lng); // this.addMarker(map, location.latLng); this.moveCamera(map, location.latLng); let markerOptions: MarkerOptions = { position: this.newLocation
}; this.addMarker(map, this.newLocation);
random_line_split
ionic-native-map.ts
available_locations: Array<Object> = []; newLocation; marker; constructor(public navCtrl: NavController, public navParams: NavParams, private googleMaps: GoogleMaps, private platform: Platform, private geolocation: Geolocation, private androidPermissions: AndroidPermissions, private alertCtrl: AlertController, private popoverCtrl: PopoverController, private mapService: MapService, private authService: AuthService, private alertCntrl: AlertDialogFactory, private savedLocationService: SavedLocationService) { this.token = localStorage.getItem('x-access-token'); this.userID = localStorage.getItem('userID'); this.preGenData = navParams.get('preGenData'); localStorage.setItem("additionalInfoText", ""); // setTimeout(() => { // this.inputFieldValue = 'New Value'; // }, 3000) } ionViewDidLoad() { console.log('ionViewDidLoad IonicNativeMapPage'); // this.loadMap(); setTimeout(() => { this.loadMap(); }, 500); } ngAfterViewInit()
this.listenToSearchInput(); this.getMapLocation(location, this.latLng); } listenToSearchInput() { this.hide = false; let location: string; console.log('location1:', location) // let searchInput$ = Observable.fromEvent(this.button.nativeElement, 'keyup') // .map(e => location = e['srcElement'].value.trim()) // .distinctUntilChanged() // .switchMap(() => this.mapService.getJSON(location, this.latLng)) // searchInput$.subscribe(location => { // this.available_locations = location; // console.log(this.available_locations); // }) } getMapLocation(location, latLng) { if (location) { // let location$ = this.mapService.getJSON(location, this.latLng); // location$.subscribe(res => console.log) } } savedButtonClicked(myEvent) { this.saved = this.saved ? false : true; setTimeout(()=>{ this.saved = this.saved ? false : true; }, 200); let inputs; this.addressResponse = inputs; let URL = globalVars.getUsersAddress(this.userID); this.authService.getCall(URL). subscribe(res => { console.log(JSON.parse(res["_body"])); inputs = JSON.parse(res["_body"])["data"]["contact"]["address"]; console.log(inputs); this.addressResponse = inputs; // let result = this.alertCntrl.checkBoxAlertDialog("Saved Locations", inputs) // console.log(result); this.radioAlertDialog("Saved Locations", inputs) }) } radioAlertDialog(title: string, inputs){ this.map.setClickable(false); let alert = this.alertCtrl.create({ title: title, cssClass: 'alertTop' }); inputs.forEach(input => { alert.addInput({ type: 'radio', label: input.alias, value: input, checked: false }); }); alert.addButton({ text: 'Cancel', handler: () => { console.log('Cancel clicked.'); } }); alert.addButton({ text: 'Okay', handler: data => { console.log('Radio data:', data); // this.testCheckboxOpen = false; // this.testCheckboxResult = data; this.locationClicked(data); } }); alert.present(); alert.onDidDismiss((data) => { console.log('OnDidDismiss', data); // dataReturned = data; this.map.setClickable(true); return data || 'null'; }); } saveButtonClicked() { this.save = this.save ? false : true; setTimeout(()=>{ this.save = this.save ? false : true; }, 200); console.log("saveButtonClicked"); let userID = localStorage.getItem("userID"); let URL = globalVars.UserAddress(userID); // console.log(locationExists); let data = { alias: this.locationAlias, address: this.address, lat: this.lat, long: this.lng } if(this.validate()){ // let locationExists: boolean = false; // this.addressResponse.forEach(address => { // locationExists = locationExists || (address.alias == this.locationAlias); // console.log(address.alias, this.locationAlias); // console.log(address.alias == this.locationAlias); // }); // console.log('location Exists: ', locationExists); // if(!locationExists){ this.authService.patchCall(URL, data) .subscribe(res => { if (res.status == 200) { console.log(res['_body']); } }); }else{ this.map.setClickable(false); let result = this.alertCntrl.openAlertDialog('Location exits', 'Please enter a location.'); // result.then(value => { // if(value){ // this.map.setClickable(true); // } // }) } // }else{ // // this.alertCntrl.openAlertDialog('Error', 'Location already Exists.') // } } openAdditionalNoteDialog(myEvent) { this.map.setClickable(false); this.isModalVisible = this.isModalVisible ? false : true; setTimeout(() => { this.isModalVisible = this.isModalVisible ? false : true; }, 200); let popover = this.popoverCtrl.create(AdditionalNote, {}, { showBackdrop: true }); popover.present({ ev: myEvent }); popover.onDidDismiss(data => { if(data){ this.map.setClickable(true); console.log(data); this.additionalInfoText = data + "\n"; localStorage.setItem("additionalInfoText", this.additionalInfoText); } }) } additionButtonClicked(myEvent) { this.addition = this.addition ? false : true; console.log("additionButtonClicked"); this.openAdditionalNoteDialog(myEvent); } locationClickedBool; locationClicked(location) { console.log("You have clicked on: ", location); this.locationClickedBool = false; this.hide = true; if(!!location){ this.inputFieldValue = ''; if(!!location.name){ this.locationClickedBool = true; console.log(location); this.inputFieldValue = location.name || ''; localStorage.setItem("Location", JSON.stringify(location)); this.lat = location.geometry.location.lat; this.lng = location.geometry.location.lng; this.address = location.formatted_address; this.locationAlias = location.name; }else{ console.log('Here'); this.locationClickedBool = true; this.inputFieldValue = location.alias || ''; localStorage.setItem("Location", JSON.stringify(location)); this.lat = location.lat; this.lng = location.long; this.address = location.address; this.locationAlias = location.alias; }; setTimeout(() => { this.available_locations = []}, 200); }else{ console.log('Here'); this.locationClickedBool = true; this.inputFieldValue = location.alias || ''; localStorage.setItem("Location", JSON.stringify(location)); this.lat = location.lat; this.lng = location.long; this.address = location.address; this.locationAlias = location.alias; }; setTimeout(() => { this.available_locations = []}, 200); //gMap = new google.maps.Map(document.getElementById('map')); // this.postion = new google.maps.LatLng(this.lat, this.lng); // this.map.setCenter(this.postion); // this.addMarkerMoveCamera(this.map, new LatLng(this.lat, this.lng)); // this.map.center = new google.maps.LatLng(this.lat, this.lng); this.addMarker(this.map, new LatLng(this.lat, this.lng)); this.moveCamera(this.map, new LatLng(this.lat, this.lng)); } validate():boolean { return (this.lat != null && this.lng != null && this.address != null) ? true :false; } startNextScreen() { console.log("Next clicked!"); let valid:boolean = this.validate(); console.log(valid); if(valid === true && this.locationClickedBool == true) { console.log(this.preGenData); this.navCtrl.push(LaundryItems, { preGenData: this.preGenData, pickupDetails: { location: { lat: this.lat, lng: this.lng, address: this.address } }, }); } else{
{ console.log("ngAfterViewInit", this.newLocation); // this.androidPermissions.checkPermission(this.androidPermissions.PERMISSION.ACCESS_FINE_LOCATION).then( // success => console.log('Permission granted'), // err => this.androidPermissions.requestPermissions(this.androidPermissions.PERMISSION.ACCESS_FINE_LOCATION) // ); // this.geolocation.getCurrentPosition().then((resp) => { // console.log(resp.coords.latitude); // console.log(resp.coords.longitude); // }).catch((error) => { // console.log('Error getting location', error); // }); // this.platform.ready().then(() => { // // this.loadMap(); // });
identifier_body
ionic-native-map.ts
: GoogleMaps, private platform: Platform, private geolocation: Geolocation, private androidPermissions: AndroidPermissions, private alertCtrl: AlertController, private popoverCtrl: PopoverController, private mapService: MapService, private authService: AuthService, private alertCntrl: AlertDialogFactory, private savedLocationService: SavedLocationService) { this.token = localStorage.getItem('x-access-token'); this.userID = localStorage.getItem('userID'); this.preGenData = navParams.get('preGenData'); localStorage.setItem("additionalInfoText", ""); // setTimeout(() => { // this.inputFieldValue = 'New Value'; // }, 3000) } ionViewDidLoad() { console.log('ionViewDidLoad IonicNativeMapPage'); // this.loadMap(); setTimeout(() => { this.loadMap(); }, 500); } ngAfterViewInit(){ console.log("ngAfterViewInit", this.newLocation); // this.androidPermissions.checkPermission(this.androidPermissions.PERMISSION.ACCESS_FINE_LOCATION).then( // success => console.log('Permission granted'), // err => this.androidPermissions.requestPermissions(this.androidPermissions.PERMISSION.ACCESS_FINE_LOCATION) // ); // this.geolocation.getCurrentPosition().then((resp) => { // console.log(resp.coords.latitude); // console.log(resp.coords.longitude); // }).catch((error) => { // console.log('Error getting location', error); // }); // this.platform.ready().then(() => { // // this.loadMap(); // }); this.listenToSearchInput(); this.getMapLocation(location, this.latLng); } listenToSearchInput() { this.hide = false; let location: string; console.log('location1:', location) // let searchInput$ = Observable.fromEvent(this.button.nativeElement, 'keyup') // .map(e => location = e['srcElement'].value.trim()) // .distinctUntilChanged() // .switchMap(() => this.mapService.getJSON(location, this.latLng)) // searchInput$.subscribe(location => { // this.available_locations = location; // console.log(this.available_locations); // }) } getMapLocation(location, latLng) { if (location) { // let location$ = this.mapService.getJSON(location, this.latLng); // location$.subscribe(res => console.log) } } savedButtonClicked(myEvent) { this.saved = this.saved ? false : true; setTimeout(()=>{ this.saved = this.saved ? false : true; }, 200); let inputs; this.addressResponse = inputs; let URL = globalVars.getUsersAddress(this.userID); this.authService.getCall(URL). subscribe(res => { console.log(JSON.parse(res["_body"])); inputs = JSON.parse(res["_body"])["data"]["contact"]["address"]; console.log(inputs); this.addressResponse = inputs; // let result = this.alertCntrl.checkBoxAlertDialog("Saved Locations", inputs) // console.log(result); this.radioAlertDialog("Saved Locations", inputs) }) } radioAlertDialog(title: string, inputs){ this.map.setClickable(false); let alert = this.alertCtrl.create({ title: title, cssClass: 'alertTop' }); inputs.forEach(input => { alert.addInput({ type: 'radio', label: input.alias, value: input, checked: false }); }); alert.addButton({ text: 'Cancel', handler: () => { console.log('Cancel clicked.'); } }); alert.addButton({ text: 'Okay', handler: data => { console.log('Radio data:', data); // this.testCheckboxOpen = false; // this.testCheckboxResult = data; this.locationClicked(data); } }); alert.present(); alert.onDidDismiss((data) => { console.log('OnDidDismiss', data); // dataReturned = data; this.map.setClickable(true); return data || 'null'; }); } saveButtonClicked() { this.save = this.save ? false : true; setTimeout(()=>{ this.save = this.save ? false : true; }, 200); console.log("saveButtonClicked"); let userID = localStorage.getItem("userID"); let URL = globalVars.UserAddress(userID); // console.log(locationExists); let data = { alias: this.locationAlias, address: this.address, lat: this.lat, long: this.lng } if(this.validate()){ // let locationExists: boolean = false; // this.addressResponse.forEach(address => { // locationExists = locationExists || (address.alias == this.locationAlias); // console.log(address.alias, this.locationAlias); // console.log(address.alias == this.locationAlias); // }); // console.log('location Exists: ', locationExists); // if(!locationExists){ this.authService.patchCall(URL, data) .subscribe(res => { if (res.status == 200) { console.log(res['_body']); } }); }else{ this.map.setClickable(false); let result = this.alertCntrl.openAlertDialog('Location exits', 'Please enter a location.'); // result.then(value => { // if(value){ // this.map.setClickable(true); // } // }) } // }else{ // // this.alertCntrl.openAlertDialog('Error', 'Location already Exists.') // } } openAdditionalNoteDialog(myEvent) { this.map.setClickable(false); this.isModalVisible = this.isModalVisible ? false : true; setTimeout(() => { this.isModalVisible = this.isModalVisible ? false : true; }, 200); let popover = this.popoverCtrl.create(AdditionalNote, {}, { showBackdrop: true }); popover.present({ ev: myEvent }); popover.onDidDismiss(data => { if(data){ this.map.setClickable(true); console.log(data); this.additionalInfoText = data + "\n"; localStorage.setItem("additionalInfoText", this.additionalInfoText); } }) } additionButtonClicked(myEvent) { this.addition = this.addition ? false : true; console.log("additionButtonClicked"); this.openAdditionalNoteDialog(myEvent); } locationClickedBool; locationClicked(location) { console.log("You have clicked on: ", location); this.locationClickedBool = false; this.hide = true; if(!!location){ this.inputFieldValue = ''; if(!!location.name){ this.locationClickedBool = true; console.log(location); this.inputFieldValue = location.name || ''; localStorage.setItem("Location", JSON.stringify(location)); this.lat = location.geometry.location.lat; this.lng = location.geometry.location.lng; this.address = location.formatted_address; this.locationAlias = location.name; }else{ console.log('Here'); this.locationClickedBool = true; this.inputFieldValue = location.alias || ''; localStorage.setItem("Location", JSON.stringify(location)); this.lat = location.lat; this.lng = location.long; this.address = location.address; this.locationAlias = location.alias; }; setTimeout(() => { this.available_locations = []}, 200); }else{ console.log('Here'); this.locationClickedBool = true; this.inputFieldValue = location.alias || ''; localStorage.setItem("Location", JSON.stringify(location)); this.lat = location.lat; this.lng = location.long; this.address = location.address; this.locationAlias = location.alias; }; setTimeout(() => { this.available_locations = []}, 200); //gMap = new google.maps.Map(document.getElementById('map')); // this.postion = new google.maps.LatLng(this.lat, this.lng); // this.map.setCenter(this.postion); // this.addMarkerMoveCamera(this.map, new LatLng(this.lat, this.lng)); // this.map.center = new google.maps.LatLng(this.lat, this.lng); this.addMarker(this.map, new LatLng(this.lat, this.lng)); this.moveCamera(this.map, new LatLng(this.lat, this.lng)); } validate():boolean { return (this.lat != null && this.lng != null && this.address != null) ? true :false; } startNextScreen() { console.log("Next clicked!"); let valid:boolean = this.validate(); console.log(valid); if(valid === true && this.locationClickedBool == true) { console.log(this.preGenData); this.navCtrl.push(LaundryItems, { preGenData: this.preGenData, pickupDetails: { location: { lat: this.lat, lng: this.lng, address: this.address } }, }); } else{ this.map.setClickable(false); this.alertCntrl.openAlertDialog("What's missing?","No location selected."); } }
loadMap
identifier_name
ionic-native-map.ts
available_locations: Array<Object> = []; newLocation; marker; constructor(public navCtrl: NavController, public navParams: NavParams, private googleMaps: GoogleMaps, private platform: Platform, private geolocation: Geolocation, private androidPermissions: AndroidPermissions, private alertCtrl: AlertController, private popoverCtrl: PopoverController, private mapService: MapService, private authService: AuthService, private alertCntrl: AlertDialogFactory, private savedLocationService: SavedLocationService) { this.token = localStorage.getItem('x-access-token'); this.userID = localStorage.getItem('userID'); this.preGenData = navParams.get('preGenData'); localStorage.setItem("additionalInfoText", ""); // setTimeout(() => { // this.inputFieldValue = 'New Value'; // }, 3000) } ionViewDidLoad() { console.log('ionViewDidLoad IonicNativeMapPage'); // this.loadMap(); setTimeout(() => { this.loadMap(); }, 500); } ngAfterViewInit(){ console.log("ngAfterViewInit", this.newLocation); // this.androidPermissions.checkPermission(this.androidPermissions.PERMISSION.ACCESS_FINE_LOCATION).then( // success => console.log('Permission granted'), // err => this.androidPermissions.requestPermissions(this.androidPermissions.PERMISSION.ACCESS_FINE_LOCATION) // ); // this.geolocation.getCurrentPosition().then((resp) => { // console.log(resp.coords.latitude); // console.log(resp.coords.longitude); // }).catch((error) => { // console.log('Error getting location', error); // }); // this.platform.ready().then(() => { // // this.loadMap(); // }); this.listenToSearchInput(); this.getMapLocation(location, this.latLng); } listenToSearchInput() { this.hide = false; let location: string; console.log('location1:', location) // let searchInput$ = Observable.fromEvent(this.button.nativeElement, 'keyup') // .map(e => location = e['srcElement'].value.trim()) // .distinctUntilChanged() // .switchMap(() => this.mapService.getJSON(location, this.latLng)) // searchInput$.subscribe(location => { // this.available_locations = location; // console.log(this.available_locations); // }) } getMapLocation(location, latLng) { if (location)
} savedButtonClicked(myEvent) { this.saved = this.saved ? false : true; setTimeout(()=>{ this.saved = this.saved ? false : true; }, 200); let inputs; this.addressResponse = inputs; let URL = globalVars.getUsersAddress(this.userID); this.authService.getCall(URL). subscribe(res => { console.log(JSON.parse(res["_body"])); inputs = JSON.parse(res["_body"])["data"]["contact"]["address"]; console.log(inputs); this.addressResponse = inputs; // let result = this.alertCntrl.checkBoxAlertDialog("Saved Locations", inputs) // console.log(result); this.radioAlertDialog("Saved Locations", inputs) }) } radioAlertDialog(title: string, inputs){ this.map.setClickable(false); let alert = this.alertCtrl.create({ title: title, cssClass: 'alertTop' }); inputs.forEach(input => { alert.addInput({ type: 'radio', label: input.alias, value: input, checked: false }); }); alert.addButton({ text: 'Cancel', handler: () => { console.log('Cancel clicked.'); } }); alert.addButton({ text: 'Okay', handler: data => { console.log('Radio data:', data); // this.testCheckboxOpen = false; // this.testCheckboxResult = data; this.locationClicked(data); } }); alert.present(); alert.onDidDismiss((data) => { console.log('OnDidDismiss', data); // dataReturned = data; this.map.setClickable(true); return data || 'null'; }); } saveButtonClicked() { this.save = this.save ? false : true; setTimeout(()=>{ this.save = this.save ? false : true; }, 200); console.log("saveButtonClicked"); let userID = localStorage.getItem("userID"); let URL = globalVars.UserAddress(userID); // console.log(locationExists); let data = { alias: this.locationAlias, address: this.address, lat: this.lat, long: this.lng } if(this.validate()){ // let locationExists: boolean = false; // this.addressResponse.forEach(address => { // locationExists = locationExists || (address.alias == this.locationAlias); // console.log(address.alias, this.locationAlias); // console.log(address.alias == this.locationAlias); // }); // console.log('location Exists: ', locationExists); // if(!locationExists){ this.authService.patchCall(URL, data) .subscribe(res => { if (res.status == 200) { console.log(res['_body']); } }); }else{ this.map.setClickable(false); let result = this.alertCntrl.openAlertDialog('Location exits', 'Please enter a location.'); // result.then(value => { // if(value){ // this.map.setClickable(true); // } // }) } // }else{ // // this.alertCntrl.openAlertDialog('Error', 'Location already Exists.') // } } openAdditionalNoteDialog(myEvent) { this.map.setClickable(false); this.isModalVisible = this.isModalVisible ? false : true; setTimeout(() => { this.isModalVisible = this.isModalVisible ? false : true; }, 200); let popover = this.popoverCtrl.create(AdditionalNote, {}, { showBackdrop: true }); popover.present({ ev: myEvent }); popover.onDidDismiss(data => { if(data){ this.map.setClickable(true); console.log(data); this.additionalInfoText = data + "\n"; localStorage.setItem("additionalInfoText", this.additionalInfoText); } }) } additionButtonClicked(myEvent) { this.addition = this.addition ? false : true; console.log("additionButtonClicked"); this.openAdditionalNoteDialog(myEvent); } locationClickedBool; locationClicked(location) { console.log("You have clicked on: ", location); this.locationClickedBool = false; this.hide = true; if(!!location){ this.inputFieldValue = ''; if(!!location.name){ this.locationClickedBool = true; console.log(location); this.inputFieldValue = location.name || ''; localStorage.setItem("Location", JSON.stringify(location)); this.lat = location.geometry.location.lat; this.lng = location.geometry.location.lng; this.address = location.formatted_address; this.locationAlias = location.name; }else{ console.log('Here'); this.locationClickedBool = true; this.inputFieldValue = location.alias || ''; localStorage.setItem("Location", JSON.stringify(location)); this.lat = location.lat; this.lng = location.long; this.address = location.address; this.locationAlias = location.alias; }; setTimeout(() => { this.available_locations = []}, 200); }else{ console.log('Here'); this.locationClickedBool = true; this.inputFieldValue = location.alias || ''; localStorage.setItem("Location", JSON.stringify(location)); this.lat = location.lat; this.lng = location.long; this.address = location.address; this.locationAlias = location.alias; }; setTimeout(() => { this.available_locations = []}, 200); //gMap = new google.maps.Map(document.getElementById('map')); // this.postion = new google.maps.LatLng(this.lat, this.lng); // this.map.setCenter(this.postion); // this.addMarkerMoveCamera(this.map, new LatLng(this.lat, this.lng)); // this.map.center = new google.maps.LatLng(this.lat, this.lng); this.addMarker(this.map, new LatLng(this.lat, this.lng)); this.moveCamera(this.map, new LatLng(this.lat, this.lng)); } validate():boolean { return (this.lat != null && this.lng != null && this.address != null) ? true :false; } startNextScreen() { console.log("Next clicked!"); let valid:boolean = this.validate(); console.log(valid); if(valid === true && this.locationClickedBool == true) { console.log(this.preGenData); this.navCtrl.push(LaundryItems, { preGenData: this.preGenData, pickupDetails: { location: { lat: this.lat, lng: this.lng, address: this.address } }, }); } else{
{ // let location$ = this.mapService.getJSON(location, this.latLng); // location$.subscribe(res => console.log) }
conditional_block
lib.rs
" BASIS, // See the License for the specific language governing permissions and // limitations under the License. // // Parts of this work are derived from the `protoc-rust-grpc` crate by // Stepan Koltsov <[email protected]>. // // Copyright 2016, Stepan Koltsov <[email protected]>. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // See the License for the specific language governing permissions and // limitations under the License. #![deny(warnings)] #![warn(missing_docs)] //! An API for programmatically invoking the grpcio gRPC compiler in the same vein as the //! [rust-protoc-grpc](https://crates.io/crates/protoc-rust-grpc) crate from Stepan Koltsov. extern crate grpcio_compiler; #[macro_use] extern crate anyhow; extern crate tempfile; extern crate protobuf; extern crate protobuf_codegen; extern crate protoc; use std::convert::AsRef; use std::fs::File; use std::io::{Read, Write}; use std::iter::Iterator; use std::path::{Path, PathBuf}; use std::vec::Vec; use anyhow::Context; use tempfile::NamedTempFile; use protobuf::{compiler_plugin, descriptor, Message}; use protobuf_codegen::Customize; use protoc::{DescriptorSetOutArgs, Protoc}; /// Custom error type used throughout this crate. pub type CompileError = ::anyhow::Error; /// Custom result type used throughout this crate. pub type CompileResult<T> = Result<T, CompileError>; fn stringify_paths<Paths>(paths: Paths) -> CompileResult<Vec<String>> where Paths: IntoIterator, Paths::Item: AsRef<Path>, { paths .into_iter() .map(|input| match input.as_ref().to_str() { Some(s) => Ok(s.to_owned()), None => Err(format_err!( "failed to convert {:?} to string", input.as_ref() )), }) .collect() } fn write_out_generated_files<P>( generation_results: Vec<compiler_plugin::GenResult>, output_dir: P, ) -> CompileResult<()> where P: AsRef<Path>, { for result in generation_results { let file = output_dir.as_ref().join(result.name); File::create(&file) .context(format!("failed to create {:?}", &file))? .write_all(&result.content) .context(format!("failed to write {:?}", &file))?; } Ok(()) } fn absolutize<P>(path: P) -> CompileResult<PathBuf> where P: AsRef<Path>, { let p = path.as_ref(); if p.is_relative() { match std::env::current_dir() { Ok(cwd) => Ok(cwd.join(p)), Err(err) => Err(format_err!( "Failed to determine CWD needed to absolutize a relative path: {:?}", err )), } } else { Ok(PathBuf::from(p)) } } fn normalize<Paths, Bases>( paths: Paths, bases: Bases, ) -> CompileResult<(Vec<PathBuf>, Vec<PathBuf>, Vec<PathBuf>)> where Paths: IntoIterator, Paths::Item: AsRef<Path>, Bases: IntoIterator, Bases::Item: AsRef<Path>, { let absolutized_bases = bases .into_iter() .map(absolutize) .collect::<CompileResult<Vec<PathBuf>>>()?; // We deal with the following cases: // a.) absolute paths // b.) paths relative to CWD // c.) paths relative to bases // // We take the strategy of transforming the relative path cases (b & c) into absolute paths (a) // and use the strip_prefix API from there. let absolutized_paths = paths .into_iter() .map(|p| { let rel_path = p.as_ref().to_path_buf(); let absolute_path = absolutize(&rel_path)?; Ok((rel_path, absolute_path)) }) // TODO(John Sirois): Use `.flatten()` pending https://github.com/rust-lang/rust/issues/48213 .flat_map(|r: CompileResult<(PathBuf, PathBuf)>| r) .map(|(rel_path, abs_path)| { if abs_path.exists() { // Case a or b. Ok(abs_path) } else { // Case c. for b in &absolutized_bases { let absolutized_path = b.join(&rel_path); if absolutized_path.exists() { return Ok(absolutized_path); } } Err(format_err!( "Failed to find the absolute path of input {:?}", rel_path )) } }) .collect::<CompileResult<Vec<PathBuf>>>()?; let relativized_paths: Vec<PathBuf> = absolutized_paths .iter() .map(|p| { for b in &absolutized_bases { if let Ok(rel_path) = p.strip_prefix(&b) { return Ok(PathBuf::from(rel_path)); } } Err(format_err!( "The input path {:?} is not contained by any of the include paths {:?}", p, absolutized_bases )) }) .collect::<CompileResult<Vec<PathBuf>>>()?; Ok((absolutized_bases, absolutized_paths, relativized_paths)) } /// Compiles a list a gRPC definitions to rust modules. /// /// # Arguments /// /// * `inputs` - A list of protobuf definition paths to compile. Paths can be specified as absolute, /// relative to the CWD or relative to one of the `includes` paths. Note that the directory each /// member of `inputs` is found under must be included in the `includes` parameter. /// * `includes` - A list of of include directory paths to pass to `protoc`. Include paths can be /// specified either as absolute or relative to the CWD. Note that the directory each member of /// `inputs` is found under must be included in this parameter. /// * `output` - Directory to place the generated rust modules into. /// * `customizations` - An Option<protobuf_codegen::Customize> allowing customization options to be /// passed to protobuf_codegen pub fn compile_grpc_protos<Inputs, Includes, Output>( inputs: Inputs, includes: Includes, output: Output, customizations: Option<Customize>, ) -> CompileResult<()> where Inputs: IntoIterator, Inputs::Item: AsRef<Path>, Includes: IntoIterator, Includes::Item: AsRef<Path>, Output: AsRef<Path>, { let protoc = Protoc::from_env_path(); protoc .check() .context("failed to find `protoc`, `protoc` must be availabe in `PATH`")?; let (absolutized_includes, absolutized_paths, relativized_inputs) = normalize(inputs, includes)?; let stringified_inputs_absolute = stringify_paths(absolutized_paths)?; let stringified_inputs = stringify_paths(relativized_inputs)?; let stringified_includes = stringify_paths(absolutized_includes)?; let descriptor_set = NamedTempFile::new()?; protoc .write_descriptor_set(DescriptorSetOutArgs { out: match descriptor_set.as_ref().to_str() { Some(s) => s, None => bail!("failed to convert descriptor set path to string"), }, input: stringified_inputs_absolute .iter() .map(String::as_str) .collect::<Vec<&str>>() .as_slice(), includes: stringified_includes .iter() .map(String::as_str) .collect::<Vec<&str>>() .as_slice(), include_imports: true, }) .context("failed to write descriptor set")?; let mut serialized_descriptor_set = Vec::new(); File::open(&descriptor_set) .context("failed to open descriptor set")? .read_to_end(&mut serialized_descriptor_set) .context("failed to read descriptor set")?; let descriptor_set = descriptor::FileDescriptorSet::parse_from_bytes(&serialized_descriptor_set) .context("failed to parse descriptor set")?; let customize = customizations.unwrap_or_default(); write_out_generated_files( grpcio_compiler::codegen::gen(descriptor_set.get_file(), stringified_inputs.as_slice()), &output, ) .context("failed to write generated grpc definitions")?; write_out_generated_files( protobuf_codegen::gen( descriptor_set.get_file(), stringified_inputs.as_slice(), &customize, ), &output, ) .context("failed to write out generated protobuf definitions")?; Ok(()) } #[cfg(test)] mod tests { use super::*; use std::path::PathBuf; use tempfile::tempdir; fn
<Input, Output>(input: Input, expected_outputs: Output) where Input: AsRef<Path>,
assert_compile_grpc_protos
identifier_name
lib.rs
" BASIS, // See the License for the specific language governing permissions and // limitations under the License. // // Parts of this work are derived from the `protoc-rust-grpc` crate by // Stepan Koltsov <[email protected]>. // // Copyright 2016, Stepan Koltsov <[email protected]>. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // See the License for the specific language governing permissions and // limitations under the License. #![deny(warnings)] #![warn(missing_docs)] //! An API for programmatically invoking the grpcio gRPC compiler in the same vein as the //! [rust-protoc-grpc](https://crates.io/crates/protoc-rust-grpc) crate from Stepan Koltsov. extern crate grpcio_compiler; #[macro_use] extern crate anyhow; extern crate tempfile; extern crate protobuf; extern crate protobuf_codegen; extern crate protoc; use std::convert::AsRef; use std::fs::File; use std::io::{Read, Write}; use std::iter::Iterator; use std::path::{Path, PathBuf}; use std::vec::Vec; use anyhow::Context; use tempfile::NamedTempFile; use protobuf::{compiler_plugin, descriptor, Message}; use protobuf_codegen::Customize; use protoc::{DescriptorSetOutArgs, Protoc}; /// Custom error type used throughout this crate. pub type CompileError = ::anyhow::Error; /// Custom result type used throughout this crate. pub type CompileResult<T> = Result<T, CompileError>; fn stringify_paths<Paths>(paths: Paths) -> CompileResult<Vec<String>> where Paths: IntoIterator, Paths::Item: AsRef<Path>, { paths .into_iter() .map(|input| match input.as_ref().to_str() { Some(s) => Ok(s.to_owned()), None => Err(format_err!( "failed to convert {:?} to string", input.as_ref() )), }) .collect() } fn write_out_generated_files<P>( generation_results: Vec<compiler_plugin::GenResult>, output_dir: P, ) -> CompileResult<()> where P: AsRef<Path>, { for result in generation_results { let file = output_dir.as_ref().join(result.name); File::create(&file) .context(format!("failed to create {:?}", &file))? .write_all(&result.content) .context(format!("failed to write {:?}", &file))?; } Ok(()) } fn absolutize<P>(path: P) -> CompileResult<PathBuf> where P: AsRef<Path>,
fn normalize<Paths, Bases>( paths: Paths, bases: Bases, ) -> CompileResult<(Vec<PathBuf>, Vec<PathBuf>, Vec<PathBuf>)> where Paths: IntoIterator, Paths::Item: AsRef<Path>, Bases: IntoIterator, Bases::Item: AsRef<Path>, { let absolutized_bases = bases .into_iter() .map(absolutize) .collect::<CompileResult<Vec<PathBuf>>>()?; // We deal with the following cases: // a.) absolute paths // b.) paths relative to CWD // c.) paths relative to bases // // We take the strategy of transforming the relative path cases (b & c) into absolute paths (a) // and use the strip_prefix API from there. let absolutized_paths = paths .into_iter() .map(|p| { let rel_path = p.as_ref().to_path_buf(); let absolute_path = absolutize(&rel_path)?; Ok((rel_path, absolute_path)) }) // TODO(John Sirois): Use `.flatten()` pending https://github.com/rust-lang/rust/issues/48213 .flat_map(|r: CompileResult<(PathBuf, PathBuf)>| r) .map(|(rel_path, abs_path)| { if abs_path.exists() { // Case a or b. Ok(abs_path) } else { // Case c. for b in &absolutized_bases { let absolutized_path = b.join(&rel_path); if absolutized_path.exists() { return Ok(absolutized_path); } } Err(format_err!( "Failed to find the absolute path of input {:?}", rel_path )) } }) .collect::<CompileResult<Vec<PathBuf>>>()?; let relativized_paths: Vec<PathBuf> = absolutized_paths .iter() .map(|p| { for b in &absolutized_bases { if let Ok(rel_path) = p.strip_prefix(&b) { return Ok(PathBuf::from(rel_path)); } } Err(format_err!( "The input path {:?} is not contained by any of the include paths {:?}", p, absolutized_bases )) }) .collect::<CompileResult<Vec<PathBuf>>>()?; Ok((absolutized_bases, absolutized_paths, relativized_paths)) } /// Compiles a list a gRPC definitions to rust modules. /// /// # Arguments /// /// * `inputs` - A list of protobuf definition paths to compile. Paths can be specified as absolute, /// relative to the CWD or relative to one of the `includes` paths. Note that the directory each /// member of `inputs` is found under must be included in the `includes` parameter. /// * `includes` - A list of of include directory paths to pass to `protoc`. Include paths can be /// specified either as absolute or relative to the CWD. Note that the directory each member of /// `inputs` is found under must be included in this parameter. /// * `output` - Directory to place the generated rust modules into. /// * `customizations` - An Option<protobuf_codegen::Customize> allowing customization options to be /// passed to protobuf_codegen pub fn compile_grpc_protos<Inputs, Includes, Output>( inputs: Inputs, includes: Includes, output: Output, customizations: Option<Customize>, ) -> CompileResult<()> where Inputs: IntoIterator, Inputs::Item: AsRef<Path>, Includes: IntoIterator, Includes::Item: AsRef<Path>, Output: AsRef<Path>, { let protoc = Protoc::from_env_path(); protoc .check() .context("failed to find `protoc`, `protoc` must be availabe in `PATH`")?; let (absolutized_includes, absolutized_paths, relativized_inputs) = normalize(inputs, includes)?; let stringified_inputs_absolute = stringify_paths(absolutized_paths)?; let stringified_inputs = stringify_paths(relativized_inputs)?; let stringified_includes = stringify_paths(absolutized_includes)?; let descriptor_set = NamedTempFile::new()?; protoc .write_descriptor_set(DescriptorSetOutArgs { out: match descriptor_set.as_ref().to_str() { Some(s) => s, None => bail!("failed to convert descriptor set path to string"), }, input: stringified_inputs_absolute .iter() .map(String::as_str) .collect::<Vec<&str>>() .as_slice(), includes: stringified_includes .iter() .map(String::as_str) .collect::<Vec<&str>>() .as_slice(), include_imports: true, }) .context("failed to write descriptor set")?; let mut serialized_descriptor_set = Vec::new(); File::open(&descriptor_set) .context("failed to open descriptor set")? .read_to_end(&mut serialized_descriptor_set) .context("failed to read descriptor set")?; let descriptor_set = descriptor::FileDescriptorSet::parse_from_bytes(&serialized_descriptor_set) .context("failed to parse descriptor set")?; let customize = customizations.unwrap_or_default(); write_out_generated_files( grpcio_compiler::codegen::gen(descriptor_set.get_file(), stringified_inputs.as_slice()), &output, ) .context("failed to write generated grpc definitions")?; write_out_generated_files( protobuf_codegen::gen( descriptor_set.get_file(), stringified_inputs.as_slice(), &customize, ), &output, ) .context("failed to write out generated protobuf definitions")?; Ok(()) } #[cfg(test)] mod tests { use super::*; use std::path::PathBuf; use tempfile::tempdir; fn assert_compile_grpc_protos<Input, Output>(input: Input, expected_outputs: Output) where Input: AsRef<Path>,
{ let p = path.as_ref(); if p.is_relative() { match std::env::current_dir() { Ok(cwd) => Ok(cwd.join(p)), Err(err) => Err(format_err!( "Failed to determine CWD needed to absolutize a relative path: {:?}", err )), } } else { Ok(PathBuf::from(p)) } }
identifier_body
lib.rs
IS" BASIS, // See the License for the specific language governing permissions and // limitations under the License. // // Parts of this work are derived from the `protoc-rust-grpc` crate by // Stepan Koltsov <[email protected]>. // // Copyright 2016, Stepan Koltsov <[email protected]>. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // See the License for the specific language governing permissions and // limitations under the License. #![deny(warnings)] #![warn(missing_docs)] //! An API for programmatically invoking the grpcio gRPC compiler in the same vein as the //! [rust-protoc-grpc](https://crates.io/crates/protoc-rust-grpc) crate from Stepan Koltsov. extern crate grpcio_compiler; #[macro_use] extern crate anyhow; extern crate tempfile; extern crate protobuf; extern crate protobuf_codegen; extern crate protoc; use std::convert::AsRef; use std::fs::File; use std::io::{Read, Write}; use std::iter::Iterator; use std::path::{Path, PathBuf}; use std::vec::Vec; use anyhow::Context; use tempfile::NamedTempFile; use protobuf::{compiler_plugin, descriptor, Message}; use protobuf_codegen::Customize; use protoc::{DescriptorSetOutArgs, Protoc}; /// Custom error type used throughout this crate. pub type CompileError = ::anyhow::Error; /// Custom result type used throughout this crate. pub type CompileResult<T> = Result<T, CompileError>; fn stringify_paths<Paths>(paths: Paths) -> CompileResult<Vec<String>> where Paths: IntoIterator, Paths::Item: AsRef<Path>, { paths .into_iter() .map(|input| match input.as_ref().to_str() { Some(s) => Ok(s.to_owned()), None => Err(format_err!( "failed to convert {:?} to string", input.as_ref() )), }) .collect() } fn write_out_generated_files<P>( generation_results: Vec<compiler_plugin::GenResult>, output_dir: P, ) -> CompileResult<()> where P: AsRef<Path>, { for result in generation_results { let file = output_dir.as_ref().join(result.name); File::create(&file) .context(format!("failed to create {:?}", &file))? .write_all(&result.content) .context(format!("failed to write {:?}", &file))?; } Ok(()) } fn absolutize<P>(path: P) -> CompileResult<PathBuf> where P: AsRef<Path>, { let p = path.as_ref(); if p.is_relative() { match std::env::current_dir() { Ok(cwd) => Ok(cwd.join(p)), Err(err) => Err(format_err!( "Failed to determine CWD needed to absolutize a relative path: {:?}", err )), } } else { Ok(PathBuf::from(p)) } } fn normalize<Paths, Bases>( paths: Paths, bases: Bases, ) -> CompileResult<(Vec<PathBuf>, Vec<PathBuf>, Vec<PathBuf>)> where Paths: IntoIterator, Paths::Item: AsRef<Path>, Bases: IntoIterator, Bases::Item: AsRef<Path>, { let absolutized_bases = bases .into_iter() .map(absolutize) .collect::<CompileResult<Vec<PathBuf>>>()?; // We deal with the following cases: // a.) absolute paths // b.) paths relative to CWD // c.) paths relative to bases // // We take the strategy of transforming the relative path cases (b & c) into absolute paths (a) // and use the strip_prefix API from there. let absolutized_paths = paths .into_iter() .map(|p| { let rel_path = p.as_ref().to_path_buf(); let absolute_path = absolutize(&rel_path)?; Ok((rel_path, absolute_path)) }) // TODO(John Sirois): Use `.flatten()` pending https://github.com/rust-lang/rust/issues/48213 .flat_map(|r: CompileResult<(PathBuf, PathBuf)>| r) .map(|(rel_path, abs_path)| { if abs_path.exists() { // Case a or b. Ok(abs_path) } else { // Case c. for b in &absolutized_bases { let absolutized_path = b.join(&rel_path); if absolutized_path.exists() { return Ok(absolutized_path); } } Err(format_err!( "Failed to find the absolute path of input {:?}", rel_path )) } }) .collect::<CompileResult<Vec<PathBuf>>>()?; let relativized_paths: Vec<PathBuf> = absolutized_paths .iter() .map(|p| { for b in &absolutized_bases { if let Ok(rel_path) = p.strip_prefix(&b) { return Ok(PathBuf::from(rel_path)); } } Err(format_err!( "The input path {:?} is not contained by any of the include paths {:?}", p, absolutized_bases )) }) .collect::<CompileResult<Vec<PathBuf>>>()?; Ok((absolutized_bases, absolutized_paths, relativized_paths)) } /// Compiles a list a gRPC definitions to rust modules. /// /// # Arguments /// /// * `inputs` - A list of protobuf definition paths to compile. Paths can be specified as absolute, /// relative to the CWD or relative to one of the `includes` paths. Note that the directory each /// member of `inputs` is found under must be included in the `includes` parameter. /// * `includes` - A list of of include directory paths to pass to `protoc`. Include paths can be /// specified either as absolute or relative to the CWD. Note that the directory each member of /// `inputs` is found under must be included in this parameter. /// * `output` - Directory to place the generated rust modules into. /// * `customizations` - An Option<protobuf_codegen::Customize> allowing customization options to be /// passed to protobuf_codegen pub fn compile_grpc_protos<Inputs, Includes, Output>( inputs: Inputs, includes: Includes, output: Output, customizations: Option<Customize>, ) -> CompileResult<()> where Inputs: IntoIterator, Inputs::Item: AsRef<Path>, Includes: IntoIterator, Includes::Item: AsRef<Path>, Output: AsRef<Path>, { let protoc = Protoc::from_env_path(); protoc .check() .context("failed to find `protoc`, `protoc` must be availabe in `PATH`")?; let (absolutized_includes, absolutized_paths, relativized_inputs) = normalize(inputs, includes)?; let stringified_inputs_absolute = stringify_paths(absolutized_paths)?; let stringified_inputs = stringify_paths(relativized_inputs)?; let stringified_includes = stringify_paths(absolutized_includes)?; let descriptor_set = NamedTempFile::new()?; protoc .write_descriptor_set(DescriptorSetOutArgs { out: match descriptor_set.as_ref().to_str() { Some(s) => s, None => bail!("failed to convert descriptor set path to string"), }, input: stringified_inputs_absolute
.as_slice(), includes: stringified_includes .iter() .map(String::as_str) .collect::<Vec<&str>>() .as_slice(), include_imports: true, }) .context("failed to write descriptor set")?; let mut serialized_descriptor_set = Vec::new(); File::open(&descriptor_set) .context("failed to open descriptor set")? .read_to_end(&mut serialized_descriptor_set) .context("failed to read descriptor set")?; let descriptor_set = descriptor::FileDescriptorSet::parse_from_bytes(&serialized_descriptor_set) .context("failed to parse descriptor set")?; let customize = customizations.unwrap_or_default(); write_out_generated_files( grpcio_compiler::codegen::gen(descriptor_set.get_file(), stringified_inputs.as_slice()), &output, ) .context("failed to write generated grpc definitions")?; write_out_generated_files( protobuf_codegen::gen( descriptor_set.get_file(), stringified_inputs.as_slice(), &customize, ), &output, ) .context("failed to write out generated protobuf definitions")?; Ok(()) } #[cfg(test)] mod tests { use super::*; use std::path::PathBuf; use tempfile::tempdir; fn assert_compile_grpc_protos<Input, Output>(input: Input, expected_outputs: Output) where Input: AsRef<Path>, Output
.iter() .map(String::as_str) .collect::<Vec<&str>>()
random_line_split
baselines_utils.py
class SetFromFlat(object): def __init__(self, var_list, dtype=tf.float32): assigns = [] shapes = list(map(var_shape, var_list)) total_size = np.sum([intprod(shape) for shape in shapes]) self.theta = theta = tf.placeholder(dtype, [total_size]) start = 0 assigns = [] for (shape, v) in zip(shapes, var_list): size = intprod(shape) assigns.append(tf.assign(v, tf.reshape(theta[start:start + size], shape))) start += size self.op = tf.group(*assigns) def __call__(self, theta): tf.get_default_session().run(self.op, feed_dict={self.theta: theta}) class GetFlat(object): def __init__(self, var_list): self.op = tf.concat(axis=0, values=[tf.reshape(v, [numel(v)]) for v in var_list]) def __call__(self): return tf.get_default_session().run(self.op) _PLACEHOLDER_CACHE = {} # name -> (placeholder, dtype, shape) def get_placeholder(name, dtype, shape): if name in _PLACEHOLDER_CACHE: out, dtype1, shape1 = _PLACEHOLDER_CACHE[name] assert dtype1 == dtype and shape1 == shape return out else: out = tf.placeholder(dtype=dtype, shape=shape, name=name) _PLACEHOLDER_CACHE[name] = (out, dtype, shape) return out def get_placeholder_cached(name): return _PLACEHOLDER_CACHE[name][0] def flattenallbut0(x): return tf.reshape(x, [-1, intprod(x.get_shape().as_list()[1:])]) # ================================================================ # Diagnostics # ================================================================ def display_var_info(vars): from baselines import logger count_params = 0 for v in vars: name = v.name if "/Adam" in name or "beta1_power" in name or "beta2_power" in name: continue count_params += np.prod(v.shape.as_list()) if "/b:" in name: continue # Wx+b, bias is not interesting to look at => count params, but not print logger.info(" %s%s%s" % (name, " "*(55-len(name)), str(v.shape))) logger.info("Total model parameters: %0.1f million" % (count_params*1e-6)) class LinearSchedule(object): def __init__(self, schedule_timesteps, final_p, initial_p=1.0): """Linear interpolation between initial_p and final_p over schedule_timesteps. After this many timesteps pass final_p is returned. Parameters ---------- schedule_timesteps: int Number of timesteps for which to linearly anneal initial_p to final_p initial_p: float initial output value final_p: float final output value """ self.schedule_timesteps = schedule_timesteps self.final_p = final_p self.initial_p = initial_p def value(self, t): """See Schedule.value""" fraction = min(float(t) / self.schedule_timesteps, 1.0) return self.initial_p + fraction * (self.final_p - self.initial_p) # deep q import os import tensorflow as tf # ================================================================ # Saving variables # ================================================================ def load_state(fname): saver = tf.train.Saver() saver.restore(tf.get_default_session(), fname) def save_state(fname): os.makedirs(os.path.dirname(fname), exist_ok=True) saver = tf.train.Saver() saver.save(tf.get_default_session(), fname) # ================================================================ # Placeholders # ================================================================ class TfInput(object): def __init__(self, name="(unnamed)"): """Generalized Tensorflow placeholder. The main differences are: - possibly uses multiple placeholders internally and returns multiple values - can apply light postprocessing to the value feed to placeholder. """ self.name = name def get(self): """Return the tf variable(s) representing the possibly postprocessed value of placeholder(s). """ raise NotImplemented() def make_feed_dict(data): """Given data input it to the placeholder(s).""" raise NotImplemented() class PlaceholderTfInput(TfInput): def __init__(self, placeholder): """Wrapper for regular tensorflow placeholder.""" super().__init__(placeholder.name) self._placeholder = placeholder def get(self): return self._placeholder def make_feed_dict(self, data): return {self._placeholder: data} class BatchInput(PlaceholderTfInput): def __init__(self, shape, dtype=tf.float32, name=None): """Creates a placeholder for a batch of tensors of a given shape and dtype Parameters ---------- shape: [int] shape of a single elemenet of the batch dtype: tf.dtype number representation used for tensor contents name: str name of the underlying placeholder """ super().__init__(tf.placeholder(dtype, [None] + list(shape), name=name)) class Uint8Input(PlaceholderTfInput): def __init__(self, shape, name=None): """Takes input in uint8 format which is cast to float32 and divided by 255 before passing it to the model. On GPU this ensures lower data transfer times. Parameters ---------- shape: [int] shape of the tensor. name: str name of the underlying placeholder """ super().__init__(tf.placeholder(tf.uint8, [None] + list(shape), name=name)) self._shape = shape self._output = tf.cast(super().get(), tf.float32) / 255.0 def get(self): return self._output # Tree import operator class SegmentTree(object): def __init__(self, capacity, operation, neutral_element): """Build a Segment Tree data structure. https://en.wikipedia.org/wiki/Segment_tree Can be used as regular array, but with two important differences: a) setting item's value is slightly slower. It is O(lg capacity) instead of O(1). b) user has access to an efficient `reduce` operation which reduces `operation` over a contiguous subsequence of items in the array. Paramters --------- capacity: int Total size of the array - must be a power of two. operation: lambda obj, obj -> obj and operation for combining elements (eg. sum, max) must for a mathematical group together with the set of possible values for array elements. neutral_element: obj neutral element for the operation above. eg. float('-inf') for max and 0 for sum. """ assert capacity > 0 and capacity & (capacity - 1) == 0, "capacity must be positive and a power of 2." self._capacity = capacity self._value = [neutral_element for _ in range(2 * capacity)] self._operation = operation def _reduce_helper(self, start, end, node, node_start, node_end): if start == node_start and end == node_end: return self._value[node] mid = (node_start + node_end) // 2 if end <= mid: return self._reduce_helper(start, end, 2 * node, node_start, mid) else: if mid + 1 <= start: return self._reduce_helper(start, end, 2 * node + 1, mid + 1, node_end) else: return self._operation( self._reduce_helper(start, mid, 2 * node, node_start, mid), self._reduce_helper(mid + 1, end, 2 * node + 1, mid + 1, node_end) ) def reduce(self, start=0, end=None): """Returns result of applying `self.operation` to a contiguous subsequence of the array. self.operation(arr[start], operation(arr[start+1], operation(... arr[end]))) Parameters ---------- start: int beginning of the subsequence end: int end of the subsequences Returns ------- reduced: obj result of reducing self.operation over the specified range of array elements. """ if end is None: end = self._capacity if end < 0: end += self._capacity end -= 1 return self._reduce_helper(start, end, 1, 0, self._capacity - 1) def __setitem__(self, idx, val): # index of the leaf idx += self._capacity self._value[idx] = val idx //= 2 while idx >= 1: self._value[idx] = self._operation( self._value[2 * idx], self._value[2 * idx + 1] ) idx //= 2 def __getitem__(self, idx): assert 0 <= idx < self._capacity return self._value[self._capacity + idx] class SumSegmentTree(SegmentTree): def __init__(self, capacity):
super(SumSegmentTree, self).__init__( capacity=capacity, operation=operator.add, neutral_element=0.0 )
identifier_body
baselines_utils.py
else: feed_dict[inpt] = value def __call__(self, *args): assert len(args) <= len(self.inputs), "Too many arguments provided" feed_dict = {} # Update the args for inpt, value in zip(self.inputs, args): self._feed_input(feed_dict, inpt, value) # Update feed dict with givens. for inpt in self.givens: feed_dict[inpt] = feed_dict.get(inpt, self.givens[inpt]) results = tf.get_default_session().run(self.outputs_update, feed_dict=feed_dict)[:-1] return results # ================================================================ # Flat vectors # ================================================================ def var_shape(x): out = x.get_shape().as_list() assert all(isinstance(a, int) for a in out), \ "shape function assumes that shape is fully known" return out def numel(x): return intprod(var_shape(x)) def intprod(x): return int(np.prod(x)) def flatgrad(loss, var_list, clip_norm=None): grads = tf.gradients(loss, var_list) if clip_norm is not None: grads = [tf.clip_by_norm(grad, clip_norm=clip_norm) for grad in grads] return tf.concat(axis=0, values=[ tf.reshape(grad if grad is not None else tf.zeros_like(v), [numel(v)]) for (v, grad) in zip(var_list, grads) ]) class SetFromFlat(object): def __init__(self, var_list, dtype=tf.float32): assigns = [] shapes = list(map(var_shape, var_list)) total_size = np.sum([intprod(shape) for shape in shapes]) self.theta = theta = tf.placeholder(dtype, [total_size]) start = 0 assigns = [] for (shape, v) in zip(shapes, var_list): size = intprod(shape) assigns.append(tf.assign(v, tf.reshape(theta[start:start + size], shape))) start += size self.op = tf.group(*assigns) def __call__(self, theta): tf.get_default_session().run(self.op, feed_dict={self.theta: theta}) class GetFlat(object): def __init__(self, var_list): self.op = tf.concat(axis=0, values=[tf.reshape(v, [numel(v)]) for v in var_list]) def __call__(self): return tf.get_default_session().run(self.op) _PLACEHOLDER_CACHE = {} # name -> (placeholder, dtype, shape) def get_placeholder(name, dtype, shape): if name in _PLACEHOLDER_CACHE: out, dtype1, shape1 = _PLACEHOLDER_CACHE[name] assert dtype1 == dtype and shape1 == shape return out else: out = tf.placeholder(dtype=dtype, shape=shape, name=name) _PLACEHOLDER_CACHE[name] = (out, dtype, shape) return out def get_placeholder_cached(name): return _PLACEHOLDER_CACHE[name][0] def flattenallbut0(x): return tf.reshape(x, [-1, intprod(x.get_shape().as_list()[1:])]) # ================================================================ # Diagnostics # ================================================================ def display_var_info(vars): from baselines import logger count_params = 0 for v in vars: name = v.name if "/Adam" in name or "beta1_power" in name or "beta2_power" in name: continue count_params += np.prod(v.shape.as_list()) if "/b:" in name: continue # Wx+b, bias is not interesting to look at => count params, but not print logger.info(" %s%s%s" % (name, " "*(55-len(name)), str(v.shape))) logger.info("Total model parameters: %0.1f million" % (count_params*1e-6)) class LinearSchedule(object): def __init__(self, schedule_timesteps, final_p, initial_p=1.0): """Linear interpolation between initial_p and final_p over schedule_timesteps. After this many timesteps pass final_p is returned. Parameters ---------- schedule_timesteps: int Number of timesteps for which to linearly anneal initial_p to final_p initial_p: float initial output value final_p: float final output value """ self.schedule_timesteps = schedule_timesteps self.final_p = final_p self.initial_p = initial_p def value(self, t): """See Schedule.value""" fraction = min(float(t) / self.schedule_timesteps, 1.0) return self.initial_p + fraction * (self.final_p - self.initial_p) # deep q import os import tensorflow as tf # ================================================================ # Saving variables # ================================================================ def load_state(fname): saver = tf.train.Saver() saver.restore(tf.get_default_session(), fname) def save_state(fname): os.makedirs(os.path.dirname(fname), exist_ok=True) saver = tf.train.Saver() saver.save(tf.get_default_session(), fname) # ================================================================ # Placeholders # ================================================================ class TfInput(object): def __init__(self, name="(unnamed)"): """Generalized Tensorflow placeholder. The main differences are: - possibly uses multiple placeholders internally and returns multiple values - can apply light postprocessing to the value feed to placeholder. """ self.name = name def get(self): """Return the tf variable(s) representing the possibly postprocessed value of placeholder(s). """ raise NotImplemented() def make_feed_dict(data): """Given data input it to the placeholder(s).""" raise NotImplemented() class PlaceholderTfInput(TfInput): def __init__(self, placeholder): """Wrapper for regular tensorflow placeholder.""" super().__init__(placeholder.name) self._placeholder = placeholder def get(self): return self._placeholder def make_feed_dict(self, data): return {self._placeholder: data} class BatchInput(PlaceholderTfInput): def __init__(self, shape, dtype=tf.float32, name=None): """Creates a placeholder for a batch of tensors of a given shape and dtype Parameters ---------- shape: [int] shape of a single elemenet of the batch dtype: tf.dtype number representation used for tensor contents name: str name of the underlying placeholder """ super().__init__(tf.placeholder(dtype, [None] + list(shape), name=name)) class Uint8Input(PlaceholderTfInput): def __init__(self, shape, name=None): """Takes input in uint8 format which is cast to float32 and divided by 255 before passing it to the model. On GPU this ensures lower data transfer times. Parameters ---------- shape: [int] shape of the tensor. name: str name of the underlying placeholder """ super().__init__(tf.placeholder(tf.uint8, [None] + list(shape), name=name)) self._shape = shape self._output = tf.cast(super().get(), tf.float32) / 255.0 def get(self): return self._output # Tree import operator class SegmentTree(object): def __init__(self, capacity, operation, neutral_element): """Build a Segment Tree data structure. https://en.wikipedia.org/wiki/Segment_tree Can be used as regular array, but with two important differences: a) setting item's value is slightly slower. It is O(lg capacity) instead of O(1). b) user has access to an efficient `reduce` operation which reduces `operation` over a contiguous subsequence of items in the array. Paramters --------- capacity: int Total size of the array - must be a power of two. operation: lambda obj, obj -> obj and operation for combining elements (eg. sum, max) must for a mathematical group together with the set of possible values for array elements. neutral_element: obj neutral element for the operation above. eg. float('-inf') for max and 0 for sum. """ assert capacity > 0 and capacity & (capacity - 1) == 0, "capacity must be positive and a power of 2." self._capacity = capacity self._value = [neutral_element for _ in range(2 * capacity)] self._operation = operation def _reduce_helper(self, start, end, node, node_start, node_end): if start == node_start and end == node_end: return self._value[node] mid = (node_start + node_end) // 2 if end <= mid: return self._reduce_helper(start, end, 2 * node, node_start, mid) else: if mid + 1 <= start: return self._reduce_helper(start, end, 2 * node + 1, mid + 1, node_end) else: return self._operation( self._reduce_helper(start, mid, 2 * node, node_start, mid), self._reduce_helper(mid + 1, end, 2 * node + 1, mid + 1, node_end)
) def reduce(self, start=0, end=None):
random_line_split
baselines_utils.py
summary_tag=None): with tf.variable_scope(name): stride_shape = [1, stride[0], stride[1], 1] filter_shape = [filter_size[0], filter_size[1], int(x.get_shape()[3]), num_filters] # there are "num input feature maps * filter height * filter width" # inputs to each hidden unit fan_in = intprod(filter_shape[:3]) # each unit in the lower layer receives a gradient from: # "num output feature maps * filter height * filter width" / # pooling size fan_out = intprod(filter_shape[:2]) * num_filters # initialize weights with random weights w_bound = np.sqrt(6. / (fan_in + fan_out)) w = tf.get_variable("W", filter_shape, dtype, tf.random_uniform_initializer(-w_bound, w_bound), collections=collections) b = tf.get_variable("b", [1, 1, 1, num_filters], initializer=tf.zeros_initializer(), collections=collections) if summary_tag is not None: tf.summary.image(summary_tag, tf.transpose(tf.reshape(w, [filter_size[0], filter_size[1], -1, 1]), [2, 0, 1, 3]), max_images=10) return tf.nn.conv2d(x, w, stride_shape, pad) + b # ================================================================ # Theano-like Function # ================================================================ def function(inputs, outputs, updates=None, givens=None): """Just like Theano function. Take a bunch of tensorflow placeholders and expressions computed based on those placeholders and produces f(inputs) -> outputs. Function f takes values to be fed to the input's placeholders and produces the values of the expressions in outputs. Input values can be passed in the same order as inputs or can be provided as kwargs based on placeholder name (passed to constructor or accessible via placeholder.op.name). Example: x = tf.placeholder(tf.int32, (), name="x") y = tf.placeholder(tf.int32, (), name="y") z = 3 * x + 2 * y lin = function([x, y], z, givens={y: 0}) with single_threaded_session(): initialize() assert lin(2) == 6 assert lin(x=3) == 9 assert lin(2, 2) == 10 assert lin(x=2, y=3) == 12 Parameters ---------- inputs: [tf.placeholder, tf.constant, or object with make_feed_dict method] list of input arguments outputs: [tf.Variable] or tf.Variable list of outputs or a single output to be returned from function. Returned value will also have the same shape. """ if isinstance(outputs, list): return _Function(inputs, outputs, updates, givens=givens) elif isinstance(outputs, (dict, collections.OrderedDict)): f = _Function(inputs, outputs.values(), updates, givens=givens) return lambda *args, **kwargs: type(outputs)(zip(outputs.keys(), f(*args, **kwargs))) else: f = _Function(inputs, [outputs], updates, givens=givens) return lambda *args, **kwargs: f(*args, **kwargs)[0] class _Function(object): def __init__(self, inputs, outputs, updates, givens): for inpt in inputs: if not hasattr(inpt, 'make_feed_dict') and not (type(inpt) is tf.Tensor and len(inpt.op.inputs) == 0): assert False, "inputs should all be placeholders, constants, or have a make_feed_dict method" self.inputs = inputs updates = updates or [] self.update_group = tf.group(*updates) self.outputs_update = list(outputs) + [self.update_group] self.givens = {} if givens is None else givens def _feed_input(self, feed_dict, inpt, value): if hasattr(inpt, 'make_feed_dict'): feed_dict.update(inpt.make_feed_dict(value)) else: feed_dict[inpt] = value def __call__(self, *args): assert len(args) <= len(self.inputs), "Too many arguments provided" feed_dict = {} # Update the args for inpt, value in zip(self.inputs, args): self._feed_input(feed_dict, inpt, value) # Update feed dict with givens. for inpt in self.givens: feed_dict[inpt] = feed_dict.get(inpt, self.givens[inpt]) results = tf.get_default_session().run(self.outputs_update, feed_dict=feed_dict)[:-1] return results # ================================================================ # Flat vectors # ================================================================ def
(x): out = x.get_shape().as_list() assert all(isinstance(a, int) for a in out), \ "shape function assumes that shape is fully known" return out def numel(x): return intprod(var_shape(x)) def intprod(x): return int(np.prod(x)) def flatgrad(loss, var_list, clip_norm=None): grads = tf.gradients(loss, var_list) if clip_norm is not None: grads = [tf.clip_by_norm(grad, clip_norm=clip_norm) for grad in grads] return tf.concat(axis=0, values=[ tf.reshape(grad if grad is not None else tf.zeros_like(v), [numel(v)]) for (v, grad) in zip(var_list, grads) ]) class SetFromFlat(object): def __init__(self, var_list, dtype=tf.float32): assigns = [] shapes = list(map(var_shape, var_list)) total_size = np.sum([intprod(shape) for shape in shapes]) self.theta = theta = tf.placeholder(dtype, [total_size]) start = 0 assigns = [] for (shape, v) in zip(shapes, var_list): size = intprod(shape) assigns.append(tf.assign(v, tf.reshape(theta[start:start + size], shape))) start += size self.op = tf.group(*assigns) def __call__(self, theta): tf.get_default_session().run(self.op, feed_dict={self.theta: theta}) class GetFlat(object): def __init__(self, var_list): self.op = tf.concat(axis=0, values=[tf.reshape(v, [numel(v)]) for v in var_list]) def __call__(self): return tf.get_default_session().run(self.op) _PLACEHOLDER_CACHE = {} # name -> (placeholder, dtype, shape) def get_placeholder(name, dtype, shape): if name in _PLACEHOLDER_CACHE: out, dtype1, shape1 = _PLACEHOLDER_CACHE[name] assert dtype1 == dtype and shape1 == shape return out else: out = tf.placeholder(dtype=dtype, shape=shape, name=name) _PLACEHOLDER_CACHE[name] = (out, dtype, shape) return out def get_placeholder_cached(name): return _PLACEHOLDER_CACHE[name][0] def flattenallbut0(x): return tf.reshape(x, [-1, intprod(x.get_shape().as_list()[1:])]) # ================================================================ # Diagnostics # ================================================================ def display_var_info(vars): from baselines import logger count_params = 0 for v in vars: name = v.name if "/Adam" in name or "beta1_power" in name or "beta2_power" in name: continue count_params += np.prod(v.shape.as_list()) if "/b:" in name: continue # Wx+b, bias is not interesting to look at => count params, but not print logger.info(" %s%s%s" % (name, " "*(55-len(name)), str(v.shape))) logger.info("Total model parameters: %0.1f million" % (count_params*1e-6)) class LinearSchedule(object): def __init__(self, schedule_timesteps, final_p, initial_p=1.0): """Linear interpolation between initial_p and final_p over schedule_timesteps. After this many timesteps pass final_p is returned. Parameters ---------- schedule_timesteps: int Number of timesteps for which to linearly anneal initial_p to final_p initial_p: float initial output value final_p: float final output value """ self.schedule_timesteps = schedule_timesteps self.final_p = final_p self.initial_p = initial_p def value(self, t): """See Schedule.value""" fraction = min(float(t) / self.schedule_timesteps, 1.0) return self.initial_p + fraction * (self.final_p - self.initial_p) # deep q import os import tensorflow as tf # ================================================================ # Saving variables # ================================================================ def load_state(fname): saver = tf.train.Saver() saver.restore(tf.get_default_session(), fname) def save_state(fname): os.makedirs(os.path.dirname(fname), exist_ok=True) saver = tf.train.Saver() saver.save(tf.get_default_session(), fname) # ================================================================ # Placeholders # ================================================================ class TfInput(object): def __init__(self, name="(unnamed)
var_shape
identifier_name
baselines_utils.py
summary_tag=None): with tf.variable_scope(name): stride_shape = [1, stride[0], stride[1], 1] filter_shape = [filter_size[0], filter_size[1], int(x.get_shape()[3]), num_filters] # there are "num input feature maps * filter height * filter width" # inputs to each hidden unit fan_in = intprod(filter_shape[:3]) # each unit in the lower layer receives a gradient from: # "num output feature maps * filter height * filter width" / # pooling size fan_out = intprod(filter_shape[:2]) * num_filters # initialize weights with random weights w_bound = np.sqrt(6. / (fan_in + fan_out)) w = tf.get_variable("W", filter_shape, dtype, tf.random_uniform_initializer(-w_bound, w_bound), collections=collections) b = tf.get_variable("b", [1, 1, 1, num_filters], initializer=tf.zeros_initializer(), collections=collections) if summary_tag is not None: tf.summary.image(summary_tag, tf.transpose(tf.reshape(w, [filter_size[0], filter_size[1], -1, 1]), [2, 0, 1, 3]), max_images=10) return tf.nn.conv2d(x, w, stride_shape, pad) + b # ================================================================ # Theano-like Function # ================================================================ def function(inputs, outputs, updates=None, givens=None): """Just like Theano function. Take a bunch of tensorflow placeholders and expressions computed based on those placeholders and produces f(inputs) -> outputs. Function f takes values to be fed to the input's placeholders and produces the values of the expressions in outputs. Input values can be passed in the same order as inputs or can be provided as kwargs based on placeholder name (passed to constructor or accessible via placeholder.op.name). Example: x = tf.placeholder(tf.int32, (), name="x") y = tf.placeholder(tf.int32, (), name="y") z = 3 * x + 2 * y lin = function([x, y], z, givens={y: 0}) with single_threaded_session(): initialize() assert lin(2) == 6 assert lin(x=3) == 9 assert lin(2, 2) == 10 assert lin(x=2, y=3) == 12 Parameters ---------- inputs: [tf.placeholder, tf.constant, or object with make_feed_dict method] list of input arguments outputs: [tf.Variable] or tf.Variable list of outputs or a single output to be returned from function. Returned value will also have the same shape. """ if isinstance(outputs, list): return _Function(inputs, outputs, updates, givens=givens) elif isinstance(outputs, (dict, collections.OrderedDict)): f = _Function(inputs, outputs.values(), updates, givens=givens) return lambda *args, **kwargs: type(outputs)(zip(outputs.keys(), f(*args, **kwargs))) else: f = _Function(inputs, [outputs], updates, givens=givens) return lambda *args, **kwargs: f(*args, **kwargs)[0] class _Function(object): def __init__(self, inputs, outputs, updates, givens): for inpt in inputs: if not hasattr(inpt, 'make_feed_dict') and not (type(inpt) is tf.Tensor and len(inpt.op.inputs) == 0): assert False, "inputs should all be placeholders, constants, or have a make_feed_dict method" self.inputs = inputs updates = updates or [] self.update_group = tf.group(*updates) self.outputs_update = list(outputs) + [self.update_group] self.givens = {} if givens is None else givens def _feed_input(self, feed_dict, inpt, value): if hasattr(inpt, 'make_feed_dict'): feed_dict.update(inpt.make_feed_dict(value)) else: feed_dict[inpt] = value def __call__(self, *args): assert len(args) <= len(self.inputs), "Too many arguments provided" feed_dict = {} # Update the args for inpt, value in zip(self.inputs, args): self._feed_input(feed_dict, inpt, value) # Update feed dict with givens. for inpt in self.givens: feed_dict[inpt] = feed_dict.get(inpt, self.givens[inpt]) results = tf.get_default_session().run(self.outputs_update, feed_dict=feed_dict)[:-1] return results # ================================================================ # Flat vectors # ================================================================ def var_shape(x): out = x.get_shape().as_list() assert all(isinstance(a, int) for a in out), \ "shape function assumes that shape is fully known" return out def numel(x): return intprod(var_shape(x)) def intprod(x): return int(np.prod(x)) def flatgrad(loss, var_list, clip_norm=None): grads = tf.gradients(loss, var_list) if clip_norm is not None: grads = [tf.clip_by_norm(grad, clip_norm=clip_norm) for grad in grads] return tf.concat(axis=0, values=[ tf.reshape(grad if grad is not None else tf.zeros_like(v), [numel(v)]) for (v, grad) in zip(var_list, grads) ]) class SetFromFlat(object): def __init__(self, var_list, dtype=tf.float32): assigns = [] shapes = list(map(var_shape, var_list)) total_size = np.sum([intprod(shape) for shape in shapes]) self.theta = theta = tf.placeholder(dtype, [total_size]) start = 0 assigns = [] for (shape, v) in zip(shapes, var_list):
self.op = tf.group(*assigns) def __call__(self, theta): tf.get_default_session().run(self.op, feed_dict={self.theta: theta}) class GetFlat(object): def __init__(self, var_list): self.op = tf.concat(axis=0, values=[tf.reshape(v, [numel(v)]) for v in var_list]) def __call__(self): return tf.get_default_session().run(self.op) _PLACEHOLDER_CACHE = {} # name -> (placeholder, dtype, shape) def get_placeholder(name, dtype, shape): if name in _PLACEHOLDER_CACHE: out, dtype1, shape1 = _PLACEHOLDER_CACHE[name] assert dtype1 == dtype and shape1 == shape return out else: out = tf.placeholder(dtype=dtype, shape=shape, name=name) _PLACEHOLDER_CACHE[name] = (out, dtype, shape) return out def get_placeholder_cached(name): return _PLACEHOLDER_CACHE[name][0] def flattenallbut0(x): return tf.reshape(x, [-1, intprod(x.get_shape().as_list()[1:])]) # ================================================================ # Diagnostics # ================================================================ def display_var_info(vars): from baselines import logger count_params = 0 for v in vars: name = v.name if "/Adam" in name or "beta1_power" in name or "beta2_power" in name: continue count_params += np.prod(v.shape.as_list()) if "/b:" in name: continue # Wx+b, bias is not interesting to look at => count params, but not print logger.info(" %s%s%s" % (name, " "*(55-len(name)), str(v.shape))) logger.info("Total model parameters: %0.1f million" % (count_params*1e-6)) class LinearSchedule(object): def __init__(self, schedule_timesteps, final_p, initial_p=1.0): """Linear interpolation between initial_p and final_p over schedule_timesteps. After this many timesteps pass final_p is returned. Parameters ---------- schedule_timesteps: int Number of timesteps for which to linearly anneal initial_p to final_p initial_p: float initial output value final_p: float final output value """ self.schedule_timesteps = schedule_timesteps self.final_p = final_p self.initial_p = initial_p def value(self, t): """See Schedule.value""" fraction = min(float(t) / self.schedule_timesteps, 1.0) return self.initial_p + fraction * (self.final_p - self.initial_p) # deep q import os import tensorflow as tf # ================================================================ # Saving variables # ================================================================ def load_state(fname): saver = tf.train.Saver() saver.restore(tf.get_default_session(), fname) def save_state(fname): os.makedirs(os.path.dirname(fname), exist_ok=True) saver = tf.train.Saver() saver.save(tf.get_default_session(), fname) # ================================================================ # Placeholders # ================================================================ class TfInput(object): def __init__(self, name="(unnamed)"):
size = intprod(shape) assigns.append(tf.assign(v, tf.reshape(theta[start:start + size], shape))) start += size
conditional_block
encoder.rs
::Error> for EncodingError { fn from(err: io::Error) -> EncodingError { EncodingError::IoError(err) } } impl From<EncodingError> for io::Error { fn from(err: EncodingError) -> io::Error { io::Error::new(io::ErrorKind::Other, (&err as &error::Error).description()) } } /// PNG Encoder pub struct Encoder<W: Write> { w: W, info: Info, } impl<W: Write> Encoder<W> { pub fn new(w: W, width: u32, height: u32) -> Encoder<W> { let mut info = Info::default(); info.width = width; info.height = height; Encoder { w: w, info: info } } pub fn new_animated_with_frame_rate(w: W, width: u32, height: u32, frames: u32, delay_num: u16, delay_den: u16) -> Result<Encoder<W>> { let mut enc = Encoder::new_animated(w, width, height, frames)?; let mut frame_ctl = enc.info.frame_control.unwrap(); frame_ctl.delay_num = delay_num; frame_ctl.delay_den = delay_den; enc.info.frame_control = Some(frame_ctl); Ok(enc) } pub fn new_animated(w: W, width: u32, height: u32, frames: u32) -> Result<Encoder<W>> { if frames > 0
else { Err(EncodingError::Format("invalid number of frames for an animated PNG".into())) } } pub fn write_header(self) -> Result<Writer<W>> { Writer::new(self.w, self.info).init() } } impl<W: Write> HasParameters for Encoder<W> {} impl<W: Write> Parameter<Encoder<W>> for ColorType { fn set_param(self, this: &mut Encoder<W>) { this.info.color_type = self } } impl<W: Write> Parameter<Encoder<W>> for BitDepth { fn set_param(self, this: &mut Encoder<W>) { this.info.bit_depth = self } } /// PNG writer pub struct Writer<W: Write> { w: W, info: Info, separate_default_image: bool, } impl<W: Write> Writer<W> { fn new(w: W, info: Info) -> Writer<W> { let w = Writer { w: w, info: info, separate_default_image: false }; w } fn init(mut self) -> Result<Self> { try!(self.w.write(&[137, 80, 78, 71, 13, 10, 26, 10])); let mut data = [0; 13]; try!((&mut data[..]).write_be(self.info.width)); try!((&mut data[4..]).write_be(self.info.height)); data[8] = self.info.bit_depth as u8; data[9] = self.info.color_type as u8; data[12] = if self.info.interlaced { 1 } else { 0 }; try!(self.write_chunk(chunk::IHDR, &data)); match self.info { Info { animation_control: Some(anim_ctl), frame_control: Some(_), ..} => { let mut data = [0; 8]; try!((&mut data[..]).write_be(anim_ctl.num_frames)); try!((&mut data[4..]).write_be(anim_ctl.num_plays)); try!(self.write_chunk(chunk::acTL, &data)); } _ => {} }; Ok(self) } pub fn write_chunk_with_fields(&mut self, name: [u8; 4], data: &[u8], fields: Option<&[u8]>) -> Result<()> { self.w.write_be(data.len() as u32 + (if fields.is_some() { fields.unwrap().len() as u32 } else { 0 }))?; self.w.write(&name)?; if fields.is_some() { try!(self.w.write(fields.unwrap())); } self.w.write(data)?; let mut crc = Crc32::new(); crc.update(&name); if fields.is_some() { crc.update(fields.unwrap()); } crc.update(data); self.w.write_be(crc.checksum())?; Ok(()) } pub fn write_chunk(&mut self, name: [u8; 4], data: &[u8]) -> Result<()> { self.write_chunk_with_fields(name, data, None) } /// Writes the image data. pub fn write_image_data(&mut self, data: &[u8]) -> Result<()> { let zlib = self.get_image_data(data)?; self.write_chunk(chunk::IDAT, &try!(zlib.finish())) } fn get_image_data(&mut self, data: &[u8]) -> Result<deflate::write::ZlibEncoder<Vec<u8>>> { let bpp = self.info.bytes_per_pixel(); let in_len = self.info.raw_row_length() - 1; let mut prev = vec![0; in_len]; let mut current = vec![0; in_len]; let data_size = in_len * self.info.height as usize; if data.len() < data_size || data_size == 0 { return Err(EncodingError::Format("not enough image data provided".into())); } let mut zlib = deflate::write::ZlibEncoder::new(Vec::new(), deflate::Compression::Fast); let filter_method = FilterType::Sub; for line in data.chunks(in_len) { current.copy_from_slice(&line); try!(zlib.write_all(&[filter_method as u8])); filter(filter_method, bpp, &prev, &mut current); try!(zlib.write_all(&current)); mem::swap(&mut prev, &mut current); } Ok(zlib) } pub fn write_separate_default_image(&mut self, data: &[u8]) -> Result<()> { match self.info { Info { animation_control: Some(_), frame_control: Some(frame_control), ..} => { if frame_control.sequence_number != 0 { Err(EncodingError::Format("separate default image provided after frame sequence has begun".into())) } else if self.separate_default_image { Err(EncodingError::Format("default image already written".into())) } else { self.separate_default_image = true; self.write_image_data(data) } } _ => { Err(EncodingError::Format("default image provided for a non-animated PNG".into())) } } } #[allow(non_snake_case)] fn write_fcTL(&mut self) -> Result<()> { let frame_ctl = self.info.frame_control.ok_or(EncodingError::Format("cannot write fcTL for a non-animated PNG".into()))?; let mut data = [0u8; 26]; (&mut data[..]).write_be(frame_ctl.sequence_number)?; (&mut data[4..]).write_be(frame_ctl.width)?; (&mut data[8..]).write_be(frame_ctl.height)?; (&mut data[12..]).write_be(frame_ctl.x_offset)?; (&mut data[16..]).write_be(frame_ctl.y_offset)?; (&mut data[20..]).write_be(frame_ctl.delay_num)?; (&mut data[22..]).write_be(frame_ctl.delay_den)?; data[24] = frame_ctl.dispose_op as u8; data[25] = frame_ctl.blend_op as u8; self.write_chunk(chunk::fcTL, &data) } #[allow(non_snake_case)] fn write_fdAT(&mut self, data: &[u8]) -> Result<()> { // println!("Writing fdAT:{:?}", self.info.frame_control.unwrap().sequence_number+1); let zlib = self.get_image_data(data)?; let mut data = [0u8; 4]; (&mut data[..]).write_be(self.info.frame_control .ok_or(EncodingError::Format("cannot write fdAT for a non-animated PNG".into()))?.sequence_number+1u32)?; self.write_chunk_with_fields(chunk::fdAT, &zlib.finish()?, Some(&data)) } pub fn write_frame(&mut self, data: &[u8]) -> Result<()> { // println!("{:?}", self.info.frame_control.unwrap().sequence_number); match self.info { Info { animation_control: Some(AnimationControl { num_frames: 0, ..}) , frame_control: Some(_), ..} => { Err(EncodingError::Format("exceeded number of frames specified".into())) }, Info { animation_control: Some(anim_ctl), frame_control: Some(frame_control), ..} => { match frame_control.sequence_number { 0 => { let ret = if self.separate_default_image { // If we've already written the default image we can write frames the normal way
{ let mut encoder = Encoder::new(w, width, height); let animation_ctl = AnimationControl { num_frames: frames, num_plays: 0 }; let mut frame_ctl = FrameControl::default(); frame_ctl.width = width; frame_ctl.height = height; encoder.info.animation_control = Some(animation_ctl); encoder.info.frame_control = Some(frame_ctl); Ok(encoder) }
conditional_block
encoder.rs
use self::EncodingError::*; match *self { IoError(ref err) => err.description(), Format(ref desc) => &desc, } } } impl fmt::Display for EncodingError { fn fmt(&self, fmt: &mut fmt::Formatter) -> result::Result<(), fmt::Error> { write!(fmt, "{}", (self as &error::Error).description()) } } impl From<io::Error> for EncodingError { fn from(err: io::Error) -> EncodingError { EncodingError::IoError(err) } } impl From<EncodingError> for io::Error { fn from(err: EncodingError) -> io::Error { io::Error::new(io::ErrorKind::Other, (&err as &error::Error).description()) } } /// PNG Encoder pub struct Encoder<W: Write> { w: W, info: Info, } impl<W: Write> Encoder<W> { pub fn new(w: W, width: u32, height: u32) -> Encoder<W> { let mut info = Info::default(); info.width = width; info.height = height; Encoder { w: w, info: info } } pub fn new_animated_with_frame_rate(w: W, width: u32, height: u32, frames: u32, delay_num: u16, delay_den: u16) -> Result<Encoder<W>> { let mut enc = Encoder::new_animated(w, width, height, frames)?; let mut frame_ctl = enc.info.frame_control.unwrap(); frame_ctl.delay_num = delay_num; frame_ctl.delay_den = delay_den; enc.info.frame_control = Some(frame_ctl); Ok(enc) } pub fn new_animated(w: W, width: u32, height: u32, frames: u32) -> Result<Encoder<W>> { if frames > 0 { let mut encoder = Encoder::new(w, width, height); let animation_ctl = AnimationControl { num_frames: frames, num_plays: 0 }; let mut frame_ctl = FrameControl::default(); frame_ctl.width = width; frame_ctl.height = height; encoder.info.animation_control = Some(animation_ctl); encoder.info.frame_control = Some(frame_ctl); Ok(encoder) } else { Err(EncodingError::Format("invalid number of frames for an animated PNG".into())) } } pub fn write_header(self) -> Result<Writer<W>> { Writer::new(self.w, self.info).init() } } impl<W: Write> HasParameters for Encoder<W> {} impl<W: Write> Parameter<Encoder<W>> for ColorType { fn set_param(self, this: &mut Encoder<W>) { this.info.color_type = self } } impl<W: Write> Parameter<Encoder<W>> for BitDepth { fn set_param(self, this: &mut Encoder<W>) { this.info.bit_depth = self } } /// PNG writer pub struct Writer<W: Write> { w: W, info: Info, separate_default_image: bool, } impl<W: Write> Writer<W> { fn new(w: W, info: Info) -> Writer<W> { let w = Writer { w: w, info: info, separate_default_image: false }; w } fn init(mut self) -> Result<Self> { try!(self.w.write(&[137, 80, 78, 71, 13, 10, 26, 10])); let mut data = [0; 13]; try!((&mut data[..]).write_be(self.info.width)); try!((&mut data[4..]).write_be(self.info.height)); data[8] = self.info.bit_depth as u8; data[9] = self.info.color_type as u8; data[12] = if self.info.interlaced { 1 } else { 0 }; try!(self.write_chunk(chunk::IHDR, &data)); match self.info { Info { animation_control: Some(anim_ctl), frame_control: Some(_), ..} => { let mut data = [0; 8]; try!((&mut data[..]).write_be(anim_ctl.num_frames)); try!((&mut data[4..]).write_be(anim_ctl.num_plays)); try!(self.write_chunk(chunk::acTL, &data)); } _ => {} }; Ok(self) } pub fn write_chunk_with_fields(&mut self, name: [u8; 4], data: &[u8], fields: Option<&[u8]>) -> Result<()> { self.w.write_be(data.len() as u32 + (if fields.is_some() { fields.unwrap().len() as u32 } else { 0 }))?; self.w.write(&name)?; if fields.is_some() { try!(self.w.write(fields.unwrap())); } self.w.write(data)?; let mut crc = Crc32::new(); crc.update(&name); if fields.is_some() { crc.update(fields.unwrap()); } crc.update(data); self.w.write_be(crc.checksum())?; Ok(()) } pub fn write_chunk(&mut self, name: [u8; 4], data: &[u8]) -> Result<()> { self.write_chunk_with_fields(name, data, None) } /// Writes the image data. pub fn write_image_data(&mut self, data: &[u8]) -> Result<()> { let zlib = self.get_image_data(data)?; self.write_chunk(chunk::IDAT, &try!(zlib.finish())) } fn get_image_data(&mut self, data: &[u8]) -> Result<deflate::write::ZlibEncoder<Vec<u8>>> { let bpp = self.info.bytes_per_pixel(); let in_len = self.info.raw_row_length() - 1; let mut prev = vec![0; in_len]; let mut current = vec![0; in_len]; let data_size = in_len * self.info.height as usize; if data.len() < data_size || data_size == 0 { return Err(EncodingError::Format("not enough image data provided".into())); } let mut zlib = deflate::write::ZlibEncoder::new(Vec::new(), deflate::Compression::Fast); let filter_method = FilterType::Sub; for line in data.chunks(in_len) { current.copy_from_slice(&line); try!(zlib.write_all(&[filter_method as u8])); filter(filter_method, bpp, &prev, &mut current); try!(zlib.write_all(&current)); mem::swap(&mut prev, &mut current); } Ok(zlib) } pub fn write_separate_default_image(&mut self, data: &[u8]) -> Result<()> { match self.info { Info { animation_control: Some(_), frame_control: Some(frame_control), ..} => { if frame_control.sequence_number != 0 { Err(EncodingError::Format("separate default image provided after frame sequence has begun".into())) } else if self.separate_default_image { Err(EncodingError::Format("default image already written".into())) } else { self.separate_default_image = true; self.write_image_data(data) } } _ => { Err(EncodingError::Format("default image provided for a non-animated PNG".into())) } } } #[allow(non_snake_case)] fn write_fcTL(&mut self) -> Result<()> { let frame_ctl = self.info.frame_control.ok_or(EncodingError::Format("cannot write fcTL for a non-animated PNG".into()))?; let mut data = [0u8; 26]; (&mut data[..]).write_be(frame_ctl.sequence_number)?; (&mut data[4..]).write_be(frame_ctl.width)?; (&mut data[8..]).write_be(frame_ctl.height)?; (&mut data[12..]).write_be(frame_ctl.x_offset)?; (&mut data[16..]).write_be(frame_ctl.y_offset)?; (&mut data[20..]).write_be(frame_ctl.delay_num)?; (&mut data[22..]).write_be(frame_ctl.delay_den)?; data[24] = frame_ctl.dispose_op as u8; data[25] = frame_ctl.blend_op as u8; self.write_chunk(chunk::fcTL, &data) } #[allow(non_snake_case)] fn write_fdAT(&mut self, data: &[u8]) -> Result<()> { // println!("Writing fdAT:{:?}", self.info.frame_control.unwrap().sequence_number+1); let zlib = self.get_image_data(data)?; let mut data = [0u8; 4]; (&mut data[..]).write_be(self.info.frame_control .ok_or(EncodingError::Format("cannot write fdAT for a non-animated PNG".into()))?.sequence_number+1u32)?; self.write_chunk_with_fields(chunk::fdAT, &zlib.finish()?, Some(&data)) } pub fn write_frame(&mut self, data: &[u8]) -> Result<()> { // println!("{:?}", self.info.frame_control.unwrap().sequence_number); match self.info { Info { animation_control: Some(AnimationControl
fn description(&self) -> &str {
random_line_split
encoder.rs
>> { if frames > 0 { let mut encoder = Encoder::new(w, width, height); let animation_ctl = AnimationControl { num_frames: frames, num_plays: 0 }; let mut frame_ctl = FrameControl::default(); frame_ctl.width = width; frame_ctl.height = height; encoder.info.animation_control = Some(animation_ctl); encoder.info.frame_control = Some(frame_ctl); Ok(encoder) } else { Err(EncodingError::Format("invalid number of frames for an animated PNG".into())) } } pub fn write_header(self) -> Result<Writer<W>> { Writer::new(self.w, self.info).init() } } impl<W: Write> HasParameters for Encoder<W> {} impl<W: Write> Parameter<Encoder<W>> for ColorType { fn set_param(self, this: &mut Encoder<W>) { this.info.color_type = self } } impl<W: Write> Parameter<Encoder<W>> for BitDepth { fn set_param(self, this: &mut Encoder<W>) { this.info.bit_depth = self } } /// PNG writer pub struct Writer<W: Write> { w: W, info: Info, separate_default_image: bool, } impl<W: Write> Writer<W> { fn new(w: W, info: Info) -> Writer<W> { let w = Writer { w: w, info: info, separate_default_image: false }; w } fn init(mut self) -> Result<Self> { try!(self.w.write(&[137, 80, 78, 71, 13, 10, 26, 10])); let mut data = [0; 13]; try!((&mut data[..]).write_be(self.info.width)); try!((&mut data[4..]).write_be(self.info.height)); data[8] = self.info.bit_depth as u8; data[9] = self.info.color_type as u8; data[12] = if self.info.interlaced { 1 } else { 0 }; try!(self.write_chunk(chunk::IHDR, &data)); match self.info { Info { animation_control: Some(anim_ctl), frame_control: Some(_), ..} => { let mut data = [0; 8]; try!((&mut data[..]).write_be(anim_ctl.num_frames)); try!((&mut data[4..]).write_be(anim_ctl.num_plays)); try!(self.write_chunk(chunk::acTL, &data)); } _ => {} }; Ok(self) } pub fn write_chunk_with_fields(&mut self, name: [u8; 4], data: &[u8], fields: Option<&[u8]>) -> Result<()> { self.w.write_be(data.len() as u32 + (if fields.is_some() { fields.unwrap().len() as u32 } else { 0 }))?; self.w.write(&name)?; if fields.is_some() { try!(self.w.write(fields.unwrap())); } self.w.write(data)?; let mut crc = Crc32::new(); crc.update(&name); if fields.is_some() { crc.update(fields.unwrap()); } crc.update(data); self.w.write_be(crc.checksum())?; Ok(()) } pub fn write_chunk(&mut self, name: [u8; 4], data: &[u8]) -> Result<()> { self.write_chunk_with_fields(name, data, None) } /// Writes the image data. pub fn write_image_data(&mut self, data: &[u8]) -> Result<()> { let zlib = self.get_image_data(data)?; self.write_chunk(chunk::IDAT, &try!(zlib.finish())) } fn get_image_data(&mut self, data: &[u8]) -> Result<deflate::write::ZlibEncoder<Vec<u8>>> { let bpp = self.info.bytes_per_pixel(); let in_len = self.info.raw_row_length() - 1; let mut prev = vec![0; in_len]; let mut current = vec![0; in_len]; let data_size = in_len * self.info.height as usize; if data.len() < data_size || data_size == 0 { return Err(EncodingError::Format("not enough image data provided".into())); } let mut zlib = deflate::write::ZlibEncoder::new(Vec::new(), deflate::Compression::Fast); let filter_method = FilterType::Sub; for line in data.chunks(in_len) { current.copy_from_slice(&line); try!(zlib.write_all(&[filter_method as u8])); filter(filter_method, bpp, &prev, &mut current); try!(zlib.write_all(&current)); mem::swap(&mut prev, &mut current); } Ok(zlib) } pub fn write_separate_default_image(&mut self, data: &[u8]) -> Result<()> { match self.info { Info { animation_control: Some(_), frame_control: Some(frame_control), ..} => { if frame_control.sequence_number != 0 { Err(EncodingError::Format("separate default image provided after frame sequence has begun".into())) } else if self.separate_default_image { Err(EncodingError::Format("default image already written".into())) } else { self.separate_default_image = true; self.write_image_data(data) } } _ => { Err(EncodingError::Format("default image provided for a non-animated PNG".into())) } } } #[allow(non_snake_case)] fn write_fcTL(&mut self) -> Result<()> { let frame_ctl = self.info.frame_control.ok_or(EncodingError::Format("cannot write fcTL for a non-animated PNG".into()))?; let mut data = [0u8; 26]; (&mut data[..]).write_be(frame_ctl.sequence_number)?; (&mut data[4..]).write_be(frame_ctl.width)?; (&mut data[8..]).write_be(frame_ctl.height)?; (&mut data[12..]).write_be(frame_ctl.x_offset)?; (&mut data[16..]).write_be(frame_ctl.y_offset)?; (&mut data[20..]).write_be(frame_ctl.delay_num)?; (&mut data[22..]).write_be(frame_ctl.delay_den)?; data[24] = frame_ctl.dispose_op as u8; data[25] = frame_ctl.blend_op as u8; self.write_chunk(chunk::fcTL, &data) } #[allow(non_snake_case)] fn write_fdAT(&mut self, data: &[u8]) -> Result<()> { // println!("Writing fdAT:{:?}", self.info.frame_control.unwrap().sequence_number+1); let zlib = self.get_image_data(data)?; let mut data = [0u8; 4]; (&mut data[..]).write_be(self.info.frame_control .ok_or(EncodingError::Format("cannot write fdAT for a non-animated PNG".into()))?.sequence_number+1u32)?; self.write_chunk_with_fields(chunk::fdAT, &zlib.finish()?, Some(&data)) } pub fn write_frame(&mut self, data: &[u8]) -> Result<()> { // println!("{:?}", self.info.frame_control.unwrap().sequence_number); match self.info { Info { animation_control: Some(AnimationControl { num_frames: 0, ..}) , frame_control: Some(_), ..} => { Err(EncodingError::Format("exceeded number of frames specified".into())) }, Info { animation_control: Some(anim_ctl), frame_control: Some(frame_control), ..} => { match frame_control.sequence_number { 0 => { let ret = if self.separate_default_image { // If we've already written the default image we can write frames the normal way // fcTL + fdAT self.write_fcTL().and(self.write_fdAT(data)) } else { // If not we'll have to set the first frame to be both: // fcTL + first frame (IDAT) self.write_fcTL().and(self.write_image_data(data)) }; let mut fc = self.info.frame_control.unwrap(); fc.inc_seq_num(1); self.info.frame_control = Some(fc); ret }, x if x == 2 * anim_ctl.num_frames - 1 => { // println!("We're done, boss"); // This is the last frame: // Do the usual and also set AnimationControl to no remaining frames: let ret = self.write_fcTL().and(self.write_fdAT(data)); let mut fc = self.info.frame_control.unwrap(); fc.set_seq_num(0); self.info.frame_control = Some(fc); ret }, _ => { // Usual case: // fcTL + fdAT // println!("Buisness as usual"); let ret = self.write_fcTL().and(self.write_fdAT(data)); let mut fc = self.info.frame_control.unwrap(); fc.inc_seq_num(2); self.info.frame_control = Some(fc); ret } } }, _ => { Err(EncodingError::Format("frame provided for a non-animated PNG".into())) } } } } impl<W: Write> Drop for Writer<W> { fn
drop
identifier_name
encoder.rs
::Error> for EncodingError { fn from(err: io::Error) -> EncodingError { EncodingError::IoError(err) } } impl From<EncodingError> for io::Error { fn from(err: EncodingError) -> io::Error
} /// PNG Encoder pub struct Encoder<W: Write> { w: W, info: Info, } impl<W: Write> Encoder<W> { pub fn new(w: W, width: u32, height: u32) -> Encoder<W> { let mut info = Info::default(); info.width = width; info.height = height; Encoder { w: w, info: info } } pub fn new_animated_with_frame_rate(w: W, width: u32, height: u32, frames: u32, delay_num: u16, delay_den: u16) -> Result<Encoder<W>> { let mut enc = Encoder::new_animated(w, width, height, frames)?; let mut frame_ctl = enc.info.frame_control.unwrap(); frame_ctl.delay_num = delay_num; frame_ctl.delay_den = delay_den; enc.info.frame_control = Some(frame_ctl); Ok(enc) } pub fn new_animated(w: W, width: u32, height: u32, frames: u32) -> Result<Encoder<W>> { if frames > 0 { let mut encoder = Encoder::new(w, width, height); let animation_ctl = AnimationControl { num_frames: frames, num_plays: 0 }; let mut frame_ctl = FrameControl::default(); frame_ctl.width = width; frame_ctl.height = height; encoder.info.animation_control = Some(animation_ctl); encoder.info.frame_control = Some(frame_ctl); Ok(encoder) } else { Err(EncodingError::Format("invalid number of frames for an animated PNG".into())) } } pub fn write_header(self) -> Result<Writer<W>> { Writer::new(self.w, self.info).init() } } impl<W: Write> HasParameters for Encoder<W> {} impl<W: Write> Parameter<Encoder<W>> for ColorType { fn set_param(self, this: &mut Encoder<W>) { this.info.color_type = self } } impl<W: Write> Parameter<Encoder<W>> for BitDepth { fn set_param(self, this: &mut Encoder<W>) { this.info.bit_depth = self } } /// PNG writer pub struct Writer<W: Write> { w: W, info: Info, separate_default_image: bool, } impl<W: Write> Writer<W> { fn new(w: W, info: Info) -> Writer<W> { let w = Writer { w: w, info: info, separate_default_image: false }; w } fn init(mut self) -> Result<Self> { try!(self.w.write(&[137, 80, 78, 71, 13, 10, 26, 10])); let mut data = [0; 13]; try!((&mut data[..]).write_be(self.info.width)); try!((&mut data[4..]).write_be(self.info.height)); data[8] = self.info.bit_depth as u8; data[9] = self.info.color_type as u8; data[12] = if self.info.interlaced { 1 } else { 0 }; try!(self.write_chunk(chunk::IHDR, &data)); match self.info { Info { animation_control: Some(anim_ctl), frame_control: Some(_), ..} => { let mut data = [0; 8]; try!((&mut data[..]).write_be(anim_ctl.num_frames)); try!((&mut data[4..]).write_be(anim_ctl.num_plays)); try!(self.write_chunk(chunk::acTL, &data)); } _ => {} }; Ok(self) } pub fn write_chunk_with_fields(&mut self, name: [u8; 4], data: &[u8], fields: Option<&[u8]>) -> Result<()> { self.w.write_be(data.len() as u32 + (if fields.is_some() { fields.unwrap().len() as u32 } else { 0 }))?; self.w.write(&name)?; if fields.is_some() { try!(self.w.write(fields.unwrap())); } self.w.write(data)?; let mut crc = Crc32::new(); crc.update(&name); if fields.is_some() { crc.update(fields.unwrap()); } crc.update(data); self.w.write_be(crc.checksum())?; Ok(()) } pub fn write_chunk(&mut self, name: [u8; 4], data: &[u8]) -> Result<()> { self.write_chunk_with_fields(name, data, None) } /// Writes the image data. pub fn write_image_data(&mut self, data: &[u8]) -> Result<()> { let zlib = self.get_image_data(data)?; self.write_chunk(chunk::IDAT, &try!(zlib.finish())) } fn get_image_data(&mut self, data: &[u8]) -> Result<deflate::write::ZlibEncoder<Vec<u8>>> { let bpp = self.info.bytes_per_pixel(); let in_len = self.info.raw_row_length() - 1; let mut prev = vec![0; in_len]; let mut current = vec![0; in_len]; let data_size = in_len * self.info.height as usize; if data.len() < data_size || data_size == 0 { return Err(EncodingError::Format("not enough image data provided".into())); } let mut zlib = deflate::write::ZlibEncoder::new(Vec::new(), deflate::Compression::Fast); let filter_method = FilterType::Sub; for line in data.chunks(in_len) { current.copy_from_slice(&line); try!(zlib.write_all(&[filter_method as u8])); filter(filter_method, bpp, &prev, &mut current); try!(zlib.write_all(&current)); mem::swap(&mut prev, &mut current); } Ok(zlib) } pub fn write_separate_default_image(&mut self, data: &[u8]) -> Result<()> { match self.info { Info { animation_control: Some(_), frame_control: Some(frame_control), ..} => { if frame_control.sequence_number != 0 { Err(EncodingError::Format("separate default image provided after frame sequence has begun".into())) } else if self.separate_default_image { Err(EncodingError::Format("default image already written".into())) } else { self.separate_default_image = true; self.write_image_data(data) } } _ => { Err(EncodingError::Format("default image provided for a non-animated PNG".into())) } } } #[allow(non_snake_case)] fn write_fcTL(&mut self) -> Result<()> { let frame_ctl = self.info.frame_control.ok_or(EncodingError::Format("cannot write fcTL for a non-animated PNG".into()))?; let mut data = [0u8; 26]; (&mut data[..]).write_be(frame_ctl.sequence_number)?; (&mut data[4..]).write_be(frame_ctl.width)?; (&mut data[8..]).write_be(frame_ctl.height)?; (&mut data[12..]).write_be(frame_ctl.x_offset)?; (&mut data[16..]).write_be(frame_ctl.y_offset)?; (&mut data[20..]).write_be(frame_ctl.delay_num)?; (&mut data[22..]).write_be(frame_ctl.delay_den)?; data[24] = frame_ctl.dispose_op as u8; data[25] = frame_ctl.blend_op as u8; self.write_chunk(chunk::fcTL, &data) } #[allow(non_snake_case)] fn write_fdAT(&mut self, data: &[u8]) -> Result<()> { // println!("Writing fdAT:{:?}", self.info.frame_control.unwrap().sequence_number+1); let zlib = self.get_image_data(data)?; let mut data = [0u8; 4]; (&mut data[..]).write_be(self.info.frame_control .ok_or(EncodingError::Format("cannot write fdAT for a non-animated PNG".into()))?.sequence_number+1u32)?; self.write_chunk_with_fields(chunk::fdAT, &zlib.finish()?, Some(&data)) } pub fn write_frame(&mut self, data: &[u8]) -> Result<()> { // println!("{:?}", self.info.frame_control.unwrap().sequence_number); match self.info { Info { animation_control: Some(AnimationControl { num_frames: 0, ..}) , frame_control: Some(_), ..} => { Err(EncodingError::Format("exceeded number of frames specified".into())) }, Info { animation_control: Some(anim_ctl), frame_control: Some(frame_control), ..} => { match frame_control.sequence_number { 0 => { let ret = if self.separate_default_image { // If we've already written the default image we can write frames the normal way
{ io::Error::new(io::ErrorKind::Other, (&err as &error::Error).description()) }
identifier_body
conn.rs
Ok(t) => t, Err(e) => return Err(e) }; Ok(Connection { addr: addr, host: host, sock: Some(Plain(sock)), name: name, term: t }) } pub fn status(&mut self) { self.send_handshake(false); // Send the status request self.write_packet(Packet::new_out(0x0)); // and read back the response let (packet_id, mut packet) = self.read_packet(); // Make sure we got the right response assert_eq!(packet_id, 0x0); // Get the JSON let json = ExtraJSON::new(json::from_str(packet.read_string()).unwrap()); println!("Minecraft Server Status [{}:{}]", self.host, self.addr.port); println!("Version: {}", json["version"]["name"].string()); println!("Protocol: {}", json["version"]["protocol"].as_int()); println!("Description: {}", json["description"].string()); println!("Players: ({}/{})", json["players"]["online"].as_int(), json["players"]["max"].as_int()); let players = json["players"]["sample"].list(); for player in players.iter() { println!("\t{} ({})", player["name"].string(), player["id"].string()); } } pub fn run(mut self) { // If the server is in online-mode // we need to do authentication and // enable encryption self.login(); // Get a port to read messages from stdin let msgs = self.read_messages(); // Yay, all good. // Now we just loop and read in all the packets we can // We don't actually do anything for most of them except // for chat and keep alives. loop { // Got a message in the queue to send? 'msg: loop { match msgs.try_recv() { comm::Data(msg) => { if msg.is_empty() { continue; } else if msg.len() > 100 { println!("Message too long."); continue; } // Send the message! let mut p = Packet::new_out(0x1); p.write_string(msg); self.write_packet(p); } comm::Empty => break 'msg, comm::Disconnected => fail!("input stream disconnected") } } // Read in and handle a packet let (packet_id, mut packet) = self.read_packet(); self.handle_message(packet_id, &mut packet); } } fn handle_message(&mut self, packet_id: i32, packet: &mut packet::InPacket) { // Keep Alive if packet_id == 0x0
else if packet_id == 0x2 { let json = packet.read_string(); debug!("Got chat message: {}", json); // Let's wrap up the Json so that we can // deal with it more easily let j = json::from_str(json).unwrap(); let j = ExtraJSON::new(j); let ty = j["translate"].string(); // Player Chat if "chat.type.text" == ty { let user = j["with"][0]["text"].string(); let msg = j["with"][1].string(); self.term.attr(term::attr::ForegroundColor(term::color::BRIGHT_GREEN)); write!(&mut self.term as &mut Writer, "<{}> ", user); self.term.reset(); self.term.write(msg.as_bytes()); self.term.write(bytes!("\n")); // Server Message } else if "chat.type.announcement" == ty { let msg = j["with"][1]["extra"].list_map(|x| x.string()).concat(); self.term.attr(term::attr::ForegroundColor(term::color::BRIGHT_YELLOW)); self.term.write(bytes!("[Server] ")); self.term.reset(); self.term.write(msg.as_bytes()); self.term.write(bytes!("\n")); } } } fn login(&mut self) { self.send_handshake(true); self.send_username(); // Read the next packet and find out whether we need // to do authentication and encryption let (mut packet_id, mut packet) = self.read_packet(); debug!("Packet ID: {}", packet_id); if packet_id == 0x1 { // Encryption Request // online-mode = true self.enable_encryption(&mut packet); // Read the next packet... let (pi, p) = self.read_packet(); packet_id = pi; packet = p; } if packet_id == 0x0 { // Disconnect let reason = packet.read_string(); debug!("Reason: {}", reason); fail!("Received disconnect."); } // Login Success assert_eq!(packet_id, 0x2); let uuid = packet.read_string(); let username = packet.read_string(); debug!("UUID: {}", uuid); debug!("Username: {}", username); } fn enable_encryption(&mut self, packet: &mut packet::InPacket) { // Get all the data from the Encryption Request packet let server_id = packet.read_string(); let key_len = packet.read_be_i16().unwrap(); let public_key = packet.read_exact(key_len as uint).unwrap(); let token_len = packet.read_be_i16().unwrap(); let verify_token = packet.read_exact(token_len as uint).unwrap(); // Server's public key let pk = crypto::RSAPublicKey::from_bytes(public_key.as_slice()).unwrap(); // Generate random 16 byte key let mut key = [0u8, ..16]; rand::task_rng().fill_bytes(key); // Encrypt shared secret with server's public key let ekey = pk.encrypt(key).unwrap(); // Encrypt verify token with server's public key let etoken = pk.encrypt(verify_token.as_slice()).unwrap(); // Generate the server id hash let mut sha1 = crypto::SHA1::new(); sha1.update(server_id.as_bytes()); sha1.update(key); sha1.update(public_key.as_slice()); let hash = sha1.special_digest(); debug!("Hash: {}", hash); // Do client auth self.authenticate(hash); // Create Encryption Response Packet let mut erp = Packet::new_out(0x1); // Write encrypted shared secret erp.write_be_i16(ekey.len() as i16); erp.write(ekey); // Write encrypted verify token erp.write_be_i16(etoken.len() as i16); erp.write(etoken); // Send self.write_packet(erp); // Create AES cipher with shared secret let aes = crypto::AES::new(key.to_owned(), key.to_owned()).unwrap(); // Get the plain TCP stream let sock = match self.sock.take_unwrap() { Plain(s) => s, _ => fail!("Expected plain socket!") }; // and wwrap it in an AES Stream let sock = crypto::AesStream::new(sock, aes); // and put the new encrypted stream back // everything form this point is encrypted self.sock = Some(Encrypted(sock)); } fn authenticate(&mut self, hash: ~str) { let url = ~"https://authserver.mojang.com/authenticate"; let c = process::ProcessConfig { program: "/usr/bin/curl", args: &[~"-d", ~"@-", ~"-H", ~"Content-Type:application/json", url], env: None, cwd: None, stdin: process::CreatePipe(true, false), stdout: process::CreatePipe(false, true), .. process::ProcessConfig::new() }; let mut p = process::Process::configure(c).unwrap(); // write json to stdin and close it write!(p.stdin.get_mut_ref() as &mut Writer, r#" \{ "agent": \{ "name": "Minecraft", "version": 1 \}, "username": "{}", "password": "{}" \}"#, "USER", "PASS"); // XXX: Don't hardcode these... p.stdin = None; // read response let out = p.wait_with_output().output; let out = str::from_utf8_owned(out.move_iter().collect()).unwrap(); debug!("Got - {}", out); let json = ExtraJSON::new(json::from_str(out).unwrap()); let token = json["accessToken"].string(); let profile = json["selectedProfile"]["id"].string(); let url = ~"https://sessionserver.mojang.com/session/minecraft/join"; let c = process::ProcessConfig { program: "/usr/bin/curl", args: &[~"-d", ~"@-", ~"-H", ~"Content-Type:application/json", url], env: None, cwd: None, stdin: process::CreatePipe(true, false), stdout: process::CreatePipe(false, true), .. process::ProcessConfig::new() }; let
{ let x = packet.read_be_i32().unwrap(); // Need to respond let mut resp = Packet::new_out(0x0); resp.write_be_i32(x); self.write_packet(resp); // Chat Message }
conditional_block
conn.rs
"].list_map(|x| x.string()).concat(); self.term.attr(term::attr::ForegroundColor(term::color::BRIGHT_YELLOW)); self.term.write(bytes!("[Server] ")); self.term.reset(); self.term.write(msg.as_bytes()); self.term.write(bytes!("\n")); } } } fn login(&mut self) { self.send_handshake(true); self.send_username(); // Read the next packet and find out whether we need // to do authentication and encryption let (mut packet_id, mut packet) = self.read_packet(); debug!("Packet ID: {}", packet_id); if packet_id == 0x1 { // Encryption Request // online-mode = true self.enable_encryption(&mut packet); // Read the next packet... let (pi, p) = self.read_packet(); packet_id = pi; packet = p; } if packet_id == 0x0 { // Disconnect let reason = packet.read_string(); debug!("Reason: {}", reason); fail!("Received disconnect."); } // Login Success assert_eq!(packet_id, 0x2); let uuid = packet.read_string(); let username = packet.read_string(); debug!("UUID: {}", uuid); debug!("Username: {}", username); } fn enable_encryption(&mut self, packet: &mut packet::InPacket) { // Get all the data from the Encryption Request packet let server_id = packet.read_string(); let key_len = packet.read_be_i16().unwrap(); let public_key = packet.read_exact(key_len as uint).unwrap(); let token_len = packet.read_be_i16().unwrap(); let verify_token = packet.read_exact(token_len as uint).unwrap(); // Server's public key let pk = crypto::RSAPublicKey::from_bytes(public_key.as_slice()).unwrap(); // Generate random 16 byte key let mut key = [0u8, ..16]; rand::task_rng().fill_bytes(key); // Encrypt shared secret with server's public key let ekey = pk.encrypt(key).unwrap(); // Encrypt verify token with server's public key let etoken = pk.encrypt(verify_token.as_slice()).unwrap(); // Generate the server id hash let mut sha1 = crypto::SHA1::new(); sha1.update(server_id.as_bytes()); sha1.update(key); sha1.update(public_key.as_slice()); let hash = sha1.special_digest(); debug!("Hash: {}", hash); // Do client auth self.authenticate(hash); // Create Encryption Response Packet let mut erp = Packet::new_out(0x1); // Write encrypted shared secret erp.write_be_i16(ekey.len() as i16); erp.write(ekey); // Write encrypted verify token erp.write_be_i16(etoken.len() as i16); erp.write(etoken); // Send self.write_packet(erp); // Create AES cipher with shared secret let aes = crypto::AES::new(key.to_owned(), key.to_owned()).unwrap(); // Get the plain TCP stream let sock = match self.sock.take_unwrap() { Plain(s) => s, _ => fail!("Expected plain socket!") }; // and wwrap it in an AES Stream let sock = crypto::AesStream::new(sock, aes); // and put the new encrypted stream back // everything form this point is encrypted self.sock = Some(Encrypted(sock)); } fn authenticate(&mut self, hash: ~str) { let url = ~"https://authserver.mojang.com/authenticate"; let c = process::ProcessConfig { program: "/usr/bin/curl", args: &[~"-d", ~"@-", ~"-H", ~"Content-Type:application/json", url], env: None, cwd: None, stdin: process::CreatePipe(true, false), stdout: process::CreatePipe(false, true), .. process::ProcessConfig::new() }; let mut p = process::Process::configure(c).unwrap(); // write json to stdin and close it write!(p.stdin.get_mut_ref() as &mut Writer, r#" \{ "agent": \{ "name": "Minecraft", "version": 1 \}, "username": "{}", "password": "{}" \}"#, "USER", "PASS"); // XXX: Don't hardcode these... p.stdin = None; // read response let out = p.wait_with_output().output; let out = str::from_utf8_owned(out.move_iter().collect()).unwrap(); debug!("Got - {}", out); let json = ExtraJSON::new(json::from_str(out).unwrap()); let token = json["accessToken"].string(); let profile = json["selectedProfile"]["id"].string(); let url = ~"https://sessionserver.mojang.com/session/minecraft/join"; let c = process::ProcessConfig { program: "/usr/bin/curl", args: &[~"-d", ~"@-", ~"-H", ~"Content-Type:application/json", url], env: None, cwd: None, stdin: process::CreatePipe(true, false), stdout: process::CreatePipe(false, true), .. process::ProcessConfig::new() }; let mut p = process::Process::configure(c).unwrap(); // write json to stdin and close it write!(p.stdin.get_mut_ref() as &mut Writer, r#" \{ "accessToken": "{}", "selectedProfile": "{}", "serverId": "{}" \}"#, token, profile, hash); p.stdin = None; // read response let out = p.wait_with_output().output; let out = str::from_utf8_owned(out.move_iter().collect()).unwrap(); debug!("Got - {}", out); } fn read_messages(&self) -> Receiver<~str> { let (chan, port) = comm::channel(); spawn(proc() { println!("Type message and then [ENTER] to send:"); let mut stdin = BufferedReader::new(io::stdin()); for line in stdin.lines() { chan.send(line.unwrap().trim().to_owned()); } }); port } fn write_packet(&mut self, p: packet::OutPacket) { // Get the actual buffer let buf = p.buf(); // Write out the packet length self.sock.write_varint(buf.len() as i32); // and the actual payload self.sock.write(buf.as_slice()); } fn read_packet(&mut self) -> (i32, packet::InPacket) { // Read the packet length let len = self.sock.read_varint(); // Now the payload let buf = self.sock.read_exact(len as uint).unwrap(); let mut p = Packet::new_in(buf); // Get the packet id let id = p.read_varint(); (id, p) } fn send_handshake(&mut self, login: bool) { let mut p = Packet::new_out(0x0); // Protocol Version p.write_varint(4); // Server host p.write_string(self.host); // Server port p.write_be_u16(self.addr.port); // State // 1 - status, 2 - login p.write_varint(if login { 2 } else { 1 }); self.write_packet(p); } fn send_username(&mut self) { let mut p = Packet::new_out(0x0); p.write_string(self.name); self.write_packet(p); } } impl Reader for Sock { fn read(&mut self, buf: &mut [u8]) -> io::IoResult<uint> { match *self { Plain(ref mut s) => s.read(buf), Encrypted(ref mut s) => s.read(buf) } } } impl Writer for Sock { fn write(&mut self, buf: &[u8]) -> io::IoResult<()> { match *self { Plain(ref mut s) => s.write(buf), Encrypted(ref mut s) => s.write(buf) } } fn flush(&mut self) -> io::IoResult<()> { match *self { Plain(ref mut s) => s.flush(), Encrypted(ref mut s) => s.flush() } } } impl Reader for Option<Sock> { fn read(&mut self, buf: &mut [u8]) -> io::IoResult<uint> { match *self { Some(ref mut s) => s.read(buf), None => Err(io::standard_error(io::OtherIoError)) } } } impl Writer for Option<Sock> { fn write(&mut self, buf: &[u8]) -> io::IoResult<()> { match *self { Some(ref mut s) => s.write(buf), None => Err(io::standard_error(io::OtherIoError)) } } fn flush(&mut self) -> io::IoResult<()>
{ match *self { Some(ref mut s) => s.flush(), None => Err(io::standard_error(io::OtherIoError)) } }
identifier_body
conn.rs
{ addr: SocketAddr, host: ~str, sock: Option<Sock>, name: ~str, term: term::Terminal<LineBufferedWriter<StdWriter>> } impl Connection { pub fn new(name: ~str, host: ~str, port: u16) -> Result<Connection, ~str> { // Resolve host let addr = match addrinfo::get_host_addresses(host) { Ok(a) => a[0], Err(e) => return Err(e.to_str()) }; let addr = SocketAddr { ip: addr, port: port }; debug!("Connecting to server at {}.", addr.to_str()); let sock = TcpStream::connect(addr); let sock = match sock { Ok(s) => s, Err(e) => return Err(format!("{} - {}", e.kind.to_str(), e.desc)) }; debug!("Successfully connected to server."); let t = match term::Terminal::new(io::stdout()) { Ok(t) => t, Err(e) => return Err(e) }; Ok(Connection { addr: addr, host: host, sock: Some(Plain(sock)), name: name, term: t }) } pub fn status(&mut self) { self.send_handshake(false); // Send the status request self.write_packet(Packet::new_out(0x0)); // and read back the response let (packet_id, mut packet) = self.read_packet(); // Make sure we got the right response assert_eq!(packet_id, 0x0); // Get the JSON let json = ExtraJSON::new(json::from_str(packet.read_string()).unwrap()); println!("Minecraft Server Status [{}:{}]", self.host, self.addr.port); println!("Version: {}", json["version"]["name"].string()); println!("Protocol: {}", json["version"]["protocol"].as_int()); println!("Description: {}", json["description"].string()); println!("Players: ({}/{})", json["players"]["online"].as_int(), json["players"]["max"].as_int()); let players = json["players"]["sample"].list(); for player in players.iter() { println!("\t{} ({})", player["name"].string(), player["id"].string()); } } pub fn run(mut self) { // If the server is in online-mode // we need to do authentication and // enable encryption self.login(); // Get a port to read messages from stdin let msgs = self.read_messages(); // Yay, all good. // Now we just loop and read in all the packets we can // We don't actually do anything for most of them except // for chat and keep alives. loop { // Got a message in the queue to send? 'msg: loop { match msgs.try_recv() { comm::Data(msg) => { if msg.is_empty() { continue; } else if msg.len() > 100 { println!("Message too long."); continue; } // Send the message! let mut p = Packet::new_out(0x1); p.write_string(msg); self.write_packet(p); } comm::Empty => break 'msg, comm::Disconnected => fail!("input stream disconnected") } } // Read in and handle a packet let (packet_id, mut packet) = self.read_packet(); self.handle_message(packet_id, &mut packet); } } fn handle_message(&mut self, packet_id: i32, packet: &mut packet::InPacket) { // Keep Alive if packet_id == 0x0 { let x = packet.read_be_i32().unwrap(); // Need to respond let mut resp = Packet::new_out(0x0); resp.write_be_i32(x); self.write_packet(resp); // Chat Message } else if packet_id == 0x2 { let json = packet.read_string(); debug!("Got chat message: {}", json); // Let's wrap up the Json so that we can // deal with it more easily let j = json::from_str(json).unwrap(); let j = ExtraJSON::new(j); let ty = j["translate"].string(); // Player Chat if "chat.type.text" == ty { let user = j["with"][0]["text"].string(); let msg = j["with"][1].string(); self.term.attr(term::attr::ForegroundColor(term::color::BRIGHT_GREEN)); write!(&mut self.term as &mut Writer, "<{}> ", user); self.term.reset(); self.term.write(msg.as_bytes()); self.term.write(bytes!("\n")); // Server Message } else if "chat.type.announcement" == ty { let msg = j["with"][1]["extra"].list_map(|x| x.string()).concat(); self.term.attr(term::attr::ForegroundColor(term::color::BRIGHT_YELLOW)); self.term.write(bytes!("[Server] ")); self.term.reset(); self.term.write(msg.as_bytes()); self.term.write(bytes!("\n")); } } } fn login(&mut self) { self.send_handshake(true); self.send_username(); // Read the next packet and find out whether we need // to do authentication and encryption let (mut packet_id, mut packet) = self.read_packet(); debug!("Packet ID: {}", packet_id); if packet_id == 0x1 { // Encryption Request // online-mode = true self.enable_encryption(&mut packet); // Read the next packet... let (pi, p) = self.read_packet(); packet_id = pi; packet = p; } if packet_id == 0x0 { // Disconnect let reason = packet.read_string(); debug!("Reason: {}", reason); fail!("Received disconnect."); } // Login Success assert_eq!(packet_id, 0x2); let uuid = packet.read_string(); let username = packet.read_string(); debug!("UUID: {}", uuid); debug!("Username: {}", username); } fn enable_encryption(&mut self, packet: &mut packet::InPacket) { // Get all the data from the Encryption Request packet let server_id = packet.read_string(); let key_len = packet.read_be_i16().unwrap(); let public_key = packet.read_exact(key_len as uint).unwrap(); let token_len = packet.read_be_i16().unwrap(); let verify_token = packet.read_exact(token_len as uint).unwrap(); // Server's public key let pk = crypto::RSAPublicKey::from_bytes(public_key.as_slice()).unwrap(); // Generate random 16 byte key let mut key = [0u8, ..16]; rand::task_rng().fill_bytes(key); // Encrypt shared secret with server's public key let ekey = pk.encrypt(key).unwrap(); // Encrypt verify token with server's public key let etoken = pk.encrypt(verify_token.as_slice()).unwrap(); // Generate the server id hash let mut sha1 = crypto::SHA1::new(); sha1.update(server_id.as_bytes()); sha1.update(key); sha1.update(public_key.as_slice()); let hash = sha1.special_digest(); debug!("Hash: {}", hash); // Do client auth self.authenticate(hash); // Create Encryption Response Packet let mut erp = Packet::new_out(0x1); // Write encrypted shared secret erp.write_be_i16(ekey.len() as i16); erp.write(ekey); // Write encrypted verify token erp.write_be_i16(etoken.len() as i16); erp.write(etoken); // Send self.write_packet(erp); // Create AES cipher with shared secret let aes = crypto::AES::new(key.to_owned(), key.to_owned()).unwrap(); // Get the plain TCP stream let sock = match self.sock.take_unwrap() { Plain(s) => s, _ => fail!("Expected plain socket!") }; // and wwrap it in an AES Stream let sock = crypto::AesStream::new(sock, aes); // and put the new encrypted stream back // everything form this point is encrypted self.sock = Some(Encrypted(sock)); } fn authenticate(&mut self, hash: ~str) { let url = ~"https://authserver.mojang.com/authenticate"; let c = process::ProcessConfig { program: "/usr/bin/curl", args: &[~"-d", ~"@-", ~"-H", ~"Content-Type:application/json", url], env: None, cwd: None, stdin: process::CreatePipe(true, false), stdout: process::CreatePipe(false, true), .. process::ProcessConfig::new() }; let mut p = process::Process::configure(c).unwrap(); // write json to stdin and close it write!(p.stdin.get_mut_ref() as &mut Writer, r#" \{ "agent": \{ "name": "Minecraft", "version": 1 \}, "username": "{}", "password": "{}"
Connection
identifier_name
conn.rs
}) } pub fn status(&mut self) { self.send_handshake(false); // Send the status request self.write_packet(Packet::new_out(0x0)); // and read back the response let (packet_id, mut packet) = self.read_packet(); // Make sure we got the right response assert_eq!(packet_id, 0x0); // Get the JSON let json = ExtraJSON::new(json::from_str(packet.read_string()).unwrap()); println!("Minecraft Server Status [{}:{}]", self.host, self.addr.port); println!("Version: {}", json["version"]["name"].string()); println!("Protocol: {}", json["version"]["protocol"].as_int()); println!("Description: {}", json["description"].string()); println!("Players: ({}/{})", json["players"]["online"].as_int(), json["players"]["max"].as_int()); let players = json["players"]["sample"].list(); for player in players.iter() { println!("\t{} ({})", player["name"].string(), player["id"].string()); } } pub fn run(mut self) { // If the server is in online-mode // we need to do authentication and // enable encryption self.login(); // Get a port to read messages from stdin let msgs = self.read_messages(); // Yay, all good. // Now we just loop and read in all the packets we can // We don't actually do anything for most of them except // for chat and keep alives. loop { // Got a message in the queue to send? 'msg: loop { match msgs.try_recv() { comm::Data(msg) => { if msg.is_empty() { continue; } else if msg.len() > 100 { println!("Message too long."); continue; } // Send the message! let mut p = Packet::new_out(0x1); p.write_string(msg); self.write_packet(p); } comm::Empty => break 'msg, comm::Disconnected => fail!("input stream disconnected") } } // Read in and handle a packet let (packet_id, mut packet) = self.read_packet(); self.handle_message(packet_id, &mut packet); } } fn handle_message(&mut self, packet_id: i32, packet: &mut packet::InPacket) { // Keep Alive if packet_id == 0x0 { let x = packet.read_be_i32().unwrap(); // Need to respond let mut resp = Packet::new_out(0x0); resp.write_be_i32(x); self.write_packet(resp); // Chat Message } else if packet_id == 0x2 { let json = packet.read_string(); debug!("Got chat message: {}", json); // Let's wrap up the Json so that we can // deal with it more easily let j = json::from_str(json).unwrap(); let j = ExtraJSON::new(j); let ty = j["translate"].string(); // Player Chat if "chat.type.text" == ty { let user = j["with"][0]["text"].string(); let msg = j["with"][1].string(); self.term.attr(term::attr::ForegroundColor(term::color::BRIGHT_GREEN)); write!(&mut self.term as &mut Writer, "<{}> ", user); self.term.reset(); self.term.write(msg.as_bytes()); self.term.write(bytes!("\n")); // Server Message } else if "chat.type.announcement" == ty { let msg = j["with"][1]["extra"].list_map(|x| x.string()).concat(); self.term.attr(term::attr::ForegroundColor(term::color::BRIGHT_YELLOW)); self.term.write(bytes!("[Server] ")); self.term.reset(); self.term.write(msg.as_bytes()); self.term.write(bytes!("\n")); } } } fn login(&mut self) { self.send_handshake(true); self.send_username(); // Read the next packet and find out whether we need // to do authentication and encryption let (mut packet_id, mut packet) = self.read_packet(); debug!("Packet ID: {}", packet_id); if packet_id == 0x1 { // Encryption Request // online-mode = true self.enable_encryption(&mut packet); // Read the next packet... let (pi, p) = self.read_packet(); packet_id = pi; packet = p; } if packet_id == 0x0 { // Disconnect let reason = packet.read_string(); debug!("Reason: {}", reason); fail!("Received disconnect."); } // Login Success assert_eq!(packet_id, 0x2); let uuid = packet.read_string(); let username = packet.read_string(); debug!("UUID: {}", uuid); debug!("Username: {}", username); } fn enable_encryption(&mut self, packet: &mut packet::InPacket) { // Get all the data from the Encryption Request packet let server_id = packet.read_string(); let key_len = packet.read_be_i16().unwrap(); let public_key = packet.read_exact(key_len as uint).unwrap(); let token_len = packet.read_be_i16().unwrap(); let verify_token = packet.read_exact(token_len as uint).unwrap(); // Server's public key let pk = crypto::RSAPublicKey::from_bytes(public_key.as_slice()).unwrap(); // Generate random 16 byte key let mut key = [0u8, ..16]; rand::task_rng().fill_bytes(key); // Encrypt shared secret with server's public key let ekey = pk.encrypt(key).unwrap(); // Encrypt verify token with server's public key let etoken = pk.encrypt(verify_token.as_slice()).unwrap(); // Generate the server id hash let mut sha1 = crypto::SHA1::new(); sha1.update(server_id.as_bytes()); sha1.update(key); sha1.update(public_key.as_slice()); let hash = sha1.special_digest(); debug!("Hash: {}", hash); // Do client auth self.authenticate(hash); // Create Encryption Response Packet let mut erp = Packet::new_out(0x1); // Write encrypted shared secret erp.write_be_i16(ekey.len() as i16); erp.write(ekey); // Write encrypted verify token erp.write_be_i16(etoken.len() as i16); erp.write(etoken); // Send self.write_packet(erp); // Create AES cipher with shared secret let aes = crypto::AES::new(key.to_owned(), key.to_owned()).unwrap(); // Get the plain TCP stream let sock = match self.sock.take_unwrap() { Plain(s) => s, _ => fail!("Expected plain socket!") }; // and wwrap it in an AES Stream let sock = crypto::AesStream::new(sock, aes); // and put the new encrypted stream back // everything form this point is encrypted self.sock = Some(Encrypted(sock)); } fn authenticate(&mut self, hash: ~str) { let url = ~"https://authserver.mojang.com/authenticate"; let c = process::ProcessConfig { program: "/usr/bin/curl", args: &[~"-d", ~"@-", ~"-H", ~"Content-Type:application/json", url], env: None, cwd: None, stdin: process::CreatePipe(true, false), stdout: process::CreatePipe(false, true), .. process::ProcessConfig::new() }; let mut p = process::Process::configure(c).unwrap(); // write json to stdin and close it write!(p.stdin.get_mut_ref() as &mut Writer, r#" \{ "agent": \{ "name": "Minecraft", "version": 1 \}, "username": "{}", "password": "{}" \}"#, "USER", "PASS"); // XXX: Don't hardcode these... p.stdin = None; // read response let out = p.wait_with_output().output; let out = str::from_utf8_owned(out.move_iter().collect()).unwrap(); debug!("Got - {}", out); let json = ExtraJSON::new(json::from_str(out).unwrap()); let token = json["accessToken"].string(); let profile = json["selectedProfile"]["id"].string(); let url = ~"https://sessionserver.mojang.com/session/minecraft/join"; let c = process::ProcessConfig { program: "/usr/bin/curl", args: &[~"-d", ~"@-", ~"-H", ~"Content-Type:application/json", url], env: None, cwd: None, stdin: process::CreatePipe(true, false), stdout: process::CreatePipe(false, true), .. process::ProcessConfig::new() }; let mut p = process::Process::configure(c).unwrap();
// write json to stdin and close it write!(p.stdin.get_mut_ref() as &mut Writer, r#" \{ "accessToken": "{}",
random_line_split
prefix_code.rs
0f0f) << 4; v = (v & 0xcccc) >> 2 | (v & 0x3333) << 2; v = (v & 0xaaaa) >> 1 | (v & 0x5555) << 1; v } pub fn generate_codes(sizes: &[u8], codes: &mut [PrefixCode]) -> bool { let mut num_codes: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1]; let mut next_code: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1]; for &s in sizes.iter() { num_codes[s as uint] += 1; } let mut code = 0u32; for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) { next_code[i] = code; code += num_codes[i]; code <<= 1; } if code != (1 << (MAX_EXPECTED_CODE_SIZE + 1)) { let mut t = 0u32; for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) { t += num_codes[i]; if t > 1 { //return false; // Error, sizes don't add up fail!("Code sizes don't add up"); } } } for i in range(0, sizes.len()) { let c = sizes[i]; let code = next_code[c as uint]; next_code[c as uint] += 1; let rev_code = reverse_u16(code) >> (16 - c as uint); codes[i] = PrefixCode::new(rev_code, c); } true } pub fn generate_codes_for_decode( sizes: &[u8], codes: &mut [PrefixCode], dec_first_offset: &mut [u16, ..17], dec_max_code: &mut [u32, ..18], dec_offset_to_sym: &mut [u16], decoder_table: &mut [u16], max_code_size: u32) -> bool { let mut num_codes: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1]; let mut next_code: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1]; for &s in sizes.iter() { num_codes[s as uint] += 1; } let mut code = 0u32; let mut offset = 0u32; for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) { next_code[i] = code; dec_first_offset[i] = offset as u16 - code as u16; code += num_codes[i]; dec_max_code[i] = code << (16 - i); code <<= 1; offset += num_codes[i]; } dec_max_code[17] = 0x10000; if code != (1 << (MAX_EXPECTED_CODE_SIZE + 1)) { let mut t = 0u32; for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) { t += num_codes[i]; if t > 1
} } for p in decoder_table.mut_iter() { *p = 0xffff; } for i in range(0, sizes.len()) { let s = sizes[i] as uint; let code = next_code[s]; next_code[s] += 1; let offset = (code as u16 + dec_first_offset[s]) as uint; dec_offset_to_sym[offset] = i as u16; let rev_code = reverse_u16(code) >> (16 - s); codes[i] = PrefixCode::new(rev_code, s as u8); if s as u32 <= max_code_size { let step = 1 << s; let code = rev_code; for p in range_step(code, 1 << max_code_size as uint, step) { decoder_table[p as uint] = i as u16; } } } true } pub fn generate_decoder_table(codes: &[PrefixCode], decoder_table: &mut [u16], max_code_size: u32) { assert!(decoder_table.len() == (1 << max_code_size as uint)); for p in decoder_table.mut_iter() { *p = 0xffff; } for i in range(0, codes.len()) { if codes[i].size() as u32 <= max_code_size { assert!(codes[i].size() > 0); let step = 1 << codes[i].size() as uint; let code = codes[i].code(); for p in range_step(code, 1 << max_code_size as uint, step) { decoder_table[p as uint] = i as u16; } } } } static POLAR_MAX_SYMBOLS: u32 = 256; pub fn polar_code_lengths(symbols: &[OrdFreq], sizes: &mut [u8]) -> u32 { unsafe { let mut tmp_freq: [u32, ..POLAR_MAX_SYMBOLS] = mem::uninitialized(); let mut orig_total_freq = 0; let mut cur_total = 0; let mut start_index = 0; let mut max_code_size = 0; let num_syms = symbols.len() as u32; for i in range(0, symbols.len()) { let sym_freq = symbols[symbols.len() - 1 - i].freq(); //let sym_freq = symbols[i].freq(); let sym_len = 31 - ctlz32(sym_freq); let adjusted_sym_freq = 1 << sym_len as uint; orig_total_freq += sym_freq; tmp_freq[i] = adjusted_sym_freq; cur_total += adjusted_sym_freq; } let mut tree_total = 1 << (31 - ctlz32(orig_total_freq)) as uint; if tree_total < orig_total_freq { tree_total <<= 1; } while cur_total < tree_total && start_index < num_syms { let mut i = start_index; while i < num_syms { let freq = tmp_freq[i as uint]; if cur_total + freq <= tree_total { tmp_freq[i as uint] += freq; cur_total += freq; if cur_total == tree_total { break; } } else { start_index = i + 1; } i += 1; } } assert_eq!(cur_total, tree_total); let tree_total_bits = 32 - ctlz32(tree_total); for i in range(0, symbols.len()) { let codesize = tree_total_bits - (32 - ctlz32(tmp_freq[i])); max_code_size = max(max_code_size, codesize); sizes[symbols[symbols.len() - 1 - i].sym() as uint] = codesize as u8; //sizes[symbols[i].sym() as uint] = codesize as u8; } max_code_size } } pub trait PrefixModel { fn incr(&mut self, sym: u32); fn update(&mut self, for_encoding: bool); fn write<BW: ::prefix_code::BitWriter>(&mut self, bw: &mut BW, sym: u32); fn read<BR: ::prefix_code::BitReader>(&mut self, br: &mut BR) -> u32; } pub trait BitWriter { fn push_bits_uni(&mut self, bits: u32, count: u32); } pub trait BitReader { fn pull_bits_uni(&mut self, count: u32) -> u32; fn peek_bits_uni16(&self) -> u16; fn skip_bits_uni(&mut self, count: u32); } #[deriving(Copy)] pub struct Foo { f: [u32, ..256] } impl Clone for Foo { fn clone(&self) -> Foo { Foo { f: self.f } } } macro_rules! define_polar_model( ($name: ident, $symbol_count: expr) => { //#[deriving(Clone)] pub struct $name { freq: [u32, ..$symbol_count], codes: [::prefix_code::PrefixCode, ..$symbol_count], decoder_table: [u16, ..(1 << 9)], sum: u32, next_rebuild: u32, dec_max_code: [u32, ..18], dec_first_offset: [u16, ..17], dec_offset_to_sym: [u16, ..$symbol_count]
{ //return false; // Error, sizes don't add up fail!("Code sizes don't add up"); }
conditional_block
prefix_code.rs
(self) -> u16 { self.s } } pub fn sort_symbols2<'a>(mut first: &'a mut [OrdFreq], mut second: &'a mut [OrdFreq]) -> &'a mut [OrdFreq] { let mut hist = [0u32, ..256 * 2]; for &s in first.iter() { let f = s.freq(); hist[ (f & 0xff) as uint] += 1; hist[256 + ((f >> 8) & 0xff) as uint] += 1; } let num_syms = first.len(); // If all radix-1 digits are zero, we only need one pass let passes = if hist[256] == num_syms as u32 { 1 } else { 2 }; for pass in range(0, passes) { let c = &mut first[0] as *mut _; let n = &mut second[0] as *mut _; let histp = &mut hist[pass << 8] as *mut _; let mut offsets: [u32, ..256] = unsafe { mem::uninitialized() }; let mut cur_ofs = 0; for i in range_step(0u, 256, 2) { offsets[i] = cur_ofs; cur_ofs += unsafe { *histp.offset(i as int) }; offsets[i + 1] = cur_ofs; cur_ofs += unsafe { *histp.offset(i as int + 1) }; } let pass_shift = pass << 3; let mut p = c; let endp = unsafe { c.offset(num_syms as int) }; while p != endp { let mut f = unsafe { *p }.freq(); f = (f >> pass_shift) & 0xff; let dst_offset = offsets[f as uint]; offsets[f as uint] += 1; unsafe { *n.offset(dst_offset as int) = *p; p = p.offset(1); } } mem::swap(&mut first, &mut second); } let mut prev = 0; for i in range(0, num_syms) { assert!(first[i].freq() >= prev); prev = first[i].freq(); } first } #[deriving(Clone)] pub struct PrefixCode(u32); impl PrefixCode { #[inline] pub fn new(code: u32, size: u8) -> PrefixCode { PrefixCode(code + (size as u32 << 16)) } pub fn code(self) -> u32 { let PrefixCode(v) = self; v & 0xffff } pub fn size(self) -> u32 { let PrefixCode(v) = self; v >> 16 } } #[inline] pub fn reverse_u16(mut v: u32) -> u32 { v = (v & 0xff00) >> 8 | (v & 0x00ff) << 8; v = (v & 0xf0f0) >> 4 | (v & 0x0f0f) << 4; v = (v & 0xcccc) >> 2 | (v & 0x3333) << 2; v = (v & 0xaaaa) >> 1 | (v & 0x5555) << 1; v } pub fn generate_codes(sizes: &[u8], codes: &mut [PrefixCode]) -> bool { let mut num_codes: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1]; let mut next_code: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1]; for &s in sizes.iter() { num_codes[s as uint] += 1; } let mut code = 0u32; for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) { next_code[i] = code; code += num_codes[i]; code <<= 1; } if code != (1 << (MAX_EXPECTED_CODE_SIZE + 1)) { let mut t = 0u32; for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) { t += num_codes[i]; if t > 1 { //return false; // Error, sizes don't add up fail!("Code sizes don't add up"); } } } for i in range(0, sizes.len()) { let c = sizes[i]; let code = next_code[c as uint]; next_code[c as uint] += 1; let rev_code = reverse_u16(code) >> (16 - c as uint); codes[i] = PrefixCode::new(rev_code, c); } true } pub fn generate_codes_for_decode( sizes: &[u8], codes: &mut [PrefixCode], dec_first_offset: &mut [u16, ..17], dec_max_code: &mut [u32, ..18], dec_offset_to_sym: &mut [u16], decoder_table: &mut [u16], max_code_size: u32) -> bool { let mut num_codes: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1]; let mut next_code: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1]; for &s in sizes.iter() { num_codes[s as uint] += 1; } let mut code = 0u32; let mut offset = 0u32; for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) { next_code[i] = code; dec_first_offset[i] = offset as u16 - code as u16; code += num_codes[i]; dec_max_code[i] = code << (16 - i); code <<= 1; offset += num_codes[i]; } dec_max_code[17] = 0x10000; if code != (1 << (MAX_EXPECTED_CODE_SIZE + 1)) { let mut t = 0u32; for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) { t += num_codes[i]; if t > 1 { //return false; // Error, sizes don't add up fail!("Code sizes don't add up"); } } } for p in decoder_table.mut_iter() { *p = 0xffff; } for i in range(0, sizes.len()) { let s = sizes[i] as uint; let code = next_code[s]; next_code[s] += 1; let offset = (code as u16 + dec_first_offset[s]) as uint; dec_offset_to_sym[offset] = i as u16; let rev_code = reverse_u16(code) >> (16 - s); codes[i] = PrefixCode::new(rev_code, s as u8); if s as u32 <= max_code_size { let step = 1 << s; let code = rev_code; for p in range_step(code, 1 << max_code_size as uint, step) { decoder_table[p as uint] = i as u16; } } } true } pub fn generate_decoder_table(codes: &[PrefixCode], decoder_table: &mut [u16], max_code_size: u32) { assert!(decoder_table.len() == (1 << max_code_size as uint)); for p in decoder_table.mut_iter() { *p = 0xffff; } for i in range(0, codes.len()) { if codes[i].size() as u32 <= max_code_size { assert!(codes[i].size() > 0); let step = 1 << codes[i].size() as uint; let code = codes[i].code(); for p in range_step(code, 1 << max_code_size as uint, step) { decoder_table[p as uint] = i as u16; } } } } static POLAR_MAX_SYMBOLS: u32 = 256; pub fn polar_code_lengths(symbols: &[OrdFreq], sizes: &mut [u8]) -> u32 { unsafe { let mut tmp_freq: [u32, ..POLAR_MAX_SYMBOLS] = mem::uninitialized(); let mut orig_total_freq = 0; let mut cur_total = 0; let mut start_index = 0; let mut max_code_size = 0; let num_syms = symbols.len() as u32; for i in range(0, symbols.len()) { let sym_freq = symbols[symbols.len() - 1 - i].freq();
sym
identifier_name
prefix_code.rs
0f0f) << 4; v = (v & 0xcccc) >> 2 | (v & 0x3333) << 2; v = (v & 0xaaaa) >> 1 | (v & 0x5555) << 1; v } pub fn generate_codes(sizes: &[u8], codes: &mut [PrefixCode]) -> bool { let mut num_codes: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1]; let mut next_code: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1]; for &s in sizes.iter() { num_codes[s as uint] += 1; } let mut code = 0u32; for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) { next_code[i] = code; code += num_codes[i]; code <<= 1; } if code != (1 << (MAX_EXPECTED_CODE_SIZE + 1)) { let mut t = 0u32; for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) { t += num_codes[i]; if t > 1 { //return false; // Error, sizes don't add up fail!("Code sizes don't add up"); } } } for i in range(0, sizes.len()) { let c = sizes[i]; let code = next_code[c as uint]; next_code[c as uint] += 1; let rev_code = reverse_u16(code) >> (16 - c as uint); codes[i] = PrefixCode::new(rev_code, c); } true } pub fn generate_codes_for_decode( sizes: &[u8], codes: &mut [PrefixCode], dec_first_offset: &mut [u16, ..17], dec_max_code: &mut [u32, ..18], dec_offset_to_sym: &mut [u16], decoder_table: &mut [u16], max_code_size: u32) -> bool { let mut num_codes: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1];
let mut code = 0u32; let mut offset = 0u32; for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) { next_code[i] = code; dec_first_offset[i] = offset as u16 - code as u16; code += num_codes[i]; dec_max_code[i] = code << (16 - i); code <<= 1; offset += num_codes[i]; } dec_max_code[17] = 0x10000; if code != (1 << (MAX_EXPECTED_CODE_SIZE + 1)) { let mut t = 0u32; for i in range(1, MAX_EXPECTED_CODE_SIZE + 1) { t += num_codes[i]; if t > 1 { //return false; // Error, sizes don't add up fail!("Code sizes don't add up"); } } } for p in decoder_table.mut_iter() { *p = 0xffff; } for i in range(0, sizes.len()) { let s = sizes[i] as uint; let code = next_code[s]; next_code[s] += 1; let offset = (code as u16 + dec_first_offset[s]) as uint; dec_offset_to_sym[offset] = i as u16; let rev_code = reverse_u16(code) >> (16 - s); codes[i] = PrefixCode::new(rev_code, s as u8); if s as u32 <= max_code_size { let step = 1 << s; let code = rev_code; for p in range_step(code, 1 << max_code_size as uint, step) { decoder_table[p as uint] = i as u16; } } } true } pub fn generate_decoder_table(codes: &[PrefixCode], decoder_table: &mut [u16], max_code_size: u32) { assert!(decoder_table.len() == (1 << max_code_size as uint)); for p in decoder_table.mut_iter() { *p = 0xffff; } for i in range(0, codes.len()) { if codes[i].size() as u32 <= max_code_size { assert!(codes[i].size() > 0); let step = 1 << codes[i].size() as uint; let code = codes[i].code(); for p in range_step(code, 1 << max_code_size as uint, step) { decoder_table[p as uint] = i as u16; } } } } static POLAR_MAX_SYMBOLS: u32 = 256; pub fn polar_code_lengths(symbols: &[OrdFreq], sizes: &mut [u8]) -> u32 { unsafe { let mut tmp_freq: [u32, ..POLAR_MAX_SYMBOLS] = mem::uninitialized(); let mut orig_total_freq = 0; let mut cur_total = 0; let mut start_index = 0; let mut max_code_size = 0; let num_syms = symbols.len() as u32; for i in range(0, symbols.len()) { let sym_freq = symbols[symbols.len() - 1 - i].freq(); //let sym_freq = symbols[i].freq(); let sym_len = 31 - ctlz32(sym_freq); let adjusted_sym_freq = 1 << sym_len as uint; orig_total_freq += sym_freq; tmp_freq[i] = adjusted_sym_freq; cur_total += adjusted_sym_freq; } let mut tree_total = 1 << (31 - ctlz32(orig_total_freq)) as uint; if tree_total < orig_total_freq { tree_total <<= 1; } while cur_total < tree_total && start_index < num_syms { let mut i = start_index; while i < num_syms { let freq = tmp_freq[i as uint]; if cur_total + freq <= tree_total { tmp_freq[i as uint] += freq; cur_total += freq; if cur_total == tree_total { break; } } else { start_index = i + 1; } i += 1; } } assert_eq!(cur_total, tree_total); let tree_total_bits = 32 - ctlz32(tree_total); for i in range(0, symbols.len()) { let codesize = tree_total_bits - (32 - ctlz32(tmp_freq[i])); max_code_size = max(max_code_size, codesize); sizes[symbols[symbols.len() - 1 - i].sym() as uint] = codesize as u8; //sizes[symbols[i].sym() as uint] = codesize as u8; } max_code_size } } pub trait PrefixModel { fn incr(&mut self, sym: u32); fn update(&mut self, for_encoding: bool); fn write<BW: ::prefix_code::BitWriter>(&mut self, bw: &mut BW, sym: u32); fn read<BR: ::prefix_code::BitReader>(&mut self, br: &mut BR) -> u32; } pub trait BitWriter { fn push_bits_uni(&mut self, bits: u32, count: u32); } pub trait BitReader { fn pull_bits_uni(&mut self, count: u32) -> u32; fn peek_bits_uni16(&self) -> u16; fn skip_bits_uni(&mut self, count: u32); } #[deriving(Copy)] pub struct Foo { f: [u32, ..256] } impl Clone for Foo { fn clone(&self) -> Foo { Foo { f: self.f } } } macro_rules! define_polar_model( ($name: ident, $symbol_count: expr) => { //#[deriving(Clone)] pub struct $name { freq: [u32, ..$symbol_count], codes: [::prefix_code::PrefixCode, ..$symbol_count], decoder_table: [u16, ..(1 << 9)], sum: u32, next_rebuild: u32, dec_max_code: [u32, ..18], dec_first_offset: [u16, ..17], dec_offset_to_sym: [u16, ..$symbol_count] }
let mut next_code: [u32, ..MAX_EXPECTED_CODE_SIZE + 1] = [0, ..MAX_EXPECTED_CODE_SIZE + 1]; for &s in sizes.iter() { num_codes[s as uint] += 1; }
random_line_split
fma_utils.py
r.raise_for_status() tracks = [] artists = [] date_created = [] for track in r.json()['aTracks']:
return tracks, artists, date_created def _get_data(self, dataset, fma_id, fields=None): url = self.BASE_URL + dataset + 's.json?' url += dataset + '_id=' + str(fma_id) + '&api_key=' + self.api_key # print(url) r = requests.get(url) r.raise_for_status() if r.json()['errors']: raise Exception(r.json()['errors']) data = r.json()['dataset'][0] r_id = data[dataset + '_id'] if r_id != str(fma_id): raise Exception('The received id {} does not correspond to' 'the requested one {}'.format(r_id, fma_id)) if fields is None: return data if type(fields) is list: ret = {} for field in fields: ret[field] = data[field] return ret else: return data[fields] def get_track(self, track_id, fields=None): return self._get_data('track', track_id, fields) def get_album(self, album_id, fields=None): return self._get_data('album', album_id, fields) def get_artist(self, artist_id, fields=None): return self._get_data('artist', artist_id, fields) def get_all(self, dataset, id_range): index = dataset + '_id' id_ = 2 if dataset is 'track' else 1 row = self._get_data(dataset, id_) df = pd.DataFrame(columns=row.keys()) df.set_index(index, inplace=True) not_found_ids = [] for id_ in id_range: try: row = self._get_data(dataset, id_) except: not_found_ids.append(id_) continue row.pop(index) df.loc[id_] = row return df, not_found_ids def download_track(self, track_file, path): url = 'https://files.freemusicarchive.org/' + track_file r = requests.get(url, stream=True) r.raise_for_status() with open(path, 'wb') as f: shutil.copyfileobj(r.raw, f) def get_track_genres(self, track_id): genres = self.get_track(track_id, 'track_genres') genre_ids = [] genre_titles = [] for genre in genres: genre_ids.append(genre['genre_id']) genre_titles.append(genre['genre_title']) return genre_ids, genre_titles def get_all_genres(self): df = pd.DataFrame(columns=['genre_parent_id', 'genre_title', 'genre_handle', 'genre_color']) df.index.rename('genre_id', inplace=True) page = 1 while True: url = self.BASE_URL + 'genres.json?limit=50' url += '&page={}&api_key={}'.format(page, self.api_key) r = requests.get(url) for genre in r.json()['dataset']: genre_id = int(genre.pop(df.index.name)) df.loc[genre_id] = genre assert (r.json()['page'] == str(page)) page += 1 if page > r.json()['total_pages']: break return df class Genres: def __init__(self, genres_df): self.df = genres_df def create_tree(self, roots, depth=None): if type(roots) is not list: roots = [roots] graph = pydot.Dot(graph_type='digraph', strict=True) def create_node(genre_id): title = self.df.at[genre_id, 'title'] ntracks = self.df.at[genre_id, '#tracks'] #name = self.df.at[genre_id, 'title'] + '\n' + str(genre_id) name = '"{}\n{} / {}"'.format(title, genre_id, ntracks) return pydot.Node(name) def create_tree(root_id, node_p, depth): if depth == 0: return children = self.df[self.df['parent'] == root_id] for child in children.iterrows(): genre_id = child[0] node_c = create_node(genre_id) graph.add_edge(pydot.Edge(node_p, node_c)) create_tree(genre_id, node_c, depth-1 if depth is not None else None) for root in roots: node_p = create_node(root) graph.add_node(node_p) create_tree(root, node_p, depth) return graph def find_roots(self): roots = [] for gid, row in self.df.iterrows(): parent = row['parent'] title = row['title'] if parent == 0: roots.append(gid) elif parent not in self.df.index: msg = '{} ({}) has parent {} which is missing'.format( gid, title, parent) raise RuntimeError(msg) return roots def load(filepath): filename = os.path.basename(filepath) if 'features' in filename: return pd.read_csv(filepath, index_col=0, header=[0, 1, 2]) if 'echonest' in filename: return pd.read_csv(filepath, index_col=0, header=[0, 1, 2]) if 'genres' in filename: return pd.read_csv(filepath, index_col=0) if 'tracks' in filename: tracks = pd.read_csv(filepath, index_col=0, header=[0, 1]) COLUMNS = [('track', 'tags'), ('album', 'tags'), ('artist', 'tags'), ('track', 'genres'), ('track', 'genres_all')] for column in COLUMNS: tracks[column] = tracks[column].map(ast.literal_eval) COLUMNS = [('track', 'date_created'), ('track', 'date_recorded'), ('album', 'date_created'), ('album', 'date_released'), ('artist', 'date_created'), ('artist', 'active_year_begin'), ('artist', 'active_year_end')] for column in COLUMNS: tracks[column] = pd.to_datetime(tracks[column]) SUBSETS = ('small', 'medium', 'large') tracks['set', 'subset'] = tracks['set', 'subset'].astype( 'category', categories=SUBSETS, ordered=True) COLUMNS = [('track', 'genre_top'), ('track', 'license'), ('album', 'type'), ('album', 'information'), ('artist', 'bio')] for column in COLUMNS: tracks[column] = tracks[column].astype('category') return tracks def get_audio_path(audio_dir, track_id): tid_str = '{:06d}'.format(track_id) return os.path.join(audio_dir, tid_str[:3], tid_str + '.mp3') class Loader: def load(self, filepath): raise NotImplemented() class RawAudioLoader(Loader): def __init__(self, sampling_rate=SAMPLING_RATE): self.sampling_rate = sampling_rate self.shape = (NB_AUDIO_SAMPLES * sampling_rate // SAMPLING_RATE, ) def load(self, filepath): return self._load(filepath)[:self.shape[0]] class LibrosaLoader(RawAudioLoader): def _load(self, filepath): import librosa sr = self.sampling_rate if self.sampling_rate != SAMPLING_RATE else None # kaiser_fast is 3x faster than kaiser_best #x, sr = librosa.load(filepath, sr=sr, res_type='kaiser_fast') x, sr = librosa.load(filepath, sr=sr) return x class AudioreadLoader(RawAudioLoader): def _load(self, filepath): import audioread a = audioread.audio_open(filepath) a.read_data() class PydubLoader(RawAudioLoader): def _load(self, filepath): from pydub import AudioSegment song = AudioSegment.from_file(filepath) song = song.set_channels(1) x = song.get_array_of_samples() # print(filepath) if song.channels != 2 else None return np.array(x) class FfmpegLoader(RawAudioLoader): def _load(self, filepath): """Fastest and less CPU intensive loading method.""" import subprocess as sp command = ['ffmpeg', '-i', filepath, '-f', 's16le', '-acodec', 'pcm_s16le', '-ac', '1'] # channels: 2 for stereo, 1 for mono if self.sampling_rate != SAMPLING_RATE: command.extend(['-ar', str(self.sampling_rate)]) command.append('-') # 30s at 44.1 kHz ~= 1.3e6 proc = sp.run(command, stdout=sp.PIPE, bufsize=10**7, stderr=sp.DEVNULL, check=True) return np.fromstring(proc.stdout, dtype="int16") def build_sample_loader(audio_dir, Y, loader): class SampleLoader: def __init__(self, tids, batch_size=4): self.lock1 = multiprocessing.Lock()
tracks.append(track['track_id']) artists.append(track['artist_name']) date_created.append(track['track_date_created'])
conditional_block
fma_utils.py
r.raise_for_status() tracks = [] artists = [] date_created = [] for track in r.json()['aTracks']: tracks.append(track['track_id']) artists.append(track['artist_name']) date_created.append(track['track_date_created']) return tracks, artists, date_created def _get_data(self, dataset, fma_id, fields=None): url = self.BASE_URL + dataset + 's.json?' url += dataset + '_id=' + str(fma_id) + '&api_key=' + self.api_key # print(url) r = requests.get(url) r.raise_for_status() if r.json()['errors']: raise Exception(r.json()['errors']) data = r.json()['dataset'][0] r_id = data[dataset + '_id'] if r_id != str(fma_id): raise Exception('The received id {} does not correspond to' 'the requested one {}'.format(r_id, fma_id)) if fields is None: return data if type(fields) is list: ret = {} for field in fields: ret[field] = data[field] return ret else: return data[fields] def get_track(self, track_id, fields=None): return self._get_data('track', track_id, fields) def get_album(self, album_id, fields=None): return self._get_data('album', album_id, fields) def get_artist(self, artist_id, fields=None): return self._get_data('artist', artist_id, fields) def get_all(self, dataset, id_range): index = dataset + '_id' id_ = 2 if dataset is 'track' else 1 row = self._get_data(dataset, id_) df = pd.DataFrame(columns=row.keys()) df.set_index(index, inplace=True) not_found_ids = [] for id_ in id_range: try: row = self._get_data(dataset, id_) except: not_found_ids.append(id_) continue row.pop(index) df.loc[id_] = row return df, not_found_ids def download_track(self, track_file, path): url = 'https://files.freemusicarchive.org/' + track_file r = requests.get(url, stream=True) r.raise_for_status() with open(path, 'wb') as f: shutil.copyfileobj(r.raw, f) def get_track_genres(self, track_id): genres = self.get_track(track_id, 'track_genres') genre_ids = [] genre_titles = [] for genre in genres: genre_ids.append(genre['genre_id']) genre_titles.append(genre['genre_title']) return genre_ids, genre_titles def get_all_genres(self): df = pd.DataFrame(columns=['genre_parent_id', 'genre_title', 'genre_handle', 'genre_color']) df.index.rename('genre_id', inplace=True) page = 1 while True: url = self.BASE_URL + 'genres.json?limit=50' url += '&page={}&api_key={}'.format(page, self.api_key) r = requests.get(url) for genre in r.json()['dataset']: genre_id = int(genre.pop(df.index.name)) df.loc[genre_id] = genre assert (r.json()['page'] == str(page)) page += 1 if page > r.json()['total_pages']: break return df class Genres: def __init__(self, genres_df): self.df = genres_df def create_tree(self, roots, depth=None): if type(roots) is not list: roots = [roots] graph = pydot.Dot(graph_type='digraph', strict=True) def create_node(genre_id): title = self.df.at[genre_id, 'title'] ntracks = self.df.at[genre_id, '#tracks'] #name = self.df.at[genre_id, 'title'] + '\n' + str(genre_id) name = '"{}\n{} / {}"'.format(title, genre_id, ntracks) return pydot.Node(name) def create_tree(root_id, node_p, depth): if depth == 0: return children = self.df[self.df['parent'] == root_id] for child in children.iterrows(): genre_id = child[0] node_c = create_node(genre_id) graph.add_edge(pydot.Edge(node_p, node_c)) create_tree(genre_id, node_c, depth-1 if depth is not None else None) for root in roots: node_p = create_node(root) graph.add_node(node_p) create_tree(root, node_p, depth) return graph def find_roots(self): roots = [] for gid, row in self.df.iterrows(): parent = row['parent'] title = row['title'] if parent == 0: roots.append(gid) elif parent not in self.df.index: msg = '{} ({}) has parent {} which is missing'.format( gid, title, parent) raise RuntimeError(msg) return roots def load(filepath): filename = os.path.basename(filepath) if 'features' in filename: return pd.read_csv(filepath, index_col=0, header=[0, 1, 2]) if 'echonest' in filename: return pd.read_csv(filepath, index_col=0, header=[0, 1, 2]) if 'genres' in filename: return pd.read_csv(filepath, index_col=0) if 'tracks' in filename: tracks = pd.read_csv(filepath, index_col=0, header=[0, 1]) COLUMNS = [('track', 'tags'), ('album', 'tags'), ('artist', 'tags'), ('track', 'genres'), ('track', 'genres_all')] for column in COLUMNS: tracks[column] = tracks[column].map(ast.literal_eval) COLUMNS = [('track', 'date_created'), ('track', 'date_recorded'), ('album', 'date_created'), ('album', 'date_released'), ('artist', 'date_created'), ('artist', 'active_year_begin'), ('artist', 'active_year_end')] for column in COLUMNS: tracks[column] = pd.to_datetime(tracks[column]) SUBSETS = ('small', 'medium', 'large') tracks['set', 'subset'] = tracks['set', 'subset'].astype( 'category', categories=SUBSETS, ordered=True) COLUMNS = [('track', 'genre_top'), ('track', 'license'), ('album', 'type'), ('album', 'information'), ('artist', 'bio')] for column in COLUMNS: tracks[column] = tracks[column].astype('category') return tracks def get_audio_path(audio_dir, track_id): tid_str = '{:06d}'.format(track_id) return os.path.join(audio_dir, tid_str[:3], tid_str + '.mp3') class
: def load(self, filepath): raise NotImplemented() class RawAudioLoader(Loader): def __init__(self, sampling_rate=SAMPLING_RATE): self.sampling_rate = sampling_rate self.shape = (NB_AUDIO_SAMPLES * sampling_rate // SAMPLING_RATE, ) def load(self, filepath): return self._load(filepath)[:self.shape[0]] class LibrosaLoader(RawAudioLoader): def _load(self, filepath): import librosa sr = self.sampling_rate if self.sampling_rate != SAMPLING_RATE else None # kaiser_fast is 3x faster than kaiser_best #x, sr = librosa.load(filepath, sr=sr, res_type='kaiser_fast') x, sr = librosa.load(filepath, sr=sr) return x class AudioreadLoader(RawAudioLoader): def _load(self, filepath): import audioread a = audioread.audio_open(filepath) a.read_data() class PydubLoader(RawAudioLoader): def _load(self, filepath): from pydub import AudioSegment song = AudioSegment.from_file(filepath) song = song.set_channels(1) x = song.get_array_of_samples() # print(filepath) if song.channels != 2 else None return np.array(x) class FfmpegLoader(RawAudioLoader): def _load(self, filepath): """Fastest and less CPU intensive loading method.""" import subprocess as sp command = ['ffmpeg', '-i', filepath, '-f', 's16le', '-acodec', 'pcm_s16le', '-ac', '1'] # channels: 2 for stereo, 1 for mono if self.sampling_rate != SAMPLING_RATE: command.extend(['-ar', str(self.sampling_rate)]) command.append('-') # 30s at 44.1 kHz ~= 1.3e6 proc = sp.run(command, stdout=sp.PIPE, bufsize=10**7, stderr=sp.DEVNULL, check=True) return np.fromstring(proc.stdout, dtype="int16") def build_sample_loader(audio_dir, Y, loader): class SampleLoader: def __init__(self, tids, batch_size=4): self.lock1 = multiprocessing.Lock()
Loader
identifier_name
fma_utils.py
) r.raise_for_status() tracks = [] artists = [] date_created = [] for track in r.json()['aTracks']: tracks.append(track['track_id']) artists.append(track['artist_name']) date_created.append(track['track_date_created']) return tracks, artists, date_created def _get_data(self, dataset, fma_id, fields=None):
return data[fields] def get_track(self, track_id, fields=None): return self._get_data('track', track_id, fields) def get_album(self, album_id, fields=None): return self._get_data('album', album_id, fields) def get_artist(self, artist_id, fields=None): return self._get_data('artist', artist_id, fields) def get_all(self, dataset, id_range): index = dataset + '_id' id_ = 2 if dataset is 'track' else 1 row = self._get_data(dataset, id_) df = pd.DataFrame(columns=row.keys()) df.set_index(index, inplace=True) not_found_ids = [] for id_ in id_range: try: row = self._get_data(dataset, id_) except: not_found_ids.append(id_) continue row.pop(index) df.loc[id_] = row return df, not_found_ids def download_track(self, track_file, path): url = 'https://files.freemusicarchive.org/' + track_file r = requests.get(url, stream=True) r.raise_for_status() with open(path, 'wb') as f: shutil.copyfileobj(r.raw, f) def get_track_genres(self, track_id): genres = self.get_track(track_id, 'track_genres') genre_ids = [] genre_titles = [] for genre in genres: genre_ids.append(genre['genre_id']) genre_titles.append(genre['genre_title']) return genre_ids, genre_titles def get_all_genres(self): df = pd.DataFrame(columns=['genre_parent_id', 'genre_title', 'genre_handle', 'genre_color']) df.index.rename('genre_id', inplace=True) page = 1 while True: url = self.BASE_URL + 'genres.json?limit=50' url += '&page={}&api_key={}'.format(page, self.api_key) r = requests.get(url) for genre in r.json()['dataset']: genre_id = int(genre.pop(df.index.name)) df.loc[genre_id] = genre assert (r.json()['page'] == str(page)) page += 1 if page > r.json()['total_pages']: break return df class Genres: def __init__(self, genres_df): self.df = genres_df def create_tree(self, roots, depth=None): if type(roots) is not list: roots = [roots] graph = pydot.Dot(graph_type='digraph', strict=True) def create_node(genre_id): title = self.df.at[genre_id, 'title'] ntracks = self.df.at[genre_id, '#tracks'] #name = self.df.at[genre_id, 'title'] + '\n' + str(genre_id) name = '"{}\n{} / {}"'.format(title, genre_id, ntracks) return pydot.Node(name) def create_tree(root_id, node_p, depth): if depth == 0: return children = self.df[self.df['parent'] == root_id] for child in children.iterrows(): genre_id = child[0] node_c = create_node(genre_id) graph.add_edge(pydot.Edge(node_p, node_c)) create_tree(genre_id, node_c, depth-1 if depth is not None else None) for root in roots: node_p = create_node(root) graph.add_node(node_p) create_tree(root, node_p, depth) return graph def find_roots(self): roots = [] for gid, row in self.df.iterrows(): parent = row['parent'] title = row['title'] if parent == 0: roots.append(gid) elif parent not in self.df.index: msg = '{} ({}) has parent {} which is missing'.format( gid, title, parent) raise RuntimeError(msg) return roots def load(filepath): filename = os.path.basename(filepath) if 'features' in filename: return pd.read_csv(filepath, index_col=0, header=[0, 1, 2]) if 'echonest' in filename: return pd.read_csv(filepath, index_col=0, header=[0, 1, 2]) if 'genres' in filename: return pd.read_csv(filepath, index_col=0) if 'tracks' in filename: tracks = pd.read_csv(filepath, index_col=0, header=[0, 1]) COLUMNS = [('track', 'tags'), ('album', 'tags'), ('artist', 'tags'), ('track', 'genres'), ('track', 'genres_all')] for column in COLUMNS: tracks[column] = tracks[column].map(ast.literal_eval) COLUMNS = [('track', 'date_created'), ('track', 'date_recorded'), ('album', 'date_created'), ('album', 'date_released'), ('artist', 'date_created'), ('artist', 'active_year_begin'), ('artist', 'active_year_end')] for column in COLUMNS: tracks[column] = pd.to_datetime(tracks[column]) SUBSETS = ('small', 'medium', 'large') tracks['set', 'subset'] = tracks['set', 'subset'].astype( 'category', categories=SUBSETS, ordered=True) COLUMNS = [('track', 'genre_top'), ('track', 'license'), ('album', 'type'), ('album', 'information'), ('artist', 'bio')] for column in COLUMNS: tracks[column] = tracks[column].astype('category') return tracks def get_audio_path(audio_dir, track_id): tid_str = '{:06d}'.format(track_id) return os.path.join(audio_dir, tid_str[:3], tid_str + '.mp3') class Loader: def load(self, filepath): raise NotImplemented() class RawAudioLoader(Loader): def __init__(self, sampling_rate=SAMPLING_RATE): self.sampling_rate = sampling_rate self.shape = (NB_AUDIO_SAMPLES * sampling_rate // SAMPLING_RATE, ) def load(self, filepath): return self._load(filepath)[:self.shape[0]] class LibrosaLoader(RawAudioLoader): def _load(self, filepath): import librosa sr = self.sampling_rate if self.sampling_rate != SAMPLING_RATE else None # kaiser_fast is 3x faster than kaiser_best #x, sr = librosa.load(filepath, sr=sr, res_type='kaiser_fast') x, sr = librosa.load(filepath, sr=sr) return x class AudioreadLoader(RawAudioLoader): def _load(self, filepath): import audioread a = audioread.audio_open(filepath) a.read_data() class PydubLoader(RawAudioLoader): def _load(self, filepath): from pydub import AudioSegment song = AudioSegment.from_file(filepath) song = song.set_channels(1) x = song.get_array_of_samples() # print(filepath) if song.channels != 2 else None return np.array(x) class FfmpegLoader(RawAudioLoader): def _load(self, filepath): """Fastest and less CPU intensive loading method.""" import subprocess as sp command = ['ffmpeg', '-i', filepath, '-f', 's16le', '-acodec', 'pcm_s16le', '-ac', '1'] # channels: 2 for stereo, 1 for mono if self.sampling_rate != SAMPLING_RATE: command.extend(['-ar', str(self.sampling_rate)]) command.append('-') # 30s at 44.1 kHz ~= 1.3e6 proc = sp.run(command, stdout=sp.PIPE, bufsize=10**7, stderr=sp.DEVNULL, check=True) return np.fromstring(proc.stdout, dtype="int16") def build_sample_loader(audio_dir, Y, loader): class SampleLoader: def __init__(self, tids, batch_size=4): self.lock1 = multiprocessing.Lock()
url = self.BASE_URL + dataset + 's.json?' url += dataset + '_id=' + str(fma_id) + '&api_key=' + self.api_key # print(url) r = requests.get(url) r.raise_for_status() if r.json()['errors']: raise Exception(r.json()['errors']) data = r.json()['dataset'][0] r_id = data[dataset + '_id'] if r_id != str(fma_id): raise Exception('The received id {} does not correspond to' 'the requested one {}'.format(r_id, fma_id)) if fields is None: return data if type(fields) is list: ret = {} for field in fields: ret[field] = data[field] return ret else:
identifier_body
fma_utils.py
) r.raise_for_status() tracks = [] artists = [] date_created = [] for track in r.json()['aTracks']: tracks.append(track['track_id']) artists.append(track['artist_name']) date_created.append(track['track_date_created']) return tracks, artists, date_created def _get_data(self, dataset, fma_id, fields=None): url = self.BASE_URL + dataset + 's.json?' url += dataset + '_id=' + str(fma_id) + '&api_key=' + self.api_key # print(url) r = requests.get(url) r.raise_for_status() if r.json()['errors']: raise Exception(r.json()['errors']) data = r.json()['dataset'][0] r_id = data[dataset + '_id'] if r_id != str(fma_id): raise Exception('The received id {} does not correspond to' 'the requested one {}'.format(r_id, fma_id)) if fields is None: return data if type(fields) is list: ret = {} for field in fields: ret[field] = data[field] return ret else: return data[fields] def get_track(self, track_id, fields=None): return self._get_data('track', track_id, fields) def get_album(self, album_id, fields=None): return self._get_data('album', album_id, fields) def get_artist(self, artist_id, fields=None): return self._get_data('artist', artist_id, fields) def get_all(self, dataset, id_range): index = dataset + '_id' id_ = 2 if dataset is 'track' else 1 row = self._get_data(dataset, id_) df = pd.DataFrame(columns=row.keys()) df.set_index(index, inplace=True) not_found_ids = [] for id_ in id_range: try: row = self._get_data(dataset, id_) except: not_found_ids.append(id_) continue row.pop(index) df.loc[id_] = row return df, not_found_ids def download_track(self, track_file, path): url = 'https://files.freemusicarchive.org/' + track_file r = requests.get(url, stream=True) r.raise_for_status() with open(path, 'wb') as f: shutil.copyfileobj(r.raw, f) def get_track_genres(self, track_id): genres = self.get_track(track_id, 'track_genres') genre_ids = [] genre_titles = [] for genre in genres: genre_ids.append(genre['genre_id']) genre_titles.append(genre['genre_title']) return genre_ids, genre_titles def get_all_genres(self): df = pd.DataFrame(columns=['genre_parent_id', 'genre_title', 'genre_handle', 'genre_color']) df.index.rename('genre_id', inplace=True) page = 1 while True: url = self.BASE_URL + 'genres.json?limit=50' url += '&page={}&api_key={}'.format(page, self.api_key) r = requests.get(url) for genre in r.json()['dataset']: genre_id = int(genre.pop(df.index.name)) df.loc[genre_id] = genre assert (r.json()['page'] == str(page)) page += 1 if page > r.json()['total_pages']: break return df class Genres: def __init__(self, genres_df): self.df = genres_df def create_tree(self, roots, depth=None): if type(roots) is not list: roots = [roots] graph = pydot.Dot(graph_type='digraph', strict=True) def create_node(genre_id): title = self.df.at[genre_id, 'title'] ntracks = self.df.at[genre_id, '#tracks']
def create_tree(root_id, node_p, depth): if depth == 0: return children = self.df[self.df['parent'] == root_id] for child in children.iterrows(): genre_id = child[0] node_c = create_node(genre_id) graph.add_edge(pydot.Edge(node_p, node_c)) create_tree(genre_id, node_c, depth-1 if depth is not None else None) for root in roots: node_p = create_node(root) graph.add_node(node_p) create_tree(root, node_p, depth) return graph def find_roots(self): roots = [] for gid, row in self.df.iterrows(): parent = row['parent'] title = row['title'] if parent == 0: roots.append(gid) elif parent not in self.df.index: msg = '{} ({}) has parent {} which is missing'.format( gid, title, parent) raise RuntimeError(msg) return roots def load(filepath): filename = os.path.basename(filepath) if 'features' in filename: return pd.read_csv(filepath, index_col=0, header=[0, 1, 2]) if 'echonest' in filename: return pd.read_csv(filepath, index_col=0, header=[0, 1, 2]) if 'genres' in filename: return pd.read_csv(filepath, index_col=0) if 'tracks' in filename: tracks = pd.read_csv(filepath, index_col=0, header=[0, 1]) COLUMNS = [('track', 'tags'), ('album', 'tags'), ('artist', 'tags'), ('track', 'genres'), ('track', 'genres_all')] for column in COLUMNS: tracks[column] = tracks[column].map(ast.literal_eval) COLUMNS = [('track', 'date_created'), ('track', 'date_recorded'), ('album', 'date_created'), ('album', 'date_released'), ('artist', 'date_created'), ('artist', 'active_year_begin'), ('artist', 'active_year_end')] for column in COLUMNS: tracks[column] = pd.to_datetime(tracks[column]) SUBSETS = ('small', 'medium', 'large') tracks['set', 'subset'] = tracks['set', 'subset'].astype( 'category', categories=SUBSETS, ordered=True) COLUMNS = [('track', 'genre_top'), ('track', 'license'), ('album', 'type'), ('album', 'information'), ('artist', 'bio')] for column in COLUMNS: tracks[column] = tracks[column].astype('category') return tracks def get_audio_path(audio_dir, track_id): tid_str = '{:06d}'.format(track_id) return os.path.join(audio_dir, tid_str[:3], tid_str + '.mp3') class Loader: def load(self, filepath): raise NotImplemented() class RawAudioLoader(Loader): def __init__(self, sampling_rate=SAMPLING_RATE): self.sampling_rate = sampling_rate self.shape = (NB_AUDIO_SAMPLES * sampling_rate // SAMPLING_RATE, ) def load(self, filepath): return self._load(filepath)[:self.shape[0]] class LibrosaLoader(RawAudioLoader): def _load(self, filepath): import librosa sr = self.sampling_rate if self.sampling_rate != SAMPLING_RATE else None # kaiser_fast is 3x faster than kaiser_best #x, sr = librosa.load(filepath, sr=sr, res_type='kaiser_fast') x, sr = librosa.load(filepath, sr=sr) return x class AudioreadLoader(RawAudioLoader): def _load(self, filepath): import audioread a = audioread.audio_open(filepath) a.read_data() class PydubLoader(RawAudioLoader): def _load(self, filepath): from pydub import AudioSegment song = AudioSegment.from_file(filepath) song = song.set_channels(1) x = song.get_array_of_samples() # print(filepath) if song.channels != 2 else None return np.array(x) class FfmpegLoader(RawAudioLoader): def _load(self, filepath): """Fastest and less CPU intensive loading method.""" import subprocess as sp command = ['ffmpeg', '-i', filepath, '-f', 's16le', '-acodec', 'pcm_s16le', '-ac', '1'] # channels: 2 for stereo, 1 for mono if self.sampling_rate != SAMPLING_RATE: command.extend(['-ar', str(self.sampling_rate)]) command.append('-') # 30s at 44.1 kHz ~= 1.3e6 proc = sp.run(command, stdout=sp.PIPE, bufsize=10**7, stderr=sp.DEVNULL, check=True) return np.fromstring(proc.stdout, dtype="int16") def build_sample_loader(audio_dir, Y, loader): class SampleLoader: def __init__(self, tids, batch_size=4): self.lock1 = multiprocessing.Lock()
#name = self.df.at[genre_id, 'title'] + '\n' + str(genre_id) name = '"{}\n{} / {}"'.format(title, genre_id, ntracks) return pydot.Node(name)
random_line_split
main.rs
.x = 5; *r.y = 6; } assert_eq!(5, point.x); assert_eq!(6, point.y); point = Point { x: 0, ..point}; assert_eq!(6, point.y); struct Color(i32, i32, i32); let black = Color(17, 0, 0); let Color(r, _, _) = black; println!("{}", r); enum Message { Quit, ChangeColor(i32, i32, i32), Move { x: i32, y: i32 }, Write(String), } let v = vec!["Hello".to_string(), "World".to_string()]; let v1: Vec<Message> = v.into_iter().map(Message::Write).collect(); let x = 5; match x { 1 => println!("one"), 2 => println!("two"), 3 => println!("three"), 4 => println!("four"), 5 => println!("five"), 6 | 7 => println!("six or seven"), _ => println!("something else"), } let number = match x { 1 => "one", 2 => "two", 3 => "three", 4 => "four", 5 => "five", _ => "something else", }; let message = Message::Quit; match message { Message::Quit => println!("quit"), Message::ChangeColor(r, g, b) => println!("color"), Message::Move { x, y: new_name_for_y } => println!("move"), Message::Write(s) => println!("write"), }; let x = 1; let c = 'c'; match c { x => println!("x: {} c: {}", x, c), } println!("x: {}", x); let origin = Point { x: 0, y: 0 }; let Point { x, y } = origin; let tuple = (5, String::from("five")); let (x, _) = tuple; //string is not moved thanks to _ println!("Tuple is: {:?}", tuple); let (x, ..) = tuple; let mut x = 5; match x { ref name @ 1 ... 5 if *name < 5 => println!("one through four {}", name), ref name @ 1 ... 5 if *name >= 5 => println!("five {}", name), ref mut mr => println!("Got a mutable reference to {}", mr), } struct Circle { x: f64, y: f64, radius: f64, }
impl Circle { fn area(&self) -> f64 { std::f64::consts::PI * (self.radius * self.radius) } fn reference(&self) -> &Circle{ println!("taking self by reference!"); self } fn mutable_reference(&mut self) { println!("taking self by mutable reference!"); } fn takes_ownership(self) { println!("taking ownership of self!"); } fn new(x: f64, y: f64, radius: f64) -> Circle { Circle { x: x, y: y, radius: radius, } } } struct CircleBuilder { x: f64, y: f64, radius: f64, } let mut c = Circle { x: 0.0, y: 0.0, radius: 2.0 }; c = Circle::new(0.0, 0.0, 2.0); println!("{}", c.reference().area()); impl CircleBuilder { fn new() -> CircleBuilder { CircleBuilder { x: 0.0, y: 0.0, radius: 1.0, } } fn x(&mut self, coordinate: f64) -> &mut CircleBuilder { self.x = coordinate; self } fn y(&mut self, coordinate: f64) -> &mut CircleBuilder { self.y = coordinate; self } fn radius(&mut self, radius: f64) -> &mut CircleBuilder { self.radius = radius; self } fn finalize(&self) -> Circle { Circle { x: self.x, y: self.y, radius: self.radius } } } c = CircleBuilder::new().x(1.0) .y(2.0) .radius(2.0) .finalize();; println!("{}", c.reference().area()); let greeting = "Hello there."; // greeting: &'static str let mut s = "Hello".to_string(); // mut s: String fn takes_slice(slice: &str) { println!("Got: {}", slice); } takes_slice(&s); for c in s.chars() { print!("{}, ", c); } let c = s.chars().nth(0); let sl = { let tmp = &s[0..5]; println!("{}", tmp); }; let mut concat = s + "foo"; println!("{}", concat); let concat2 = "bar".to_string() + &concat; println!("{}", concat2); let x: Option<i32> = Some(5); fn takes_anything<T>(x: T) { // Do something with `x`. } takes_anything(concat2); struct PointGeneric<T> { x: T, y: T, } impl<T> PointGeneric<T> { fn swap(&mut self) { std::mem::swap(&mut self.x, &mut self.y); } } let int_origin = PointGeneric { x: 0, y: 0 }; let float_origin = PointGeneric { x: 0.0, y: 0.0 }; trait HasArea { fn area(&self) -> f64; fn is_larger(&self, &Self) -> bool; } impl HasArea for Circle { fn area(&self) -> f64 { std::f64::consts::PI * (self.radius * self.radius) } fn is_larger(&self, other: &Self) -> bool { self.area() > other.area() } } use std::fmt::Debug; fn print_area<T: HasArea>(shape: T) { println!("This shape has an area of {}", shape.area()); } fn test <T: HasArea + Debug>(){ } fn test2 <T>() where T : HasArea + Debug{ } let c = Circle { x: 0.0f64, y: 0.0f64, radius: 1.0f64, }; print_area(c); trait bar : HasArea { fn is_valid(&self) -> bool; fn is_invalid(&self) -> bool { !self.is_valid() } } #[derive(Debug)] struct deriving; impl Drop for Circle { fn drop(&mut self) { println!("Dropping!"); } } fn main() { let x = Circle { x: 0.0, y: 0.0, radius: 2.0 }; // Do stuff. } let option: Option<i32> = Some(5); match option { Some(x) => { println!("match!"); }, None => {}, } if option.is_some() { let x = option.unwrap(); println!("match!"); } if let Some(x) = option { println!("match!"); } trait FooBar { fn method(&self) -> String; } impl FooBar for u8 { fn method(&self) -> String { format!("u8: {}", *self) } } impl FooBar for String { fn method(&self) -> String { format!("string: {}", *self) } } fn do_something<T: FooBar>(x: T) { x.method(); } let x = 5u8; let y = "Hello".to_string(); do_something(x); do_something(y); fn do_something2(x: &FooBar) { x.method(); } let x = 5u8; //casting do_something2(&x as &FooBar); //coercing do_something2(&x); let add = |x| x + 1; println!("{}", add(2)); let mut num = 5; { let mut add_num = |x: i32| num += x; add_num(5); } assert_eq!(10, num); //move closure let mut num = 5; { let mut add_num = move |x: i32| num += x; add_num(5); } assert_eq!(5, num); fn call_with_one<F>(closure : F) -> i32 where F: Fn(i32) -> i32{ closure(1) } let answer = call_with_one(|x| x + 2); assert_eq!(3, answer); fn call_with_one2(some_closure: &Fn(i32) -> i32) -> i32 { some_closure(1) } let
random_line_split
main.rs
.x = 5; *r.y = 6; } assert_eq!(5, point.x); assert_eq!(6, point.y); point = Point { x: 0, ..point}; assert_eq!(6, point.y); struct Color(i32, i32, i32); let black = Color(17, 0, 0); let Color(r, _, _) = black; println!("{}", r); enum Message { Quit, ChangeColor(i32, i32, i32), Move { x: i32, y: i32 }, Write(String), } let v = vec!["Hello".to_string(), "World".to_string()]; let v1: Vec<Message> = v.into_iter().map(Message::Write).collect(); let x = 5; match x { 1 => println!("one"), 2 => println!("two"), 3 => println!("three"), 4 => println!("four"), 5 => println!("five"), 6 | 7 => println!("six or seven"), _ => println!("something else"), } let number = match x { 1 => "one", 2 => "two", 3 => "three", 4 => "four", 5 => "five", _ => "something else", }; let message = Message::Quit; match message { Message::Quit => println!("quit"), Message::ChangeColor(r, g, b) => println!("color"), Message::Move { x, y: new_name_for_y } => println!("move"), Message::Write(s) => println!("write"), }; let x = 1; let c = 'c'; match c { x => println!("x: {} c: {}", x, c), } println!("x: {}", x); let origin = Point { x: 0, y: 0 }; let Point { x, y } = origin; let tuple = (5, String::from("five")); let (x, _) = tuple; //string is not moved thanks to _ println!("Tuple is: {:?}", tuple); let (x, ..) = tuple; let mut x = 5; match x { ref name @ 1 ... 5 if *name < 5 => println!("one through four {}", name), ref name @ 1 ... 5 if *name >= 5 => println!("five {}", name), ref mut mr => println!("Got a mutable reference to {}", mr), } struct Circle { x: f64, y: f64, radius: f64, } impl Circle { fn area(&self) -> f64 { std::f64::consts::PI * (self.radius * self.radius) } fn reference(&self) -> &Circle{ println!("taking self by reference!"); self } fn mutable_reference(&mut self) { println!("taking self by mutable reference!"); } fn takes_ownership(self) { println!("taking ownership of self!"); } fn new(x: f64, y: f64, radius: f64) -> Circle { Circle { x: x, y: y, radius: radius, } } } struct CircleBuilder { x: f64, y: f64, radius: f64, } let mut c = Circle { x: 0.0, y: 0.0, radius: 2.0 }; c = Circle::new(0.0, 0.0, 2.0); println!("{}", c.reference().area()); impl CircleBuilder { fn new() -> CircleBuilder { CircleBuilder { x: 0.0, y: 0.0, radius: 1.0, } } fn x(&mut self, coordinate: f64) -> &mut CircleBuilder { self.x = coordinate; self } fn y(&mut self, coordinate: f64) -> &mut CircleBuilder { self.y = coordinate; self } fn radius(&mut self, radius: f64) -> &mut CircleBuilder { self.radius = radius; self } fn finalize(&self) -> Circle { Circle { x: self.x, y: self.y, radius: self.radius } } } c = CircleBuilder::new().x(1.0) .y(2.0) .radius(2.0) .finalize();; println!("{}", c.reference().area()); let greeting = "Hello there."; // greeting: &'static str let mut s = "Hello".to_string(); // mut s: String fn takes_slice(slice: &str) { println!("Got: {}", slice); } takes_slice(&s); for c in s.chars() { print!("{}, ", c); } let c = s.chars().nth(0); let sl = { let tmp = &s[0..5]; println!("{}", tmp); }; let mut concat = s + "foo"; println!("{}", concat); let concat2 = "bar".to_string() + &concat; println!("{}", concat2); let x: Option<i32> = Some(5); fn takes_anything<T>(x: T) { // Do something with `x`. } takes_anything(concat2); struct PointGeneric<T> { x: T, y: T, } impl<T> PointGeneric<T> { fn swap(&mut self) { std::mem::swap(&mut self.x, &mut self.y); } } let int_origin = PointGeneric { x: 0, y: 0 }; let float_origin = PointGeneric { x: 0.0, y: 0.0 }; trait HasArea { fn area(&self) -> f64; fn is_larger(&self, &Self) -> bool; } impl HasArea for Circle { fn area(&self) -> f64 { std::f64::consts::PI * (self.radius * self.radius) } fn is_larger(&self, other: &Self) -> bool { self.area() > other.area() } } use std::fmt::Debug; fn print_area<T: HasArea>(shape: T) { println!("This shape has an area of {}", shape.area()); } fn test <T: HasArea + Debug>(){ } fn test2 <T>() where T : HasArea + Debug{ } let c = Circle { x: 0.0f64, y: 0.0f64, radius: 1.0f64, }; print_area(c); trait bar : HasArea { fn is_valid(&self) -> bool; fn is_invalid(&self) -> bool { !self.is_valid() } } #[derive(Debug)] struct deriving; impl Drop for Circle { fn drop(&mut self) { println!("Dropping!"); } } fn main() { let x = Circle { x: 0.0, y: 0.0, radius: 2.0 }; // Do stuff. } let option: Option<i32> = Some(5); match option { Some(x) => { println!("match!"); }, None => {}, } if option.is_some() { let x = option.unwrap(); println!("match!"); } if let Some(x) = option { println!("match!"); } trait FooBar { fn method(&self) -> String; } impl FooBar for u8 { fn method(&self) -> String { format!("u8: {}", *self) } } impl FooBar for String { fn
(&self) -> String { format!("string: {}", *self) } } fn do_something<T: FooBar>(x: T) { x.method(); } let x = 5u8; let y = "Hello".to_string(); do_something(x); do_something(y); fn do_something2(x: &FooBar) { x.method(); } let x = 5u8; //casting do_something2(&x as &FooBar); //coercing do_something2(&x); let add = |x| x + 1; println!("{}", add(2)); let mut num = 5; { let mut add_num = |x: i32| num += x; add_num(5); } assert_eq!(10, num); //move closure let mut num = 5; { let mut add_num = move |x: i32| num += x; add_num(5); } assert_eq!(5, num); fn call_with_one<F>(closure : F) -> i32 where F: Fn(i32) -> i32{ closure(1) } let answer = call_with_one(|x| x + 2); assert_eq!(3, answer); fn call_with_one2(some_closure: &Fn(i32) -> i32) -> i32 { some_closure(1) }
method
identifier_name
main.rs
!("taking self by mutable reference!"); } fn takes_ownership(self) { println!("taking ownership of self!"); } fn new(x: f64, y: f64, radius: f64) -> Circle { Circle { x: x, y: y, radius: radius, } } } struct CircleBuilder { x: f64, y: f64, radius: f64, } let mut c = Circle { x: 0.0, y: 0.0, radius: 2.0 }; c = Circle::new(0.0, 0.0, 2.0); println!("{}", c.reference().area()); impl CircleBuilder { fn new() -> CircleBuilder { CircleBuilder { x: 0.0, y: 0.0, radius: 1.0, } } fn x(&mut self, coordinate: f64) -> &mut CircleBuilder { self.x = coordinate; self } fn y(&mut self, coordinate: f64) -> &mut CircleBuilder { self.y = coordinate; self } fn radius(&mut self, radius: f64) -> &mut CircleBuilder { self.radius = radius; self } fn finalize(&self) -> Circle { Circle { x: self.x, y: self.y, radius: self.radius } } } c = CircleBuilder::new().x(1.0) .y(2.0) .radius(2.0) .finalize();; println!("{}", c.reference().area()); let greeting = "Hello there."; // greeting: &'static str let mut s = "Hello".to_string(); // mut s: String fn takes_slice(slice: &str) { println!("Got: {}", slice); } takes_slice(&s); for c in s.chars() { print!("{}, ", c); } let c = s.chars().nth(0); let sl = { let tmp = &s[0..5]; println!("{}", tmp); }; let mut concat = s + "foo"; println!("{}", concat); let concat2 = "bar".to_string() + &concat; println!("{}", concat2); let x: Option<i32> = Some(5); fn takes_anything<T>(x: T) { // Do something with `x`. } takes_anything(concat2); struct PointGeneric<T> { x: T, y: T, } impl<T> PointGeneric<T> { fn swap(&mut self) { std::mem::swap(&mut self.x, &mut self.y); } } let int_origin = PointGeneric { x: 0, y: 0 }; let float_origin = PointGeneric { x: 0.0, y: 0.0 }; trait HasArea { fn area(&self) -> f64; fn is_larger(&self, &Self) -> bool; } impl HasArea for Circle { fn area(&self) -> f64 { std::f64::consts::PI * (self.radius * self.radius) } fn is_larger(&self, other: &Self) -> bool { self.area() > other.area() } } use std::fmt::Debug; fn print_area<T: HasArea>(shape: T) { println!("This shape has an area of {}", shape.area()); } fn test <T: HasArea + Debug>(){ } fn test2 <T>() where T : HasArea + Debug{ } let c = Circle { x: 0.0f64, y: 0.0f64, radius: 1.0f64, }; print_area(c); trait bar : HasArea { fn is_valid(&self) -> bool; fn is_invalid(&self) -> bool { !self.is_valid() } } #[derive(Debug)] struct deriving; impl Drop for Circle { fn drop(&mut self) { println!("Dropping!"); } } fn main() { let x = Circle { x: 0.0, y: 0.0, radius: 2.0 }; // Do stuff. } let option: Option<i32> = Some(5); match option { Some(x) => { println!("match!"); }, None => {}, } if option.is_some() { let x = option.unwrap(); println!("match!"); } if let Some(x) = option { println!("match!"); } trait FooBar { fn method(&self) -> String; } impl FooBar for u8 { fn method(&self) -> String { format!("u8: {}", *self) } } impl FooBar for String { fn method(&self) -> String { format!("string: {}", *self) } } fn do_something<T: FooBar>(x: T) { x.method(); } let x = 5u8; let y = "Hello".to_string(); do_something(x); do_something(y); fn do_something2(x: &FooBar) { x.method(); } let x = 5u8; //casting do_something2(&x as &FooBar); //coercing do_something2(&x); let add = |x| x + 1; println!("{}", add(2)); let mut num = 5; { let mut add_num = |x: i32| num += x; add_num(5); } assert_eq!(10, num); //move closure let mut num = 5; { let mut add_num = move |x: i32| num += x; add_num(5); } assert_eq!(5, num); fn call_with_one<F>(closure : F) -> i32 where F: Fn(i32) -> i32{ closure(1) } let answer = call_with_one(|x| x + 2); assert_eq!(3, answer); fn call_with_one2(some_closure: &Fn(i32) -> i32) -> i32 { some_closure(1) } let answer = call_with_one2(&|x| x + 2); assert_eq!(3, answer); fn call_with_ref<F>(some_closure:F) -> i32 where F: for<'a> Fn(&'a i32) -> i32 { let value = 0; some_closure(&value) } fn add_one(i: i32) -> i32 { i + 1 } let f = add_one; call_with_one2(&f); fn factory() -> Box<Fn(i32) -> i32> { let num = 5; Box::new(move |x| x + num) } let f = factory(); let answer = f(1); assert_eq!(6, answer); trait Foo2 { fn f(&self); } trait Bar2 { fn f(&self); } struct Baz; impl Foo2 for Baz { fn f(&self) { println!("Baz’s impl of Foo"); } } impl Bar2 for Baz { fn f(&self) { println!("Baz’s impl of Bar"); } } let b = Baz; Foo2::f(&b); Bar2::f(&b); println!("Hello in English: {}", phrases::english::greetings::hello()); println!("Hello in English: {}", hi()); //inline, several memory address //better than static const TOTO: i32 = 12; //same address for all use static mut TOTO2: i32 = 12; unsafe { TOTO2 = 2; } #[test] fn check() { assert_eq!(2, 1 + 1); } #[cfg(target_os = "macos")] mod macos_only { } type mytype = String; let s:mytype = "toto".to_string(); use std::result; enum ConcreteError { Foo, Bar, } type Result<T> = result::Result<T, ConcreteError>; let casty = TOTO as i64; use std::mem; unsafe { let a = [0u8, 1u8, 0u8, 0u8]; let b = mem::transmute::<[u8; 4], u32>(a); println!("{}", b); } trait Graph { type N; type E; fn has_edge(&self, &Self::N, &Self::N) -> bool; fn edges(&self, &Self::N) -> Vec<Self::E>; } struct Node; struct Edge; struct MyGraph; impl Graph for MyGraph { type N = Node; type E = Edge; fn has_edge(&self, n1: &Node, n2: &Node) -> bool { true } fn edges(&self, n: &Node) -> Vec<Edge> {
Vec::new() } }
identifier_body
static.go
{ return err } return nil } // ValidateHeader validates Header. // Returns the first error encountered. func ValidateHeader(header *Header) error { if header == nil { return consts.ErrNilHeader } tokenType := header.TokenTyp if tokenType < NoType || tokenType > Jet { return consts.ErrUnknownTokenType } alg := header.Alg if alg < NoAlg || alg > Hs512 { return consts.ErrUnknownAlgorithm } return nil } // ValidateBody validates Body. // Checks if token string has expired. // Returns the first error encountered. func ValidateBody(body *Body) error { if body == nil { return consts.ErrNilBody } if err := validation.ValidateUserUUID(body.UUID); err != nil { return err } permission := body.Permission if permission < NoPermission || permission > Admin { return consts.ErrUnknownPermission } if isExpired(body.ExpirationTimestamp) { return consts.ErrExpiredBody } return nil } // ValidateSecret checks if the secret is still valid and has not expired. // Returns an error if the Secret is not valid and has expired. func ValidateSecret(secret *pbauth.Secret) error { if secret == nil { return consts.ErrNilSecret } if strings.TrimSpace(secret.Key) == "" { return consts.ErrEmptySecret } createTime := secret.CreatedTimestamp if createTime == 0 || createTime > time.Now().UTC().Unix() { return consts.ErrInvalidSecretCreateTimestamp } if isExpired(secret.ExpirationTimestamp) { return consts.ErrExpiredSecret } return nil } func isExpired(timestamp int64) bool { if timestamp <= 0 || time.Now().UTC().Unix() >= timestamp { return true } return false } // NewToken generates token string using a header, body, and secret. // Return error if an error exists during signing. func NewToken(header *Header, body *Body, secret *pbauth.Secret) (string, error) { if err := ValidateHeader(header); err != nil { return "", err } if err := ValidateBody(body); err != nil
if err := ValidateSecret(secret); err != nil { return "", err } if body.Permission == Admin && header.Alg != Hs512 { return "", consts.ErrInvalidPermission } // Currently supports JWT, JET if header.TokenTyp != Jwt && header.TokenTyp != Jet { return "", consts.ErrUnknownTokenType } tokenString, err := getTokenSignature(header, body, secret) if err != nil { return "", err } return tokenString, nil } // getTokenSignature gets the token signature using the encoded header, body, and secret key. // Return error if an error exists during signing. func getTokenSignature(header *Header, body *Body, secret *pbauth.Secret) (string, error) { if err := ValidateHeader(header); err != nil { return "", err } if err := ValidateBody(body); err != nil { return "", err } if err := ValidateSecret(secret); err != nil { return "", err } if body.Permission == Admin && header.Alg != Hs512 { return "", consts.ErrInvalidPermission } if header.TokenTyp != Jwt && header.TokenTyp != Jet { return "", consts.ErrUnknownTokenType } // Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)> // 1. Encode the header encodedHeader, err := base64Encode(header) if err != nil { return "", err } // 2. Encode the body encodedBody, err := base64Encode(body) if err != nil { return "", err } // 3. Build <encoded header>.<encoded body> // 4. Build <hashed(<encoded header>.<encoded body>)> // 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)> return buildTokenSignature(encodedHeader, encodedBody, header.Alg, secret) } // buildTokenSignature builds the token signature using the encoded header, body, selected algorithm, and secret key. // Return error if an error exists during signing. func buildTokenSignature(encodedHeader string, encodedBody string, alg Algorithm, secret *pbauth.Secret) (string, error) { if strings.TrimSpace(encodedHeader) == "" { return "", consts.ErrInvalidEncodedHeader } if strings.TrimSpace(encodedBody) == "" { return "", consts.ErrInvalidEncodedBody } if err := ValidateSecret(secret); err != nil { return "", err } // 3. Build <encoded header>.<encoded body> var bufferHeaderBody bytes.Buffer bufferHeaderBody.WriteString(encodedHeader) bufferHeaderBody.WriteString(".") bufferHeaderBody.WriteString(encodedBody) encodedHeaderBody := bufferHeaderBody.String() // 4. Build <hashed(<encoded header>.<encoded body>)> encodedSignature, err := hashSignature(alg, encodedHeaderBody, secret) if err != nil { return "", nil } // 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)> var bufferTokenSignature bytes.Buffer bufferTokenSignature.WriteString(encodedHeaderBody) bufferTokenSignature.WriteString(".") bufferTokenSignature.WriteString(encodedSignature) return bufferTokenSignature.String(), nil } // base64Encode takes in a interface and encodes it as a string. // Returns a base 64 encoded string or error during marshalling. func base64Encode(src interface{}) (string, error) { if src == nil { return "", consts.ErrNilInterface } srcMarshal, err := json.Marshal(src) if err != nil { return "", err } srcString := string(srcMarshal) // TODO maybe use Trim return strings.TrimRight(base64.URLEncoding.EncodeToString([]byte(srcString)), "="), nil } // base64Encode takes in a base 64 encoded string. // Returns the actual string or an error of it fails to decode the string. func base64Decode(src string) (string, error) { if strings.TrimSpace(src) == "" { return "", consts.ErrEmptyString } if l := len(src) % 4; l > 0 { src += strings.Repeat("=", 4-l) } decoded, err := base64.URLEncoding.DecodeString(src) if err != nil { errMsg := fmt.Errorf("decoding error %s", err) return "", errMsg } return string(decoded), nil } // hashSignature generates a HMAC hash of a string using a secret func hashSignature(alg Algorithm, signatureValue string, secret *pbauth.Secret) (string, error) { if strings.TrimSpace(signatureValue) == "" { return "", consts.ErrInvalidSignatureValue } if err := ValidateSecret(secret); err != nil { return "", err } key := []byte(secret.Key) var h hash.Hash switch alg { case Hs256: h = hmac.New(sha256.New, key) case Hs512: h = hmac.New(sha512.New, key) default: return "", consts.ErrNoHashAlgorithm } h.Write([]byte(signatureValue)) return base64.URLEncoding.EncodeToString(h.Sum(nil)), nil } // isEquivalentHash validates a hash against a value func isEquivalentHash(alg Algorithm, signatureValue string, secret *pbauth.Secret, hashedValue string) bool { if err := ValidateSecret(secret); err != nil { return false } /* hashSignature cannot be reversed all you can do is hash the same character and compare it with a hashed value. If it evaluates to true, then the character is a what is in the hash. The isValidHash function only hashes the value with the secret and compared it with the hash. */ actualHashedValue, err := hashSignature(alg, signatureValue, secret) if err != nil { return false } return hashedValue == actualHashedValue } // ExtractUUID takes in a token string and extracts the UUID from the body. // Returns the uuid or an empty string due to an error. func ExtractUUID(tokenString string) string { tokenSignature := strings.Split(tokenString, ".") if len(tokenSignature) != 3 { return "" } decodedBody, err := base64Decode(tokenSignature[1]) if err != nil { return "" } body := &Body{} if err := json.Unmarshal([]byte(decodedBody), body); err != nil { return "" } if body == nil { return "" } if err := validation.ValidateUserUUID(body.UUID); err != nil { return "" } return body.UUID } // GenerateSecretKey generates a base64 URL-safe string // built from securely generated random bytes. // Number of bytes is determined by tokenSize. // Return error if system's secure random number generator fails. func GenerateSecretKey(tokenSize int) (string, error) { if tokenSize <= 0 { return "", consts.ErrInvalidTokenSize } keyGenLocker.Lock() defer keyGenLocker.Unlock() randomBytes := make([]byte, tokenSize) _, err := cryptorand.Read(randomBytes) if err != nil { return "", err } return base64.URLEncoding.EncodeToString(randomBytes), nil } // GenerateExpirationTimestamp returns
{ return "", err }
conditional_block
static.go
{ return err } return nil } // ValidateHeader validates Header. // Returns the first error encountered. func ValidateHeader(header *Header) error { if header == nil { return consts.ErrNilHeader } tokenType := header.TokenTyp if tokenType < NoType || tokenType > Jet { return consts.ErrUnknownTokenType } alg := header.Alg if alg < NoAlg || alg > Hs512 { return consts.ErrUnknownAlgorithm } return nil } // ValidateBody validates Body. // Checks if token string has expired. // Returns the first error encountered. func ValidateBody(body *Body) error { if body == nil { return consts.ErrNilBody } if err := validation.ValidateUserUUID(body.UUID); err != nil { return err } permission := body.Permission if permission < NoPermission || permission > Admin { return consts.ErrUnknownPermission } if isExpired(body.ExpirationTimestamp) { return consts.ErrExpiredBody } return nil } // ValidateSecret checks if the secret is still valid and has not expired. // Returns an error if the Secret is not valid and has expired. func ValidateSecret(secret *pbauth.Secret) error { if secret == nil { return consts.ErrNilSecret } if strings.TrimSpace(secret.Key) == "" { return consts.ErrEmptySecret } createTime := secret.CreatedTimestamp if createTime == 0 || createTime > time.Now().UTC().Unix() { return consts.ErrInvalidSecretCreateTimestamp } if isExpired(secret.ExpirationTimestamp) { return consts.ErrExpiredSecret } return nil } func isExpired(timestamp int64) bool { if timestamp <= 0 || time.Now().UTC().Unix() >= timestamp { return true } return false } // NewToken generates token string using a header, body, and secret. // Return error if an error exists during signing. func NewToken(header *Header, body *Body, secret *pbauth.Secret) (string, error) { if err := ValidateHeader(header); err != nil { return "", err } if err := ValidateBody(body); err != nil { return "", err } if err := ValidateSecret(secret); err != nil { return "", err } if body.Permission == Admin && header.Alg != Hs512 { return "", consts.ErrInvalidPermission } // Currently supports JWT, JET if header.TokenTyp != Jwt && header.TokenTyp != Jet { return "", consts.ErrUnknownTokenType } tokenString, err := getTokenSignature(header, body, secret) if err != nil { return "", err } return tokenString, nil } // getTokenSignature gets the token signature using the encoded header, body, and secret key. // Return error if an error exists during signing. func getTokenSignature(header *Header, body *Body, secret *pbauth.Secret) (string, error) { if err := ValidateHeader(header); err != nil { return "", err } if err := ValidateBody(body); err != nil { return "", err } if err := ValidateSecret(secret); err != nil { return "", err } if body.Permission == Admin && header.Alg != Hs512 { return "", consts.ErrInvalidPermission } if header.TokenTyp != Jwt && header.TokenTyp != Jet { return "", consts.ErrUnknownTokenType } // Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)> // 1. Encode the header encodedHeader, err := base64Encode(header) if err != nil { return "", err } // 2. Encode the body encodedBody, err := base64Encode(body) if err != nil { return "", err } // 3. Build <encoded header>.<encoded body> // 4. Build <hashed(<encoded header>.<encoded body>)> // 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)> return buildTokenSignature(encodedHeader, encodedBody, header.Alg, secret) } // buildTokenSignature builds the token signature using the encoded header, body, selected algorithm, and secret key. // Return error if an error exists during signing. func buildTokenSignature(encodedHeader string, encodedBody string, alg Algorithm, secret *pbauth.Secret) (string, error) { if strings.TrimSpace(encodedHeader) == "" { return "", consts.ErrInvalidEncodedHeader } if strings.TrimSpace(encodedBody) == "" { return "", consts.ErrInvalidEncodedBody } if err := ValidateSecret(secret); err != nil { return "", err } // 3. Build <encoded header>.<encoded body> var bufferHeaderBody bytes.Buffer bufferHeaderBody.WriteString(encodedHeader) bufferHeaderBody.WriteString(".") bufferHeaderBody.WriteString(encodedBody)
encodedSignature, err := hashSignature(alg, encodedHeaderBody, secret) if err != nil { return "", nil } // 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)> var bufferTokenSignature bytes.Buffer bufferTokenSignature.WriteString(encodedHeaderBody) bufferTokenSignature.WriteString(".") bufferTokenSignature.WriteString(encodedSignature) return bufferTokenSignature.String(), nil } // base64Encode takes in a interface and encodes it as a string. // Returns a base 64 encoded string or error during marshalling. func base64Encode(src interface{}) (string, error) { if src == nil { return "", consts.ErrNilInterface } srcMarshal, err := json.Marshal(src) if err != nil { return "", err } srcString := string(srcMarshal) // TODO maybe use Trim return strings.TrimRight(base64.URLEncoding.EncodeToString([]byte(srcString)), "="), nil } // base64Encode takes in a base 64 encoded string. // Returns the actual string or an error of it fails to decode the string. func base64Decode(src string) (string, error) { if strings.TrimSpace(src) == "" { return "", consts.ErrEmptyString } if l := len(src) % 4; l > 0 { src += strings.Repeat("=", 4-l) } decoded, err := base64.URLEncoding.DecodeString(src) if err != nil { errMsg := fmt.Errorf("decoding error %s", err) return "", errMsg } return string(decoded), nil } // hashSignature generates a HMAC hash of a string using a secret func hashSignature(alg Algorithm, signatureValue string, secret *pbauth.Secret) (string, error) { if strings.TrimSpace(signatureValue) == "" { return "", consts.ErrInvalidSignatureValue } if err := ValidateSecret(secret); err != nil { return "", err } key := []byte(secret.Key) var h hash.Hash switch alg { case Hs256: h = hmac.New(sha256.New, key) case Hs512: h = hmac.New(sha512.New, key) default: return "", consts.ErrNoHashAlgorithm } h.Write([]byte(signatureValue)) return base64.URLEncoding.EncodeToString(h.Sum(nil)), nil } // isEquivalentHash validates a hash against a value func isEquivalentHash(alg Algorithm, signatureValue string, secret *pbauth.Secret, hashedValue string) bool { if err := ValidateSecret(secret); err != nil { return false } /* hashSignature cannot be reversed all you can do is hash the same character and compare it with a hashed value. If it evaluates to true, then the character is a what is in the hash. The isValidHash function only hashes the value with the secret and compared it with the hash. */ actualHashedValue, err := hashSignature(alg, signatureValue, secret) if err != nil { return false } return hashedValue == actualHashedValue } // ExtractUUID takes in a token string and extracts the UUID from the body. // Returns the uuid or an empty string due to an error. func ExtractUUID(tokenString string) string { tokenSignature := strings.Split(tokenString, ".") if len(tokenSignature) != 3 { return "" } decodedBody, err := base64Decode(tokenSignature[1]) if err != nil { return "" } body := &Body{} if err := json.Unmarshal([]byte(decodedBody), body); err != nil { return "" } if body == nil { return "" } if err := validation.ValidateUserUUID(body.UUID); err != nil { return "" } return body.UUID } // GenerateSecretKey generates a base64 URL-safe string // built from securely generated random bytes. // Number of bytes is determined by tokenSize. // Return error if system's secure random number generator fails. func GenerateSecretKey(tokenSize int) (string, error) { if tokenSize <= 0 { return "", consts.ErrInvalidTokenSize } keyGenLocker.Lock() defer keyGenLocker.Unlock() randomBytes := make([]byte, tokenSize) _, err := cryptorand.Read(randomBytes) if err != nil { return "", err } return base64.URLEncoding.EncodeToString(randomBytes), nil } // GenerateExpirationTimestamp returns the expiration
encodedHeaderBody := bufferHeaderBody.String() // 4. Build <hashed(<encoded header>.<encoded body>)>
random_line_split
static.go
{ return err } return nil } // ValidateHeader validates Header. // Returns the first error encountered. func ValidateHeader(header *Header) error { if header == nil { return consts.ErrNilHeader } tokenType := header.TokenTyp if tokenType < NoType || tokenType > Jet { return consts.ErrUnknownTokenType } alg := header.Alg if alg < NoAlg || alg > Hs512 { return consts.ErrUnknownAlgorithm } return nil } // ValidateBody validates Body. // Checks if token string has expired. // Returns the first error encountered. func ValidateBody(body *Body) error { if body == nil { return consts.ErrNilBody } if err := validation.ValidateUserUUID(body.UUID); err != nil { return err } permission := body.Permission if permission < NoPermission || permission > Admin { return consts.ErrUnknownPermission } if isExpired(body.ExpirationTimestamp) { return consts.ErrExpiredBody } return nil } // ValidateSecret checks if the secret is still valid and has not expired. // Returns an error if the Secret is not valid and has expired. func ValidateSecret(secret *pbauth.Secret) error
func isExpired(timestamp int64) bool { if timestamp <= 0 || time.Now().UTC().Unix() >= timestamp { return true } return false } // NewToken generates token string using a header, body, and secret. // Return error if an error exists during signing. func NewToken(header *Header, body *Body, secret *pbauth.Secret) (string, error) { if err := ValidateHeader(header); err != nil { return "", err } if err := ValidateBody(body); err != nil { return "", err } if err := ValidateSecret(secret); err != nil { return "", err } if body.Permission == Admin && header.Alg != Hs512 { return "", consts.ErrInvalidPermission } // Currently supports JWT, JET if header.TokenTyp != Jwt && header.TokenTyp != Jet { return "", consts.ErrUnknownTokenType } tokenString, err := getTokenSignature(header, body, secret) if err != nil { return "", err } return tokenString, nil } // getTokenSignature gets the token signature using the encoded header, body, and secret key. // Return error if an error exists during signing. func getTokenSignature(header *Header, body *Body, secret *pbauth.Secret) (string, error) { if err := ValidateHeader(header); err != nil { return "", err } if err := ValidateBody(body); err != nil { return "", err } if err := ValidateSecret(secret); err != nil { return "", err } if body.Permission == Admin && header.Alg != Hs512 { return "", consts.ErrInvalidPermission } if header.TokenTyp != Jwt && header.TokenTyp != Jet { return "", consts.ErrUnknownTokenType } // Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)> // 1. Encode the header encodedHeader, err := base64Encode(header) if err != nil { return "", err } // 2. Encode the body encodedBody, err := base64Encode(body) if err != nil { return "", err } // 3. Build <encoded header>.<encoded body> // 4. Build <hashed(<encoded header>.<encoded body>)> // 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)> return buildTokenSignature(encodedHeader, encodedBody, header.Alg, secret) } // buildTokenSignature builds the token signature using the encoded header, body, selected algorithm, and secret key. // Return error if an error exists during signing. func buildTokenSignature(encodedHeader string, encodedBody string, alg Algorithm, secret *pbauth.Secret) (string, error) { if strings.TrimSpace(encodedHeader) == "" { return "", consts.ErrInvalidEncodedHeader } if strings.TrimSpace(encodedBody) == "" { return "", consts.ErrInvalidEncodedBody } if err := ValidateSecret(secret); err != nil { return "", err } // 3. Build <encoded header>.<encoded body> var bufferHeaderBody bytes.Buffer bufferHeaderBody.WriteString(encodedHeader) bufferHeaderBody.WriteString(".") bufferHeaderBody.WriteString(encodedBody) encodedHeaderBody := bufferHeaderBody.String() // 4. Build <hashed(<encoded header>.<encoded body>)> encodedSignature, err := hashSignature(alg, encodedHeaderBody, secret) if err != nil { return "", nil } // 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)> var bufferTokenSignature bytes.Buffer bufferTokenSignature.WriteString(encodedHeaderBody) bufferTokenSignature.WriteString(".") bufferTokenSignature.WriteString(encodedSignature) return bufferTokenSignature.String(), nil } // base64Encode takes in a interface and encodes it as a string. // Returns a base 64 encoded string or error during marshalling. func base64Encode(src interface{}) (string, error) { if src == nil { return "", consts.ErrNilInterface } srcMarshal, err := json.Marshal(src) if err != nil { return "", err } srcString := string(srcMarshal) // TODO maybe use Trim return strings.TrimRight(base64.URLEncoding.EncodeToString([]byte(srcString)), "="), nil } // base64Encode takes in a base 64 encoded string. // Returns the actual string or an error of it fails to decode the string. func base64Decode(src string) (string, error) { if strings.TrimSpace(src) == "" { return "", consts.ErrEmptyString } if l := len(src) % 4; l > 0 { src += strings.Repeat("=", 4-l) } decoded, err := base64.URLEncoding.DecodeString(src) if err != nil { errMsg := fmt.Errorf("decoding error %s", err) return "", errMsg } return string(decoded), nil } // hashSignature generates a HMAC hash of a string using a secret func hashSignature(alg Algorithm, signatureValue string, secret *pbauth.Secret) (string, error) { if strings.TrimSpace(signatureValue) == "" { return "", consts.ErrInvalidSignatureValue } if err := ValidateSecret(secret); err != nil { return "", err } key := []byte(secret.Key) var h hash.Hash switch alg { case Hs256: h = hmac.New(sha256.New, key) case Hs512: h = hmac.New(sha512.New, key) default: return "", consts.ErrNoHashAlgorithm } h.Write([]byte(signatureValue)) return base64.URLEncoding.EncodeToString(h.Sum(nil)), nil } // isEquivalentHash validates a hash against a value func isEquivalentHash(alg Algorithm, signatureValue string, secret *pbauth.Secret, hashedValue string) bool { if err := ValidateSecret(secret); err != nil { return false } /* hashSignature cannot be reversed all you can do is hash the same character and compare it with a hashed value. If it evaluates to true, then the character is a what is in the hash. The isValidHash function only hashes the value with the secret and compared it with the hash. */ actualHashedValue, err := hashSignature(alg, signatureValue, secret) if err != nil { return false } return hashedValue == actualHashedValue } // ExtractUUID takes in a token string and extracts the UUID from the body. // Returns the uuid or an empty string due to an error. func ExtractUUID(tokenString string) string { tokenSignature := strings.Split(tokenString, ".") if len(tokenSignature) != 3 { return "" } decodedBody, err := base64Decode(tokenSignature[1]) if err != nil { return "" } body := &Body{} if err := json.Unmarshal([]byte(decodedBody), body); err != nil { return "" } if body == nil { return "" } if err := validation.ValidateUserUUID(body.UUID); err != nil { return "" } return body.UUID } // GenerateSecretKey generates a base64 URL-safe string // built from securely generated random bytes. // Number of bytes is determined by tokenSize. // Return error if system's secure random number generator fails. func GenerateSecretKey(tokenSize int) (string, error) { if tokenSize <= 0 { return "", consts.ErrInvalidTokenSize } keyGenLocker.Lock() defer keyGenLocker.Unlock() randomBytes := make([]byte, tokenSize) _, err := cryptorand.Read(randomBytes) if err != nil { return "", err } return base64.URLEncoding.EncodeToString(randomBytes), nil } // GenerateExpirationTimestamp returns
{ if secret == nil { return consts.ErrNilSecret } if strings.TrimSpace(secret.Key) == "" { return consts.ErrEmptySecret } createTime := secret.CreatedTimestamp if createTime == 0 || createTime > time.Now().UTC().Unix() { return consts.ErrInvalidSecretCreateTimestamp } if isExpired(secret.ExpirationTimestamp) { return consts.ErrExpiredSecret } return nil }
identifier_body
static.go
{ return err } return nil } // ValidateHeader validates Header. // Returns the first error encountered. func ValidateHeader(header *Header) error { if header == nil { return consts.ErrNilHeader } tokenType := header.TokenTyp if tokenType < NoType || tokenType > Jet { return consts.ErrUnknownTokenType } alg := header.Alg if alg < NoAlg || alg > Hs512 { return consts.ErrUnknownAlgorithm } return nil } // ValidateBody validates Body. // Checks if token string has expired. // Returns the first error encountered. func ValidateBody(body *Body) error { if body == nil { return consts.ErrNilBody } if err := validation.ValidateUserUUID(body.UUID); err != nil { return err } permission := body.Permission if permission < NoPermission || permission > Admin { return consts.ErrUnknownPermission } if isExpired(body.ExpirationTimestamp) { return consts.ErrExpiredBody } return nil } // ValidateSecret checks if the secret is still valid and has not expired. // Returns an error if the Secret is not valid and has expired. func ValidateSecret(secret *pbauth.Secret) error { if secret == nil { return consts.ErrNilSecret } if strings.TrimSpace(secret.Key) == "" { return consts.ErrEmptySecret } createTime := secret.CreatedTimestamp if createTime == 0 || createTime > time.Now().UTC().Unix() { return consts.ErrInvalidSecretCreateTimestamp } if isExpired(secret.ExpirationTimestamp) { return consts.ErrExpiredSecret } return nil } func isExpired(timestamp int64) bool { if timestamp <= 0 || time.Now().UTC().Unix() >= timestamp { return true } return false } // NewToken generates token string using a header, body, and secret. // Return error if an error exists during signing. func NewToken(header *Header, body *Body, secret *pbauth.Secret) (string, error) { if err := ValidateHeader(header); err != nil { return "", err } if err := ValidateBody(body); err != nil { return "", err } if err := ValidateSecret(secret); err != nil { return "", err } if body.Permission == Admin && header.Alg != Hs512 { return "", consts.ErrInvalidPermission } // Currently supports JWT, JET if header.TokenTyp != Jwt && header.TokenTyp != Jet { return "", consts.ErrUnknownTokenType } tokenString, err := getTokenSignature(header, body, secret) if err != nil { return "", err } return tokenString, nil } // getTokenSignature gets the token signature using the encoded header, body, and secret key. // Return error if an error exists during signing. func
(header *Header, body *Body, secret *pbauth.Secret) (string, error) { if err := ValidateHeader(header); err != nil { return "", err } if err := ValidateBody(body); err != nil { return "", err } if err := ValidateSecret(secret); err != nil { return "", err } if body.Permission == Admin && header.Alg != Hs512 { return "", consts.ErrInvalidPermission } if header.TokenTyp != Jwt && header.TokenTyp != Jet { return "", consts.ErrUnknownTokenType } // Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)> // 1. Encode the header encodedHeader, err := base64Encode(header) if err != nil { return "", err } // 2. Encode the body encodedBody, err := base64Encode(body) if err != nil { return "", err } // 3. Build <encoded header>.<encoded body> // 4. Build <hashed(<encoded header>.<encoded body>)> // 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)> return buildTokenSignature(encodedHeader, encodedBody, header.Alg, secret) } // buildTokenSignature builds the token signature using the encoded header, body, selected algorithm, and secret key. // Return error if an error exists during signing. func buildTokenSignature(encodedHeader string, encodedBody string, alg Algorithm, secret *pbauth.Secret) (string, error) { if strings.TrimSpace(encodedHeader) == "" { return "", consts.ErrInvalidEncodedHeader } if strings.TrimSpace(encodedBody) == "" { return "", consts.ErrInvalidEncodedBody } if err := ValidateSecret(secret); err != nil { return "", err } // 3. Build <encoded header>.<encoded body> var bufferHeaderBody bytes.Buffer bufferHeaderBody.WriteString(encodedHeader) bufferHeaderBody.WriteString(".") bufferHeaderBody.WriteString(encodedBody) encodedHeaderBody := bufferHeaderBody.String() // 4. Build <hashed(<encoded header>.<encoded body>)> encodedSignature, err := hashSignature(alg, encodedHeaderBody, secret) if err != nil { return "", nil } // 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)> var bufferTokenSignature bytes.Buffer bufferTokenSignature.WriteString(encodedHeaderBody) bufferTokenSignature.WriteString(".") bufferTokenSignature.WriteString(encodedSignature) return bufferTokenSignature.String(), nil } // base64Encode takes in a interface and encodes it as a string. // Returns a base 64 encoded string or error during marshalling. func base64Encode(src interface{}) (string, error) { if src == nil { return "", consts.ErrNilInterface } srcMarshal, err := json.Marshal(src) if err != nil { return "", err } srcString := string(srcMarshal) // TODO maybe use Trim return strings.TrimRight(base64.URLEncoding.EncodeToString([]byte(srcString)), "="), nil } // base64Encode takes in a base 64 encoded string. // Returns the actual string or an error of it fails to decode the string. func base64Decode(src string) (string, error) { if strings.TrimSpace(src) == "" { return "", consts.ErrEmptyString } if l := len(src) % 4; l > 0 { src += strings.Repeat("=", 4-l) } decoded, err := base64.URLEncoding.DecodeString(src) if err != nil { errMsg := fmt.Errorf("decoding error %s", err) return "", errMsg } return string(decoded), nil } // hashSignature generates a HMAC hash of a string using a secret func hashSignature(alg Algorithm, signatureValue string, secret *pbauth.Secret) (string, error) { if strings.TrimSpace(signatureValue) == "" { return "", consts.ErrInvalidSignatureValue } if err := ValidateSecret(secret); err != nil { return "", err } key := []byte(secret.Key) var h hash.Hash switch alg { case Hs256: h = hmac.New(sha256.New, key) case Hs512: h = hmac.New(sha512.New, key) default: return "", consts.ErrNoHashAlgorithm } h.Write([]byte(signatureValue)) return base64.URLEncoding.EncodeToString(h.Sum(nil)), nil } // isEquivalentHash validates a hash against a value func isEquivalentHash(alg Algorithm, signatureValue string, secret *pbauth.Secret, hashedValue string) bool { if err := ValidateSecret(secret); err != nil { return false } /* hashSignature cannot be reversed all you can do is hash the same character and compare it with a hashed value. If it evaluates to true, then the character is a what is in the hash. The isValidHash function only hashes the value with the secret and compared it with the hash. */ actualHashedValue, err := hashSignature(alg, signatureValue, secret) if err != nil { return false } return hashedValue == actualHashedValue } // ExtractUUID takes in a token string and extracts the UUID from the body. // Returns the uuid or an empty string due to an error. func ExtractUUID(tokenString string) string { tokenSignature := strings.Split(tokenString, ".") if len(tokenSignature) != 3 { return "" } decodedBody, err := base64Decode(tokenSignature[1]) if err != nil { return "" } body := &Body{} if err := json.Unmarshal([]byte(decodedBody), body); err != nil { return "" } if body == nil { return "" } if err := validation.ValidateUserUUID(body.UUID); err != nil { return "" } return body.UUID } // GenerateSecretKey generates a base64 URL-safe string // built from securely generated random bytes. // Number of bytes is determined by tokenSize. // Return error if system's secure random number generator fails. func GenerateSecretKey(tokenSize int) (string, error) { if tokenSize <= 0 { return "", consts.ErrInvalidTokenSize } keyGenLocker.Lock() defer keyGenLocker.Unlock() randomBytes := make([]byte, tokenSize) _, err := cryptorand.Read(randomBytes) if err != nil { return "", err } return base64.URLEncoding.EncodeToString(randomBytes), nil } // GenerateExpirationTimestamp returns the
getTokenSignature
identifier_name
reader.rs
Context { fn validate(&self) -> Result<()> { if self.max_concurrency == 0 { return Err(Error::msg("max concurrency must be at least 1.")); } Ok(()) } } impl ReaderContext { fn default_reader_threads() -> usize { 1 } } /// A given query payload that describes how the reader should /// search the index. #[derive(Debug, Deserialize)] pub struct QueryPayload { /// The query(s) itself. query: QuerySelector, /// The amount of results to limit by. #[serde(default = "QueryPayload::default_limit")] limit: usize, /// The amount of documents to skip before getting the results. #[serde(default)] offset: usize, /// A specified field to order results by, this defaults to the /// score of the indexed documents (relevancy). order_by: Option<String>, /// How to sort the data (asc/desc). #[serde(default)] sort: Sort, } impl QueryPayload { fn default_limit() -> usize { 20 } } /// What order to sort the returned data. #[derive(Debug, Copy, Clone, Deserialize)] #[serde(rename_all = "lowercase")] pub enum Sort { /// Sort the data in ascending order. Asc, /// Sort the data in descending order. (Default) Desc, }
#[derive(Debug, Serialize)] pub struct QueryResults { /// The retrieved documents. pub(crate) hits: Vec<DocumentHit>, /// The total amount of documents matching the search count: usize, /// The amount of time taken to search in seconds. time_taken: f32, } impl QueryResults { #[inline] pub fn len(&self) -> usize { self.hits.len() } } /// Attaches an order by clause to the collector. /// /// This collected the values with be returned in the order according to the /// given field value. fn order_and_search<R: AsScore + tantivy::fastfield::FastValue>( searcher: &Searcher, field: Field, query: &dyn Query, collector: TopDocs, executor: &Executor, ) -> Result<(Vec<(R, DocAddress)>, usize)> { let collector = collector.order_by_fast_field(field); searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from) } /// Performs the search operation and processes the returned results. fn process_search<S: AsScore>( searcher: &Searcher, schema: &Schema, top_docs: Vec<(S, DocAddress)>, ) -> Result<Vec<DocumentHit>> { let mut hits = Vec::with_capacity(top_docs.len()); for (ratio, ref_address) in top_docs { let retrieved_doc = searcher.doc(ref_address)?; let mut doc = schema.to_named_doc(&retrieved_doc); let id = doc.0 .remove("_id") .ok_or_else(|| Error::msg("document has been missed labeled (missing primary key '_id'), the dataset is invalid"))?; if let Value::U64(v) = id[0] { hits.push(DocumentHit { doc, document_id: v, score: ratio.as_score(), }); } else { return Err(Error::msg("document has been missed labeled (missing identifier tag), the dataset is invalid")); } } Ok(hits) } /// Orders the search results by the given field with a given sort (ASC, DESC) /// /// This function is super messy just because of all the type inference /// so any contributions to clean this up would be very appreciated. fn order_or_sort( sort: Sort, field: Field, query: &dyn Query, schema: &Schema, searcher: &Searcher, collector: TopDocs, executor: &Executor, ) -> Result<(Vec<DocumentHit>, usize)> { let field_type = schema.get_field_entry(field).field_type(); if let Sort::Desc = sort { return match field_type { FieldType::I64(_) => { let out: (Vec<(i64, DocAddress)>, usize) = order_and_search(&searcher, field, query, collector, executor)?; Ok((process_search(&searcher, schema, out.0)?, out.1)) }, FieldType::U64(_) => { let out: (Vec<(u64, DocAddress)>, usize) = order_and_search(&searcher, field, query, collector, executor)?; Ok((process_search(&searcher, schema, out.0)?, out.1)) }, FieldType::F64(_) => { let out: (Vec<(f64, DocAddress)>, usize) = order_and_search(&searcher, field, query, collector, executor)?; Ok((process_search(&searcher, schema, out.0)?, out.1)) }, FieldType::Date(_) => { let out: (Vec<(DateTime, DocAddress)>, usize) = order_and_search(&searcher, field, query, collector, executor)?; Ok((process_search(&searcher, schema, out.0)?, out.1)) }, _ => Err(Error::msg("field is not a fast field")), }; } let out = match field_type { FieldType::I64(_) => { let collector = collector.custom_score(move |segment_reader: &SegmentReader| { let reader = segment_reader .fast_fields() .i64(field) .expect("field exists"); move |doc: DocId| { let value: i64 = reader.get(doc); std::cmp::Reverse(value) } }); let out: (Vec<(Reverse<i64>, DocAddress)>, usize) = searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from)?; (process_search(&searcher, schema, out.0)?, out.1) }, FieldType::U64(_) => { let collector = collector.custom_score(move |segment_reader: &SegmentReader| { let reader = segment_reader .fast_fields() .u64(field) .expect("field exists"); move |doc: DocId| { let value: u64 = reader.get(doc); std::cmp::Reverse(value) } }); let out: (Vec<(Reverse<u64>, DocAddress)>, usize) = searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from)?; (process_search(&searcher, schema, out.0)?, out.1) }, FieldType::F64(_) => { let collector = collector.custom_score(move |segment_reader: &SegmentReader| { let reader = segment_reader .fast_fields() .f64(field) .expect("field exists"); move |doc: DocId| { let value: f64 = reader.get(doc); std::cmp::Reverse(value) } }); let out: (Vec<(Reverse<f64>, DocAddress)>, usize) = searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from)?; (process_search(&searcher, schema, out.0)?, out.1) }, FieldType::Date(_) => { let collector = collector.custom_score(move |segment_reader: &SegmentReader| { let reader = segment_reader .fast_fields() .date(field) .expect("field exists"); move |doc: DocId| { let value: DateTime = reader.get(doc); std::cmp::Reverse(value) } }); let out: (Vec<(Reverse<DateTime>, DocAddress)>, usize) = searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from)?; (process_search(&searcher, schema, out.0)?, out.1) }, _ => return Err(Error::msg("field is not a fast field")), }; return Ok(out); } /// The reader of the given index. /// /// This manages all searches on the index which encompasses the concurrency /// limiters and thread pool execution. /// /// Each index should only have on `Reader` instance. pub(crate) struct Reader { /// The executor pool. pool: crate::ReaderExecutor, /// The query factory system. query_handler: QueryBuilder, } impl Reader { /// Creates a new reader from the given index context. pub(crate) async fn create(ctx: &IndexContext) -> Result<Self> { let reader: IndexReader = ctx .index .reader_builder() .reload_policy(ReloadPolicy::OnCommit) .num_searchers(ctx.reader_ctx.max_concurrency) .try_into()?; info!( "[ READER @ {} ] index reader created with reload policy=OnCommit, num_searchers={}", &ctx.name, ctx.reader_ctx.max_concurrency, ); let pool = { let pool = SearcherExecutorPool::create( reader, ctx.reader_ctx.reader_threads, ctx.reader_ctx.max_concurrency, ) .await?; Arc::new(pool) }; info!( "[ READER @ {} ]
impl Default for Sort { fn default() -> Self { Self::Desc } }
random_line_split
reader.rs
Context { fn validate(&self) -> Result<()> { if self.max_concurrency == 0 { return Err(Error::msg("max concurrency must be at least 1.")); } Ok(()) } } impl ReaderContext { fn default_reader_threads() -> usize { 1 } } /// A given query payload that describes how the reader should /// search the index. #[derive(Debug, Deserialize)] pub struct QueryPayload { /// The query(s) itself. query: QuerySelector, /// The amount of results to limit by. #[serde(default = "QueryPayload::default_limit")] limit: usize, /// The amount of documents to skip before getting the results. #[serde(default)] offset: usize, /// A specified field to order results by, this defaults to the /// score of the indexed documents (relevancy). order_by: Option<String>, /// How to sort the data (asc/desc). #[serde(default)] sort: Sort, } impl QueryPayload { fn default_limit() -> usize { 20 } } /// What order to sort the returned data. #[derive(Debug, Copy, Clone, Deserialize)] #[serde(rename_all = "lowercase")] pub enum Sort { /// Sort the data in ascending order. Asc, /// Sort the data in descending order. (Default) Desc, } impl Default for Sort { fn default() -> Self { Self::Desc } } #[derive(Debug, Serialize)] pub struct QueryResults { /// The retrieved documents. pub(crate) hits: Vec<DocumentHit>, /// The total amount of documents matching the search count: usize, /// The amount of time taken to search in seconds. time_taken: f32, } impl QueryResults { #[inline] pub fn len(&self) -> usize { self.hits.len() } } /// Attaches an order by clause to the collector. /// /// This collected the values with be returned in the order according to the /// given field value. fn order_and_search<R: AsScore + tantivy::fastfield::FastValue>( searcher: &Searcher, field: Field, query: &dyn Query, collector: TopDocs, executor: &Executor, ) -> Result<(Vec<(R, DocAddress)>, usize)> { let collector = collector.order_by_fast_field(field); searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from) } /// Performs the search operation and processes the returned results. fn process_search<S: AsScore>( searcher: &Searcher, schema: &Schema, top_docs: Vec<(S, DocAddress)>, ) -> Result<Vec<DocumentHit>> { let mut hits = Vec::with_capacity(top_docs.len()); for (ratio, ref_address) in top_docs { let retrieved_doc = searcher.doc(ref_address)?; let mut doc = schema.to_named_doc(&retrieved_doc); let id = doc.0 .remove("_id") .ok_or_else(|| Error::msg("document has been missed labeled (missing primary key '_id'), the dataset is invalid"))?; if let Value::U64(v) = id[0] { hits.push(DocumentHit { doc, document_id: v, score: ratio.as_score(), }); } else { return Err(Error::msg("document has been missed labeled (missing identifier tag), the dataset is invalid")); } } Ok(hits) } /// Orders the search results by the given field with a given sort (ASC, DESC) /// /// This function is super messy just because of all the type inference /// so any contributions to clean this up would be very appreciated. fn order_or_sort( sort: Sort, field: Field, query: &dyn Query, schema: &Schema, searcher: &Searcher, collector: TopDocs, executor: &Executor, ) -> Result<(Vec<DocumentHit>, usize)> { let field_type = schema.get_field_entry(field).field_type(); if let Sort::Desc = sort { return match field_type { FieldType::I64(_) => { let out: (Vec<(i64, DocAddress)>, usize) = order_and_search(&searcher, field, query, collector, executor)?; Ok((process_search(&searcher, schema, out.0)?, out.1)) }, FieldType::U64(_) => { let out: (Vec<(u64, DocAddress)>, usize) = order_and_search(&searcher, field, query, collector, executor)?; Ok((process_search(&searcher, schema, out.0)?, out.1)) }, FieldType::F64(_) => { let out: (Vec<(f64, DocAddress)>, usize) = order_and_search(&searcher, field, query, collector, executor)?; Ok((process_search(&searcher, schema, out.0)?, out.1)) }, FieldType::Date(_) =>
, _ => Err(Error::msg("field is not a fast field")), }; } let out = match field_type { FieldType::I64(_) => { let collector = collector.custom_score(move |segment_reader: &SegmentReader| { let reader = segment_reader .fast_fields() .i64(field) .expect("field exists"); move |doc: DocId| { let value: i64 = reader.get(doc); std::cmp::Reverse(value) } }); let out: (Vec<(Reverse<i64>, DocAddress)>, usize) = searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from)?; (process_search(&searcher, schema, out.0)?, out.1) }, FieldType::U64(_) => { let collector = collector.custom_score(move |segment_reader: &SegmentReader| { let reader = segment_reader .fast_fields() .u64(field) .expect("field exists"); move |doc: DocId| { let value: u64 = reader.get(doc); std::cmp::Reverse(value) } }); let out: (Vec<(Reverse<u64>, DocAddress)>, usize) = searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from)?; (process_search(&searcher, schema, out.0)?, out.1) }, FieldType::F64(_) => { let collector = collector.custom_score(move |segment_reader: &SegmentReader| { let reader = segment_reader .fast_fields() .f64(field) .expect("field exists"); move |doc: DocId| { let value: f64 = reader.get(doc); std::cmp::Reverse(value) } }); let out: (Vec<(Reverse<f64>, DocAddress)>, usize) = searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from)?; (process_search(&searcher, schema, out.0)?, out.1) }, FieldType::Date(_) => { let collector = collector.custom_score(move |segment_reader: &SegmentReader| { let reader = segment_reader .fast_fields() .date(field) .expect("field exists"); move |doc: DocId| { let value: DateTime = reader.get(doc); std::cmp::Reverse(value) } }); let out: (Vec<(Reverse<DateTime>, DocAddress)>, usize) = searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from)?; (process_search(&searcher, schema, out.0)?, out.1) }, _ => return Err(Error::msg("field is not a fast field")), }; return Ok(out); } /// The reader of the given index. /// /// This manages all searches on the index which encompasses the concurrency /// limiters and thread pool execution. /// /// Each index should only have on `Reader` instance. pub(crate) struct Reader { /// The executor pool. pool: crate::ReaderExecutor, /// The query factory system. query_handler: QueryBuilder, } impl Reader { /// Creates a new reader from the given index context. pub(crate) async fn create(ctx: &IndexContext) -> Result<Self> { let reader: IndexReader = ctx .index .reader_builder() .reload_policy(ReloadPolicy::OnCommit) .num_searchers(ctx.reader_ctx.max_concurrency) .try_into()?; info!( "[ READER @ {} ] index reader created with reload policy=OnCommit, num_searchers={}", &ctx.name, ctx.reader_ctx.max_concurrency, ); let pool = { let pool = SearcherExecutorPool::create( reader, ctx.reader_ctx.reader_threads, ctx.reader_ctx.max_concurrency, ) .await?; Arc::new(pool) }; info!( "[ READER @ {}
{ let out: (Vec<(DateTime, DocAddress)>, usize) = order_and_search(&searcher, field, query, collector, executor)?; Ok((process_search(&searcher, schema, out.0)?, out.1)) }
conditional_block
reader.rs
Context { fn
(&self) -> Result<()> { if self.max_concurrency == 0 { return Err(Error::msg("max concurrency must be at least 1.")); } Ok(()) } } impl ReaderContext { fn default_reader_threads() -> usize { 1 } } /// A given query payload that describes how the reader should /// search the index. #[derive(Debug, Deserialize)] pub struct QueryPayload { /// The query(s) itself. query: QuerySelector, /// The amount of results to limit by. #[serde(default = "QueryPayload::default_limit")] limit: usize, /// The amount of documents to skip before getting the results. #[serde(default)] offset: usize, /// A specified field to order results by, this defaults to the /// score of the indexed documents (relevancy). order_by: Option<String>, /// How to sort the data (asc/desc). #[serde(default)] sort: Sort, } impl QueryPayload { fn default_limit() -> usize { 20 } } /// What order to sort the returned data. #[derive(Debug, Copy, Clone, Deserialize)] #[serde(rename_all = "lowercase")] pub enum Sort { /// Sort the data in ascending order. Asc, /// Sort the data in descending order. (Default) Desc, } impl Default for Sort { fn default() -> Self { Self::Desc } } #[derive(Debug, Serialize)] pub struct QueryResults { /// The retrieved documents. pub(crate) hits: Vec<DocumentHit>, /// The total amount of documents matching the search count: usize, /// The amount of time taken to search in seconds. time_taken: f32, } impl QueryResults { #[inline] pub fn len(&self) -> usize { self.hits.len() } } /// Attaches an order by clause to the collector. /// /// This collected the values with be returned in the order according to the /// given field value. fn order_and_search<R: AsScore + tantivy::fastfield::FastValue>( searcher: &Searcher, field: Field, query: &dyn Query, collector: TopDocs, executor: &Executor, ) -> Result<(Vec<(R, DocAddress)>, usize)> { let collector = collector.order_by_fast_field(field); searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from) } /// Performs the search operation and processes the returned results. fn process_search<S: AsScore>( searcher: &Searcher, schema: &Schema, top_docs: Vec<(S, DocAddress)>, ) -> Result<Vec<DocumentHit>> { let mut hits = Vec::with_capacity(top_docs.len()); for (ratio, ref_address) in top_docs { let retrieved_doc = searcher.doc(ref_address)?; let mut doc = schema.to_named_doc(&retrieved_doc); let id = doc.0 .remove("_id") .ok_or_else(|| Error::msg("document has been missed labeled (missing primary key '_id'), the dataset is invalid"))?; if let Value::U64(v) = id[0] { hits.push(DocumentHit { doc, document_id: v, score: ratio.as_score(), }); } else { return Err(Error::msg("document has been missed labeled (missing identifier tag), the dataset is invalid")); } } Ok(hits) } /// Orders the search results by the given field with a given sort (ASC, DESC) /// /// This function is super messy just because of all the type inference /// so any contributions to clean this up would be very appreciated. fn order_or_sort( sort: Sort, field: Field, query: &dyn Query, schema: &Schema, searcher: &Searcher, collector: TopDocs, executor: &Executor, ) -> Result<(Vec<DocumentHit>, usize)> { let field_type = schema.get_field_entry(field).field_type(); if let Sort::Desc = sort { return match field_type { FieldType::I64(_) => { let out: (Vec<(i64, DocAddress)>, usize) = order_and_search(&searcher, field, query, collector, executor)?; Ok((process_search(&searcher, schema, out.0)?, out.1)) }, FieldType::U64(_) => { let out: (Vec<(u64, DocAddress)>, usize) = order_and_search(&searcher, field, query, collector, executor)?; Ok((process_search(&searcher, schema, out.0)?, out.1)) }, FieldType::F64(_) => { let out: (Vec<(f64, DocAddress)>, usize) = order_and_search(&searcher, field, query, collector, executor)?; Ok((process_search(&searcher, schema, out.0)?, out.1)) }, FieldType::Date(_) => { let out: (Vec<(DateTime, DocAddress)>, usize) = order_and_search(&searcher, field, query, collector, executor)?; Ok((process_search(&searcher, schema, out.0)?, out.1)) }, _ => Err(Error::msg("field is not a fast field")), }; } let out = match field_type { FieldType::I64(_) => { let collector = collector.custom_score(move |segment_reader: &SegmentReader| { let reader = segment_reader .fast_fields() .i64(field) .expect("field exists"); move |doc: DocId| { let value: i64 = reader.get(doc); std::cmp::Reverse(value) } }); let out: (Vec<(Reverse<i64>, DocAddress)>, usize) = searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from)?; (process_search(&searcher, schema, out.0)?, out.1) }, FieldType::U64(_) => { let collector = collector.custom_score(move |segment_reader: &SegmentReader| { let reader = segment_reader .fast_fields() .u64(field) .expect("field exists"); move |doc: DocId| { let value: u64 = reader.get(doc); std::cmp::Reverse(value) } }); let out: (Vec<(Reverse<u64>, DocAddress)>, usize) = searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from)?; (process_search(&searcher, schema, out.0)?, out.1) }, FieldType::F64(_) => { let collector = collector.custom_score(move |segment_reader: &SegmentReader| { let reader = segment_reader .fast_fields() .f64(field) .expect("field exists"); move |doc: DocId| { let value: f64 = reader.get(doc); std::cmp::Reverse(value) } }); let out: (Vec<(Reverse<f64>, DocAddress)>, usize) = searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from)?; (process_search(&searcher, schema, out.0)?, out.1) }, FieldType::Date(_) => { let collector = collector.custom_score(move |segment_reader: &SegmentReader| { let reader = segment_reader .fast_fields() .date(field) .expect("field exists"); move |doc: DocId| { let value: DateTime = reader.get(doc); std::cmp::Reverse(value) } }); let out: (Vec<(Reverse<DateTime>, DocAddress)>, usize) = searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from)?; (process_search(&searcher, schema, out.0)?, out.1) }, _ => return Err(Error::msg("field is not a fast field")), }; return Ok(out); } /// The reader of the given index. /// /// This manages all searches on the index which encompasses the concurrency /// limiters and thread pool execution. /// /// Each index should only have on `Reader` instance. pub(crate) struct Reader { /// The executor pool. pool: crate::ReaderExecutor, /// The query factory system. query_handler: QueryBuilder, } impl Reader { /// Creates a new reader from the given index context. pub(crate) async fn create(ctx: &IndexContext) -> Result<Self> { let reader: IndexReader = ctx .index .reader_builder() .reload_policy(ReloadPolicy::OnCommit) .num_searchers(ctx.reader_ctx.max_concurrency) .try_into()?; info!( "[ READER @ {} ] index reader created with reload policy=OnCommit, num_searchers={}", &ctx.name, ctx.reader_ctx.max_concurrency, ); let pool = { let pool = SearcherExecutorPool::create( reader, ctx.reader_ctx.reader_threads, ctx.reader_ctx.max_concurrency, ) .await?; Arc::new(pool) }; info!( "[ READER @ {} ]
validate
identifier_name
reader.rs
pub(crate) hits: Vec<DocumentHit>, /// The total amount of documents matching the search count: usize, /// The amount of time taken to search in seconds. time_taken: f32, } impl QueryResults { #[inline] pub fn len(&self) -> usize { self.hits.len() } } /// Attaches an order by clause to the collector. /// /// This collected the values with be returned in the order according to the /// given field value. fn order_and_search<R: AsScore + tantivy::fastfield::FastValue>( searcher: &Searcher, field: Field, query: &dyn Query, collector: TopDocs, executor: &Executor, ) -> Result<(Vec<(R, DocAddress)>, usize)> { let collector = collector.order_by_fast_field(field); searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from) } /// Performs the search operation and processes the returned results. fn process_search<S: AsScore>( searcher: &Searcher, schema: &Schema, top_docs: Vec<(S, DocAddress)>, ) -> Result<Vec<DocumentHit>> { let mut hits = Vec::with_capacity(top_docs.len()); for (ratio, ref_address) in top_docs { let retrieved_doc = searcher.doc(ref_address)?; let mut doc = schema.to_named_doc(&retrieved_doc); let id = doc.0 .remove("_id") .ok_or_else(|| Error::msg("document has been missed labeled (missing primary key '_id'), the dataset is invalid"))?; if let Value::U64(v) = id[0] { hits.push(DocumentHit { doc, document_id: v, score: ratio.as_score(), }); } else { return Err(Error::msg("document has been missed labeled (missing identifier tag), the dataset is invalid")); } } Ok(hits) } /// Orders the search results by the given field with a given sort (ASC, DESC) /// /// This function is super messy just because of all the type inference /// so any contributions to clean this up would be very appreciated. fn order_or_sort( sort: Sort, field: Field, query: &dyn Query, schema: &Schema, searcher: &Searcher, collector: TopDocs, executor: &Executor, ) -> Result<(Vec<DocumentHit>, usize)> { let field_type = schema.get_field_entry(field).field_type(); if let Sort::Desc = sort { return match field_type { FieldType::I64(_) => { let out: (Vec<(i64, DocAddress)>, usize) = order_and_search(&searcher, field, query, collector, executor)?; Ok((process_search(&searcher, schema, out.0)?, out.1)) }, FieldType::U64(_) => { let out: (Vec<(u64, DocAddress)>, usize) = order_and_search(&searcher, field, query, collector, executor)?; Ok((process_search(&searcher, schema, out.0)?, out.1)) }, FieldType::F64(_) => { let out: (Vec<(f64, DocAddress)>, usize) = order_and_search(&searcher, field, query, collector, executor)?; Ok((process_search(&searcher, schema, out.0)?, out.1)) }, FieldType::Date(_) => { let out: (Vec<(DateTime, DocAddress)>, usize) = order_and_search(&searcher, field, query, collector, executor)?; Ok((process_search(&searcher, schema, out.0)?, out.1)) }, _ => Err(Error::msg("field is not a fast field")), }; } let out = match field_type { FieldType::I64(_) => { let collector = collector.custom_score(move |segment_reader: &SegmentReader| { let reader = segment_reader .fast_fields() .i64(field) .expect("field exists"); move |doc: DocId| { let value: i64 = reader.get(doc); std::cmp::Reverse(value) } }); let out: (Vec<(Reverse<i64>, DocAddress)>, usize) = searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from)?; (process_search(&searcher, schema, out.0)?, out.1) }, FieldType::U64(_) => { let collector = collector.custom_score(move |segment_reader: &SegmentReader| { let reader = segment_reader .fast_fields() .u64(field) .expect("field exists"); move |doc: DocId| { let value: u64 = reader.get(doc); std::cmp::Reverse(value) } }); let out: (Vec<(Reverse<u64>, DocAddress)>, usize) = searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from)?; (process_search(&searcher, schema, out.0)?, out.1) }, FieldType::F64(_) => { let collector = collector.custom_score(move |segment_reader: &SegmentReader| { let reader = segment_reader .fast_fields() .f64(field) .expect("field exists"); move |doc: DocId| { let value: f64 = reader.get(doc); std::cmp::Reverse(value) } }); let out: (Vec<(Reverse<f64>, DocAddress)>, usize) = searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from)?; (process_search(&searcher, schema, out.0)?, out.1) }, FieldType::Date(_) => { let collector = collector.custom_score(move |segment_reader: &SegmentReader| { let reader = segment_reader .fast_fields() .date(field) .expect("field exists"); move |doc: DocId| { let value: DateTime = reader.get(doc); std::cmp::Reverse(value) } }); let out: (Vec<(Reverse<DateTime>, DocAddress)>, usize) = searcher .search_with_executor(query, &(collector, Count), executor) .map_err(Error::from)?; (process_search(&searcher, schema, out.0)?, out.1) }, _ => return Err(Error::msg("field is not a fast field")), }; return Ok(out); } /// The reader of the given index. /// /// This manages all searches on the index which encompasses the concurrency /// limiters and thread pool execution. /// /// Each index should only have on `Reader` instance. pub(crate) struct Reader { /// The executor pool. pool: crate::ReaderExecutor, /// The query factory system. query_handler: QueryBuilder, } impl Reader { /// Creates a new reader from the given index context. pub(crate) async fn create(ctx: &IndexContext) -> Result<Self> { let reader: IndexReader = ctx .index .reader_builder() .reload_policy(ReloadPolicy::OnCommit) .num_searchers(ctx.reader_ctx.max_concurrency) .try_into()?; info!( "[ READER @ {} ] index reader created with reload policy=OnCommit, num_searchers={}", &ctx.name, ctx.reader_ctx.max_concurrency, ); let pool = { let pool = SearcherExecutorPool::create( reader, ctx.reader_ctx.reader_threads, ctx.reader_ctx.max_concurrency, ) .await?; Arc::new(pool) }; info!( "[ READER @ {} ] executor pool has successfully started! max_concurrency={}, total_threads={}", &ctx.name, ctx.reader_ctx.max_concurrency, ctx.reader_ctx.max_concurrency * ctx.reader_ctx.reader_threads ); let query_ctx = ctx.query_ctx.clone(); let query_handler = QueryBuilder::new( query_ctx, ctx.stop_words.clone(), ctx.correction_manager.clone(), &ctx.index, pool.clone(), ); info!( "[ QUERY-BUILDER @ {} ] query builder constructed with config: fast-fuzzy={} strip-stop-words={}.", &ctx.name, ctx.query_ctx.use_fast_fuzzy, ctx.query_ctx.strip_stop_words, ); Ok(Self { pool, query_handler, }) } /// Gets a singular document from the given id. /// /// If no document is found an error is raised without context. pub(crate) async fn get_document(&self, id: DocumentId) -> Result<DocumentHit>
{ let id_field = self.query_handler.id_field(); let document = self .pool .spawn(move |searcher, executor| { let qry = TermQuery::new( Term::from_field_u64(id_field, id), IndexRecordOption::Basic, ); let mut results = searcher.search_with_executor( &qry, &TopDocs::with_limit(1), executor, )?; if results.len() == 0 { return Err(Error::msg(format!( "no document exists with id: '{}'", id )));
identifier_body
LibphysGRU.py
d) self.identity = tf.eye(Sd) return E, U, W, V, b, c, [None, None] def GRUnn(self, out_prev, x_t): E, U, W, V, b, c = self.E, self.U, self.W, self.V, self.b, self.c Hd, Sd, Bd = self.signal2model.hidden_dim, self.signal2model.signal_dim, tf.shape(x_t)[0] coversion_ones = tf.ones((1, Bd), dtype=tf.float32, name="conversion_matrix") # s_prev, o_prev, l_prev = out_prev s_prev, o_prev = out_prev def GRU(last_input, gru_params): s_g_prev, u, w, b = gru_params z = tf.nn.sigmoid(tf.matmul(u[0], last_input) + tf.matmul(w[0], s_g_prev) + tf.matmul(tf.reshape(b[0], (Hd, 1)), coversion_ones)) r = tf.nn.sigmoid(tf.matmul(u[1], last_input) + tf.matmul(w[1], s_g_prev) + tf.matmul(tf.reshape(b[1], (Hd, 1)), coversion_ones)) value = tf.matmul(u[2], last_input) + tf.matmul(w[2], s_g_prev * r) + \ tf.matmul(tf.reshape(b[2], (Hd, 1)), coversion_ones) s_candidate = tf.nn.tanh(value) output = tf.add(((tf.ones_like(z) - z) * s_candidate), (z * s_g_prev), name="out_GRU") return output # x_e -> (Hd x Mb) x_e = tf.gather(self.E, x_t, axis=1) s_t_ = [] s_t_.append(GRU(x_e, [s_prev[0], U[0], W[0], b[0]])) s_t_.append(GRU(s_t_[0], [s_prev[1], U[1], W[1], b[1]])) s_t_.append(GRU(s_t_[1], [s_prev[2], U[2], W[2], b[2]])) s_t = tf.stack(s_t_) # tf.scan(GRU, (s_prev, self.U, self.W, self.b), initializer=x_e, parallel_iterations=1, name="states") logits = tf.matmul(self.V, s_t[-1]) + tf.matmul(tf.reshape(self.c, (Sd, 1)), coversion_ones) o_t = tf.nn.softmax(logits, axis=2) return [s_t, o_t]#, logits] def feed_forward_predict(self, x_batch): initial_s = tf.zeros((self.signal2model.n_grus, self.signal2model.hidden_dim, tf.shape(x_batch)[1]), dtype=np.float32) initial_out = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32) # initial_l = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32) # x_batch = (N x Bd) - N (samples); Bd - Batch dimension # [s, o, l] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out, initial_l], parallel_iterations=1, # name="network_output") [_, o] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out], parallel_iterations=1, name="network_output") return o def feed_forward_predict_with_states(self, x_batch): initial_s = tf.zeros((self.signal2model.n_grus, self.signal2model.hidden_dim, tf.shape(x_batch)[1]), dtype=np.float32) initial_out = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32) # initial_l = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32) # x_batch = (N x Bd) - N (samples); Bd - Batch dimension
# [s, o, l] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out, initial_l], parallel_iterations=1, # name="network_output") [s, o] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out], parallel_iterations=1, name="network_output") return [s, o] def calculate_predictions(self): # MAP NOT WORKING: # shape(X)[0] -> Windows # shape(X)[1] -> Samples # n_batches = int(signal2model.batch_size / self.signal2model.mini_batch_size) # N = tf.shape(self.X)[1] # print(X) # get the matrices from E with tf.gather(E, X, axis=1, name="X_e") # transpose these matrices for (batch_size, HD, N) # reshape to enter map, where each minibatch is entered at the same time (n_batches, mini_batch, HD, N) # transpose to enter the DNN inside -> (n_batches, N, mini_batch) return self.feed_forward_predict(tf.transpose(self.X)) def to_one_hot_vector_in_mini_batches(self, matrix): return self.get_one_hot(matrix) def get_one_hot(self, columns): return tf.gather(self.identity, columns) def calculate_cross_entropy(self): return None # logits = tf.transpose(self.logits, perm=[2, 0, 1]) # n_batches = int(self.signal2model.batch_size / self.signal2model.mini_batch_size) # y = tf.reshape(self.Y, (n_batches, self.signal2model.mini_batch_size, tf.shape(self.Y)[1])) # self.full_loss = tf.losses.sparse_softmax_cross_entropy_with_logits(labels=y, logits=logits) # return self.full_loss def calculate_mse(self): _y = self.to_one_hot_vector_in_mini_batches(self.Y) return tf.reduce_mean(tf.subtract(tf.transpose(self.out, perm=[2, 0, 1]), _y) ** 2, axis=2, name="mse") def calculate_mse_vector_loss(self, x, y): with tf.variable_scope('vector_loss'): return tf.reduce_mean(self.calculate_minibatch_mse(x, y), axis=0, name="vector_loss") def calculate_mse_loss(self): return tf.reduce_mean(self.calculate_minibatch_mse(), axis=0, name="loss") @property def loss_op(self): """ An Operation that takes one optimization step. """ return self.loss def init_optimizer(self): trainables = self.parameters grads = tf.gradients(self.loss, trainables) grad_var_pairs = zip(grads, trainables) # grads, _ = tf.clip_by_global_norm(grads, clip_norm=1.0) # self.learning_rate = tf.train.exponential_decay( # self.signal2model.learning_rate_val, self.epoch, self.signal2model.count_to_break_max, # self.signal2model.decay, staircase=True) # with tf.device('/gpu:1'): optimizer = tf.train.RMSPropOptimizer(self.learning_rate_gpu) # optimizer = tf.train.AdamOptimizer(self.learning_rate) self._optimize_op = optimizer.apply_gradients(grad_var_pairs) @property def optimize_op(self): """ An Operation that takes one optimization step. """ return self._optimize_op def train(self, X, Y, signal2model=None): self.batch_size += np.shape(X)[0] self.init_time = time.time() plt.ion() if signal2model is not None: self.signal2model = signal2model plt.ion() condition_not_met = True history = [] self.epoch = 0 self.loss_history = [] tf.summary.scalar('loss', self.loss) merged = tf.summary.merge_all() train_writer = tf.summary.FileWriter('train', self.session.graph) # run_options = tf.RunOptions(report_tensor_allocations_upon_oom=True) # db_url = 'postgres://belo:passsword@localhost/postgres' # experiments, steps, model_params = labnotebook.initialize(db_url) # model_desc = {'loss': 0.} # experiment = labnotebook.start_experiment(model_desc=model_desc) tf.global_variables_initializer() while condition_not_met: self.epoch += 1 # tic = time.time() random_indexes = np.random.permutation(self.signal2model.batch_size) groups = np.reshape(random_indexes, (int(self.signal2model.batch_size/self.signal2model.mini_batch_size), self.signal2model.mini_batch_size)) for group in groups: dictionary = self.shuffle(X, Y, group) op, group_loss = self.session.run( [self.optimize_op, self.loss_op], feed_dict=dictionary)#, options=run_options) # labnotebook.step_experiment(experiment, # timestep=str(self.epoch), # trainacc=0, # valacc=0, # trainloss=str(group_loss)) # print("toc: {0} secs".format(time
random_line_split
LibphysGRU.py
, U, W, V, b, c, [None, None] def GRUnn(self, out_prev, x_t): E, U, W, V, b, c = self.E, self.U, self.W, self.V, self.b, self.c Hd, Sd, Bd = self.signal2model.hidden_dim, self.signal2model.signal_dim, tf.shape(x_t)[0] coversion_ones = tf.ones((1, Bd), dtype=tf.float32, name="conversion_matrix") # s_prev, o_prev, l_prev = out_prev s_prev, o_prev = out_prev def GRU(last_input, gru_params): s_g_prev, u, w, b = gru_params z = tf.nn.sigmoid(tf.matmul(u[0], last_input) + tf.matmul(w[0], s_g_prev) + tf.matmul(tf.reshape(b[0], (Hd, 1)), coversion_ones)) r = tf.nn.sigmoid(tf.matmul(u[1], last_input) + tf.matmul(w[1], s_g_prev) + tf.matmul(tf.reshape(b[1], (Hd, 1)), coversion_ones)) value = tf.matmul(u[2], last_input) + tf.matmul(w[2], s_g_prev * r) + \ tf.matmul(tf.reshape(b[2], (Hd, 1)), coversion_ones) s_candidate = tf.nn.tanh(value) output = tf.add(((tf.ones_like(z) - z) * s_candidate), (z * s_g_prev), name="out_GRU") return output # x_e -> (Hd x Mb) x_e = tf.gather(self.E, x_t, axis=1) s_t_ = [] s_t_.append(GRU(x_e, [s_prev[0], U[0], W[0], b[0]])) s_t_.append(GRU(s_t_[0], [s_prev[1], U[1], W[1], b[1]])) s_t_.append(GRU(s_t_[1], [s_prev[2], U[2], W[2], b[2]])) s_t = tf.stack(s_t_) # tf.scan(GRU, (s_prev, self.U, self.W, self.b), initializer=x_e, parallel_iterations=1, name="states") logits = tf.matmul(self.V, s_t[-1]) + tf.matmul(tf.reshape(self.c, (Sd, 1)), coversion_ones) o_t = tf.nn.softmax(logits, axis=2) return [s_t, o_t]#, logits] def feed_forward_predict(self, x_batch): initial_s = tf.zeros((self.signal2model.n_grus, self.signal2model.hidden_dim, tf.shape(x_batch)[1]), dtype=np.float32) initial_out = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32) # initial_l = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32) # x_batch = (N x Bd) - N (samples); Bd - Batch dimension # [s, o, l] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out, initial_l], parallel_iterations=1, # name="network_output") [_, o] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out], parallel_iterations=1, name="network_output") return o def feed_forward_predict_with_states(self, x_batch): initial_s = tf.zeros((self.signal2model.n_grus, self.signal2model.hidden_dim, tf.shape(x_batch)[1]), dtype=np.float32) initial_out = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32) # initial_l = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32) # x_batch = (N x Bd) - N (samples); Bd - Batch dimension # [s, o, l] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out, initial_l], parallel_iterations=1, # name="network_output") [s, o] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out], parallel_iterations=1, name="network_output") return [s, o] def calculate_predictions(self): # MAP NOT WORKING: # shape(X)[0] -> Windows # shape(X)[1] -> Samples # n_batches = int(signal2model.batch_size / self.signal2model.mini_batch_size) # N = tf.shape(self.X)[1] # print(X) # get the matrices from E with tf.gather(E, X, axis=1, name="X_e") # transpose these matrices for (batch_size, HD, N) # reshape to enter map, where each minibatch is entered at the same time (n_batches, mini_batch, HD, N) # transpose to enter the DNN inside -> (n_batches, N, mini_batch) return self.feed_forward_predict(tf.transpose(self.X)) def to_one_hot_vector_in_mini_batches(self, matrix): return self.get_one_hot(matrix) def get_one_hot(self, columns): return tf.gather(self.identity, columns) def calculate_cross_entropy(self): return None # logits = tf.transpose(self.logits, perm=[2, 0, 1]) # n_batches = int(self.signal2model.batch_size / self.signal2model.mini_batch_size) # y = tf.reshape(self.Y, (n_batches, self.signal2model.mini_batch_size, tf.shape(self.Y)[1])) # self.full_loss = tf.losses.sparse_softmax_cross_entropy_with_logits(labels=y, logits=logits) # return self.full_loss def calculate_mse(self): _y = self.to_one_hot_vector_in_mini_batches(self.Y) return tf.reduce_mean(tf.subtract(tf.transpose(self.out, perm=[2, 0, 1]), _y) ** 2, axis=2, name="mse") def calculate_mse_vector_loss(self, x, y): with tf.variable_scope('vector_loss'): return tf.reduce_mean(self.calculate_minibatch_mse(x, y), axis=0, name="vector_loss") def calculate_mse_loss(self): return tf.reduce_mean(self.calculate_minibatch_mse(), axis=0, name="loss") @property def loss_op(self): """ An Operation that takes one optimization step. """ return self.loss def init_optimizer(self): trainables = self.parameters grads = tf.gradients(self.loss, trainables) grad_var_pairs = zip(grads, trainables) # grads, _ = tf.clip_by_global_norm(grads, clip_norm=1.0) # self.learning_rate = tf.train.exponential_decay( # self.signal2model.learning_rate_val, self.epoch, self.signal2model.count_to_break_max, # self.signal2model.decay, staircase=True) # with tf.device('/gpu:1'): optimizer = tf.train.RMSPropOptimizer(self.learning_rate_gpu) # optimizer = tf.train.AdamOptimizer(self.learning_rate) self._optimize_op = optimizer.apply_gradients(grad_var_pairs) @property def optimize_op(self): """ An Operation that takes one optimization step. """ return self._optimize_op def train(self, X, Y, signal2model=None): self.batch_size += np.shape(X)[0] self.init_time = time.time() plt.ion() if signal2model is not None: self.signal2model = signal2model plt.ion() condition_not_met = True history = [] self.epoch = 0 self.loss_history = [] tf.summary.scalar('loss', self.loss) merged = tf.summary.merge_all() train_writer = tf.summary.FileWriter('train', self.session.graph) # run_options = tf.RunOptions(report_tensor_allocations_upon_oom=True) # db_url = 'postgres://belo:passsword@localhost/postgres' # experiments, steps, model_params = labnotebook.initialize(db_url) # model_desc = {'loss': 0.} # experiment = labnotebook.start_experiment(model_desc=model_desc) tf.global_variables_initializer() while condition_not_met:
self.epoch += 1 # tic = time.time() random_indexes = np.random.permutation(self.signal2model.batch_size) groups = np.reshape(random_indexes, (int(self.signal2model.batch_size/self.signal2model.mini_batch_size), self.signal2model.mini_batch_size)) for group in groups: dictionary = self.shuffle(X, Y, group) op, group_loss = self.session.run( [self.optimize_op, self.loss_op], feed_dict=dictionary)#, options=run_options) # labnotebook.step_experiment(experiment, # timestep=str(self.epoch), # trainacc=0, # valacc=0, # trainloss=str(group_loss)) # print("toc: {0} secs".format(time.time()-tic)) # new_tic = time.time()
conditional_block
LibphysGRU.py
) self.identity = tf.eye(Sd) return E, U, W, V, b, c, [None, None] def GRUnn(self, out_prev, x_t): E, U, W, V, b, c = self.E, self.U, self.W, self.V, self.b, self.c Hd, Sd, Bd = self.signal2model.hidden_dim, self.signal2model.signal_dim, tf.shape(x_t)[0] coversion_ones = tf.ones((1, Bd), dtype=tf.float32, name="conversion_matrix") # s_prev, o_prev, l_prev = out_prev s_prev, o_prev = out_prev def GRU(last_input, gru_params): s_g_prev, u, w, b = gru_params z = tf.nn.sigmoid(tf.matmul(u[0], last_input) + tf.matmul(w[0], s_g_prev) + tf.matmul(tf.reshape(b[0], (Hd, 1)), coversion_ones)) r = tf.nn.sigmoid(tf.matmul(u[1], last_input) + tf.matmul(w[1], s_g_prev) + tf.matmul(tf.reshape(b[1], (Hd, 1)), coversion_ones)) value = tf.matmul(u[2], last_input) + tf.matmul(w[2], s_g_prev * r) + \ tf.matmul(tf.reshape(b[2], (Hd, 1)), coversion_ones) s_candidate = tf.nn.tanh(value) output = tf.add(((tf.ones_like(z) - z) * s_candidate), (z * s_g_prev), name="out_GRU") return output # x_e -> (Hd x Mb) x_e = tf.gather(self.E, x_t, axis=1) s_t_ = [] s_t_.append(GRU(x_e, [s_prev[0], U[0], W[0], b[0]])) s_t_.append(GRU(s_t_[0], [s_prev[1], U[1], W[1], b[1]])) s_t_.append(GRU(s_t_[1], [s_prev[2], U[2], W[2], b[2]])) s_t = tf.stack(s_t_) # tf.scan(GRU, (s_prev, self.U, self.W, self.b), initializer=x_e, parallel_iterations=1, name="states") logits = tf.matmul(self.V, s_t[-1]) + tf.matmul(tf.reshape(self.c, (Sd, 1)), coversion_ones) o_t = tf.nn.softmax(logits, axis=2) return [s_t, o_t]#, logits] def feed_forward_predict(self, x_batch): initial_s = tf.zeros((self.signal2model.n_grus, self.signal2model.hidden_dim, tf.shape(x_batch)[1]), dtype=np.float32) initial_out = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32) # initial_l = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32) # x_batch = (N x Bd) - N (samples); Bd - Batch dimension # [s, o, l] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out, initial_l], parallel_iterations=1, # name="network_output") [_, o] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out], parallel_iterations=1, name="network_output") return o def feed_forward_predict_with_states(self, x_batch): initial_s = tf.zeros((self.signal2model.n_grus, self.signal2model.hidden_dim, tf.shape(x_batch)[1]), dtype=np.float32) initial_out = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32) # initial_l = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32) # x_batch = (N x Bd) - N (samples); Bd - Batch dimension # [s, o, l] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out, initial_l], parallel_iterations=1, # name="network_output") [s, o] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out], parallel_iterations=1, name="network_output") return [s, o] def calculate_predictions(self): # MAP NOT WORKING: # shape(X)[0] -> Windows # shape(X)[1] -> Samples # n_batches = int(signal2model.batch_size / self.signal2model.mini_batch_size) # N = tf.shape(self.X)[1] # print(X) # get the matrices from E with tf.gather(E, X, axis=1, name="X_e") # transpose these matrices for (batch_size, HD, N) # reshape to enter map, where each minibatch is entered at the same time (n_batches, mini_batch, HD, N) # transpose to enter the DNN inside -> (n_batches, N, mini_batch) return self.feed_forward_predict(tf.transpose(self.X)) def to_one_hot_vector_in_mini_batches(self, matrix): return self.get_one_hot(matrix) def get_one_hot(self, columns): return tf.gather(self.identity, columns) def calculate_cross_entropy(self): return None # logits = tf.transpose(self.logits, perm=[2, 0, 1]) # n_batches = int(self.signal2model.batch_size / self.signal2model.mini_batch_size) # y = tf.reshape(self.Y, (n_batches, self.signal2model.mini_batch_size, tf.shape(self.Y)[1])) # self.full_loss = tf.losses.sparse_softmax_cross_entropy_with_logits(labels=y, logits=logits) # return self.full_loss def calculate_mse(self): _y = self.to_one_hot_vector_in_mini_batches(self.Y) return tf.reduce_mean(tf.subtract(tf.transpose(self.out, perm=[2, 0, 1]), _y) ** 2, axis=2, name="mse") def calculate_mse_vector_loss(self, x, y): with tf.variable_scope('vector_loss'): return tf.reduce_mean(self.calculate_minibatch_mse(x, y), axis=0, name="vector_loss") def calculate_mse_loss(self): return tf.reduce_mean(self.calculate_minibatch_mse(), axis=0, name="loss") @property def loss_op(self): """ An Operation that takes one optimization step. """ return self.loss def init_optimizer(self):
@property def optimize_op(self): """ An Operation that takes one optimization step. """ return self._optimize_op def train(self, X, Y, signal2model=None): self.batch_size += np.shape(X)[0] self.init_time = time.time() plt.ion() if signal2model is not None: self.signal2model = signal2model plt.ion() condition_not_met = True history = [] self.epoch = 0 self.loss_history = [] tf.summary.scalar('loss', self.loss) merged = tf.summary.merge_all() train_writer = tf.summary.FileWriter('train', self.session.graph) # run_options = tf.RunOptions(report_tensor_allocations_upon_oom=True) # db_url = 'postgres://belo:passsword@localhost/postgres' # experiments, steps, model_params = labnotebook.initialize(db_url) # model_desc = {'loss': 0.} # experiment = labnotebook.start_experiment(model_desc=model_desc) tf.global_variables_initializer() while condition_not_met: self.epoch += 1 # tic = time.time() random_indexes = np.random.permutation(self.signal2model.batch_size) groups = np.reshape(random_indexes, (int(self.signal2model.batch_size/self.signal2model.mini_batch_size), self.signal2model.mini_batch_size)) for group in groups: dictionary = self.shuffle(X, Y, group) op, group_loss = self.session.run( [self.optimize_op, self.loss_op], feed_dict=dictionary)#, options=run_options) # labnotebook.step_experiment(experiment, # timestep=str(self.epoch), # trainacc=0, # valacc=0, # trainloss=str(group_loss)) # print("toc: {0} secs".format
trainables = self.parameters grads = tf.gradients(self.loss, trainables) grad_var_pairs = zip(grads, trainables) # grads, _ = tf.clip_by_global_norm(grads, clip_norm=1.0) # self.learning_rate = tf.train.exponential_decay( # self.signal2model.learning_rate_val, self.epoch, self.signal2model.count_to_break_max, # self.signal2model.decay, staircase=True) # with tf.device('/gpu:1'): optimizer = tf.train.RMSPropOptimizer(self.learning_rate_gpu) # optimizer = tf.train.AdamOptimizer(self.learning_rate) self._optimize_op = optimizer.apply_gradients(grad_var_pairs)
identifier_body
LibphysGRU.py
) self.identity = tf.eye(Sd) return E, U, W, V, b, c, [None, None] def GRUnn(self, out_prev, x_t): E, U, W, V, b, c = self.E, self.U, self.W, self.V, self.b, self.c Hd, Sd, Bd = self.signal2model.hidden_dim, self.signal2model.signal_dim, tf.shape(x_t)[0] coversion_ones = tf.ones((1, Bd), dtype=tf.float32, name="conversion_matrix") # s_prev, o_prev, l_prev = out_prev s_prev, o_prev = out_prev def GRU(last_input, gru_params): s_g_prev, u, w, b = gru_params z = tf.nn.sigmoid(tf.matmul(u[0], last_input) + tf.matmul(w[0], s_g_prev) + tf.matmul(tf.reshape(b[0], (Hd, 1)), coversion_ones)) r = tf.nn.sigmoid(tf.matmul(u[1], last_input) + tf.matmul(w[1], s_g_prev) + tf.matmul(tf.reshape(b[1], (Hd, 1)), coversion_ones)) value = tf.matmul(u[2], last_input) + tf.matmul(w[2], s_g_prev * r) + \ tf.matmul(tf.reshape(b[2], (Hd, 1)), coversion_ones) s_candidate = tf.nn.tanh(value) output = tf.add(((tf.ones_like(z) - z) * s_candidate), (z * s_g_prev), name="out_GRU") return output # x_e -> (Hd x Mb) x_e = tf.gather(self.E, x_t, axis=1) s_t_ = [] s_t_.append(GRU(x_e, [s_prev[0], U[0], W[0], b[0]])) s_t_.append(GRU(s_t_[0], [s_prev[1], U[1], W[1], b[1]])) s_t_.append(GRU(s_t_[1], [s_prev[2], U[2], W[2], b[2]])) s_t = tf.stack(s_t_) # tf.scan(GRU, (s_prev, self.U, self.W, self.b), initializer=x_e, parallel_iterations=1, name="states") logits = tf.matmul(self.V, s_t[-1]) + tf.matmul(tf.reshape(self.c, (Sd, 1)), coversion_ones) o_t = tf.nn.softmax(logits, axis=2) return [s_t, o_t]#, logits] def
(self, x_batch): initial_s = tf.zeros((self.signal2model.n_grus, self.signal2model.hidden_dim, tf.shape(x_batch)[1]), dtype=np.float32) initial_out = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32) # initial_l = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32) # x_batch = (N x Bd) - N (samples); Bd - Batch dimension # [s, o, l] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out, initial_l], parallel_iterations=1, # name="network_output") [_, o] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out], parallel_iterations=1, name="network_output") return o def feed_forward_predict_with_states(self, x_batch): initial_s = tf.zeros((self.signal2model.n_grus, self.signal2model.hidden_dim, tf.shape(x_batch)[1]), dtype=np.float32) initial_out = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32) # initial_l = tf.zeros((self.signal2model.signal_dim, tf.shape(x_batch)[1]), dtype=np.float32) # x_batch = (N x Bd) - N (samples); Bd - Batch dimension # [s, o, l] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out, initial_l], parallel_iterations=1, # name="network_output") [s, o] = tf.scan(self.GRUnn, x_batch, initializer=[initial_s, initial_out], parallel_iterations=1, name="network_output") return [s, o] def calculate_predictions(self): # MAP NOT WORKING: # shape(X)[0] -> Windows # shape(X)[1] -> Samples # n_batches = int(signal2model.batch_size / self.signal2model.mini_batch_size) # N = tf.shape(self.X)[1] # print(X) # get the matrices from E with tf.gather(E, X, axis=1, name="X_e") # transpose these matrices for (batch_size, HD, N) # reshape to enter map, where each minibatch is entered at the same time (n_batches, mini_batch, HD, N) # transpose to enter the DNN inside -> (n_batches, N, mini_batch) return self.feed_forward_predict(tf.transpose(self.X)) def to_one_hot_vector_in_mini_batches(self, matrix): return self.get_one_hot(matrix) def get_one_hot(self, columns): return tf.gather(self.identity, columns) def calculate_cross_entropy(self): return None # logits = tf.transpose(self.logits, perm=[2, 0, 1]) # n_batches = int(self.signal2model.batch_size / self.signal2model.mini_batch_size) # y = tf.reshape(self.Y, (n_batches, self.signal2model.mini_batch_size, tf.shape(self.Y)[1])) # self.full_loss = tf.losses.sparse_softmax_cross_entropy_with_logits(labels=y, logits=logits) # return self.full_loss def calculate_mse(self): _y = self.to_one_hot_vector_in_mini_batches(self.Y) return tf.reduce_mean(tf.subtract(tf.transpose(self.out, perm=[2, 0, 1]), _y) ** 2, axis=2, name="mse") def calculate_mse_vector_loss(self, x, y): with tf.variable_scope('vector_loss'): return tf.reduce_mean(self.calculate_minibatch_mse(x, y), axis=0, name="vector_loss") def calculate_mse_loss(self): return tf.reduce_mean(self.calculate_minibatch_mse(), axis=0, name="loss") @property def loss_op(self): """ An Operation that takes one optimization step. """ return self.loss def init_optimizer(self): trainables = self.parameters grads = tf.gradients(self.loss, trainables) grad_var_pairs = zip(grads, trainables) # grads, _ = tf.clip_by_global_norm(grads, clip_norm=1.0) # self.learning_rate = tf.train.exponential_decay( # self.signal2model.learning_rate_val, self.epoch, self.signal2model.count_to_break_max, # self.signal2model.decay, staircase=True) # with tf.device('/gpu:1'): optimizer = tf.train.RMSPropOptimizer(self.learning_rate_gpu) # optimizer = tf.train.AdamOptimizer(self.learning_rate) self._optimize_op = optimizer.apply_gradients(grad_var_pairs) @property def optimize_op(self): """ An Operation that takes one optimization step. """ return self._optimize_op def train(self, X, Y, signal2model=None): self.batch_size += np.shape(X)[0] self.init_time = time.time() plt.ion() if signal2model is not None: self.signal2model = signal2model plt.ion() condition_not_met = True history = [] self.epoch = 0 self.loss_history = [] tf.summary.scalar('loss', self.loss) merged = tf.summary.merge_all() train_writer = tf.summary.FileWriter('train', self.session.graph) # run_options = tf.RunOptions(report_tensor_allocations_upon_oom=True) # db_url = 'postgres://belo:passsword@localhost/postgres' # experiments, steps, model_params = labnotebook.initialize(db_url) # model_desc = {'loss': 0.} # experiment = labnotebook.start_experiment(model_desc=model_desc) tf.global_variables_initializer() while condition_not_met: self.epoch += 1 # tic = time.time() random_indexes = np.random.permutation(self.signal2model.batch_size) groups = np.reshape(random_indexes, (int(self.signal2model.batch_size/self.signal2model.mini_batch_size), self.signal2model.mini_batch_size)) for group in groups: dictionary = self.shuffle(X, Y, group) op, group_loss = self.session.run( [self.optimize_op, self.loss_op], feed_dict=dictionary)#, options=run_options) # labnotebook.step_experiment(experiment, # timestep=str(self.epoch), # trainacc=0, # valacc=0, # trainloss=str(group_loss)) # print("toc: {0} secs".format
feed_forward_predict
identifier_name
search-table.controller.js
License is distributed on an "AS IS" BASIS, WITHOUT * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the * License for the specific language governing permissions and limitations * under the License. */ (function () { 'use strict'; /** * @ngdoc controller * @name SearchTableController * * @description * Controller for the search table. * Serves as the focal point for table actions. */ angular .module('horizon.dashboard.project.search') .controller('searchTableController', SearchTableController); SearchTableController.$inject = [ '$scope', '$filter', '$q', '$timeout', 'searchPluginResourceTypesFilter', 'horizon.framework.conf.resource-type-registry.service', 'horizon.app.core.openstack-service-api.userSession', 'horizon.dashboard.project.search.searchlightFacetUtils', 'horizon.dashboard.project.search.searchlightSearchHelper', 'horizon.dashboard.project.search.settingsService', 'horizon.dashboard.search.search.util.cache.service' ]; function SearchTableController($scope, $filter, $q, $timeout, searchPluginResourceTypesFilter, registry, userSession, searchlightFacetUtils, searchlightSearchHelper, searchSettings, cache) { var ctrl = this; ctrl.filter = $filter; ctrl.hits = []; ctrl.hitsSrc = []; ctrl.initialized = false; ctrl.searchFacets = []; ctrl.excludedTypes = ['OS::Glance::Metadef']; ctrl.searchSettings = searchSettings; ctrl.defaultResourceTypes = []; ctrl.defaultFacets = searchlightFacetUtils.defaultFacets(); ctrl.registry = registry; ctrl.refresh = searchlightSearchHelper.repeatLastSearchWithLatestSettings; ctrl.actionResultHandler = actionResultHandler; ctrl.getSearchlightKey = getSearchlightKey; ctrl.userSession = {}; var adHocPollInterval = 500; var adHocPollDuration = 5000; //ctrl.isNested; init(); //////////////////////////////// function init() { ctrl.searchSettings.initScope($scope); searchlightFacetUtils.initScope($scope); if (searchlightSearchHelper.lastSearchQueryOptions) { ctrl.searchFacets = searchlightSearchHelper.lastSearchQueryOptions.searchFacets; if (searchlightSearchHelper.lastSearchQueryOptions.queryString) { $timeout(setInput(searchlightSearchHelper.lastSearchQueryOptions.queryString)); } } else { ctrl.searchFacets = ctrl.defaultFacets; } userSession.get() .then(function onUserSessionGet(session) { ctrl.userSession = session; }); } function setInput(text) { return function() { angular.element('.search-input').val(text); }; } /*function isNested (input) { var result = angular.isArray(input) && input.length > 0 && angular.isObject(input[0]) && Object.keys(input[0]).length > 1; return result; }*/ var pluginsUpdatedWatcher = $scope.$on( ctrl.searchSettings.events.pluginsUpdatedEvent, pluginsUpdated ); function pluginsUpdated(event, plugins) { var pluginToTypesOptions = { excludedTypes: ctrl.excludedTypes, flatten: true }; ctrl.defaultResourceTypes = searchPluginResourceTypesFilter(plugins, pluginToTypesOptions); ctrl.defaultResourceTypes.forEach(function(type) { registry.initActions(type, $scope); }); searchlightFacetUtils.setTypeFacetFromResourceTypes( ctrl.defaultResourceTypes, ctrl.searchFacets); searchlightFacetUtils.broadcastFacetsChanged(searchlightSearchHelper.lastSearchQueryOptions); ctrl.initialized = true; if (searchlightSearchHelper.lastSearchQueryOptions) { searchlightSearchHelper.lastSearchQueryOptions.onSearchSuccess = onSearchResult; searchlightSearchHelper.lastSearchQueryOptions.onSearchError = onSearchResult; searchlightSearchHelper.repeatLastSearchWithLatestSettings(); } else { search(); } } var fullTextSearchTimeout; var searchUpdatedWatcher = $scope.$on('serverSearchUpdated', function (event, searchData) { // Magic search always broadcasts this at startup, so // we have to not run until we are fully initialized. if (!ctrl.initialized) { return; } function performSearch() { fullTextSearchTimeout = null; search(searchData); } if (searchData.queryStringChanged) { // This keeps the query from being executed too rapidly // when the user is performing rapid key presses. if (fullTextSearchTimeout) { $timeout.cancel(fullTextSearchTimeout); } fullTextSearchTimeout = $timeout( performSearch, ctrl.searchSettings.settings.fullTextSearch.delayInMS ); } else if (searchData.magicSearchQueryChanged) { performSearch(); } }); var checkFacetsWatcher = $scope.$on('checkFacets', function (event, selectedFacets) { //Facets are actually DOM elements. This affects the styling. $timeout(function () { angular.forEach(selectedFacets, function setIsServerTrue(facet) { facet.isServer = true; }); }); }); var searchSettingsUpdatedWatcher = $scope.$on( ctrl.searchSettings.events.settingsUpdatedEvent, searchlightSearchHelper.repeatLastSearchWithLatestSettings ); $scope.$on('$destroy', function cleanupListeners() { searchlightSearchHelper.stopSearchPolling(); checkFacetsWatcher(); searchUpdatedWatcher(); searchSettingsUpdatedWatcher(); pluginsUpdatedWatcher(); }); function search(queryOptions) { queryOptions = queryOptions || {}; queryOptions.allFacetDefinitions = ctrl.searchFacets; queryOptions.searchFacets = ctrl.searchFacets; queryOptions.defaultResourceTypes = ctrl.defaultResourceTypes; queryOptions.onSearchSuccess = onSearchResult; queryOptions.onSearchError = onSearchResult; return searchlightSearchHelper.search(queryOptions); } function onSearchResult(response) { cache.clean(adHocPollDuration * 3); ctrl.hitsSrc = response.hits.map(syncWithCache).filter(removeDeletedItems); ctrl.queryResponse = response; } function syncWithCache(searchlight_item) { return cache.sync(searchlight_item, searchlight_item._id, getSearchlightTimestamp(searchlight_item)); } function removeDeletedItems(searchlight_item) { if ( searchlight_item.deleted ) { return false; } else { return true; } } function actionResultHandler(returnValue) { return $q.when(returnValue, actionSuccessHandler, actionErrorHandler); } /* function repeatUntilChangedResults() { // For now, all we can do is poll for a period of time. searchlightSearchHelper.startAdHocPolling(adHocPollInterval, adHocPollDuration); } */ function actionSuccessHandler(result) { // For now, always poll for 5 seconds after every action. This is not // needed with default polling enabled. //repeatUntilChangedResults(); // The action has completed (for whatever "complete" means to that // action. Notice the view doesn't really need to know the semantics of the // particular action because the actions return data in a standard form. // That return includes the id and type of each created, updated, deleted // and failed item. // // This handler is also careful to check the type of each item. This // is important because actions which create non-images are launched from // the images page (like create "volume" from image). var deletedIds, updatedIds, createdIds, failedIds; if ( result ) { // Reduce the results to just image ids ignoring other types the action // may have produced deletedIds = getIdsOfType(result.deleted, undefined); updatedIds = getIdsOfType(result.updated, undefined); createdIds = getIdsOfType(result.created, undefined); failedIds = getIdsOfType(result.failed, undefined); addItemsToCache(deletedIds, true); addItemsToCache(updatedIds); addItemsToCache(createdIds); // Handle deleted images if (deletedIds.length) { // Do nothing for now } // Handle updated and created images if ( updatedIds.length || createdIds.length ) { } // Handle failed images if ( failedIds ) { // Do nothing for now } } else { // promise resolved, but no result returned. Because the action didn't // tell us what happened...reload the displayed items just in case. } } function addItemsToCache(ids, deleted) { var searchlight_item; ids.forEach(function addToCache(id) { var index = ctrl.hitsSrc.findIndex(function findItemWithId(item) { if (item._source.id === id) { return item; } }); if ( index >= 0 ) { var searchlight_item = ctrl.hitsSrc[index]; if ( deleted )
if ( searchlight_item ) { searchlight_item.dirty = true; searchlight_item.deleted = deleted; cache.add(searchlight_item, searchlight_item._id, getSearchlightTimestamp(searchlight_item)); } } }); }
{ ctrl.hitsSrc.splice(index,1); }
conditional_block
search-table.controller.js
the License is distributed on an "AS IS" BASIS, WITHOUT * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the * License for the specific language governing permissions and limitations * under the License. */ (function () { 'use strict'; /** * @ngdoc controller * @name SearchTableController * * @description * Controller for the search table. * Serves as the focal point for table actions. */ angular .module('horizon.dashboard.project.search') .controller('searchTableController', SearchTableController); SearchTableController.$inject = [ '$scope', '$filter', '$q', '$timeout', 'searchPluginResourceTypesFilter',
'horizon.framework.conf.resource-type-registry.service', 'horizon.app.core.openstack-service-api.userSession', 'horizon.dashboard.project.search.searchlightFacetUtils', 'horizon.dashboard.project.search.searchlightSearchHelper', 'horizon.dashboard.project.search.settingsService', 'horizon.dashboard.search.search.util.cache.service' ]; function SearchTableController($scope, $filter, $q, $timeout, searchPluginResourceTypesFilter, registry, userSession, searchlightFacetUtils, searchlightSearchHelper, searchSettings, cache) { var ctrl = this; ctrl.filter = $filter; ctrl.hits = []; ctrl.hitsSrc = []; ctrl.initialized = false; ctrl.searchFacets = []; ctrl.excludedTypes = ['OS::Glance::Metadef']; ctrl.searchSettings = searchSettings; ctrl.defaultResourceTypes = []; ctrl.defaultFacets = searchlightFacetUtils.defaultFacets(); ctrl.registry = registry; ctrl.refresh = searchlightSearchHelper.repeatLastSearchWithLatestSettings; ctrl.actionResultHandler = actionResultHandler; ctrl.getSearchlightKey = getSearchlightKey; ctrl.userSession = {}; var adHocPollInterval = 500; var adHocPollDuration = 5000; //ctrl.isNested; init(); //////////////////////////////// function init() { ctrl.searchSettings.initScope($scope); searchlightFacetUtils.initScope($scope); if (searchlightSearchHelper.lastSearchQueryOptions) { ctrl.searchFacets = searchlightSearchHelper.lastSearchQueryOptions.searchFacets; if (searchlightSearchHelper.lastSearchQueryOptions.queryString) { $timeout(setInput(searchlightSearchHelper.lastSearchQueryOptions.queryString)); } } else { ctrl.searchFacets = ctrl.defaultFacets; } userSession.get() .then(function onUserSessionGet(session) { ctrl.userSession = session; }); } function setInput(text) { return function() { angular.element('.search-input').val(text); }; } /*function isNested (input) { var result = angular.isArray(input) && input.length > 0 && angular.isObject(input[0]) && Object.keys(input[0]).length > 1; return result; }*/ var pluginsUpdatedWatcher = $scope.$on( ctrl.searchSettings.events.pluginsUpdatedEvent, pluginsUpdated ); function pluginsUpdated(event, plugins) { var pluginToTypesOptions = { excludedTypes: ctrl.excludedTypes, flatten: true }; ctrl.defaultResourceTypes = searchPluginResourceTypesFilter(plugins, pluginToTypesOptions); ctrl.defaultResourceTypes.forEach(function(type) { registry.initActions(type, $scope); }); searchlightFacetUtils.setTypeFacetFromResourceTypes( ctrl.defaultResourceTypes, ctrl.searchFacets); searchlightFacetUtils.broadcastFacetsChanged(searchlightSearchHelper.lastSearchQueryOptions); ctrl.initialized = true; if (searchlightSearchHelper.lastSearchQueryOptions) { searchlightSearchHelper.lastSearchQueryOptions.onSearchSuccess = onSearchResult; searchlightSearchHelper.lastSearchQueryOptions.onSearchError = onSearchResult; searchlightSearchHelper.repeatLastSearchWithLatestSettings(); } else { search(); } } var fullTextSearchTimeout; var searchUpdatedWatcher = $scope.$on('serverSearchUpdated', function (event, searchData) { // Magic search always broadcasts this at startup, so // we have to not run until we are fully initialized. if (!ctrl.initialized) { return; } function performSearch() { fullTextSearchTimeout = null; search(searchData); } if (searchData.queryStringChanged) { // This keeps the query from being executed too rapidly // when the user is performing rapid key presses. if (fullTextSearchTimeout) { $timeout.cancel(fullTextSearchTimeout); } fullTextSearchTimeout = $timeout( performSearch, ctrl.searchSettings.settings.fullTextSearch.delayInMS ); } else if (searchData.magicSearchQueryChanged) { performSearch(); } }); var checkFacetsWatcher = $scope.$on('checkFacets', function (event, selectedFacets) { //Facets are actually DOM elements. This affects the styling. $timeout(function () { angular.forEach(selectedFacets, function setIsServerTrue(facet) { facet.isServer = true; }); }); }); var searchSettingsUpdatedWatcher = $scope.$on( ctrl.searchSettings.events.settingsUpdatedEvent, searchlightSearchHelper.repeatLastSearchWithLatestSettings ); $scope.$on('$destroy', function cleanupListeners() { searchlightSearchHelper.stopSearchPolling(); checkFacetsWatcher(); searchUpdatedWatcher(); searchSettingsUpdatedWatcher(); pluginsUpdatedWatcher(); }); function search(queryOptions) { queryOptions = queryOptions || {}; queryOptions.allFacetDefinitions = ctrl.searchFacets; queryOptions.searchFacets = ctrl.searchFacets; queryOptions.defaultResourceTypes = ctrl.defaultResourceTypes; queryOptions.onSearchSuccess = onSearchResult; queryOptions.onSearchError = onSearchResult; return searchlightSearchHelper.search(queryOptions); } function onSearchResult(response) { cache.clean(adHocPollDuration * 3); ctrl.hitsSrc = response.hits.map(syncWithCache).filter(removeDeletedItems); ctrl.queryResponse = response; } function syncWithCache(searchlight_item) { return cache.sync(searchlight_item, searchlight_item._id, getSearchlightTimestamp(searchlight_item)); } function removeDeletedItems(searchlight_item) { if ( searchlight_item.deleted ) { return false; } else { return true; } } function actionResultHandler(returnValue) { return $q.when(returnValue, actionSuccessHandler, actionErrorHandler); } /* function repeatUntilChangedResults() { // For now, all we can do is poll for a period of time. searchlightSearchHelper.startAdHocPolling(adHocPollInterval, adHocPollDuration); } */ function actionSuccessHandler(result) { // For now, always poll for 5 seconds after every action. This is not // needed with default polling enabled. //repeatUntilChangedResults(); // The action has completed (for whatever "complete" means to that // action. Notice the view doesn't really need to know the semantics of the // particular action because the actions return data in a standard form. // That return includes the id and type of each created, updated, deleted // and failed item. // // This handler is also careful to check the type of each item. This // is important because actions which create non-images are launched from // the images page (like create "volume" from image). var deletedIds, updatedIds, createdIds, failedIds; if ( result ) { // Reduce the results to just image ids ignoring other types the action // may have produced deletedIds = getIdsOfType(result.deleted, undefined); updatedIds = getIdsOfType(result.updated, undefined); createdIds = getIdsOfType(result.created, undefined); failedIds = getIdsOfType(result.failed, undefined); addItemsToCache(deletedIds, true); addItemsToCache(updatedIds); addItemsToCache(createdIds); // Handle deleted images if (deletedIds.length) { // Do nothing for now } // Handle updated and created images if ( updatedIds.length || createdIds.length ) { } // Handle failed images if ( failedIds ) { // Do nothing for now } } else { // promise resolved, but no result returned. Because the action didn't // tell us what happened...reload the displayed items just in case. } } function addItemsToCache(ids, deleted) { var searchlight_item; ids.forEach(function addToCache(id) { var index = ctrl.hitsSrc.findIndex(function findItemWithId(item) { if (item._source.id === id) { return item; } }); if ( index >= 0 ) { var searchlight_item = ctrl.hitsSrc[index]; if ( deleted ) { ctrl.hitsSrc.splice(index,1); } if ( searchlight_item ) { searchlight_item.dirty = true; searchlight_item.deleted = deleted; cache.add(searchlight_item, searchlight_item._id, getSearchlightTimestamp(searchlight_item)); } } }); }
random_line_split
search-table.controller.js
description * Controller for the search table. * Serves as the focal point for table actions. */ angular .module('horizon.dashboard.project.search') .controller('searchTableController', SearchTableController); SearchTableController.$inject = [ '$scope', '$filter', '$q', '$timeout', 'searchPluginResourceTypesFilter', 'horizon.framework.conf.resource-type-registry.service', 'horizon.app.core.openstack-service-api.userSession', 'horizon.dashboard.project.search.searchlightFacetUtils', 'horizon.dashboard.project.search.searchlightSearchHelper', 'horizon.dashboard.project.search.settingsService', 'horizon.dashboard.search.search.util.cache.service' ]; function SearchTableController($scope, $filter, $q, $timeout, searchPluginResourceTypesFilter, registry, userSession, searchlightFacetUtils, searchlightSearchHelper, searchSettings, cache) { var ctrl = this; ctrl.filter = $filter; ctrl.hits = []; ctrl.hitsSrc = []; ctrl.initialized = false; ctrl.searchFacets = []; ctrl.excludedTypes = ['OS::Glance::Metadef']; ctrl.searchSettings = searchSettings; ctrl.defaultResourceTypes = []; ctrl.defaultFacets = searchlightFacetUtils.defaultFacets(); ctrl.registry = registry; ctrl.refresh = searchlightSearchHelper.repeatLastSearchWithLatestSettings; ctrl.actionResultHandler = actionResultHandler; ctrl.getSearchlightKey = getSearchlightKey; ctrl.userSession = {}; var adHocPollInterval = 500; var adHocPollDuration = 5000; //ctrl.isNested; init(); //////////////////////////////// function init() { ctrl.searchSettings.initScope($scope); searchlightFacetUtils.initScope($scope); if (searchlightSearchHelper.lastSearchQueryOptions) { ctrl.searchFacets = searchlightSearchHelper.lastSearchQueryOptions.searchFacets; if (searchlightSearchHelper.lastSearchQueryOptions.queryString) { $timeout(setInput(searchlightSearchHelper.lastSearchQueryOptions.queryString)); } } else { ctrl.searchFacets = ctrl.defaultFacets; } userSession.get() .then(function onUserSessionGet(session) { ctrl.userSession = session; }); } function setInput(text) { return function() { angular.element('.search-input').val(text); }; } /*function isNested (input) { var result = angular.isArray(input) && input.length > 0 && angular.isObject(input[0]) && Object.keys(input[0]).length > 1; return result; }*/ var pluginsUpdatedWatcher = $scope.$on( ctrl.searchSettings.events.pluginsUpdatedEvent, pluginsUpdated ); function pluginsUpdated(event, plugins) { var pluginToTypesOptions = { excludedTypes: ctrl.excludedTypes, flatten: true }; ctrl.defaultResourceTypes = searchPluginResourceTypesFilter(plugins, pluginToTypesOptions); ctrl.defaultResourceTypes.forEach(function(type) { registry.initActions(type, $scope); }); searchlightFacetUtils.setTypeFacetFromResourceTypes( ctrl.defaultResourceTypes, ctrl.searchFacets); searchlightFacetUtils.broadcastFacetsChanged(searchlightSearchHelper.lastSearchQueryOptions); ctrl.initialized = true; if (searchlightSearchHelper.lastSearchQueryOptions) { searchlightSearchHelper.lastSearchQueryOptions.onSearchSuccess = onSearchResult; searchlightSearchHelper.lastSearchQueryOptions.onSearchError = onSearchResult; searchlightSearchHelper.repeatLastSearchWithLatestSettings(); } else { search(); } } var fullTextSearchTimeout; var searchUpdatedWatcher = $scope.$on('serverSearchUpdated', function (event, searchData) { // Magic search always broadcasts this at startup, so // we have to not run until we are fully initialized. if (!ctrl.initialized) { return; } function performSearch() { fullTextSearchTimeout = null; search(searchData); } if (searchData.queryStringChanged) { // This keeps the query from being executed too rapidly // when the user is performing rapid key presses. if (fullTextSearchTimeout) { $timeout.cancel(fullTextSearchTimeout); } fullTextSearchTimeout = $timeout( performSearch, ctrl.searchSettings.settings.fullTextSearch.delayInMS ); } else if (searchData.magicSearchQueryChanged) { performSearch(); } }); var checkFacetsWatcher = $scope.$on('checkFacets', function (event, selectedFacets) { //Facets are actually DOM elements. This affects the styling. $timeout(function () { angular.forEach(selectedFacets, function setIsServerTrue(facet) { facet.isServer = true; }); }); }); var searchSettingsUpdatedWatcher = $scope.$on( ctrl.searchSettings.events.settingsUpdatedEvent, searchlightSearchHelper.repeatLastSearchWithLatestSettings ); $scope.$on('$destroy', function cleanupListeners() { searchlightSearchHelper.stopSearchPolling(); checkFacetsWatcher(); searchUpdatedWatcher(); searchSettingsUpdatedWatcher(); pluginsUpdatedWatcher(); }); function search(queryOptions) { queryOptions = queryOptions || {}; queryOptions.allFacetDefinitions = ctrl.searchFacets; queryOptions.searchFacets = ctrl.searchFacets; queryOptions.defaultResourceTypes = ctrl.defaultResourceTypes; queryOptions.onSearchSuccess = onSearchResult; queryOptions.onSearchError = onSearchResult; return searchlightSearchHelper.search(queryOptions); } function onSearchResult(response) { cache.clean(adHocPollDuration * 3); ctrl.hitsSrc = response.hits.map(syncWithCache).filter(removeDeletedItems); ctrl.queryResponse = response; } function syncWithCache(searchlight_item) { return cache.sync(searchlight_item, searchlight_item._id, getSearchlightTimestamp(searchlight_item)); } function removeDeletedItems(searchlight_item) { if ( searchlight_item.deleted ) { return false; } else { return true; } } function actionResultHandler(returnValue) { return $q.when(returnValue, actionSuccessHandler, actionErrorHandler); } /* function repeatUntilChangedResults() { // For now, all we can do is poll for a period of time. searchlightSearchHelper.startAdHocPolling(adHocPollInterval, adHocPollDuration); } */ function actionSuccessHandler(result) { // For now, always poll for 5 seconds after every action. This is not // needed with default polling enabled. //repeatUntilChangedResults(); // The action has completed (for whatever "complete" means to that // action. Notice the view doesn't really need to know the semantics of the // particular action because the actions return data in a standard form. // That return includes the id and type of each created, updated, deleted // and failed item. // // This handler is also careful to check the type of each item. This // is important because actions which create non-images are launched from // the images page (like create "volume" from image). var deletedIds, updatedIds, createdIds, failedIds; if ( result ) { // Reduce the results to just image ids ignoring other types the action // may have produced deletedIds = getIdsOfType(result.deleted, undefined); updatedIds = getIdsOfType(result.updated, undefined); createdIds = getIdsOfType(result.created, undefined); failedIds = getIdsOfType(result.failed, undefined); addItemsToCache(deletedIds, true); addItemsToCache(updatedIds); addItemsToCache(createdIds); // Handle deleted images if (deletedIds.length) { // Do nothing for now } // Handle updated and created images if ( updatedIds.length || createdIds.length ) { } // Handle failed images if ( failedIds ) { // Do nothing for now } } else { // promise resolved, but no result returned. Because the action didn't // tell us what happened...reload the displayed items just in case. } } function addItemsToCache(ids, deleted) { var searchlight_item; ids.forEach(function addToCache(id) { var index = ctrl.hitsSrc.findIndex(function findItemWithId(item) { if (item._source.id === id) { return item; } }); if ( index >= 0 ) { var searchlight_item = ctrl.hitsSrc[index]; if ( deleted ) { ctrl.hitsSrc.splice(index,1); } if ( searchlight_item ) { searchlight_item.dirty = true; searchlight_item.deleted = deleted; cache.add(searchlight_item, searchlight_item._id, getSearchlightTimestamp(searchlight_item)); } } }); } function actionErrorHandler(reason) { // eslint-disable-line no-unused-vars // Action has failed. Do nothing. } function getIdsOfType(items, type) { var result; function typeIdReduce(accumulator, item)
{ if (type === undefined || item.type === type) { accumulator.push(item.id); } return accumulator; }
identifier_body
search-table.controller.js
the License is distributed on an "AS IS" BASIS, WITHOUT * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the * License for the specific language governing permissions and limitations * under the License. */ (function () { 'use strict'; /** * @ngdoc controller * @name SearchTableController * * @description * Controller for the search table. * Serves as the focal point for table actions. */ angular .module('horizon.dashboard.project.search') .controller('searchTableController', SearchTableController); SearchTableController.$inject = [ '$scope', '$filter', '$q', '$timeout', 'searchPluginResourceTypesFilter', 'horizon.framework.conf.resource-type-registry.service', 'horizon.app.core.openstack-service-api.userSession', 'horizon.dashboard.project.search.searchlightFacetUtils', 'horizon.dashboard.project.search.searchlightSearchHelper', 'horizon.dashboard.project.search.settingsService', 'horizon.dashboard.search.search.util.cache.service' ]; function SearchTableController($scope, $filter, $q, $timeout, searchPluginResourceTypesFilter, registry, userSession, searchlightFacetUtils, searchlightSearchHelper, searchSettings, cache) { var ctrl = this; ctrl.filter = $filter; ctrl.hits = []; ctrl.hitsSrc = []; ctrl.initialized = false; ctrl.searchFacets = []; ctrl.excludedTypes = ['OS::Glance::Metadef']; ctrl.searchSettings = searchSettings; ctrl.defaultResourceTypes = []; ctrl.defaultFacets = searchlightFacetUtils.defaultFacets(); ctrl.registry = registry; ctrl.refresh = searchlightSearchHelper.repeatLastSearchWithLatestSettings; ctrl.actionResultHandler = actionResultHandler; ctrl.getSearchlightKey = getSearchlightKey; ctrl.userSession = {}; var adHocPollInterval = 500; var adHocPollDuration = 5000; //ctrl.isNested; init(); //////////////////////////////// function init() { ctrl.searchSettings.initScope($scope); searchlightFacetUtils.initScope($scope); if (searchlightSearchHelper.lastSearchQueryOptions) { ctrl.searchFacets = searchlightSearchHelper.lastSearchQueryOptions.searchFacets; if (searchlightSearchHelper.lastSearchQueryOptions.queryString) { $timeout(setInput(searchlightSearchHelper.lastSearchQueryOptions.queryString)); } } else { ctrl.searchFacets = ctrl.defaultFacets; } userSession.get() .then(function onUserSessionGet(session) { ctrl.userSession = session; }); } function setInput(text) { return function() { angular.element('.search-input').val(text); }; } /*function isNested (input) { var result = angular.isArray(input) && input.length > 0 && angular.isObject(input[0]) && Object.keys(input[0]).length > 1; return result; }*/ var pluginsUpdatedWatcher = $scope.$on( ctrl.searchSettings.events.pluginsUpdatedEvent, pluginsUpdated ); function pluginsUpdated(event, plugins) { var pluginToTypesOptions = { excludedTypes: ctrl.excludedTypes, flatten: true }; ctrl.defaultResourceTypes = searchPluginResourceTypesFilter(plugins, pluginToTypesOptions); ctrl.defaultResourceTypes.forEach(function(type) { registry.initActions(type, $scope); }); searchlightFacetUtils.setTypeFacetFromResourceTypes( ctrl.defaultResourceTypes, ctrl.searchFacets); searchlightFacetUtils.broadcastFacetsChanged(searchlightSearchHelper.lastSearchQueryOptions); ctrl.initialized = true; if (searchlightSearchHelper.lastSearchQueryOptions) { searchlightSearchHelper.lastSearchQueryOptions.onSearchSuccess = onSearchResult; searchlightSearchHelper.lastSearchQueryOptions.onSearchError = onSearchResult; searchlightSearchHelper.repeatLastSearchWithLatestSettings(); } else { search(); } } var fullTextSearchTimeout; var searchUpdatedWatcher = $scope.$on('serverSearchUpdated', function (event, searchData) { // Magic search always broadcasts this at startup, so // we have to not run until we are fully initialized. if (!ctrl.initialized) { return; } function performSearch() { fullTextSearchTimeout = null; search(searchData); } if (searchData.queryStringChanged) { // This keeps the query from being executed too rapidly // when the user is performing rapid key presses. if (fullTextSearchTimeout) { $timeout.cancel(fullTextSearchTimeout); } fullTextSearchTimeout = $timeout( performSearch, ctrl.searchSettings.settings.fullTextSearch.delayInMS ); } else if (searchData.magicSearchQueryChanged) { performSearch(); } }); var checkFacetsWatcher = $scope.$on('checkFacets', function (event, selectedFacets) { //Facets are actually DOM elements. This affects the styling. $timeout(function () { angular.forEach(selectedFacets, function setIsServerTrue(facet) { facet.isServer = true; }); }); }); var searchSettingsUpdatedWatcher = $scope.$on( ctrl.searchSettings.events.settingsUpdatedEvent, searchlightSearchHelper.repeatLastSearchWithLatestSettings ); $scope.$on('$destroy', function cleanupListeners() { searchlightSearchHelper.stopSearchPolling(); checkFacetsWatcher(); searchUpdatedWatcher(); searchSettingsUpdatedWatcher(); pluginsUpdatedWatcher(); }); function search(queryOptions) { queryOptions = queryOptions || {}; queryOptions.allFacetDefinitions = ctrl.searchFacets; queryOptions.searchFacets = ctrl.searchFacets; queryOptions.defaultResourceTypes = ctrl.defaultResourceTypes; queryOptions.onSearchSuccess = onSearchResult; queryOptions.onSearchError = onSearchResult; return searchlightSearchHelper.search(queryOptions); } function onSearchResult(response) { cache.clean(adHocPollDuration * 3); ctrl.hitsSrc = response.hits.map(syncWithCache).filter(removeDeletedItems); ctrl.queryResponse = response; } function
(searchlight_item) { return cache.sync(searchlight_item, searchlight_item._id, getSearchlightTimestamp(searchlight_item)); } function removeDeletedItems(searchlight_item) { if ( searchlight_item.deleted ) { return false; } else { return true; } } function actionResultHandler(returnValue) { return $q.when(returnValue, actionSuccessHandler, actionErrorHandler); } /* function repeatUntilChangedResults() { // For now, all we can do is poll for a period of time. searchlightSearchHelper.startAdHocPolling(adHocPollInterval, adHocPollDuration); } */ function actionSuccessHandler(result) { // For now, always poll for 5 seconds after every action. This is not // needed with default polling enabled. //repeatUntilChangedResults(); // The action has completed (for whatever "complete" means to that // action. Notice the view doesn't really need to know the semantics of the // particular action because the actions return data in a standard form. // That return includes the id and type of each created, updated, deleted // and failed item. // // This handler is also careful to check the type of each item. This // is important because actions which create non-images are launched from // the images page (like create "volume" from image). var deletedIds, updatedIds, createdIds, failedIds; if ( result ) { // Reduce the results to just image ids ignoring other types the action // may have produced deletedIds = getIdsOfType(result.deleted, undefined); updatedIds = getIdsOfType(result.updated, undefined); createdIds = getIdsOfType(result.created, undefined); failedIds = getIdsOfType(result.failed, undefined); addItemsToCache(deletedIds, true); addItemsToCache(updatedIds); addItemsToCache(createdIds); // Handle deleted images if (deletedIds.length) { // Do nothing for now } // Handle updated and created images if ( updatedIds.length || createdIds.length ) { } // Handle failed images if ( failedIds ) { // Do nothing for now } } else { // promise resolved, but no result returned. Because the action didn't // tell us what happened...reload the displayed items just in case. } } function addItemsToCache(ids, deleted) { var searchlight_item; ids.forEach(function addToCache(id) { var index = ctrl.hitsSrc.findIndex(function findItemWithId(item) { if (item._source.id === id) { return item; } }); if ( index >= 0 ) { var searchlight_item = ctrl.hitsSrc[index]; if ( deleted ) { ctrl.hitsSrc.splice(index,1); } if ( searchlight_item ) { searchlight_item.dirty = true; searchlight_item.deleted = deleted; cache.add(searchlight_item, searchlight_item._id, getSearchlightTimestamp(searchlight_item)); } } }); }
syncWithCache
identifier_name
modal-gallery.component.ts
this.description = description; this.extUrl = extUrl; this.selected = selected; } } /** * Enum `Keyboard` with keys and their relative key codes. */ export enum Keyboard { ESC = 27, LEFT_ARROW = 37, RIGHT_ARROW = 39, UP_ARROW = 38, DOWN_ARROW = 40 } /** * Interface `Description` to change the description, either with a full custom * description or with a small and simple customization. */ export interface Description { customFullDescription?: string; imageText?: string; numberSeparator?: string; beforeTextDescription?: string; } // /** // * Interface `ButtonsConfig` to show/hide buttons. // */ // export interface ButtonsConfig { // download?: boolean; // extUrl?: boolean; // close?: boolean; // } /** * Interface `KeyboardConfig` to assign custom keyCodes to ESC, RIGHT and LEFT keyboard's actions. */ export interface KeyboardConfig { esc?: number; right?: number; left?: number; } /** * Interface `SlideConfig` to configure sliding features of modal gallery. */ export interface SlideConfig { infinite?: boolean; } /** * Main Component of this library with the modal gallery. */ @Component({ selector: 'modal-gallery', exportAs: 'modalGallery', styleUrls: ['modal-gallery.scss'], templateUrl: 'modal-gallery.html' }) export class AngularModalGalleryComponent implements OnInit, OnDestroy, OnChanges { /** * Array or Observable input that represents a list of Images used to show both * thumbs and the modal gallery. */ @Input() modalImages: Observable<Array<Image>> | Array<Image>; @Input() selectionLimit: number; /** * Number to open the modal gallery (passing a value >=0) showing the image with the * imagePointer's index. * * Be careful, because this feature will be probably deprecated/changed in version 4.0.0 */ @Input() imagePointer: number; /** * Boolean required to enable image download with both ctrl+s/cmd+s and download button. * If you want to show enable button, this is not enough. You have to use also `buttonsConfig`. */ @Input() downloadable: boolean = false; /** * Description object with the configuration to show image descriptions. */ @Input() description: Description; /** * Object of type `ButtonsConfig` to show/hide buttons. * This is used only inside `ngOnInit()` to create `configButtons` */ @Input() selectionInfo: SelectionInfo; @Input() buttonsConfig: ButtonsConfig; /** * Object of type `KeyboardConfig` to assign custom keys to ESC, RIGHT and LEFT keyboard's actions. */ @Input() keyboardConfig: KeyboardConfig; /** * enableCloseOutside's input to enable modal-gallery close's behaviour while clicking * on the semi-transparent background. Disabled by default. */ @Input() enableCloseOutside: boolean = false; /** * Object of type `SlideConfig` to configure sliding of modal gallery. */ @Input() slideConfig: SlideConfig; /** * DEPRECATED * -----REMOVE THIS IN 4.0.0----- deprecated both showDownloadButton and showExtUrlButton */ @Input() showDownloadButton: boolean = false; // deprecated /** * DEPRECATED * -----REMOVE THIS IN 4.0.0----- deprecated both showDownloadButton and showExtUrlButton */ @Input() showExtUrlButton: boolean = false; // deprecated @Output() close: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>(); @Output() show: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>(); @Output() firstImage: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>(); @Output() lastImage: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>(); @Output() hasData: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>(); @Output() selectChanged: EventEmitter<Image> = new EventEmitter<Image>(); selectedImageCount: number = 0; /** * Boolean that it is true if the modal gallery is visible */ opened: boolean = false; /** * Boolean that it is true if an image of the modal gallery is still loading */ loading: boolean = false; /** * Boolean to open the modal gallery. Closed by default. */ showGallery: boolean = false; /** * Array of `Image` that represent the model of this library with all images, thumbs and so on. */ images: Image[]; /** * `Image` currently visible. */ currentImage: Image; /** * Number that represents the index of the current image. */ currentImageIndex: number = 0; /** * Object of type `ButtonsConfig` used to configure buttons visibility. This is a temporary value * initialized by the real `buttonsConfig`'s input */ configButtons: ButtonsConfig; /** * Enum of type `Action` used to pass a click action when you click on the modal image. * Declared here to be used inside the template. */ clickAction: Action = Action.CLICK; /** * Boolean that it's true when you are watching the first image (currently visible). */ isFirstImage: boolean = false; /** * Boolean that it's true when you are watching the last image (currently visible). */ isLastImage: boolean = false; canSelectImage: boolean = false; /** * Paging related variables */ totalImageCount: number = 0; pageSize: number = 20; pageCount: number = 0; currentPage: number = 0; /** * Private SWIPE_ACTION to define all swipe actions used by hammerjs. */ private SWIPE_ACTION = { LEFT: 'swipeleft', RIGHT: 'swiperight', UP: 'swipeup', DOWN: 'swipedown' }; /** * When you pass an Observable of `Image`s as `modalImages`, you have to subscribe to that * Observable. So, to prevent memory leaks, you must store the subscription and call `unsubscribe` in * OnDestroy. */ private subscription: Subscription; /** * Listener to catch keyboard's events and call the right method based on the key. * For instance, pressing esc, this will call `closeGallery(Action.KEYBOARD)` and so on. * If you passed a valid `keyboardConfig` esc, right and left buttons will be customized based on your data. * @param e KeyboardEvent caught by the listener. */ @HostListener('window:keydown', ['$event']) onKeyDown(e: KeyboardEvent) { if (!this.opened) { return; } const esc: number = this.keyboardConfig && this.keyboardConfig.esc ? this.keyboardConfig.esc : Keyboard.ESC; const right: number = this.keyboardConfig && this.keyboardConfig.right ? this.keyboardConfig.right : Keyboard.RIGHT_ARROW; const left: number = this.keyboardConfig && this.keyboardConfig.left ? this.keyboardConfig.left : Keyboard.LEFT_ARROW; switch (e.keyCode) { case esc: this.closeGallery(Action.KEYBOARD); break; case right: this.nextImage(Action.KEYBOARD); break; case left: this.prevImage(Action.KEYBOARD); break; } } /** * Constructor with the injection of ´KeyboardService´ that initialize some description fields * based on default values. */ constructor(private keyboardService: KeyboardService) { // if description isn't provided initialize it with a default object if (!this.description) { this.description = { imageText: 'Image ', numberSeparator: '/', beforeTextDescription: ' - ' }; } // if one of the Description fields isn't initialized, provide a default value this.description.imageText = this.description.imageText || 'Image '; this.description.numberSeparator = this.description.numberSeparator || '/'; this.description.beforeTextDescription = this.description.beforeTextDescription || ' - '; } /** * This method will initialize the pager when the images are loaded. */ initializePager(){ if (this.images.length > 0){ this.totalImageCount = this.images.length; this.pageCount = this.totalImageCount / this.pageSize; this.currentPage = 1; } } /** * Method ´ngOnInit´ to build `configButtons` and to call `initImages()`. * This is an Angular's lifecycle hook, so its called automatically by Angular itself. * In particular, it's called only one time!!! */ ngOnInit() { // build configButtons to use it inside upper-buttons this.configButtons = { download: this.showDownloadButton || (this.buttonsConfig && this.buttonsConfig.download), extUrl: this.showExtUrlButton || (this.buttonsConfig && this.buttonsConfig.extUrl), close: (this.buttonsConfig && this.buttonsConfig.close) }; // call initImages passing true as parameter, because I want to emit `hasData` event this.initImages(true); } /** * Method ´ngOnChanges´ to init images preventing errors. * This
this.width = width; this.height = height; this.thumb = thumb;
random_line_split
modal-gallery.component.ts
want to show enable button, this is not enough. You have to use also `buttonsConfig`. */ @Input() downloadable: boolean = false; /** * Description object with the configuration to show image descriptions. */ @Input() description: Description; /** * Object of type `ButtonsConfig` to show/hide buttons. * This is used only inside `ngOnInit()` to create `configButtons` */ @Input() selectionInfo: SelectionInfo; @Input() buttonsConfig: ButtonsConfig; /** * Object of type `KeyboardConfig` to assign custom keys to ESC, RIGHT and LEFT keyboard's actions. */ @Input() keyboardConfig: KeyboardConfig; /** * enableCloseOutside's input to enable modal-gallery close's behaviour while clicking * on the semi-transparent background. Disabled by default. */ @Input() enableCloseOutside: boolean = false; /** * Object of type `SlideConfig` to configure sliding of modal gallery. */ @Input() slideConfig: SlideConfig; /** * DEPRECATED * -----REMOVE THIS IN 4.0.0----- deprecated both showDownloadButton and showExtUrlButton */ @Input() showDownloadButton: boolean = false; // deprecated /** * DEPRECATED * -----REMOVE THIS IN 4.0.0----- deprecated both showDownloadButton and showExtUrlButton */ @Input() showExtUrlButton: boolean = false; // deprecated @Output() close: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>(); @Output() show: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>(); @Output() firstImage: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>(); @Output() lastImage: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>(); @Output() hasData: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>(); @Output() selectChanged: EventEmitter<Image> = new EventEmitter<Image>(); selectedImageCount: number = 0; /** * Boolean that it is true if the modal gallery is visible */ opened: boolean = false; /** * Boolean that it is true if an image of the modal gallery is still loading */ loading: boolean = false; /** * Boolean to open the modal gallery. Closed by default. */ showGallery: boolean = false; /** * Array of `Image` that represent the model of this library with all images, thumbs and so on. */ images: Image[]; /** * `Image` currently visible. */ currentImage: Image; /** * Number that represents the index of the current image. */ currentImageIndex: number = 0; /** * Object of type `ButtonsConfig` used to configure buttons visibility. This is a temporary value * initialized by the real `buttonsConfig`'s input */ configButtons: ButtonsConfig; /** * Enum of type `Action` used to pass a click action when you click on the modal image. * Declared here to be used inside the template. */ clickAction: Action = Action.CLICK; /** * Boolean that it's true when you are watching the first image (currently visible). */ isFirstImage: boolean = false; /** * Boolean that it's true when you are watching the last image (currently visible). */ isLastImage: boolean = false; canSelectImage: boolean = false; /** * Paging related variables */ totalImageCount: number = 0; pageSize: number = 20; pageCount: number = 0; currentPage: number = 0; /** * Private SWIPE_ACTION to define all swipe actions used by hammerjs. */ private SWIPE_ACTION = { LEFT: 'swipeleft', RIGHT: 'swiperight', UP: 'swipeup', DOWN: 'swipedown' }; /** * When you pass an Observable of `Image`s as `modalImages`, you have to subscribe to that * Observable. So, to prevent memory leaks, you must store the subscription and call `unsubscribe` in * OnDestroy. */ private subscription: Subscription; /** * Listener to catch keyboard's events and call the right method based on the key. * For instance, pressing esc, this will call `closeGallery(Action.KEYBOARD)` and so on. * If you passed a valid `keyboardConfig` esc, right and left buttons will be customized based on your data. * @param e KeyboardEvent caught by the listener. */ @HostListener('window:keydown', ['$event']) onKeyDown(e: KeyboardEvent) { if (!this.opened)
const esc: number = this.keyboardConfig && this.keyboardConfig.esc ? this.keyboardConfig.esc : Keyboard.ESC; const right: number = this.keyboardConfig && this.keyboardConfig.right ? this.keyboardConfig.right : Keyboard.RIGHT_ARROW; const left: number = this.keyboardConfig && this.keyboardConfig.left ? this.keyboardConfig.left : Keyboard.LEFT_ARROW; switch (e.keyCode) { case esc: this.closeGallery(Action.KEYBOARD); break; case right: this.nextImage(Action.KEYBOARD); break; case left: this.prevImage(Action.KEYBOARD); break; } } /** * Constructor with the injection of ´KeyboardService´ that initialize some description fields * based on default values. */ constructor(private keyboardService: KeyboardService) { // if description isn't provided initialize it with a default object if (!this.description) { this.description = { imageText: 'Image ', numberSeparator: '/', beforeTextDescription: ' - ' }; } // if one of the Description fields isn't initialized, provide a default value this.description.imageText = this.description.imageText || 'Image '; this.description.numberSeparator = this.description.numberSeparator || '/'; this.description.beforeTextDescription = this.description.beforeTextDescription || ' - '; } /** * This method will initialize the pager when the images are loaded. */ initializePager(){ if (this.images.length > 0){ this.totalImageCount = this.images.length; this.pageCount = this.totalImageCount / this.pageSize; this.currentPage = 1; } } /** * Method ´ngOnInit´ to build `configButtons` and to call `initImages()`. * This is an Angular's lifecycle hook, so its called automatically by Angular itself. * In particular, it's called only one time!!! */ ngOnInit() { // build configButtons to use it inside upper-buttons this.configButtons = { download: this.showDownloadButton || (this.buttonsConfig && this.buttonsConfig.download), extUrl: this.showExtUrlButton || (this.buttonsConfig && this.buttonsConfig.extUrl), close: (this.buttonsConfig && this.buttonsConfig.close) }; // call initImages passing true as parameter, because I want to emit `hasData` event this.initImages(true); } /** * Method ´ngOnChanges´ to init images preventing errors. * This is an Angular's lifecycle hook, so its called automatically by Angular itself. * In particular, it's called before `ngOnInit()` and whenever one or more data-bound input properties change. * @param changes `SimpleChanges` object of current and previous property values provided by Angular. */ ngOnChanges(changes: SimpleChanges) { // to prevent errors when you pass to this library // the array of images inside a subscribe block, in this way: `...subscribe(val => { this.images = arrayOfImages })` // As you can see, I'm providing examples in these situations in all official demos if (this.modalImages) { // I pass `false` as parameter, because I DON'T want to emit `hasData` // event (preventing multiple hasData events while initializing) this.initImages(false); } } getImageCountsToDisplay(){ var selectedImages = this.images.filter(image=>{ return image.selected === true; }); var selectedImageCount = selectedImages.length; var tobeselected = this.selectionLimit - selectedImageCount; this.canSelectImage = tobeselected <= 0 && !this.currentImage.selected; return "You need to select " + tobeselected + " images." } /** * Method `getDescriptionToDisplay` to get the image description based on input params. * If you provide a full description this will be the visible description, otherwise, * it will be built using the `description` object, concatenating its fields. * @returns String description to display. */ getDescriptionToDisplay() { if (this.description && this.description.customFullDescription) { return this.description.customFullDescription; } // If the current image hasn't a description, // prevent to write the ' - ' (or this.description.beforeTextDescription) if (!this.currentImage.description || this.currentImage.description === '') { return `${this.description.imageText}${this.currentImageIndex + 1}${this.description.numberSeparator}${this.images.length}`; } return `${this.description.imageText}${this.currentImageIndex + 1}${this.description.numberSeparator}${this.images.length}${this.description.beforeTextDescription}${this.currentImage.description}`; } /** * Method `swipe
{ return; }
conditional_block
modal-gallery.component.ts
want to show enable button, this is not enough. You have to use also `buttonsConfig`. */ @Input() downloadable: boolean = false; /** * Description object with the configuration to show image descriptions. */ @Input() description: Description; /** * Object of type `ButtonsConfig` to show/hide buttons. * This is used only inside `ngOnInit()` to create `configButtons` */ @Input() selectionInfo: SelectionInfo; @Input() buttonsConfig: ButtonsConfig; /** * Object of type `KeyboardConfig` to assign custom keys to ESC, RIGHT and LEFT keyboard's actions. */ @Input() keyboardConfig: KeyboardConfig; /** * enableCloseOutside's input to enable modal-gallery close's behaviour while clicking * on the semi-transparent background. Disabled by default. */ @Input() enableCloseOutside: boolean = false; /** * Object of type `SlideConfig` to configure sliding of modal gallery. */ @Input() slideConfig: SlideConfig; /** * DEPRECATED * -----REMOVE THIS IN 4.0.0----- deprecated both showDownloadButton and showExtUrlButton */ @Input() showDownloadButton: boolean = false; // deprecated /** * DEPRECATED * -----REMOVE THIS IN 4.0.0----- deprecated both showDownloadButton and showExtUrlButton */ @Input() showExtUrlButton: boolean = false; // deprecated @Output() close: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>(); @Output() show: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>(); @Output() firstImage: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>(); @Output() lastImage: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>(); @Output() hasData: EventEmitter<ImageModalEvent> = new EventEmitter<ImageModalEvent>(); @Output() selectChanged: EventEmitter<Image> = new EventEmitter<Image>(); selectedImageCount: number = 0; /** * Boolean that it is true if the modal gallery is visible */ opened: boolean = false; /** * Boolean that it is true if an image of the modal gallery is still loading */ loading: boolean = false; /** * Boolean to open the modal gallery. Closed by default. */ showGallery: boolean = false; /** * Array of `Image` that represent the model of this library with all images, thumbs and so on. */ images: Image[]; /** * `Image` currently visible. */ currentImage: Image; /** * Number that represents the index of the current image. */ currentImageIndex: number = 0; /** * Object of type `ButtonsConfig` used to configure buttons visibility. This is a temporary value * initialized by the real `buttonsConfig`'s input */ configButtons: ButtonsConfig; /** * Enum of type `Action` used to pass a click action when you click on the modal image. * Declared here to be used inside the template. */ clickAction: Action = Action.CLICK; /** * Boolean that it's true when you are watching the first image (currently visible). */ isFirstImage: boolean = false; /** * Boolean that it's true when you are watching the last image (currently visible). */ isLastImage: boolean = false; canSelectImage: boolean = false; /** * Paging related variables */ totalImageCount: number = 0; pageSize: number = 20; pageCount: number = 0; currentPage: number = 0; /** * Private SWIPE_ACTION to define all swipe actions used by hammerjs. */ private SWIPE_ACTION = { LEFT: 'swipeleft', RIGHT: 'swiperight', UP: 'swipeup', DOWN: 'swipedown' }; /** * When you pass an Observable of `Image`s as `modalImages`, you have to subscribe to that * Observable. So, to prevent memory leaks, you must store the subscription and call `unsubscribe` in * OnDestroy. */ private subscription: Subscription; /** * Listener to catch keyboard's events and call the right method based on the key. * For instance, pressing esc, this will call `closeGallery(Action.KEYBOARD)` and so on. * If you passed a valid `keyboardConfig` esc, right and left buttons will be customized based on your data. * @param e KeyboardEvent caught by the listener. */ @HostListener('window:keydown', ['$event']) onKeyDown(e: KeyboardEvent) { if (!this.opened) { return; } const esc: number = this.keyboardConfig && this.keyboardConfig.esc ? this.keyboardConfig.esc : Keyboard.ESC; const right: number = this.keyboardConfig && this.keyboardConfig.right ? this.keyboardConfig.right : Keyboard.RIGHT_ARROW; const left: number = this.keyboardConfig && this.keyboardConfig.left ? this.keyboardConfig.left : Keyboard.LEFT_ARROW; switch (e.keyCode) { case esc: this.closeGallery(Action.KEYBOARD); break; case right: this.nextImage(Action.KEYBOARD); break; case left: this.prevImage(Action.KEYBOARD); break; } } /** * Constructor with the injection of ´KeyboardService´ that initialize some description fields * based on default values. */ constructor(private keyboardService: KeyboardService) { // if description isn't provided initialize it with a default object if (!this.description) { this.description = { imageText: 'Image ', numberSeparator: '/', beforeTextDescription: ' - ' }; } // if one of the Description fields isn't initialized, provide a default value this.description.imageText = this.description.imageText || 'Image '; this.description.numberSeparator = this.description.numberSeparator || '/'; this.description.beforeTextDescription = this.description.beforeTextDescription || ' - '; } /** * This method will initialize the pager when the images are loaded. */ initializePager(){ if (this.images.length > 0){ this.totalImageCount = this.images.length; this.pageCount = this.totalImageCount / this.pageSize; this.currentPage = 1; } } /** * Method ´ngOnInit´ to build `configButtons` and to call `initImages()`. * This is an Angular's lifecycle hook, so its called automatically by Angular itself. * In particular, it's called only one time!!! */ ngOnInit() { // build configButtons to use it inside upper-buttons this.configButtons = { download: this.showDownloadButton || (this.buttonsConfig && this.buttonsConfig.download), extUrl: this.showExtUrlButton || (this.buttonsConfig && this.buttonsConfig.extUrl), close: (this.buttonsConfig && this.buttonsConfig.close) }; // call initImages passing true as parameter, because I want to emit `hasData` event this.initImages(true); } /** * Method ´ngOnChanges´ to init images preventing errors. * This is an Angular's lifecycle hook, so its called automatically by Angular itself. * In particular, it's called before `ngOnInit()` and whenever one or more data-bound input properties change. * @param changes `SimpleChanges` object of current and previous property values provided by Angular. */ ngOnChanges(changes: SimpleChanges) {
tImageCountsToDisplay(){ var selectedImages = this.images.filter(image=>{ return image.selected === true; }); var selectedImageCount = selectedImages.length; var tobeselected = this.selectionLimit - selectedImageCount; this.canSelectImage = tobeselected <= 0 && !this.currentImage.selected; return "You need to select " + tobeselected + " images." } /** * Method `getDescriptionToDisplay` to get the image description based on input params. * If you provide a full description this will be the visible description, otherwise, * it will be built using the `description` object, concatenating its fields. * @returns String description to display. */ getDescriptionToDisplay() { if (this.description && this.description.customFullDescription) { return this.description.customFullDescription; } // If the current image hasn't a description, // prevent to write the ' - ' (or this.description.beforeTextDescription) if (!this.currentImage.description || this.currentImage.description === '') { return `${this.description.imageText}${this.currentImageIndex + 1}${this.description.numberSeparator}${this.images.length}`; } return `${this.description.imageText}${this.currentImageIndex + 1}${this.description.numberSeparator}${this.images.length}${this.description.beforeTextDescription}${this.currentImage.description}`; } /** * Method `swipe
// to prevent errors when you pass to this library // the array of images inside a subscribe block, in this way: `...subscribe(val => { this.images = arrayOfImages })` // As you can see, I'm providing examples in these situations in all official demos if (this.modalImages) { // I pass `false` as parameter, because I DON'T want to emit `hasData` // event (preventing multiple hasData events while initializing) this.initImages(false); } } ge
identifier_body
modal-gallery.component.ts
of `Image`s as `modalImages`, you have to subscribe to that * Observable. So, to prevent memory leaks, you must store the subscription and call `unsubscribe` in * OnDestroy. */ private subscription: Subscription; /** * Listener to catch keyboard's events and call the right method based on the key. * For instance, pressing esc, this will call `closeGallery(Action.KEYBOARD)` and so on. * If you passed a valid `keyboardConfig` esc, right and left buttons will be customized based on your data. * @param e KeyboardEvent caught by the listener. */ @HostListener('window:keydown', ['$event']) onKeyDown(e: KeyboardEvent) { if (!this.opened) { return; } const esc: number = this.keyboardConfig && this.keyboardConfig.esc ? this.keyboardConfig.esc : Keyboard.ESC; const right: number = this.keyboardConfig && this.keyboardConfig.right ? this.keyboardConfig.right : Keyboard.RIGHT_ARROW; const left: number = this.keyboardConfig && this.keyboardConfig.left ? this.keyboardConfig.left : Keyboard.LEFT_ARROW; switch (e.keyCode) { case esc: this.closeGallery(Action.KEYBOARD); break; case right: this.nextImage(Action.KEYBOARD); break; case left: this.prevImage(Action.KEYBOARD); break; } } /** * Constructor with the injection of ´KeyboardService´ that initialize some description fields * based on default values. */ constructor(private keyboardService: KeyboardService) { // if description isn't provided initialize it with a default object if (!this.description) { this.description = { imageText: 'Image ', numberSeparator: '/', beforeTextDescription: ' - ' }; } // if one of the Description fields isn't initialized, provide a default value this.description.imageText = this.description.imageText || 'Image '; this.description.numberSeparator = this.description.numberSeparator || '/'; this.description.beforeTextDescription = this.description.beforeTextDescription || ' - '; } /** * This method will initialize the pager when the images are loaded. */ initializePager(){ if (this.images.length > 0){ this.totalImageCount = this.images.length; this.pageCount = this.totalImageCount / this.pageSize; this.currentPage = 1; } } /** * Method ´ngOnInit´ to build `configButtons` and to call `initImages()`. * This is an Angular's lifecycle hook, so its called automatically by Angular itself. * In particular, it's called only one time!!! */ ngOnInit() { // build configButtons to use it inside upper-buttons this.configButtons = { download: this.showDownloadButton || (this.buttonsConfig && this.buttonsConfig.download), extUrl: this.showExtUrlButton || (this.buttonsConfig && this.buttonsConfig.extUrl), close: (this.buttonsConfig && this.buttonsConfig.close) }; // call initImages passing true as parameter, because I want to emit `hasData` event this.initImages(true); } /** * Method ´ngOnChanges´ to init images preventing errors. * This is an Angular's lifecycle hook, so its called automatically by Angular itself. * In particular, it's called before `ngOnInit()` and whenever one or more data-bound input properties change. * @param changes `SimpleChanges` object of current and previous property values provided by Angular. */ ngOnChanges(changes: SimpleChanges) { // to prevent errors when you pass to this library // the array of images inside a subscribe block, in this way: `...subscribe(val => { this.images = arrayOfImages })` // As you can see, I'm providing examples in these situations in all official demos if (this.modalImages) { // I pass `false` as parameter, because I DON'T want to emit `hasData` // event (preventing multiple hasData events while initializing) this.initImages(false); } } getImageCountsToDisplay(){ var selectedImages = this.images.filter(image=>{ return image.selected === true; }); var selectedImageCount = selectedImages.length; var tobeselected = this.selectionLimit - selectedImageCount; this.canSelectImage = tobeselected <= 0 && !this.currentImage.selected; return "You need to select " + tobeselected + " images." } /** * Method `getDescriptionToDisplay` to get the image description based on input params. * If you provide a full description this will be the visible description, otherwise, * it will be built using the `description` object, concatenating its fields. * @returns String description to display. */ getDescriptionToDisplay() { if (this.description && this.description.customFullDescription) { return this.description.customFullDescription; } // If the current image hasn't a description, // prevent to write the ' - ' (or this.description.beforeTextDescription) if (!this.currentImage.description || this.currentImage.description === '') { return `${this.description.imageText}${this.currentImageIndex + 1}${this.description.numberSeparator}${this.images.length}`; } return `${this.description.imageText}${this.currentImageIndex + 1}${this.description.numberSeparator}${this.images.length}${this.description.beforeTextDescription}${this.currentImage.description}`; } /** * Method `swipe` used by Hammerjs to support touch gestures. * @param index Number that represent the current visible index * @param action String that represent the direction of the swipe action. 'swiperight' by default. */ swipe(index: number, action = this.SWIPE_ACTION.RIGHT) { switch (action) { case this.SWIPE_ACTION.RIGHT: this.nextImage(Action.SWIPE); break; case this.SWIPE_ACTION.LEFT: this.prevImage(Action.SWIPE); break; // case this.SWIPE_ACTION.UP: // break; // case this.SWIPE_ACTION.DOWN: // break; } } /** * Method `closeGallery` to close the modal gallery. * @param action Enum of type `Action` that represents the source * action that closed the modal gallery. NORMAL by default. */ closeGallery(action: Action = Action.NORMAL) { this.close.emit(new ImageModalEvent(action, true)); this.opened = false; this.keyboardService.reset(); } imageSelectionChangedComponent(image: any){ this.selectChanged.emit(image); } /** * Method `prevImage` to go back to the previous image shown into the modal gallery. * @param action Enum of type `Action` that represents the source * action that moved back to the previous image. NORMAL by default. */ prevImage(action: Action = Action.NORMAL) { // check if prevImage should be blocked if (this.isPreventSliding(0)) { return; } this.loading = true; this.currentImageIndex = this.getPrevIndex(action, this.currentImageIndex); this.showModalGallery(this.currentImageIndex); } /** * Method `nextImage` to go back to the previous image shown into the modal gallery. * @param action Enum of type `Action` that represents the source * action that moved to the next image. NORMAL by default. */ nextImage(action: Action = Action.NORMAL) { // check if nextImage should be blocked if (this.isPreventSliding(this.images.length - 1)) { return; } this.loading = true; this.currentImageIndex = this.getNextIndex(action, this.currentImageIndex); this.showModalGallery(this.currentImageIndex); } /** * Method `onShowModalGallery` called when you click on an image of your gallery. * The input index is the index of the clicked image thumb. * @param index Number that represents the index of the clicked image. */ onShowModalGallery(index: number) { this.showModalGallery(index); } /** * Method `showModalGallery` to show the modal gallery displaying the image with * the index specified as input parameter. * It will also register a new `keyboardService` to catch keyboard's events to download the current * image with keyboard's shortcuts. This service, will be removed when modal gallery component will be destroyed. * @param index Number that represents the index of the image to show. */ showModalGallery(index: number) { this.keyboardService.add((event: KeyboardEvent, combo: string) => { if (event.preventDefault) { event.preventDefault(); } else { // internet explorer event.returnValue = false; } this.downloadImage(); }); // enable/disable 'infinite sliding' based on @Input() slideConfig this.manageSlideConfig(index); this.currentImageIndex = index; this.opened = true; this.currentImage = this.images[this.currentImageIndex]; this.loading = false; // emit current visible image index this.show.emit(new ImageModalEvent(Action.LOAD, this.currentImageIndex + 1)); } /** * Method `downloadImage` to download the current visible image, only if `downloadable` is true. * For IE, this will navigate to the image instead of a direct download as in all modern browsers. */ downlo
adImage() {
identifier_name
lib.rs
(jit: bool) -> Box<dyn TargetIsa + 'static> { let mut flags_builder = cranelift::codegen::settings::builder(); // `simplejit` requires non-PIC code if !jit { // allow creating shared libraries flags_builder .enable("is_pic") .expect("is_pic should be a valid option"); } // use debug assertions flags_builder .enable("enable_verifier") .expect("enable_verifier should be a valid option"); // don't emit call to __cranelift_probestack flags_builder .set("enable_probestack", "false") .expect("enable_probestack should be a valid option"); let flags = Flags::new(flags_builder); cranelift::codegen::isa::lookup(TARGET) .unwrap_or_else(|_| panic!("platform not supported: {}", TARGET)) .finish(flags) } pub fn initialize_aot_module(name: String) -> Module<ObjectBackend> { let builder = ObjectBuilder::new( get_isa(false), name, cranelift_module::default_libcall_names(), ); Module::new(builder.expect("unsupported binary format or target architecture")) } enum Id { Function(FuncId), Global(DataId), Local(StackSlot), } struct Compiler<T: Backend> { module: Module<T>, debug: bool, // if false, we last saw a switch last_saw_loop: bool, strings: HashMap<Vec<u8>, DataId>, declarations: HashMap<Symbol, Id>, loops: Vec<(Block, Block)>, // switch, default, end // if default is empty once we get to the end of a switch body, // we didn't see a default case switches: Vec<(Switch, Option<Block>, Block)>, labels: HashMap<InternedStr, Block>, error_handler: ErrorHandler, } impl<B: Backend> Compiler<B> { fn new(module: Module<B>, debug: bool) -> Compiler<B> { Compiler { module, declarations: HashMap::new(), loops: Vec::new(), switches: Vec::new(), labels: HashMap::new(), // the initial value doesn't really matter last_saw_loop: true, strings: Default::default(), error_handler: Default::default(), debug, } } // we have to consider the following cases: // 1. declaration before definition // 2. 2nd declaration before definition // 3. definition // 4. declaration after definition // 1. should declare `id` a import unless specified as `static`. // 3. should always declare `id` as export or local. // 2. and 4. should be a no-op. fn declare_func(&mut self, symbol: Symbol, is_definition: bool) -> CompileResult<FuncId> { use saltwater_parser::get_str; if !is_definition { // case 2 and 4 if let Some(Id::Function(func_id)) = self.declarations.get(&symbol) { return Ok(*func_id); } } let metadata = symbol.get(); let func_type = match &metadata.ctype { Type::Function(func_type) => func_type, _ => unreachable!("bug in backend: only functions should be passed to `declare_func`"), }; let signature = func_type.signature(self.module.isa()); let linkage = match metadata.storage_class { StorageClass::Auto | StorageClass::Extern if is_definition => Linkage::Export, StorageClass::Auto | StorageClass::Extern => Linkage::Import, StorageClass::Static => Linkage::Local, StorageClass::Register | StorageClass::Typedef => unreachable!(), }; let func_id = self .module .declare_function(get_str!(metadata.id), linkage, &signature) .unwrap_or_else(|err| panic!("{}", err)); self.declarations.insert(symbol, Id::Function(func_id)); Ok(func_id) } /// declare an object on the stack fn declare_stack( &mut self, decl: Declaration, location: Location, builder: &mut FunctionBuilder, ) -> CompileResult<()> { let meta = decl.symbol.get(); if let StorageClass::Typedef = meta.storage_class { return Ok(()); } if let Type::Function(_) = &meta.ctype { self.declare_func(decl.symbol, false)?; return Ok(()); } let u64_size = match meta.ctype.sizeof() { Ok(size) => size, Err(err) => { return Err(CompileError::semantic(Locatable { data: err.into(), location, })) } }; let kind = StackSlotKind::ExplicitSlot; let size = match u32::try_from(u64_size) { Ok(size) => size, Err(_) => return Err(CompileError::semantic(Locatable { data: "cannot store items on the stack that are more than 4 GB, it will overflow the stack".into(), location, })) }; let data = StackSlotData { kind, size, offset: None, }; let stack_slot = builder.create_stack_slot(data); self.declarations.insert(decl.symbol, Id::Local(stack_slot)); if let Some(init) = decl.init { self.store_stack(init, stack_slot, builder)?; } Ok(()) } fn store_stack( &mut self, init: Initializer, stack_slot: StackSlot, builder: &mut FunctionBuilder, ) -> CompileResult<()> { match init { Initializer::Scalar(expr) => { let val = self.compile_expr(*expr, builder)?; // TODO: replace with `builder.ins().stack_store(val.ir_val, stack_slot, 0);` // when Cranelift implements stack_store for i8 and i16 let addr = builder.ins().stack_addr(Type::ptr_type(), stack_slot, 0); builder.ins().store(MemFlags::new(), val.ir_val, addr, 0); } Initializer::InitializerList(_) => unimplemented!("aggregate dynamic initialization"), Initializer::FunctionBody(_) => unreachable!("functions can't be stored on the stack"), } Ok(()) } // TODO: this is grossly inefficient, ask Cranelift devs if // there's an easier way to make parameters modifiable. fn store_stack_params( &mut self, params: &[Symbol], func_start: Block, location: &Location, builder: &mut FunctionBuilder, ) -> CompileResult<()> { // Cranelift requires that all block params are declared up front let ir_vals: Vec<_> = params .iter() .map(|param| { let ir_type = param.get().ctype.as_ir_type(); Ok(builder.append_block_param(func_start, ir_type)) }) .collect::<CompileResult<_>>()?; for (&param, ir_val) in params.iter().zip(ir_vals) { let u64_size = match param.get().ctype.sizeof() { Err(data) => semantic_err!(data.into(), *location), Ok(size) => size, }; let u32_size = match u32::try_from(u64_size) { Err(_) => semantic_err!( format!( "size {} is too large for stack (can only handle 32-bit values)", u64_size ), *location ), Ok(size) => size, }; let stack_data = StackSlotData { kind: StackSlotKind::ExplicitSlot, size: u32_size, offset: None, }; let slot = builder.create_stack_slot(stack_data); // TODO: need to take the address before storing until Cranelift implements // stores for i8 and i16 // then this can be replaced with `builder.ins().stack_store(ir_val, slot, 0);` // See https://github.com/CraneStation/cranelift/issues/433 let addr = builder.ins().stack_addr(Type::ptr_type(), slot, 0); builder.ins().store(MemFlags::new(), ir_val, addr, 0); self.declarations.insert(param, Id::Local(slot)); } Ok(()) } fn compile_func( &mut self, symbol: Symbol, func_type: &FunctionType, stmts: Vec<Stmt>, location: Location, ) -> CompileResult<()> { let func_id = self.declare_func(symbol, true)?; // TODO: make declare_func should take a `signature` after all? // This just calculates it twice, it's probably fine let signature = func_type.signature(self.module.isa()); // external name is meant to be a lookup in a symbol table, // but we just give it garbage values let mut func = Function::with_name_signature(ExternalName::user(0, 0), signature); // this context is just boiler plate let mut ctx = FunctionBuilderContext::new(); let mut builder = FunctionBuilder::new(&mut func, &mut ctx); let func_start = builder.create_block(); builder.switch_to_block(func_start); let should_ret = func_type.should_return(); if
get_isa
identifier_name
lib.rs
: true, strings: Default::default(), error_handler: Default::default(), debug, } } // we have to consider the following cases: // 1. declaration before definition // 2. 2nd declaration before definition // 3. definition // 4. declaration after definition // 1. should declare `id` a import unless specified as `static`. // 3. should always declare `id` as export or local. // 2. and 4. should be a no-op. fn declare_func(&mut self, symbol: Symbol, is_definition: bool) -> CompileResult<FuncId> { use saltwater_parser::get_str; if !is_definition { // case 2 and 4 if let Some(Id::Function(func_id)) = self.declarations.get(&symbol) { return Ok(*func_id); } } let metadata = symbol.get(); let func_type = match &metadata.ctype { Type::Function(func_type) => func_type, _ => unreachable!("bug in backend: only functions should be passed to `declare_func`"), }; let signature = func_type.signature(self.module.isa()); let linkage = match metadata.storage_class { StorageClass::Auto | StorageClass::Extern if is_definition => Linkage::Export, StorageClass::Auto | StorageClass::Extern => Linkage::Import, StorageClass::Static => Linkage::Local, StorageClass::Register | StorageClass::Typedef => unreachable!(), }; let func_id = self .module .declare_function(get_str!(metadata.id), linkage, &signature) .unwrap_or_else(|err| panic!("{}", err)); self.declarations.insert(symbol, Id::Function(func_id)); Ok(func_id) } /// declare an object on the stack fn declare_stack( &mut self, decl: Declaration, location: Location, builder: &mut FunctionBuilder, ) -> CompileResult<()> { let meta = decl.symbol.get(); if let StorageClass::Typedef = meta.storage_class { return Ok(()); } if let Type::Function(_) = &meta.ctype { self.declare_func(decl.symbol, false)?; return Ok(()); } let u64_size = match meta.ctype.sizeof() { Ok(size) => size, Err(err) => { return Err(CompileError::semantic(Locatable { data: err.into(), location, })) } }; let kind = StackSlotKind::ExplicitSlot; let size = match u32::try_from(u64_size) { Ok(size) => size, Err(_) => return Err(CompileError::semantic(Locatable { data: "cannot store items on the stack that are more than 4 GB, it will overflow the stack".into(), location, })) }; let data = StackSlotData { kind, size, offset: None, }; let stack_slot = builder.create_stack_slot(data); self.declarations.insert(decl.symbol, Id::Local(stack_slot)); if let Some(init) = decl.init { self.store_stack(init, stack_slot, builder)?; } Ok(()) } fn store_stack( &mut self, init: Initializer, stack_slot: StackSlot, builder: &mut FunctionBuilder, ) -> CompileResult<()>
// TODO: this is grossly inefficient, ask Cranelift devs if // there's an easier way to make parameters modifiable. fn store_stack_params( &mut self, params: &[Symbol], func_start: Block, location: &Location, builder: &mut FunctionBuilder, ) -> CompileResult<()> { // Cranelift requires that all block params are declared up front let ir_vals: Vec<_> = params .iter() .map(|param| { let ir_type = param.get().ctype.as_ir_type(); Ok(builder.append_block_param(func_start, ir_type)) }) .collect::<CompileResult<_>>()?; for (&param, ir_val) in params.iter().zip(ir_vals) { let u64_size = match param.get().ctype.sizeof() { Err(data) => semantic_err!(data.into(), *location), Ok(size) => size, }; let u32_size = match u32::try_from(u64_size) { Err(_) => semantic_err!( format!( "size {} is too large for stack (can only handle 32-bit values)", u64_size ), *location ), Ok(size) => size, }; let stack_data = StackSlotData { kind: StackSlotKind::ExplicitSlot, size: u32_size, offset: None, }; let slot = builder.create_stack_slot(stack_data); // TODO: need to take the address before storing until Cranelift implements // stores for i8 and i16 // then this can be replaced with `builder.ins().stack_store(ir_val, slot, 0);` // See https://github.com/CraneStation/cranelift/issues/433 let addr = builder.ins().stack_addr(Type::ptr_type(), slot, 0); builder.ins().store(MemFlags::new(), ir_val, addr, 0); self.declarations.insert(param, Id::Local(slot)); } Ok(()) } fn compile_func( &mut self, symbol: Symbol, func_type: &FunctionType, stmts: Vec<Stmt>, location: Location, ) -> CompileResult<()> { let func_id = self.declare_func(symbol, true)?; // TODO: make declare_func should take a `signature` after all? // This just calculates it twice, it's probably fine let signature = func_type.signature(self.module.isa()); // external name is meant to be a lookup in a symbol table, // but we just give it garbage values let mut func = Function::with_name_signature(ExternalName::user(0, 0), signature); // this context is just boiler plate let mut ctx = FunctionBuilderContext::new(); let mut builder = FunctionBuilder::new(&mut func, &mut ctx); let func_start = builder.create_block(); builder.switch_to_block(func_start); let should_ret = func_type.should_return(); if func_type.has_params() { self.store_stack_params( // TODO: get rid of this clone &func_type.params, func_start, &location, &mut builder, )?; } self.compile_all(stmts, &mut builder)?; if !builder.is_filled() { let id = symbol.get().id; if id == InternedStr::get_or_intern("main") { let ir_int = func_type.return_type.as_ir_type(); let zero = [builder.ins().iconst(ir_int, 0)]; builder.ins().return_(&zero); } else if should_ret { semantic_err!( format!( "expected a return statement before end of function '{}' returning {}", id, func_type.return_type ), location ); } else { // void function, return nothing builder.ins().return_(&[]); } } builder.seal_all_blocks(); builder.finalize(); let flags = settings::Flags::new(settings::builder()); if self.debug { println!("ir: {}", func); } if let Err(err) = codegen::verify_function(&func, &flags) { panic!( "verification error: {}\nnote: while compiling {}", err, func ); } let mut ctx = codegen::Context::for_function(func); let mut trap_sink = codegen::binemit::NullTrapSink {}; if let Err(err) = self .module .define_function(func_id, &mut ctx, &mut trap_sink) { panic!( "definition error: {}\nnote: while compiling {}", err, ctx.func ); } Ok(()) } } pub type Product = <cranelift_object::ObjectBackend as Backend>::Product; /// Compile and return the declarations and warnings. pub fn compile<B: Backend>(module: Module<B>, buf: &str, opt: Opt) -> Program<Module<B>> { use saltwater_parser::{check_semantics, vec_deque}; let debug_asm = opt.debug_asm; let mut program = check_semantics(buf, opt); let hir = match program.result { Ok(hir) => hir, Err(err) => { return Program { result:
{ match init { Initializer::Scalar(expr) => { let val = self.compile_expr(*expr, builder)?; // TODO: replace with `builder.ins().stack_store(val.ir_val, stack_slot, 0);` // when Cranelift implements stack_store for i8 and i16 let addr = builder.ins().stack_addr(Type::ptr_type(), stack_slot, 0); builder.ins().store(MemFlags::new(), val.ir_val, addr, 0); } Initializer::InitializerList(_) => unimplemented!("aggregate dynamic initialization"), Initializer::FunctionBody(_) => unreachable!("functions can't be stored on the stack"), } Ok(()) }
identifier_body
lib.rs
()); } let u64_size = match meta.ctype.sizeof() { Ok(size) => size, Err(err) => { return Err(CompileError::semantic(Locatable { data: err.into(), location, })) } }; let kind = StackSlotKind::ExplicitSlot; let size = match u32::try_from(u64_size) { Ok(size) => size, Err(_) => return Err(CompileError::semantic(Locatable { data: "cannot store items on the stack that are more than 4 GB, it will overflow the stack".into(), location, })) }; let data = StackSlotData { kind, size, offset: None, }; let stack_slot = builder.create_stack_slot(data); self.declarations.insert(decl.symbol, Id::Local(stack_slot)); if let Some(init) = decl.init { self.store_stack(init, stack_slot, builder)?; } Ok(()) } fn store_stack( &mut self, init: Initializer, stack_slot: StackSlot, builder: &mut FunctionBuilder, ) -> CompileResult<()> { match init { Initializer::Scalar(expr) => { let val = self.compile_expr(*expr, builder)?; // TODO: replace with `builder.ins().stack_store(val.ir_val, stack_slot, 0);` // when Cranelift implements stack_store for i8 and i16 let addr = builder.ins().stack_addr(Type::ptr_type(), stack_slot, 0); builder.ins().store(MemFlags::new(), val.ir_val, addr, 0); } Initializer::InitializerList(_) => unimplemented!("aggregate dynamic initialization"), Initializer::FunctionBody(_) => unreachable!("functions can't be stored on the stack"), } Ok(()) } // TODO: this is grossly inefficient, ask Cranelift devs if // there's an easier way to make parameters modifiable. fn store_stack_params( &mut self, params: &[Symbol], func_start: Block, location: &Location, builder: &mut FunctionBuilder, ) -> CompileResult<()> { // Cranelift requires that all block params are declared up front let ir_vals: Vec<_> = params .iter() .map(|param| { let ir_type = param.get().ctype.as_ir_type(); Ok(builder.append_block_param(func_start, ir_type)) }) .collect::<CompileResult<_>>()?; for (&param, ir_val) in params.iter().zip(ir_vals) { let u64_size = match param.get().ctype.sizeof() { Err(data) => semantic_err!(data.into(), *location), Ok(size) => size, }; let u32_size = match u32::try_from(u64_size) { Err(_) => semantic_err!( format!( "size {} is too large for stack (can only handle 32-bit values)", u64_size ), *location ), Ok(size) => size, }; let stack_data = StackSlotData { kind: StackSlotKind::ExplicitSlot, size: u32_size, offset: None, }; let slot = builder.create_stack_slot(stack_data); // TODO: need to take the address before storing until Cranelift implements // stores for i8 and i16 // then this can be replaced with `builder.ins().stack_store(ir_val, slot, 0);` // See https://github.com/CraneStation/cranelift/issues/433 let addr = builder.ins().stack_addr(Type::ptr_type(), slot, 0); builder.ins().store(MemFlags::new(), ir_val, addr, 0); self.declarations.insert(param, Id::Local(slot)); } Ok(()) } fn compile_func( &mut self, symbol: Symbol, func_type: &FunctionType, stmts: Vec<Stmt>, location: Location, ) -> CompileResult<()> { let func_id = self.declare_func(symbol, true)?; // TODO: make declare_func should take a `signature` after all? // This just calculates it twice, it's probably fine let signature = func_type.signature(self.module.isa()); // external name is meant to be a lookup in a symbol table, // but we just give it garbage values let mut func = Function::with_name_signature(ExternalName::user(0, 0), signature); // this context is just boiler plate let mut ctx = FunctionBuilderContext::new(); let mut builder = FunctionBuilder::new(&mut func, &mut ctx); let func_start = builder.create_block(); builder.switch_to_block(func_start); let should_ret = func_type.should_return(); if func_type.has_params() { self.store_stack_params( // TODO: get rid of this clone &func_type.params, func_start, &location, &mut builder, )?; } self.compile_all(stmts, &mut builder)?; if !builder.is_filled() { let id = symbol.get().id; if id == InternedStr::get_or_intern("main") { let ir_int = func_type.return_type.as_ir_type(); let zero = [builder.ins().iconst(ir_int, 0)]; builder.ins().return_(&zero); } else if should_ret { semantic_err!( format!( "expected a return statement before end of function '{}' returning {}", id, func_type.return_type ), location ); } else { // void function, return nothing builder.ins().return_(&[]); } } builder.seal_all_blocks(); builder.finalize(); let flags = settings::Flags::new(settings::builder()); if self.debug { println!("ir: {}", func); } if let Err(err) = codegen::verify_function(&func, &flags) { panic!( "verification error: {}\nnote: while compiling {}", err, func ); } let mut ctx = codegen::Context::for_function(func); let mut trap_sink = codegen::binemit::NullTrapSink {}; if let Err(err) = self .module .define_function(func_id, &mut ctx, &mut trap_sink) { panic!( "definition error: {}\nnote: while compiling {}", err, ctx.func ); } Ok(()) } } pub type Product = <cranelift_object::ObjectBackend as Backend>::Product; /// Compile and return the declarations and warnings. pub fn compile<B: Backend>(module: Module<B>, buf: &str, opt: Opt) -> Program<Module<B>> { use saltwater_parser::{check_semantics, vec_deque}; let debug_asm = opt.debug_asm; let mut program = check_semantics(buf, opt); let hir = match program.result { Ok(hir) => hir, Err(err) => { return Program { result: Err(err), warnings: program.warnings, files: program.files, } } }; // really we'd like to have all errors but that requires a refactor let mut err = None; let mut compiler = Compiler::new(module, debug_asm); for decl in hir { let meta = decl.data.symbol.get(); if let StorageClass::Typedef = meta.storage_class { continue; } let current = match &meta.ctype { Type::Function(func_type) => match decl.data.init { Some(Initializer::FunctionBody(stmts)) => { compiler.compile_func(decl.data.symbol, &func_type, stmts, decl.location) } None => compiler.declare_func(decl.data.symbol, false).map(|_| ()), _ => unreachable!("functions can only be initialized by a FunctionBody"), }, Type::Void | Type::Error => unreachable!("parser let an incomplete type through"), _ => { if let Some(Initializer::FunctionBody(_)) = &decl.data.init { unreachable!("only functions should have a function body") } compiler.store_static(decl.data.symbol, decl.data.init, decl.location) } }; if let Err(e) = current { err = Some(e); break; } } let warns = compiler.error_handler.warnings; let (result, ir_warnings) = if let Some(err) = err { (Err(err), warns) } else { (Ok(compiler.module), warns) }; program.warnings.extend(ir_warnings); Program { result: result.map_err(|errs| vec_deque![errs]), warnings: program.warnings, files: program.files, } } pub fn assemble(product: Product, output: &Path) -> Result<(), saltwater_parser::Error> { use std::fs::File; use std::io::{self, Write}; let bytes = product.emit().map_err(saltwater_parser::Error::Platform)?; File::create(output)? .write_all(&bytes) .map_err(io::Error::into) } pub fn link(obj_file: &Path, output: &Path) -> Result<(), std::io::Error> {
use std::io::{Error, ErrorKind}; use std::process::Command;
random_line_split
cef.go
func (d defaultLogger) Panicf(fmt string, args ... interface{}) { log.Panicf("[cef] " + fmt, args...) } // Sandbox is disabled. Including the "cef_sandbox.lib" // library results in lots of GCC warnings/errors. It is // compatible only with VS 2010. It would be required to // build it using GCC. Add -lcef_sandbox to LDFLAGS. // capi doesn't expose sandbox functions, you need do add // these before import "C": // void* cef_sandbox_info_create(); // void cef_sandbox_info_destroy(void* sandbox_info); var _SandboxInfo unsafe.Pointer type Settings struct { CachePath string LogSeverity int LogFile string ResourcesDirPath string LocalesDirPath string RemoteDebuggingPort int PersistSessionCookies bool IgnoreCertificateErrors int } type CefState int var ( STATE_DEFAULT CefState = 0 STATE_ENABLED CefState = 1 STATE_DISABLED CefState = 2 ) type BrowserSettings struct { StandardFontFamily string FixedFontFamily string SerifFontFamily string SansSerifFontFamily string CursiveFontFamily string FantasyFontFamily string DefaultFontSize int DefaultFixedFontSize int MinimumFontSize int MinimumLogicalFontSize int DefaultEncoding string RemoteFonts CefState Javascript CefState JavascriptOpenWindows CefState JavascriptCloseWindows CefState JavascriptAccessClipboard CefState JavascriptDomPaste CefState CaretBrowsing CefState Java CefState Plugins CefState UniversalAccessFromFileUrls CefState FileAccessFromFileUrls CefState WebSecurity CefState ImageLoading CefState ImageShrinkStandaloneToFit CefState TextAreaResize CefState TabToLinks CefState LocalStorage CefState Databases CefState ApplicationCache CefState Webgl CefState BackgroundColor uint32 } func _InitializeGlobalCStructures() { _InitializeGlobalCStructuresBase() //_InitializeGlobalCStructuresApp() /* _DisplayHandler = InitializeDisplayHandler() _DownloadHandler = InitializeDownloadHandler()*/ } func SetLogger(logger SimpleLogger) { Logger = logger } func ExecuteProcess(programHandle unsafe.Pointer, appHandler AppHandler) int { Logger.Infof("ExecuteProcess, args=%v", os.Args) _InitializeGlobalCStructures() if appHandler.GetAppHandlerT().CStruct == nil { panic("GetAppHandlerT cannot have a nil CStruct. Call NewAppHandlerT() to create one!") } FillMainArgs(_MainArgs, programHandle) // Sandbox info needs to be passed to both cef_execute_process() // and cef_initialize(). // OFF: _SandboxInfo = C.cef_sandbox_info_create() Logger.Infof("MainArgs %X _AppHanlder %X _SandboxInfo %X", _MainArgs, appHandler.GetAppHandlerT().CStruct, _SandboxInfo) go_AddRef(unsafe.Pointer(_MainArgs)) go_AddRef(unsafe.Pointer(appHandler.GetAppHandlerT().CStruct)) go_AddRef(unsafe.Pointer(_SandboxInfo)) var exitCode C.int = C.cef_execute_process(_MainArgs, appHandler.GetAppHandlerT().CStruct, _SandboxInfo) if (exitCode >= 0) { os.Exit(int(exitCode)) } return int(exitCode) } func Initialize(settings Settings, appHandler AppHandler) int { Logger.Infof("Initialize\n") if _MainArgs == nil { // _MainArgs structure is initialized and filled in ExecuteProcess. // If cef_execute_process is not called, and there is a call // to cef_initialize, then it would result in creation of infinite // number of processes. See Issue 1199 in CEF: // https://code.google.com/p/chromiumembedded/issues/detail?id=1199 Logger.Errorf("ERROR: missing a call to ExecuteProcess\n") return 0 } // Initialize cef_settings_t structure. var cefSettings *C.struct__cef_settings_t cefSettings = (*C.struct__cef_settings_t)( C.calloc(1, C.sizeof_struct__cef_settings_t)) cefSettings.size = C.sizeof_struct__cef_settings_t // cache_path // ---------- if (settings.CachePath != "") { Logger.Infof("CachePath=%s\n", settings.CachePath) } var cachePath *C.char = C.CString(settings.CachePath) defer C.free(unsafe.Pointer(cachePath)) C.cef_string_from_utf8(cachePath, C.strlen(cachePath), C.cefStringCastToCefString16(&cefSettings.cache_path)) // log_severity // ------------ cefSettings.log_severity = (C.cef_log_severity_t)(C.int(settings.LogSeverity)) // log_file // -------- if (settings.LogFile != "") { Logger.Infof("LogFile=%s\n", settings.LogFile) } var logFile *C.char = C.CString(settings.LogFile) defer C.free(unsafe.Pointer(logFile)) C.cef_string_from_utf8(logFile, C.strlen(logFile), C.cefStringCastToCefString16(&cefSettings.log_file)) // resources_dir_path // ------------------ if settings.ResourcesDirPath == "" && runtime.GOOS != "darwin" { // Setting this path is required for the tests to run fine. cwd, _ := os.Getwd() settings.ResourcesDirPath = cwd } if (settings.ResourcesDirPath != "") { Logger.Infof("ResourcesDirPath=%s\n", settings.ResourcesDirPath) } var resourcesDirPath *C.char = C.CString(settings.ResourcesDirPath) defer C.free(unsafe.Pointer(resourcesDirPath)) C.cef_string_from_utf8(resourcesDirPath, C.strlen(resourcesDirPath), C.cefStringCastToCefString16(&cefSettings.resources_dir_path)) // locales_dir_path // ---------------- if settings.LocalesDirPath == "" && runtime.GOOS != "darwin" { // Setting this path is required for the tests to run fine. cwd, _ := os.Getwd() settings.LocalesDirPath = cwd + "/locales" } if (settings.LocalesDirPath != "") { Logger.Infof("LocalesDirPath=%s\n", settings.LocalesDirPath) } var localesDirPath *C.char = C.CString(settings.LocalesDirPath) defer C.free(unsafe.Pointer(localesDirPath)) C.cef_string_from_utf8(localesDirPath, C.strlen(localesDirPath), C.cefStringCastToCefString16(&cefSettings.locales_dir_path)) if settings.PersistSessionCookies { cefSettings.persist_session_cookies = 1 } cefSettings.remote_debugging_port = C.int(settings.RemoteDebuggingPort) cefSettings.ignore_certificate_errors = C.int(settings.IgnoreCertificateErrors) // no_sandbox // ---------- cefSettings.no_sandbox = C.int(1) go_AddRef(unsafe.Pointer(_MainArgs)) go_AddRef(unsafe.Pointer(appHandler.GetAppHandlerT().CStruct)) go_AddRef(unsafe.Pointer(_SandboxInfo)) ret := C.cef_initialize(_MainArgs, cefSettings, appHandler.GetAppHandlerT().CStruct, _SandboxInfo) return int(ret) } func CreateBrowser(hwnd unsafe.Pointer, clientHandler ClientHandler, browserSettings BrowserSettings, url string) bool { Logger.Infof("CreateBrowser, url=%s\n", url) // Initialize cef_window_info_t structure. var windowInfo *C.cef_window_info_t windowInfo = (*C.cef_window_info_t)( C.calloc(1, C.sizeof_cef_window_info_t)) FillWindowInfo(windowInfo, hwnd) // url var cefUrl *C.cef_string_t cefUrl = (*C.cef_string_t)( C.calloc(1, C.sizeof_cef_string_t)) var charUrl *C.char = C.CString(url) defer C.free(unsafe.Pointer(charUrl)) C.cef_string_from_utf8(charUrl, C.strlen(charUrl), C.cefStringCastToCefString16(cefUrl)) // Initialize cef_browser_settings_t structure. cefBrowserSettings := browserSettings.toC() // Do not create the browser synchronously using the // cef_browser_host_create_browser_sync() function, as // it is unreliable. Instead obtain browser object in // life_span_handler::on_after_created. In that callback // keep CEF browser objects in a global map (cef window // handle -> cef browser) and introduce // a GetBrowserByWindowHandle() function. This
{ log.Printf("[cef] " + fmt, args...) }
identifier_body
cef.go
LogSeverity int LogFile string ResourcesDirPath string LocalesDirPath string RemoteDebuggingPort int PersistSessionCookies bool IgnoreCertificateErrors int } type CefState int var ( STATE_DEFAULT CefState = 0 STATE_ENABLED CefState = 1 STATE_DISABLED CefState = 2 ) type BrowserSettings struct { StandardFontFamily string FixedFontFamily string SerifFontFamily string SansSerifFontFamily string CursiveFontFamily string FantasyFontFamily string DefaultFontSize int DefaultFixedFontSize int MinimumFontSize int MinimumLogicalFontSize int DefaultEncoding string RemoteFonts CefState Javascript CefState JavascriptOpenWindows CefState JavascriptCloseWindows CefState JavascriptAccessClipboard CefState JavascriptDomPaste CefState CaretBrowsing CefState Java CefState Plugins CefState UniversalAccessFromFileUrls CefState FileAccessFromFileUrls CefState WebSecurity CefState ImageLoading CefState ImageShrinkStandaloneToFit CefState TextAreaResize CefState TabToLinks CefState LocalStorage CefState Databases CefState ApplicationCache CefState Webgl CefState BackgroundColor uint32 } func _InitializeGlobalCStructures() { _InitializeGlobalCStructuresBase() //_InitializeGlobalCStructuresApp() /* _DisplayHandler = InitializeDisplayHandler() _DownloadHandler = InitializeDownloadHandler()*/ } func SetLogger(logger SimpleLogger) { Logger = logger } func ExecuteProcess(programHandle unsafe.Pointer, appHandler AppHandler) int { Logger.Infof("ExecuteProcess, args=%v", os.Args) _InitializeGlobalCStructures() if appHandler.GetAppHandlerT().CStruct == nil { panic("GetAppHandlerT cannot have a nil CStruct. Call NewAppHandlerT() to create one!") } FillMainArgs(_MainArgs, programHandle) // Sandbox info needs to be passed to both cef_execute_process() // and cef_initialize(). // OFF: _SandboxInfo = C.cef_sandbox_info_create() Logger.Infof("MainArgs %X _AppHanlder %X _SandboxInfo %X", _MainArgs, appHandler.GetAppHandlerT().CStruct, _SandboxInfo) go_AddRef(unsafe.Pointer(_MainArgs)) go_AddRef(unsafe.Pointer(appHandler.GetAppHandlerT().CStruct)) go_AddRef(unsafe.Pointer(_SandboxInfo)) var exitCode C.int = C.cef_execute_process(_MainArgs, appHandler.GetAppHandlerT().CStruct, _SandboxInfo) if (exitCode >= 0) { os.Exit(int(exitCode)) } return int(exitCode) } func Initialize(settings Settings, appHandler AppHandler) int { Logger.Infof("Initialize\n") if _MainArgs == nil { // _MainArgs structure is initialized and filled in ExecuteProcess. // If cef_execute_process is not called, and there is a call // to cef_initialize, then it would result in creation of infinite // number of processes. See Issue 1199 in CEF: // https://code.google.com/p/chromiumembedded/issues/detail?id=1199 Logger.Errorf("ERROR: missing a call to ExecuteProcess\n") return 0 } // Initialize cef_settings_t structure. var cefSettings *C.struct__cef_settings_t cefSettings = (*C.struct__cef_settings_t)( C.calloc(1, C.sizeof_struct__cef_settings_t)) cefSettings.size = C.sizeof_struct__cef_settings_t // cache_path // ---------- if (settings.CachePath != "") { Logger.Infof("CachePath=%s\n", settings.CachePath) } var cachePath *C.char = C.CString(settings.CachePath) defer C.free(unsafe.Pointer(cachePath)) C.cef_string_from_utf8(cachePath, C.strlen(cachePath), C.cefStringCastToCefString16(&cefSettings.cache_path)) // log_severity // ------------ cefSettings.log_severity = (C.cef_log_severity_t)(C.int(settings.LogSeverity)) // log_file // -------- if (settings.LogFile != "") { Logger.Infof("LogFile=%s\n", settings.LogFile) } var logFile *C.char = C.CString(settings.LogFile) defer C.free(unsafe.Pointer(logFile)) C.cef_string_from_utf8(logFile, C.strlen(logFile), C.cefStringCastToCefString16(&cefSettings.log_file)) // resources_dir_path // ------------------ if settings.ResourcesDirPath == "" && runtime.GOOS != "darwin" { // Setting this path is required for the tests to run fine. cwd, _ := os.Getwd() settings.ResourcesDirPath = cwd } if (settings.ResourcesDirPath != "") { Logger.Infof("ResourcesDirPath=%s\n", settings.ResourcesDirPath) } var resourcesDirPath *C.char = C.CString(settings.ResourcesDirPath) defer C.free(unsafe.Pointer(resourcesDirPath)) C.cef_string_from_utf8(resourcesDirPath, C.strlen(resourcesDirPath), C.cefStringCastToCefString16(&cefSettings.resources_dir_path)) // locales_dir_path // ---------------- if settings.LocalesDirPath == "" && runtime.GOOS != "darwin" { // Setting this path is required for the tests to run fine. cwd, _ := os.Getwd() settings.LocalesDirPath = cwd + "/locales" } if (settings.LocalesDirPath != "") { Logger.Infof("LocalesDirPath=%s\n", settings.LocalesDirPath) } var localesDirPath *C.char = C.CString(settings.LocalesDirPath) defer C.free(unsafe.Pointer(localesDirPath)) C.cef_string_from_utf8(localesDirPath, C.strlen(localesDirPath), C.cefStringCastToCefString16(&cefSettings.locales_dir_path)) if settings.PersistSessionCookies { cefSettings.persist_session_cookies = 1 } cefSettings.remote_debugging_port = C.int(settings.RemoteDebuggingPort) cefSettings.ignore_certificate_errors = C.int(settings.IgnoreCertificateErrors) // no_sandbox // ---------- cefSettings.no_sandbox = C.int(1) go_AddRef(unsafe.Pointer(_MainArgs)) go_AddRef(unsafe.Pointer(appHandler.GetAppHandlerT().CStruct)) go_AddRef(unsafe.Pointer(_SandboxInfo)) ret := C.cef_initialize(_MainArgs, cefSettings, appHandler.GetAppHandlerT().CStruct, _SandboxInfo) return int(ret) } func CreateBrowser(hwnd unsafe.Pointer, clientHandler ClientHandler, browserSettings BrowserSettings,
Logger.Infof("CreateBrowser, url=%s\n", url) // Initialize cef_window_info_t structure. var windowInfo *C.cef_window_info_t windowInfo = (*C.cef_window_info_t)( C.calloc(1, C.sizeof_cef_window_info_t)) FillWindowInfo(windowInfo, hwnd) // url var cefUrl *C.cef_string_t cefUrl = (*C.cef_string_t)( C.calloc(1, C.sizeof_cef_string_t)) var charUrl *C.char = C.CString(url) defer C.free(unsafe.Pointer(charUrl)) C.cef_string_from_utf8(charUrl, C.strlen(charUrl), C.cefStringCastToCefString16(cefUrl)) // Initialize cef_browser_settings_t structure. cefBrowserSettings := browserSettings.toC() // Do not create the browser synchronously using the // cef_browser_host_create_browser_sync() function, as // it is unreliable. Instead obtain browser object in // life_span_handler::on_after_created. In that callback // keep CEF browser objects in a global map (cef window // handle -> cef browser) and introduce // a GetBrowserByWindowHandle() function. This function // will first guess the CEF window handle using for example // WinAPI functions and then search the global map of cef // browser objects. go_AddRef(unsafe.Pointer(clientHandler.GetClientHandlerT().CStruct)) result := C.cef_browser_host_create_browser( windowInfo, clientHandler.GetClientHandlerT().CStruct, cefUrl, cefBrowserSettings, nil, ) return result == C.int(1) } func RunMessageLoop() { Logger.Infof("RunMessageLoop\n") C.cef_run_message_loop() } func QuitMessageLoop() { Logger.Infof("QuitMessageLoop\n") C.cef_quit_message_loop() } func Shutdown() { Logger.Infof("Shutdown\n") C.cef_shutdown()
url string) bool {
random_line_split
cef.go
(fmt string, args ... interface{}) { log.Printf("[cef] " + fmt, args...) } func (d defaultLogger) Panicf(fmt string, args ... interface{}) { log.Panicf("[cef] " + fmt, args...) } // Sandbox is disabled. Including the "cef_sandbox.lib" // library results in lots of GCC warnings/errors. It is // compatible only with VS 2010. It would be required to // build it using GCC. Add -lcef_sandbox to LDFLAGS. // capi doesn't expose sandbox functions, you need do add // these before import "C": // void* cef_sandbox_info_create(); // void cef_sandbox_info_destroy(void* sandbox_info); var _SandboxInfo unsafe.Pointer type Settings struct { CachePath string LogSeverity int LogFile string ResourcesDirPath string LocalesDirPath string RemoteDebuggingPort int PersistSessionCookies bool IgnoreCertificateErrors int } type CefState int var ( STATE_DEFAULT CefState = 0 STATE_ENABLED CefState = 1 STATE_DISABLED CefState = 2 ) type BrowserSettings struct { StandardFontFamily string FixedFontFamily string SerifFontFamily string SansSerifFontFamily string CursiveFontFamily string FantasyFontFamily string DefaultFontSize int DefaultFixedFontSize int MinimumFontSize int MinimumLogicalFontSize int DefaultEncoding string RemoteFonts CefState Javascript CefState JavascriptOpenWindows CefState JavascriptCloseWindows CefState JavascriptAccessClipboard CefState JavascriptDomPaste CefState CaretBrowsing CefState Java CefState Plugins CefState UniversalAccessFromFileUrls CefState FileAccessFromFileUrls CefState WebSecurity CefState ImageLoading CefState ImageShrinkStandaloneToFit CefState TextAreaResize CefState TabToLinks CefState LocalStorage CefState Databases CefState ApplicationCache CefState Webgl CefState BackgroundColor uint32 } func _InitializeGlobalCStructures() { _InitializeGlobalCStructuresBase() //_InitializeGlobalCStructuresApp() /* _DisplayHandler = InitializeDisplayHandler() _DownloadHandler = InitializeDownloadHandler()*/ } func SetLogger(logger SimpleLogger) { Logger = logger } func ExecuteProcess(programHandle unsafe.Pointer, appHandler AppHandler) int { Logger.Infof("ExecuteProcess, args=%v", os.Args) _InitializeGlobalCStructures() if appHandler.GetAppHandlerT().CStruct == nil { panic("GetAppHandlerT cannot have a nil CStruct. Call NewAppHandlerT() to create one!") } FillMainArgs(_MainArgs, programHandle) // Sandbox info needs to be passed to both cef_execute_process() // and cef_initialize(). // OFF: _SandboxInfo = C.cef_sandbox_info_create() Logger.Infof("MainArgs %X _AppHanlder %X _SandboxInfo %X", _MainArgs, appHandler.GetAppHandlerT().CStruct, _SandboxInfo) go_AddRef(unsafe.Pointer(_MainArgs)) go_AddRef(unsafe.Pointer(appHandler.GetAppHandlerT().CStruct)) go_AddRef(unsafe.Pointer(_SandboxInfo)) var exitCode C.int = C.cef_execute_process(_MainArgs, appHandler.GetAppHandlerT().CStruct, _SandboxInfo) if (exitCode >= 0) { os.Exit(int(exitCode)) } return int(exitCode) } func Initialize(settings Settings, appHandler AppHandler) int { Logger.Infof("Initialize\n") if _MainArgs == nil { // _MainArgs structure is initialized and filled in ExecuteProcess. // If cef_execute_process is not called, and there is a call // to cef_initialize, then it would result in creation of infinite // number of processes. See Issue 1199 in CEF: // https://code.google.com/p/chromiumembedded/issues/detail?id=1199 Logger.Errorf("ERROR: missing a call to ExecuteProcess\n") return 0 } // Initialize cef_settings_t structure. var cefSettings *C.struct__cef_settings_t cefSettings = (*C.struct__cef_settings_t)( C.calloc(1, C.sizeof_struct__cef_settings_t)) cefSettings.size = C.sizeof_struct__cef_settings_t // cache_path // ---------- if (settings.CachePath != "") { Logger.Infof("CachePath=%s\n", settings.CachePath) } var cachePath *C.char = C.CString(settings.CachePath) defer C.free(unsafe.Pointer(cachePath)) C.cef_string_from_utf8(cachePath, C.strlen(cachePath), C.cefStringCastToCefString16(&cefSettings.cache_path)) // log_severity // ------------ cefSettings.log_severity = (C.cef_log_severity_t)(C.int(settings.LogSeverity)) // log_file // -------- if (settings.LogFile != "") { Logger.Infof("LogFile=%s\n", settings.LogFile) } var logFile *C.char = C.CString(settings.LogFile) defer C.free(unsafe.Pointer(logFile)) C.cef_string_from_utf8(logFile, C.strlen(logFile), C.cefStringCastToCefString16(&cefSettings.log_file)) // resources_dir_path // ------------------ if settings.ResourcesDirPath == "" && runtime.GOOS != "darwin" { // Setting this path is required for the tests to run fine. cwd, _ := os.Getwd() settings.ResourcesDirPath = cwd } if (settings.ResourcesDirPath != "") { Logger.Infof("ResourcesDirPath=%s\n", settings.ResourcesDirPath) } var resourcesDirPath *C.char = C.CString(settings.ResourcesDirPath) defer C.free(unsafe.Pointer(resourcesDirPath)) C.cef_string_from_utf8(resourcesDirPath, C.strlen(resourcesDirPath), C.cefStringCastToCefString16(&cefSettings.resources_dir_path)) // locales_dir_path // ---------------- if settings.LocalesDirPath == "" && runtime.GOOS != "darwin" { // Setting this path is required for the tests to run fine. cwd, _ := os.Getwd() settings.LocalesDirPath = cwd + "/locales" } if (settings.LocalesDirPath != "") { Logger.Infof("LocalesDirPath=%s\n", settings.LocalesDirPath) } var localesDirPath *C.char = C.CString(settings.LocalesDirPath) defer C.free(unsafe.Pointer(localesDirPath)) C.cef_string_from_utf8(localesDirPath, C.strlen(localesDirPath), C.cefStringCastToCefString16(&cefSettings.locales_dir_path)) if settings.PersistSessionCookies { cefSettings.persist_session_cookies = 1 } cefSettings.remote_debugging_port = C.int(settings.RemoteDebuggingPort) cefSettings.ignore_certificate_errors = C.int(settings.IgnoreCertificateErrors) // no_sandbox // ---------- cefSettings.no_sandbox = C.int(1) go_AddRef(unsafe.Pointer(_MainArgs)) go_AddRef(unsafe.Pointer(appHandler.GetAppHandlerT().CStruct)) go_AddRef(unsafe.Pointer(_SandboxInfo)) ret := C.cef_initialize(_MainArgs, cefSettings, appHandler.GetAppHandlerT().CStruct, _SandboxInfo) return int(ret) } func CreateBrowser(hwnd unsafe.Pointer, clientHandler ClientHandler, browserSettings BrowserSettings, url string) bool { Logger.Infof("CreateBrowser, url=%s\n", url) // Initialize cef_window_info_t structure. var windowInfo *C.cef_window_info_t windowInfo = (*C.cef_window_info_t)( C.calloc(1, C.sizeof_cef_window_info_t)) FillWindowInfo(windowInfo, hwnd) // url var cefUrl *C.cef_string_t cefUrl = (*C.cef_string_t)( C.calloc(1, C.sizeof_cef_string_t)) var charUrl *C.char = C.CString(url) defer C.free(unsafe.Pointer(charUrl)) C.cef_string_from_utf8(charUrl, C.strlen(charUrl), C.cefStringCastToCefString16(cefUrl)) // Initialize cef_browser_settings_t structure. cefBrowserSettings := browserSettings.toC() // Do not create the browser synchronously using the // cef_browser_host_create_browser_sync() function, as // it is unreliable. Instead obtain browser object in // life_span_handler::on_after_created. In that callback // keep CEF browser objects in a global map (cef window // handle -> cef browser) and introduce // a Get
Errorf
identifier_name
cef.go
efState JavascriptCloseWindows CefState JavascriptAccessClipboard CefState JavascriptDomPaste CefState CaretBrowsing CefState Java CefState Plugins CefState UniversalAccessFromFileUrls CefState FileAccessFromFileUrls CefState WebSecurity CefState ImageLoading CefState ImageShrinkStandaloneToFit CefState TextAreaResize CefState TabToLinks CefState LocalStorage CefState Databases CefState ApplicationCache CefState Webgl CefState BackgroundColor uint32 } func _InitializeGlobalCStructures() { _InitializeGlobalCStructuresBase() //_InitializeGlobalCStructuresApp() /* _DisplayHandler = InitializeDisplayHandler() _DownloadHandler = InitializeDownloadHandler()*/ } func SetLogger(logger SimpleLogger) { Logger = logger } func ExecuteProcess(programHandle unsafe.Pointer, appHandler AppHandler) int { Logger.Infof("ExecuteProcess, args=%v", os.Args) _InitializeGlobalCStructures() if appHandler.GetAppHandlerT().CStruct == nil { panic("GetAppHandlerT cannot have a nil CStruct. Call NewAppHandlerT() to create one!") } FillMainArgs(_MainArgs, programHandle) // Sandbox info needs to be passed to both cef_execute_process() // and cef_initialize(). // OFF: _SandboxInfo = C.cef_sandbox_info_create() Logger.Infof("MainArgs %X _AppHanlder %X _SandboxInfo %X", _MainArgs, appHandler.GetAppHandlerT().CStruct, _SandboxInfo) go_AddRef(unsafe.Pointer(_MainArgs)) go_AddRef(unsafe.Pointer(appHandler.GetAppHandlerT().CStruct)) go_AddRef(unsafe.Pointer(_SandboxInfo)) var exitCode C.int = C.cef_execute_process(_MainArgs, appHandler.GetAppHandlerT().CStruct, _SandboxInfo) if (exitCode >= 0) { os.Exit(int(exitCode)) } return int(exitCode) } func Initialize(settings Settings, appHandler AppHandler) int { Logger.Infof("Initialize\n") if _MainArgs == nil { // _MainArgs structure is initialized and filled in ExecuteProcess. // If cef_execute_process is not called, and there is a call // to cef_initialize, then it would result in creation of infinite // number of processes. See Issue 1199 in CEF: // https://code.google.com/p/chromiumembedded/issues/detail?id=1199 Logger.Errorf("ERROR: missing a call to ExecuteProcess\n") return 0 } // Initialize cef_settings_t structure. var cefSettings *C.struct__cef_settings_t cefSettings = (*C.struct__cef_settings_t)( C.calloc(1, C.sizeof_struct__cef_settings_t)) cefSettings.size = C.sizeof_struct__cef_settings_t // cache_path // ---------- if (settings.CachePath != "") { Logger.Infof("CachePath=%s\n", settings.CachePath) } var cachePath *C.char = C.CString(settings.CachePath) defer C.free(unsafe.Pointer(cachePath)) C.cef_string_from_utf8(cachePath, C.strlen(cachePath), C.cefStringCastToCefString16(&cefSettings.cache_path)) // log_severity // ------------ cefSettings.log_severity = (C.cef_log_severity_t)(C.int(settings.LogSeverity)) // log_file // -------- if (settings.LogFile != "") { Logger.Infof("LogFile=%s\n", settings.LogFile) } var logFile *C.char = C.CString(settings.LogFile) defer C.free(unsafe.Pointer(logFile)) C.cef_string_from_utf8(logFile, C.strlen(logFile), C.cefStringCastToCefString16(&cefSettings.log_file)) // resources_dir_path // ------------------ if settings.ResourcesDirPath == "" && runtime.GOOS != "darwin" { // Setting this path is required for the tests to run fine. cwd, _ := os.Getwd() settings.ResourcesDirPath = cwd } if (settings.ResourcesDirPath != "") { Logger.Infof("ResourcesDirPath=%s\n", settings.ResourcesDirPath) } var resourcesDirPath *C.char = C.CString(settings.ResourcesDirPath) defer C.free(unsafe.Pointer(resourcesDirPath)) C.cef_string_from_utf8(resourcesDirPath, C.strlen(resourcesDirPath), C.cefStringCastToCefString16(&cefSettings.resources_dir_path)) // locales_dir_path // ---------------- if settings.LocalesDirPath == "" && runtime.GOOS != "darwin" { // Setting this path is required for the tests to run fine. cwd, _ := os.Getwd() settings.LocalesDirPath = cwd + "/locales" } if (settings.LocalesDirPath != "") { Logger.Infof("LocalesDirPath=%s\n", settings.LocalesDirPath) } var localesDirPath *C.char = C.CString(settings.LocalesDirPath) defer C.free(unsafe.Pointer(localesDirPath)) C.cef_string_from_utf8(localesDirPath, C.strlen(localesDirPath), C.cefStringCastToCefString16(&cefSettings.locales_dir_path)) if settings.PersistSessionCookies { cefSettings.persist_session_cookies = 1 } cefSettings.remote_debugging_port = C.int(settings.RemoteDebuggingPort) cefSettings.ignore_certificate_errors = C.int(settings.IgnoreCertificateErrors) // no_sandbox // ---------- cefSettings.no_sandbox = C.int(1) go_AddRef(unsafe.Pointer(_MainArgs)) go_AddRef(unsafe.Pointer(appHandler.GetAppHandlerT().CStruct)) go_AddRef(unsafe.Pointer(_SandboxInfo)) ret := C.cef_initialize(_MainArgs, cefSettings, appHandler.GetAppHandlerT().CStruct, _SandboxInfo) return int(ret) } func CreateBrowser(hwnd unsafe.Pointer, clientHandler ClientHandler, browserSettings BrowserSettings, url string) bool { Logger.Infof("CreateBrowser, url=%s\n", url) // Initialize cef_window_info_t structure. var windowInfo *C.cef_window_info_t windowInfo = (*C.cef_window_info_t)( C.calloc(1, C.sizeof_cef_window_info_t)) FillWindowInfo(windowInfo, hwnd) // url var cefUrl *C.cef_string_t cefUrl = (*C.cef_string_t)( C.calloc(1, C.sizeof_cef_string_t)) var charUrl *C.char = C.CString(url) defer C.free(unsafe.Pointer(charUrl)) C.cef_string_from_utf8(charUrl, C.strlen(charUrl), C.cefStringCastToCefString16(cefUrl)) // Initialize cef_browser_settings_t structure. cefBrowserSettings := browserSettings.toC() // Do not create the browser synchronously using the // cef_browser_host_create_browser_sync() function, as // it is unreliable. Instead obtain browser object in // life_span_handler::on_after_created. In that callback // keep CEF browser objects in a global map (cef window // handle -> cef browser) and introduce // a GetBrowserByWindowHandle() function. This function // will first guess the CEF window handle using for example // WinAPI functions and then search the global map of cef // browser objects. go_AddRef(unsafe.Pointer(clientHandler.GetClientHandlerT().CStruct)) result := C.cef_browser_host_create_browser( windowInfo, clientHandler.GetClientHandlerT().CStruct, cefUrl, cefBrowserSettings, nil, ) return result == C.int(1) } func RunMessageLoop() { Logger.Infof("RunMessageLoop\n") C.cef_run_message_loop() } func QuitMessageLoop() { Logger.Infof("QuitMessageLoop\n") C.cef_quit_message_loop() } func Shutdown() { Logger.Infof("Shutdown\n") C.cef_shutdown() // OFF: cef_sandbox_info_destroy(_SandboxInfo) } func extractCefMultiMap(cefMapPointer C.cef_string_multimap_t) map[string][]string { numKeys := C.cef_string_multimap_size(cefMapPointer) goMap := make(map[string][]string) for i := 0; i < int(numKeys); i++ { var key *C.cef_string_utf16_t = C.cef_string_userfree_utf16_alloc() C.cef_string_multimap_key(cefMapPointer, C.int(i), C.cefString16CastToCefString(key)) charKeyUtf8 := C.cefStringToUtf8(C.cefString16CastToCefString(key)) goKey := C.GoString(charKeyUtf8.str) if _, ok := goMap[goKey]; ok
{ continue }
conditional_block
romaO.py
9, 0.76047], [0.33253, 0.59151, 0.75704], [0.32893, 0.58412, 0.75351], [0.32559, 0.57671, 0.74987], [0.32256, 0.56928, 0.74613], [0.31978, 0.56186, 0.74228], [0.31727, 0.55441, 0.7383], [0.31505, 0.54695, 0.73422], [0.31311, 0.53948, 0.73002], [0.31144, 0.53201, 0.72569], [0.31007, 0.52453, 0.72124], [0.30897, 0.51704, 0.71667], [0.30811, 0.50955, 0.71197], [0.30755, 0.50205, 0.70713], [0.30726, 0.49456, 0.70216], [0.30723, 0.48707, 0.69706], [0.30746, 0.47958, 0.69182], [0.30795, 0.4721, 0.68643], [0.3087, 0.46463, 0.6809], [0.30968, 0.45716, 0.67525], [0.31088, 0.44973, 0.66944], [0.31228, 0.44232, 0.6635], [0.31393, 0.43493, 0.65741], [0.31578, 0.42758, 0.65118], [0.3178, 0.42025, 0.64482], [0.32001, 0.41299, 0.63833], [0.32238, 0.40577, 0.6317], [0.32489, 0.39861, 0.62495], [0.32755, 0.39152, 0.61809], [0.33035, 0.38448, 0.61111], [0.33327, 0.37755, 0.60402], [0.33627, 0.37068, 0.59684], [0.33939, 0.36392, 0.58955], [0.34257, 0.35728, 0.58219], [0.3458, 0.35073, 0.57476], [0.34912, 0.34428, 0.56727], [0.35247, 0.33797, 0.55971], [0.35587, 0.33179, 0.55212], [0.35927, 0.32574, 0.54448], [0.36271, 0.31986, 0.53684], [0.36617, 0.31411, 0.52917], [0.36961, 0.30852, 0.52148], [0.37306, 0.30306, 0.51382], [0.37652, 0.2978, 0.50615], [0.37994, 0.29269, 0.49854], [0.38336, 0.28775, 0.49094], [0.38674, 0.28301, 0.48337], [0.39011, 0.27842, 0.47586], [0.39346, 0.27401, 0.4684], [0.39677, 0.26978, 0.461], [0.40006, 0.26573, 0.45366], [0.40333, 0.26185, 0.4464], [0.40655, 0.25815, 0.43921], [0.40974, 0.25466, 0.43212], [0.4129, 0.25132, 0.42509], [0.41602, 0.24817, 0.41813], [0.41912, 0.24515, 0.41128], [0.42218, 0.24235, 0.40451], [0.42522, 0.23972, 0.39784], [0.42823, 0.23728, 0.39126], [0.43121, 0.23498, 0.38475], [0.43415, 0.23282, 0.37836], [0.43708, 0.23086, 0.37204], [0.43998, 0.22907, 0.36583], [0.44286, 0.22743, 0.3597], [0.44571, 0.22596, 0.35366], [0.44855, 0.2246, 0.34773]] romaO_map = LinearSegmentedColormap.from_list('romaO', cm_data) # For use of "viscm view" test_cm = romaO_map if __name__ == "__main__":
import matplotlib.pyplot as plt import numpy as np try: from viscm import viscm viscm(romaO_map) except ImportError: print("viscm not found, falling back on simple display") plt.imshow(np.linspace(0, 100, 256)[None, :], aspect='auto', cmap=romaO_map) plt.show()
conditional_block
romaO.py
2165], [0.78242, 0.6759, 0.33066], [0.78669, 0.68481, 0.33988], [0.79087, 0.69365, 0.34929], [0.79494, 0.7024, 0.35888], [0.7989, 0.71106, 0.36867], [0.80273, 0.71961, 0.37859], [0.80642, 0.72803, 0.38866], [0.80996, 0.73631, 0.39885], [0.81334, 0.74446, 0.40916], [0.81655, 0.75244, 0.41957], [0.81956, 0.76025, 0.43004], [0.82239, 0.76787, 0.44057], [0.82501, 0.7753, 0.45115], [0.82742, 0.78252, 0.46174], [0.8296, 0.78953, 0.47235], [0.83155, 0.79631, 0.48293], [0.83326, 0.80287, 0.49349], [0.83472, 0.80919, 0.50402], [0.83592, 0.81526, 0.51449], [0.83686, 0.82109, 0.52487], [0.83753, 0.82666, 0.53517], [0.83793, 0.83198, 0.54537], [0.83805, 0.83703, 0.55546], [0.83788, 0.84182, 0.56542], [0.83744, 0.84635, 0.57525], [0.8367, 0.85061, 0.58493], [0.83567, 0.85462, 0.59446], [0.83435, 0.85835, 0.60382], [0.83274, 0.86183, 0.61301], [0.83084, 0.86504, 0.62202], [0.82864, 0.868, 0.63085], [0.82615, 0.87068, 0.63949], [0.82337, 0.87312, 0.64792], [0.8203, 0.87531, 0.65617], [0.81695, 0.87724, 0.6642], [0.81331, 0.87892, 0.67203], [0.80939, 0.88036, 0.67964], [0.80518, 0.88156, 0.68705], [0.80071, 0.8825, 0.69424], [0.79595, 0.88322, 0.70121], [0.79094, 0.8837, 0.70797], [0.78566, 0.88395, 0.7145], [0.78012, 0.88396, 0.72082], [0.77433, 0.88375, 0.72692], [0.7683, 0.88331, 0.73279], [0.76203, 0.88264, 0.73844], [0.75553, 0.88177, 0.74387], [0.74879, 0.88066, 0.74908], [0.74184, 0.87934, 0.75407], [0.73468, 0.87781, 0.75884], [0.72731, 0.87607, 0.76339], [0.71976, 0.87411, 0.76772], [0.71201, 0.87195, 0.77184], [0.70408, 0.86958, 0.77573], [0.69599, 0.86701, 0.77941], [0.68774, 0.86425, 0.78288], [0.67934, 0.86127, 0.78614], [0.67081, 0.85811, 0.78919], [0.66215, 0.85476, 0.79202], [0.65336, 0.8512, 0.79465], [0.64448, 0.84747, 0.79707], [0.6355, 0.84356, 0.7993], [0.62645, 0.83947, 0.80131], [0.61732, 0.83519, 0.80313], [0.60814, 0.83075, 0.80476], [0.59891, 0.82614, 0.80619],
[0.58965, 0.82137, 0.80743], [0.58037, 0.81644, 0.80848], [0.57108, 0.81135, 0.80935], [0.56181, 0.80612, 0.81004], [0.55255, 0.80074, 0.81055],
random_line_split
tags.rs
; } None => unreachable!(), } } let result = if tags.is_empty() { ctx.psql().get_tags_mapset(mapset_id).await } else { let db_result = match action { Action::Add => ctx.psql().add_tags_mapset(mapset_id, tags).await, Action::Remove => ctx.psql().remove_tags_mapset(mapset_id, tags).await, }; match db_result { Ok(_) => ctx.psql().get_tags_mapset(mapset_id).await, Err(err) => Err(err), } }; // Then show the final tags match result { Ok(tags) => { let content = format!("{OSU_BASE}beatmapsets/{mapset_id} is now tagged as:\n{tags}"); let builder = MessageBuilder::new().content(content); msg.create_message(&ctx, builder).await?; } Err(err) => { let _ = msg.error(&ctx, GENERAL_ISSUE).await; return Err(err); } } Ok(()) } // #[command] // #[short_desc("Help out tagging backgrounds")] // #[long_desc( // "Let me give you mapsets that still need to be tagged.\n\ // React to them properly, then lock it in by reacting with ✅.\n\ // To leave the loop, react with ❌ or just wait 10 minutes.\n\ // Mode can be specified in the first argument, defaults to std.\n\ // **You need to be verified to use this command**, feel free to \ // let Badewanne3 know if you want to help out tagging backgrounds." // )] // #[usage("[std / mna]")] // #[aliases("bgt", "bgtag")] // #[owner()] async fn bgtags(ctx: Arc<Context>, data: CommandData) -> Result<()> { let (msg, mut args) = match data { CommandData::Message { msg, args, .. } => (msg, args), CommandData::Interaction { .. } => unreachable!(), }; // Parse arguments as mode let mode = match args.next() { Some(arg) => match arg.cow_to_ascii_lowercase().as_ref() { "mna" | "mania" | "m" => GameMode::Mania, "osu" | "std" | "standard" | "o" => GameMode::Osu, _ => { let content = "Could not parse first argument as mode. \ Provide either `mna`, or `std`"; return msg.error(&ctx, content).await; } }, None => GameMode::Osu, }; let mut untagged = match ctx.psql().get_all_tags_mapset(mode).await { Ok(tags) => tags.iter().any(|tag| tag.untagged()), Err(err) => { let _ = msg.error(&ctx, GENERAL_ISSUE).await; return Err(err); } }; if !untagged { let content = "All backgrounds have been tagged, \ here are some random ones you can review again though"; let builder = MessageBuilder::new().content(content); let _ = msg.create_message(&ctx, builder).await; } let mut owner = msg.author.id; loop { // Get all mapsets for which tags are missing let tags_result = if untagged { ctx.psql().get_all_tags_mapset(mode).await } else { ctx.psql() .get_random_tags_mapset(mode) .await .map(|tags| vec![tags]) }; let mapsets = match tags_result { Ok(mut tags) => { if untagged { if tags.iter().any(|tag| tag.untagged()) { tags.retain(|tag| tag.untagged()); } else { let content = "All backgrounds have been tagged, \ here are some random ones you can review again though"; let builder = MessageBuilder::new().content(content); let _ = msg.create_message(&ctx, builder).await; untagged = false; tags.truncate(1); } } tags } Err(err) => { let _ = msg.error(&ctx, GENERAL_ISSUE).await; return Err(err); } }; let (mapset_id, img) = get_random_image(mapsets, mode).await; let content = format!( "<@{owner}> Which tags should this mapsets get: {OSU_BASE}beatmapsets/{mapset_id}\n\ ```\n\ 🍋: Easy 🎨: Weeb 😱: Hard name 🗽: English 💯: Tech\n\ 🤓: Hard 🍨: Kpop 🪀: Alternate 🌀: Streams ✅: Lock in\n\ 🤡: Meme 👨‍🌾: Farm 🟦: Blue sky 👴: Old ❌: Exit loop\n\ ```" ); let img = Attachment::from_bytes("bg_img.png".to_owned(), img); // Send response let response = ctx .http .create_message(msg.channel_id) .content(&content)? .attachments(&[img]) .unwrap() .exec() .await? .model() .await?; let msg_id = response.id; // Setup collector let reaction_stream = ctx .standby .wait_for_event_stream(move |event: &Event| match event { Event::ReactionAdd(event) => { event.message_id == msg_id && event.user_id.get() == OWNER_USER_ID } Event::ReactionRemove(event) => { event.message_id == msg_id && event.user_id.get() == OWNER_USER_ID } _ => false, }) .map(|event| match event { Event::ReactionAdd(add) => ReactionWrapper::Add(add.0), Event::ReactionRemove(remove) => ReactionWrapper::Remove(remove.0), _ => unreachable!(), }) .timeout(Duration::from_secs(600)); tokio::pin!(reaction_stream); // Add reactions let reactions = [ "🍋", "🤓", "🤡", "🎨", "🍨", "👨‍🌾", "😱", "🪀", "🟦", "🗽", "🌀", "👴", "💯", "✅", "❌", ]; for &reaction in reactions.iter() { let emote = Emote::Custom(reaction); send_reaction(&*ctx, &response, emote).await?; } let mut break_loop = true; // Run collector let mut add_tags = MapsetTags::empty(); let mut remove_tags = MapsetTags::empty(); while let Some(Ok(reaction)) = reaction_stream.next().await { let tag = if let ReactionType::Unicode { ref name } = reaction.as_deref().emoji { match name.as_str() { "🍋" => MapsetTags::Easy, "🤓" => MapsetTags::Hard, "🤡" => MapsetTags::Meme, "👴" => MapsetTags::Old, "😱" => MapsetTags::HardName, "🟦" => MapsetTags::BlueSky, "🪀" => MapsetTags::Alternate, "🗽" => MapsetTags::English, "👨‍🌾" => MapsetTags::Farm, "💯" => MapsetTags::Tech, "🎨" => MapsetTags::Weeb, "🌀" => MapsetTags::Streams, "🍨" => MapsetTags::Kpop, "✅" => { owner = reaction.as_deref().user_id; break_loop = false; break; } "❌" => break, _ => continue, } } else { continue; }; match reaction { ReactionWrapper::Add(_) => { add_tags.insert(tag); } ReactionWrapper::Remove(_) => { remove_tags.insert(tag); } } } if !add_tags.is_empty() { if let Err(err) = ctx.psql().add_tags_mapset(mapset_id, add_tags).await { warn!(?err, "Failed to add tags"); } } if !remove_tags.is_empty() { if let Err(err) = ctx.psql().remove_tags_mapset(mapset_id, remove_tags).await { warn!(?err, "Failed to remove tags"); } } // Then show the final tags match ctx.psql().get_tags_mapset(mapset_id).await { Ok(tags) => { if !tags.is_empty() { let content = format!( "{}beatmapsets/{} is now tagged as:\n{}", OSU_BASE, mapset_id, tags, ); let builder = MessageBuilder::new().content(content); msg.create_message(&ctx, builder).await?; } } Err(err) => { let _ = msg.error(&ctx, GENERAL_ISSUE).await; return Err(err); } }; if break_loop {
let builder = MessageBuilder::new().content("Exiting loop :wave:"); msg.create_message(
conditional_block
tags.rs
data: CommandData) -> Result<()> { let (msg, mut args) = match data { CommandData::Message { msg, args, .. } => (msg, args), CommandData::Interaction { .. } => unreachable!(), }; // Parse mapset id let mapset_id = match args.next().map(u32::from_str) { Some(Ok(num)) => num, Some(Err(_)) => { let content = "Could not parse mapset id. Be sure to specify it as first argument"; return msg.error(&ctx, content).await; } None => { let content = "Arguments: `[mapset id] [add/a/remove/r] [list of tags]`\n\ Example: `21662 r hard farm streams alternate hardname tech weeb bluesky`\n\ Tags: `farm, streams, alternate, old, meme, hardname, easy, hard, tech, \ weeb, bluesky, english`"; let builder = MessageBuilder::new().content(content); msg.create_message(&ctx, builder).await?; return Ok(()); } }; // Check if there is background for the given mapset id if ctx.psql().get_tags_mapset(mapset_id).await.is_err() { let content = "No background entry found with this id"; return msg.error(&ctx, content).await; } // Parse action let action = match args.next().map(Action::from_str) { Some(Ok(action)) => action, None | Some(Err(_)) => { let content = "Could not parse action. \ Be sure to specify `r`, `remove`, `a`, or `add` as second argument"; return msg.error(&ctx, content).await; } }; // Parse tags let mut tags = MapsetTags::empty(); while !args.is_empty() { match args.next().map(MapsetTags::from_str) { Some(Ok(tag)) => tags.insert(tag), Some(Err(tag)) => { let content = format!( "Could not parse tag `{tag}`.\n\ Be sure to only give these tags:\n\ `farm, streams, alternate, old, meme, hardname, \ easy, hard, tech, weeb, bluesky, english`" ); return msg.error(&ctx, content).await; } None => unreachable!(), } } let result = if tags.is_empty() { ctx.psql().get_tags_mapset(mapset_id).await } else { let db_result = match action { Action::Add => ctx.psql().add_tags_mapset(mapset_id, tags).await, Action::Remove => ctx.psql().remove_tags_mapset(mapset_id, tags).await, }; match db_result { Ok(_) => ctx.psql().get_tags_mapset(mapset_id).await, Err(err) => Err(err), } }; // Then show the final tags match result { Ok(tags) => { let content = format!("{OSU_BASE}beatmapsets/{mapset_id} is now tagged as:\n{tags}"); let builder = MessageBuilder::new().content(content); msg.create_message(&ctx, builder).await?; } Err(err) => { let _ = msg.error(&ctx, GENERAL_ISSUE).await; return Err(err); } } Ok(()) } // #[command] // #[short_desc("Help out tagging backgrounds")] // #[long_desc( // "Let me give you mapsets that still need to be tagged.\n\ // React to them properly, then lock it in by reacting with ✅.\n\ // To leave the loop, react with ❌ or just wait 10 minutes.\n\ // Mode can be specified in the first argument, defaults to std.\n\ // **You need to be verified to use this command**, feel free to \ // let Badewanne3 know if you want to help out tagging backgrounds." // )] // #[usage("[std / mna]")] // #[aliases("bgt", "bgtag")] // #[owner()] async fn bgtags(ctx: Arc<Context>, data: CommandData) -> Result<()> { let (msg, mut args) = match data { CommandData::Message { msg, args, .. } => (msg, args),
let mode = match args.next() { Some(arg) => match arg.cow_to_ascii_lowercase().as_ref() { "mna" | "mania" | "m" => GameMode::Mania, "osu" | "std" | "standard" | "o" => GameMode::Osu, _ => { let content = "Could not parse first argument as mode. \ Provide either `mna`, or `std`"; return msg.error(&ctx, content).await; } }, None => GameMode::Osu, }; let mut untagged = match ctx.psql().get_all_tags_mapset(mode).await { Ok(tags) => tags.iter().any(|tag| tag.untagged()), Err(err) => { let _ = msg.error(&ctx, GENERAL_ISSUE).await; return Err(err); } }; if !untagged { let content = "All backgrounds have been tagged, \ here are some random ones you can review again though"; let builder = MessageBuilder::new().content(content); let _ = msg.create_message(&ctx, builder).await; } let mut owner = msg.author.id; loop { // Get all mapsets for which tags are missing let tags_result = if untagged { ctx.psql().get_all_tags_mapset(mode).await } else { ctx.psql() .get_random_tags_mapset(mode) .await .map(|tags| vec![tags]) }; let mapsets = match tags_result { Ok(mut tags) => { if untagged { if tags.iter().any(|tag| tag.untagged()) { tags.retain(|tag| tag.untagged()); } else { let content = "All backgrounds have been tagged, \ here are some random ones you can review again though"; let builder = MessageBuilder::new().content(content); let _ = msg.create_message(&ctx, builder).await; untagged = false; tags.truncate(1); } } tags } Err(err) => { let _ = msg.error(&ctx, GENERAL_ISSUE).await; return Err(err); } }; let (mapset_id, img) = get_random_image(mapsets, mode).await; let content = format!( "<@{owner}> Which tags should this mapsets get: {OSU_BASE}beatmapsets/{mapset_id}\n\ ```\n\ 🍋: Easy 🎨: Weeb 😱: Hard name 🗽: English 💯: Tech\n\ 🤓: Hard 🍨: Kpop 🪀: Alternate 🌀: Streams ✅: Lock in\n\ 🤡: Meme 👨‍🌾: Farm 🟦: Blue sky 👴: Old ❌: Exit loop\n\ ```" ); let img = Attachment::from_bytes("bg_img.png".to_owned(), img); // Send response let response = ctx .http .create_message(msg.channel_id) .content(&content)? .attachments(&[img]) .unwrap() .exec() .await? .model() .await?; let msg_id = response.id; // Setup collector let reaction_stream = ctx .standby .wait_for_event_stream(move |event: &Event| match event { Event::ReactionAdd(event) => { event.message_id == msg_id && event.user_id.get() == OWNER_USER_ID } Event::ReactionRemove(event) => { event.message_id == msg_id && event.user_id.get() == OWNER_USER_ID } _ => false, }) .map(|event| match event { Event::ReactionAdd(add) => ReactionWrapper::Add(add.0), Event::ReactionRemove(remove) => ReactionWrapper::Remove(remove.0), _ => unreachable!(), }) .timeout(Duration::from_secs(600)); tokio::pin!(reaction_stream); // Add reactions let reactions = [ "🍋", "🤓", "🤡", "🎨", "🍨", "👨‍🌾", "😱", "🪀", "🟦", "🗽", "🌀", "👴", "💯", "✅", "❌", ]; for &reaction in reactions.iter() { let emote = Emote::Custom(reaction); send_reaction(&*ctx, &response, emote).await?; } let mut break_loop = true; // Run collector let mut add_tags = MapsetTags::empty(); let mut remove_tags = MapsetTags::empty(); while let Some(Ok(reaction)) = reaction_stream.next().await { let
CommandData::Interaction { .. } => unreachable!(), }; // Parse arguments as mode
random_line_split
tags.rs
, tags).await, Action::Remove => ctx.psql().remove_tags_mapset(mapset_id, tags).await, }; match db_result { Ok(_) => ctx.psql().get_tags_mapset(mapset_id).await, Err(err) => Err(err), } }; // Then show the final tags match result { Ok(tags) => { let content = format!("{OSU_BASE}beatmapsets/{mapset_id} is now tagged as:\n{tags}"); let builder = MessageBuilder::new().content(content); msg.create_message(&ctx, builder).await?; } Err(err) => { let _ = msg.error(&ctx, GENERAL_ISSUE).await; return Err(err); } } Ok(()) } // #[command] // #[short_desc("Help out tagging backgrounds")] // #[long_desc( // "Let me give you mapsets that still need to be tagged.\n\ // React to them properly, then lock it in by reacting with ✅.\n\ // To leave the loop, react with ❌ or just wait 10 minutes.\n\ // Mode can be specified in the first argument, defaults to std.\n\ // **You need to be verified to use this command**, feel free to \ // let Badewanne3 know if you want to help out tagging backgrounds." // )] // #[usage("[std / mna]")] // #[aliases("bgt", "bgtag")] // #[owner()] async fn bgtags(ctx: Arc<Context>, data: CommandData) -> Result<()> { let (msg, mut args) = match data { CommandData::Message { msg, args, .. } => (msg, args), CommandData::Interaction { .. } => unreachable!(), }; // Parse arguments as mode let mode = match args.next() { Some(arg) => match arg.cow_to_ascii_lowercase().as_ref() { "mna" | "mania" | "m" => GameMode::Mania, "osu" | "std" | "standard" | "o" => GameMode::Osu, _ => { let content = "Could not parse first argument as mode. \ Provide either `mna`, or `std`"; return msg.error(&ctx, content).await; } }, None => GameMode::Osu, }; let mut untagged = match ctx.psql().get_all_tags_mapset(mode).await { Ok(tags) => tags.iter().any(|tag| tag.untagged()), Err(err) => { let _ = msg.error(&ctx, GENERAL_ISSUE).await; return Err(err); } }; if !untagged { let content = "All backgrounds have been tagged, \ here are some random ones you can review again though"; let builder = MessageBuilder::new().content(content); let _ = msg.create_message(&ctx, builder).await; } let mut owner = msg.author.id; loop { // Get all mapsets for which tags are missing let tags_result = if untagged { ctx.psql().get_all_tags_mapset(mode).await } else { ctx.psql() .get_random_tags_mapset(mode) .await .map(|tags| vec![tags]) }; let mapsets = match tags_result { Ok(mut tags) => { if untagged { if tags.iter().any(|tag| tag.untagged()) { tags.retain(|tag| tag.untagged()); } else { let content = "All backgrounds have been tagged, \ here are some random ones you can review again though"; let builder = MessageBuilder::new().content(content); let _ = msg.create_message(&ctx, builder).await; untagged = false; tags.truncate(1); } } tags } Err(err) => { let _ = msg.error(&ctx, GENERAL_ISSUE).await; return Err(err); } }; let (mapset_id, img) = get_random_image(mapsets, mode).await; let content = format!( "<@{owner}> Which tags should this mapsets get: {OSU_BASE}beatmapsets/{mapset_id}\n\ ```\n\ 🍋: Easy 🎨: Weeb 😱: Hard name 🗽: English 💯: Tech\n\ 🤓: Hard 🍨: Kpop 🪀: Alternate 🌀: Streams ✅: Lock in\n\ 🤡: Meme 👨‍🌾: Farm 🟦: Blue sky 👴: Old ❌: Exit loop\n\ ```" ); let img = Attachment::from_bytes("bg_img.png".to_owned(), img); // Send response let response = ctx .http .create_message(msg.channel_id) .content(&content)? .attachments(&[img]) .unwrap() .exec() .await? .model() .await?; let msg_id = response.id; // Setup collector let reaction_stream = ctx .standby .wait_for_event_stream(move |event: &Event| match event { Event::ReactionAdd(event) => { event.message_id == msg_id && event.user_id.get() == OWNER_USER_ID } Event::ReactionRemove(event) => { event.message_id == msg_id && event.user_id.get() == OWNER_USER_ID } _ => false, }) .map(|event| match event { Event::ReactionAdd(add) => ReactionWrapper::Add(add.0), Event::ReactionRemove(remove) => ReactionWrapper::Remove(remove.0), _ => unreachable!(), }) .timeout(Duration::from_secs(600)); tokio::pin!(reaction_stream); // Add reactions let reactions = [ "🍋", "🤓", "🤡", "🎨", "🍨", "👨‍🌾", "😱", "🪀", "🟦", "🗽", "🌀", "👴", "💯", "✅", "❌", ]; for &reaction in reactions.iter() { let emote = Emote::Custom(reaction); send_reaction(&*ctx, &response, emote).await?; } let mut break_loop = true; // Run collector let mut add_tags = MapsetTags::empty(); let mut remove_tags = MapsetTags::empty(); while let Some(Ok(reaction)) = reaction_stream.next().await { let tag = if let ReactionType::Unicode { ref name } = reaction.as_deref().emoji { match name.as_str() { "🍋" => MapsetTags::Easy, "🤓" => MapsetTags::Hard, "🤡" => MapsetTags::Meme, "👴" => MapsetTags::Old, "😱" => MapsetTags::HardName, "🟦" => MapsetTags::BlueSky, "🪀" => MapsetTags::Alternate, "🗽" => MapsetTags::English, "👨‍🌾" => MapsetTags::Farm, "💯" => MapsetTags::Tech, "🎨" => MapsetTags::Weeb, "🌀" => MapsetTags::Streams, "🍨" => MapsetTags::Kpop, "✅" => { owner = reaction.as_deref().user_id; break_loop = false; break; } "❌" => break, _ => continue, } } else { continue; }; match reaction { ReactionWrapper::Add(_) => { add_tags.insert(tag); } ReactionWrapper::Remove(_) => { remove_tags.insert(tag); } } } if !add_tags.is_empty() { if let Err(err) = ctx.psql().add_tags_mapset(mapset_id, add_tags).await { warn!(?err, "Failed to add tags"); } } if !remove_tags.is_empty() { if let Err(err) = ctx.psql().remove_tags_mapset(mapset_id, remove_tags).await { warn!(?err, "Failed to remove tags"); } } // Then show the final tags match ctx.psql().get_tags_mapset(mapset_id).await { Ok(tags) => { if !tags.is_empty() { let content = format!( "{}beatmapsets/{} is now tagged as:\n{}", OSU_BASE, mapset_id, tags, ); let builder = MessageBuilder::new().content(content); msg.create_message(&ctx, builder).await?; } } Err(err) => { let _ = msg.error(&ctx, GENERAL_ISSUE).await; return Err(err); } }; if break_loop { let builder = MessageBuilder::new().content("Exiting loop :wave:"); msg.create_message(&ctx, builder).await?; break; } } Ok(()) } async fn get_random_image(mut mapsets: Vec<MapsetTagWrapper>, mode: GameMode) -> (u32, Vec<u8>) { let mut path = CONFIG.get().unwrap().paths.backgrounds.
clone(); ma
identifier_name
tags.rs
: CommandData) -> Result<()> { let (msg, mut args) = match data { CommandData::Message { msg, args, .. } => (msg, args), CommandData::Interaction { .. } => unreachable!(), }; // Parse mapset id let mapset_id = match args.next().map(u32::from_str) { Some(Ok(num)) => num, Some(Err(_)) => { let content = "Could not parse mapset id. Be sure to specify it as first argument"; return msg.error(&ctx, content).await; } None => { let content = "Arguments: `[mapset id] [add/a/remove/r] [list of tags]`\n\ Example: `21662 r hard farm streams alternate hardname tech weeb bluesky`\n\ Tags: `farm, streams, alternate, old, meme, hardname, easy, hard, tech, \ weeb, bluesky, english`"; let builder = MessageBuilder::new().content(content); msg.create_message(&ctx, builder).await?; return Ok(()); } }; // Check if there is background for the given mapset id if ctx.psql().get_tags_mapset(mapset_id).await.is_err() { let content = "No background entry found with this id"; return msg.error(&ctx, content).await; } // Parse action let action = match args.next().map(Action::from_str) { Some(Ok(action)) => action, None | Some(Err(_)) => { let content = "Could not parse action. \ Be sure to specify `r`, `remove`, `a`, or `add` as second argument"; return msg.error(&ctx, content).await; } }; // Parse tags let mut tags = MapsetTags::empty(); while !args.is_empty() { match args.next().map(MapsetTags::from_str) { Some(Ok(tag)) => tags.insert(tag), Some(Err(tag)) => { let content = format!( "Could not parse tag `{tag}`.\n\ Be sure to only give these tags:\n\ `farm, streams, alternate, old, meme, hardname, \ easy, hard, tech, weeb, bluesky, english`" ); return msg.error(&ctx, content).await; } None => unreachable!(), } } let result = if tags.is_empty() { ctx.psql().get_tags_mapset(mapset_id).await } else { let db_result = match action { Action::Add => ctx.psql().add_tags_mapset(mapset_id, tags).await, Action::Remove => ctx.psql().remove_tags_mapset(mapset_id, tags).await, }; match db_result { Ok(_) => ctx.psql().get_tags_mapset(mapset_id).await, Err(err) => Err(err), } }; // Then show the final tags match result { Ok(tags) => { let content = format!("{OSU_BASE}beatmapsets/{mapset_id} is now tagged as:\n{tags}"); let builder = MessageBuilder::new().content(content); msg.create_message(&ctx, builder).await?; } Err(err) => { let _ = msg.error(&ctx, GENERAL_ISSUE).await; return Err(err); } } Ok(()) } // #[command] // #[short_desc("Help out tagging backgrounds")] // #[long_desc( // "Let me give you mapsets that still need to be tagged.\n\ // React to them properly, then lock it in by reacting with ✅.\n\ // To leave the loop, react with ❌ or just wait 10 minutes.\n\ // Mode can be specified in the first argument, defaults to std.\n\ // **You need to be verified to use this command**, feel free to \ // let Badewanne3 know if you want to help out tagging backgrounds." // )] // #[usage("[std / mna]")] // #[aliases("bgt", "bgtag")] // #[owner()] async fn bgtags(ctx: Arc<Context>, data: CommandData) -> Result<()> {
let mut untagged = match ctx.psql().get_all_tags_mapset(mode).await { Ok(tags) => tags.iter().any(|tag| tag.untagged()), Err(err) => { let _ = msg.error(&ctx, GENERAL_ISSUE).await; return Err(err); } }; if !untagged { let content = "All backgrounds have been tagged, \ here are some random ones you can review again though"; let builder = MessageBuilder::new().content(content); let _ = msg.create_message(&ctx, builder).await; } let mut owner = msg.author.id; loop { // Get all mapsets for which tags are missing let tags_result = if untagged { ctx.psql().get_all_tags_mapset(mode).await } else { ctx.psql() .get_random_tags_mapset(mode) .await .map(|tags| vec![tags]) }; let mapsets = match tags_result { Ok(mut tags) => { if untagged { if tags.iter().any(|tag| tag.untagged()) { tags.retain(|tag| tag.untagged()); } else { let content = "All backgrounds have been tagged, \ here are some random ones you can review again though"; let builder = MessageBuilder::new().content(content); let _ = msg.create_message(&ctx, builder).await; untagged = false; tags.truncate(1); } } tags } Err(err) => { let _ = msg.error(&ctx, GENERAL_ISSUE).await; return Err(err); } }; let (mapset_id, img) = get_random_image(mapsets, mode).await; let content = format!( "<@{owner}> Which tags should this mapsets get: {OSU_BASE}beatmapsets/{mapset_id}\n\ ```\n\ 🍋: Easy 🎨: Weeb 😱: Hard name 🗽: English 💯: Tech\n\ 🤓: Hard 🍨: Kpop 🪀: Alternate 🌀: Streams ✅: Lock in\n\ 🤡: Meme 👨‍🌾: Farm 🟦: Blue sky 👴: Old ❌: Exit loop\n\ ```" ); let img = Attachment::from_bytes("bg_img.png".to_owned(), img); // Send response let response = ctx .http .create_message(msg.channel_id) .content(&content)? .attachments(&[img]) .unwrap() .exec() .await? .model() .await?; let msg_id = response.id; // Setup collector let reaction_stream = ctx .standby .wait_for_event_stream(move |event: &Event| match event { Event::ReactionAdd(event) => { event.message_id == msg_id && event.user_id.get() == OWNER_USER_ID } Event::ReactionRemove(event) => { event.message_id == msg_id && event.user_id.get() == OWNER_USER_ID } _ => false, }) .map(|event| match event { Event::ReactionAdd(add) => ReactionWrapper::Add(add.0), Event::ReactionRemove(remove) => ReactionWrapper::Remove(remove.0), _ => unreachable!(), }) .timeout(Duration::from_secs(600)); tokio::pin!(reaction_stream); // Add reactions let reactions = [ "🍋", "🤓", "🤡", "🎨", "🍨", "👨‍🌾", "😱", "🪀", "🟦", "🗽", "🌀", "👴", "💯", "✅", "❌", ]; for &reaction in reactions.iter() { let emote = Emote::Custom(reaction); send_reaction(&*ctx, &response, emote).await?; } let mut break_loop = true; // Run collector let mut add_tags = MapsetTags::empty(); let mut remove_tags = MapsetTags::empty(); while let Some(Ok(reaction)) = reaction_stream.next().await {
let (msg, mut args) = match data { CommandData::Message { msg, args, .. } => (msg, args), CommandData::Interaction { .. } => unreachable!(), }; // Parse arguments as mode let mode = match args.next() { Some(arg) => match arg.cow_to_ascii_lowercase().as_ref() { "mna" | "mania" | "m" => GameMode::Mania, "osu" | "std" | "standard" | "o" => GameMode::Osu, _ => { let content = "Could not parse first argument as mode. \ Provide either `mna`, or `std`"; return msg.error(&ctx, content).await; } }, None => GameMode::Osu, };
identifier_body
golem_renderer.rs
Float, Dimension::D4) )); for matrix_counter in 1 ..= shader.description.num_float_matrices { uniforms.push(Uniform::new( MATRIX_VARIABLE_NAMES[matrix_counter as usize], UniformType::Matrix(Dimension::D4) )); } for color_counter in 1 ..= shader.description.num_colors { uniforms.push(Uniform::new( COLOR_VARIABLE_NAMES[color_counter as usize], UniformType::Vector(NumberType::Float, Dimension::D4) )); } for vector_counter in 1 ..= shader.description.num_float_vectors { uniforms.push(Uniform::new( FLOAT_VECTOR_VARIABLE_NAMES[vector_counter as usize], UniformType::Vector(NumberType::Float, Dimension::D4) )); } for vector_counter in 1 ..= shader.description.num_int_vectors { uniforms.push(Uniform::new( INT_VECTOR_VARIABLE_NAMES[vector_counter as usize], UniformType::Vector(NumberType::Int, Dimension::D4) )); } for float_counter in 1 ..= shader.description.num_floats { uniforms.push(Uniform::new( FLOAT_VARIABLE_NAMES[float_counter as usize], UniformType::Scalar(NumberType::Float) )); } for int_counter in 1 ..= shader.description.num_ints { uniforms.push(Uniform::new( INT_VARIABLE_NAMES[int_counter as usize], UniformType::Scalar(NumberType::Int) )); } let shader_description = ShaderDescription { vertex_input: &[ Attribute::new("vertexInnerPosition", AttributeType::Vector(Dimension::D2)) ], fragment_input: &[ Attribute::new("innerPosition", AttributeType::Vector(Dimension::D2)), Attribute::new("outerPosition", AttributeType::Vector(Dimension::D2)) ], uniforms: &uniforms, vertex_shader: " void main() { innerPosition = 0.5 * vertexInnerPosition + 0.5; vec2 bottomLeftBounds = vertexBounds.xy; vec2 topRightBounds = vertexBounds.zw; outerPosition = bottomLeftBounds + innerPosition * (topRightBounds - bottomLeftBounds); gl_Position = vec4(2.0 * outerPosition - vec2(1.0, 1.0), 0.0, 1.0); } ", fragment_shader: &shader.description.source_code }; ShaderProgram::new(golem, shader_description) }, |shader_program| { shader_program.set_uniform("vertexBounds", UniformValue::Vector4([min_x, min_y, max_x, max_y]))?; for matrix_counter in 1 ..= shader.description.num_float_matrices { let _result = shader_program.set_uniform( &format!("matrix{}", matrix_counter), UniformValue::Matrix4(parameters.float_matrices[matrix_counter as usize - 1]) ); } for color_counter in 1 ..= shader.description.num_colors { let _result = shader_program.set_uniform( &format!("color{}", color_counter), UniformValue::Vector4(parameters.colors[color_counter as usize - 1].to_float_array()) ); } for vector_counter in 1 ..= shader.description.num_float_vectors { let _result = shader_program.set_uniform( &format!("floatVector{}", vector_counter), UniformValue::Vector4(parameters.float_vectors[vector_counter as usize - 1]) ); } for vector_counter in 1 ..= shader.description.num_int_vectors { let _result = shader_program.set_uniform( &format!("intVector{}", vector_counter), UniformValue::IVector4(parameters.int_vectors[vector_counter as usize - 1]) ); } for float_counter in 1 ..= shader.description.num_floats { let _result = shader_program.set_uniform( &format!("float{}", float_counter),
); } for int_counter in 1 ..= shader.description.num_ints { let _result = shader_program.set_uniform( &format!("int{}", int_counter), UniformValue::Int(parameters.ints[int_counter as usize - 1]) ); } unsafe { shader_program.draw( self.get_quad_vertices(), self.get_quad_indices(), 0 .. self.get_num_quad_indices(), GeometryMode::Triangles ) } } ).expect("Shader shouldn't fail"); } /// Gets the golem `Context` of this `Renderer`. Use this context to perform drawing operations /// that are not covered by the other methods of `Renderer`. Note that using this will damage /// the portability of the application since this will only work when a Golem renderer is used. pub fn get_context(&self) -> &Context { &self.context } // This will be handled internally. pub(super) fn apply_viewport_and_scissor(&self) { self.get_viewport().set_viewport(&self.context); self.get_scissor().set_scissor(&self.context); } /// Gets a reference to a `VertexBuffer` representing the basic `quad` model (simply the /// positions [(-1.0, -1.0), (1.0, -1.0), (1.0, 1.0), (-1.0, 1.0)] ). /// /// This model can be surprisingly useful for `Component`s because this simple model can be /// quite powerful in combination with the right (fragment) shader: by discarding the right /// pixels, it is easy to construct other shapes like circles. It is also great for drawing /// basic images. /// /// As explained above, it can be useful for many `Component`. It would be a slight waste of /// resources to let every component create its own quad `VertexBuffer`. To solve this issue, /// all components in need of the quad model can simply share this one. pub fn get_quad_vertices(&self) -> &VertexBuffer { &self.storage.quad_vertices } /// Gets a reference to the corresponding `ElementBuffer` of the `VertexBuffer` given by the /// `get_quad_vertices` method. (These indices are just [(0, 1, 2), (2, 3, 0)].) pub fn get_quad_indices(&self) -> &ElementBuffer { &self.storage.quad_indices } /// Gets the number of indices in the `ElementBuffer` given by the `get_quad_indices` /// method, in integers (which is just 6). pub fn get_num_quad_indices(&self) -> usize { 6 } /// Checks if the shader with the given *id* has been cached by this `Renderer`. If so, `bind`s /// that shader and calls the given *use_shader* closure. /// /// If the shader with the given *id* is **not** found in the cache, the given `create_shader` /// closure will be called to create this. Then, it will be stored in the cache and its `bind` /// function will be called. And finally, the given *use_shader* closure will be called. /// /// ## Motivation /// Caching shaders can make the implementation of the `render` methods of `Component`s easier /// while also improving performance: `Component`s often need shader(s) for rendering, and they /// either need to create it at the start of every call of their `render` method (which is very /// bad for performance). Or, they could create it lazily during their first `render` call and /// store it for later (which is annoying to program because it requires adding an extra /// `Option<ShaderProgram>` field and maintain that). That would be better for performance, but /// is still suboptimal because every `Component` will need its **own** instance of the /// shader it need(s), even if many other `Component`s need that exact same shader. /// /// When `Component`s use this method, they no longer need to worry about storing the shader /// (because the `Renderer` will take care of that), and it will automatically be shared by all /// other `Component` that use this method and the same shader **id**. pub fn use_cached_shader( &self, id: &ShaderId, create_shader: impl FnOnce(&golem::Context) -> Result<ShaderProgram, GolemError>, use_shader: impl FnOnce(&mut ShaderProgram) -> Result<(), GolemError>, ) -> Result<(), GolemError> { let mut cache = self.storage.shader_cache.borrow_mut(); cache.use_shader(id, || create_shader(&self.context), use_shader) } pub fn load_texture(&self, cpu_texture: &crate::Texture) -> Result<golem::Texture, GolemError> { let mut gpu_texture = golem::Texture::new(&self.context)?; let pixel_buffer = cpu_texture.create_pixel_buffer(); gpu_texture.set_image( Some(&pixel_buffer), cpu_texture.get_width(), cpu_texture.get_height(), ColorFormat::RGBA, ); gpu_texture.set_wrap_h(TextureWrap::ClampToEdge)?; gpu_texture.set_wrap_v(TextureWrap::ClampToEdge)?; gpu_texture.set_magnification(TextureFilter::Linear)?; gpu_texture.set_minification(TextureFilter::Linear)?; Ok(gpu_texture) } } pub(super) struct GolemRenderStorage { // Frequently
UniformValue::Float(parameters.floats[float_counter as usize - 1])
random_line_split
golem_renderer.rs
Float, Dimension::D4) )); for matrix_counter in 1 ..= shader.description.num_float_matrices { uniforms.push(Uniform::new( MATRIX_VARIABLE_NAMES[matrix_counter as usize], UniformType::Matrix(Dimension::D4) )); } for color_counter in 1 ..= shader.description.num_colors { uniforms.push(Uniform::new( COLOR_VARIABLE_NAMES[color_counter as usize], UniformType::Vector(NumberType::Float, Dimension::D4) )); } for vector_counter in 1 ..= shader.description.num_float_vectors { uniforms.push(Uniform::new( FLOAT_VECTOR_VARIABLE_NAMES[vector_counter as usize], UniformType::Vector(NumberType::Float, Dimension::D4) )); } for vector_counter in 1 ..= shader.description.num_int_vectors { uniforms.push(Uniform::new( INT_VECTOR_VARIABLE_NAMES[vector_counter as usize], UniformType::Vector(NumberType::Int, Dimension::D4) )); } for float_counter in 1 ..= shader.description.num_floats { uniforms.push(Uniform::new( FLOAT_VARIABLE_NAMES[float_counter as usize], UniformType::Scalar(NumberType::Float) )); } for int_counter in 1 ..= shader.description.num_ints { uniforms.push(Uniform::new( INT_VARIABLE_NAMES[int_counter as usize], UniformType::Scalar(NumberType::Int) )); } let shader_description = ShaderDescription { vertex_input: &[ Attribute::new("vertexInnerPosition", AttributeType::Vector(Dimension::D2)) ], fragment_input: &[ Attribute::new("innerPosition", AttributeType::Vector(Dimension::D2)), Attribute::new("outerPosition", AttributeType::Vector(Dimension::D2)) ], uniforms: &uniforms, vertex_shader: " void main() { innerPosition = 0.5 * vertexInnerPosition + 0.5; vec2 bottomLeftBounds = vertexBounds.xy; vec2 topRightBounds = vertexBounds.zw; outerPosition = bottomLeftBounds + innerPosition * (topRightBounds - bottomLeftBounds); gl_Position = vec4(2.0 * outerPosition - vec2(1.0, 1.0), 0.0, 1.0); } ", fragment_shader: &shader.description.source_code }; ShaderProgram::new(golem, shader_description) }, |shader_program| { shader_program.set_uniform("vertexBounds", UniformValue::Vector4([min_x, min_y, max_x, max_y]))?; for matrix_counter in 1 ..= shader.description.num_float_matrices { let _result = shader_program.set_uniform( &format!("matrix{}", matrix_counter), UniformValue::Matrix4(parameters.float_matrices[matrix_counter as usize - 1]) ); } for color_counter in 1 ..= shader.description.num_colors { let _result = shader_program.set_uniform( &format!("color{}", color_counter), UniformValue::Vector4(parameters.colors[color_counter as usize - 1].to_float_array()) ); } for vector_counter in 1 ..= shader.description.num_float_vectors { let _result = shader_program.set_uniform( &format!("floatVector{}", vector_counter), UniformValue::Vector4(parameters.float_vectors[vector_counter as usize - 1]) ); } for vector_counter in 1 ..= shader.description.num_int_vectors { let _result = shader_program.set_uniform( &format!("intVector{}", vector_counter), UniformValue::IVector4(parameters.int_vectors[vector_counter as usize - 1]) ); } for float_counter in 1 ..= shader.description.num_floats { let _result = shader_program.set_uniform( &format!("float{}", float_counter), UniformValue::Float(parameters.floats[float_counter as usize - 1]) ); } for int_counter in 1 ..= shader.description.num_ints { let _result = shader_program.set_uniform( &format!("int{}", int_counter), UniformValue::Int(parameters.ints[int_counter as usize - 1]) ); } unsafe { shader_program.draw( self.get_quad_vertices(), self.get_quad_indices(), 0 .. self.get_num_quad_indices(), GeometryMode::Triangles ) } } ).expect("Shader shouldn't fail"); } /// Gets the golem `Context` of this `Renderer`. Use this context to perform drawing operations /// that are not covered by the other methods of `Renderer`. Note that using this will damage /// the portability of the application since this will only work when a Golem renderer is used. pub fn get_context(&self) -> &Context { &self.context } // This will be handled internally. pub(super) fn apply_viewport_and_scissor(&self) { self.get_viewport().set_viewport(&self.context); self.get_scissor().set_scissor(&self.context); } /// Gets a reference to a `VertexBuffer` representing the basic `quad` model (simply the /// positions [(-1.0, -1.0), (1.0, -1.0), (1.0, 1.0), (-1.0, 1.0)] ). /// /// This model can be surprisingly useful for `Component`s because this simple model can be /// quite powerful in combination with the right (fragment) shader: by discarding the right /// pixels, it is easy to construct other shapes like circles. It is also great for drawing /// basic images. /// /// As explained above, it can be useful for many `Component`. It would be a slight waste of /// resources to let every component create its own quad `VertexBuffer`. To solve this issue, /// all components in need of the quad model can simply share this one. pub fn get_quad_vertices(&self) -> &VertexBuffer { &self.storage.quad_vertices } /// Gets a reference to the corresponding `ElementBuffer` of the `VertexBuffer` given by the /// `get_quad_vertices` method. (These indices are just [(0, 1, 2), (2, 3, 0)].) pub fn
(&self) -> &ElementBuffer { &self.storage.quad_indices } /// Gets the number of indices in the `ElementBuffer` given by the `get_quad_indices` /// method, in integers (which is just 6). pub fn get_num_quad_indices(&self) -> usize { 6 } /// Checks if the shader with the given *id* has been cached by this `Renderer`. If so, `bind`s /// that shader and calls the given *use_shader* closure. /// /// If the shader with the given *id* is **not** found in the cache, the given `create_shader` /// closure will be called to create this. Then, it will be stored in the cache and its `bind` /// function will be called. And finally, the given *use_shader* closure will be called. /// /// ## Motivation /// Caching shaders can make the implementation of the `render` methods of `Component`s easier /// while also improving performance: `Component`s often need shader(s) for rendering, and they /// either need to create it at the start of every call of their `render` method (which is very /// bad for performance). Or, they could create it lazily during their first `render` call and /// store it for later (which is annoying to program because it requires adding an extra /// `Option<ShaderProgram>` field and maintain that). That would be better for performance, but /// is still suboptimal because every `Component` will need its **own** instance of the /// shader it need(s), even if many other `Component`s need that exact same shader. /// /// When `Component`s use this method, they no longer need to worry about storing the shader /// (because the `Renderer` will take care of that), and it will automatically be shared by all /// other `Component` that use this method and the same shader **id**. pub fn use_cached_shader( &self, id: &ShaderId, create_shader: impl FnOnce(&golem::Context) -> Result<ShaderProgram, GolemError>, use_shader: impl FnOnce(&mut ShaderProgram) -> Result<(), GolemError>, ) -> Result<(), GolemError> { let mut cache = self.storage.shader_cache.borrow_mut(); cache.use_shader(id, || create_shader(&self.context), use_shader) } pub fn load_texture(&self, cpu_texture: &crate::Texture) -> Result<golem::Texture, GolemError> { let mut gpu_texture = golem::Texture::new(&self.context)?; let pixel_buffer = cpu_texture.create_pixel_buffer(); gpu_texture.set_image( Some(&pixel_buffer), cpu_texture.get_width(), cpu_texture.get_height(), ColorFormat::RGBA, ); gpu_texture.set_wrap_h(TextureWrap::ClampToEdge)?; gpu_texture.set_wrap_v(TextureWrap::ClampToEdge)?; gpu_texture.set_magnification(TextureFilter::Linear)?; gpu_texture.set_minification(TextureFilter::Linear)?; Ok(gpu_texture) } } pub(super) struct GolemRenderStorage { //
get_quad_indices
identifier_name
golem_renderer.rs
].to_float_array()) ); } for vector_counter in 1 ..= shader.description.num_float_vectors { let _result = shader_program.set_uniform( &format!("floatVector{}", vector_counter), UniformValue::Vector4(parameters.float_vectors[vector_counter as usize - 1]) ); } for vector_counter in 1 ..= shader.description.num_int_vectors { let _result = shader_program.set_uniform( &format!("intVector{}", vector_counter), UniformValue::IVector4(parameters.int_vectors[vector_counter as usize - 1]) ); } for float_counter in 1 ..= shader.description.num_floats { let _result = shader_program.set_uniform( &format!("float{}", float_counter), UniformValue::Float(parameters.floats[float_counter as usize - 1]) ); } for int_counter in 1 ..= shader.description.num_ints { let _result = shader_program.set_uniform( &format!("int{}", int_counter), UniformValue::Int(parameters.ints[int_counter as usize - 1]) ); } unsafe { shader_program.draw( self.get_quad_vertices(), self.get_quad_indices(), 0 .. self.get_num_quad_indices(), GeometryMode::Triangles ) } } ).expect("Shader shouldn't fail"); } /// Gets the golem `Context` of this `Renderer`. Use this context to perform drawing operations /// that are not covered by the other methods of `Renderer`. Note that using this will damage /// the portability of the application since this will only work when a Golem renderer is used. pub fn get_context(&self) -> &Context { &self.context } // This will be handled internally. pub(super) fn apply_viewport_and_scissor(&self) { self.get_viewport().set_viewport(&self.context); self.get_scissor().set_scissor(&self.context); } /// Gets a reference to a `VertexBuffer` representing the basic `quad` model (simply the /// positions [(-1.0, -1.0), (1.0, -1.0), (1.0, 1.0), (-1.0, 1.0)] ). /// /// This model can be surprisingly useful for `Component`s because this simple model can be /// quite powerful in combination with the right (fragment) shader: by discarding the right /// pixels, it is easy to construct other shapes like circles. It is also great for drawing /// basic images. /// /// As explained above, it can be useful for many `Component`. It would be a slight waste of /// resources to let every component create its own quad `VertexBuffer`. To solve this issue, /// all components in need of the quad model can simply share this one. pub fn get_quad_vertices(&self) -> &VertexBuffer { &self.storage.quad_vertices } /// Gets a reference to the corresponding `ElementBuffer` of the `VertexBuffer` given by the /// `get_quad_vertices` method. (These indices are just [(0, 1, 2), (2, 3, 0)].) pub fn get_quad_indices(&self) -> &ElementBuffer { &self.storage.quad_indices } /// Gets the number of indices in the `ElementBuffer` given by the `get_quad_indices` /// method, in integers (which is just 6). pub fn get_num_quad_indices(&self) -> usize { 6 } /// Checks if the shader with the given *id* has been cached by this `Renderer`. If so, `bind`s /// that shader and calls the given *use_shader* closure. /// /// If the shader with the given *id* is **not** found in the cache, the given `create_shader` /// closure will be called to create this. Then, it will be stored in the cache and its `bind` /// function will be called. And finally, the given *use_shader* closure will be called. /// /// ## Motivation /// Caching shaders can make the implementation of the `render` methods of `Component`s easier /// while also improving performance: `Component`s often need shader(s) for rendering, and they /// either need to create it at the start of every call of their `render` method (which is very /// bad for performance). Or, they could create it lazily during their first `render` call and /// store it for later (which is annoying to program because it requires adding an extra /// `Option<ShaderProgram>` field and maintain that). That would be better for performance, but /// is still suboptimal because every `Component` will need its **own** instance of the /// shader it need(s), even if many other `Component`s need that exact same shader. /// /// When `Component`s use this method, they no longer need to worry about storing the shader /// (because the `Renderer` will take care of that), and it will automatically be shared by all /// other `Component` that use this method and the same shader **id**. pub fn use_cached_shader( &self, id: &ShaderId, create_shader: impl FnOnce(&golem::Context) -> Result<ShaderProgram, GolemError>, use_shader: impl FnOnce(&mut ShaderProgram) -> Result<(), GolemError>, ) -> Result<(), GolemError> { let mut cache = self.storage.shader_cache.borrow_mut(); cache.use_shader(id, || create_shader(&self.context), use_shader) } pub fn load_texture(&self, cpu_texture: &crate::Texture) -> Result<golem::Texture, GolemError> { let mut gpu_texture = golem::Texture::new(&self.context)?; let pixel_buffer = cpu_texture.create_pixel_buffer(); gpu_texture.set_image( Some(&pixel_buffer), cpu_texture.get_width(), cpu_texture.get_height(), ColorFormat::RGBA, ); gpu_texture.set_wrap_h(TextureWrap::ClampToEdge)?; gpu_texture.set_wrap_v(TextureWrap::ClampToEdge)?; gpu_texture.set_magnification(TextureFilter::Linear)?; gpu_texture.set_minification(TextureFilter::Linear)?; Ok(gpu_texture) } } pub(super) struct GolemRenderStorage { // Frequently used and cheap buffers quad_vertices: VertexBuffer, quad_indices: ElementBuffer, shader_cache: RefCell<ShaderCache>, } impl GolemRenderStorage { fn new(context: &Context) -> Result<Self, GolemError> { let mut quad_vertices = VertexBuffer::new(context)?; #[rustfmt::skip] quad_vertices.set_data(&[-1.0, -1.0, 1.0, -1.0, 1.0, 1.0, -1.0, 1.0]); let mut quad_indices = ElementBuffer::new(context)?; quad_indices.set_data(&[0, 1, 2, 2, 3, 0]); // Practice will have to tell whether 200 is good. let max_cached_shaders = 200; Ok(Self { quad_vertices, quad_indices, shader_cache: RefCell::new(ShaderCache::new(max_cached_shaders)), }) } } struct ShaderCache { map: HashMap<ShaderId, CachedShader>, max_cached_shaders: usize, current_time: u64, } impl ShaderCache { fn new(max_cached_shaders: usize) -> Self { assert!(max_cached_shaders > 0); Self { map: HashMap::new(), max_cached_shaders, current_time: 0, } } fn get_existing(&mut self, id: &ShaderId) -> &mut ShaderProgram { let cached = self.map.get_mut(id).unwrap(); cached.last_used = self.current_time; return &mut cached.shader; } fn use_shader( &mut self, id: &ShaderId, create_shader: impl FnOnce() -> Result<ShaderProgram, GolemError>, use_shader: impl FnOnce(&mut ShaderProgram) -> Result<(), GolemError>, ) -> Result<(), GolemError> { self.current_time += 1; // If we have the value already, update its last_used and return it // Unfortunately, we do 2 hash lookups. I tried using only 1, but couldn't convince compiler let has_already = self.map.contains_key(id); if has_already { let shader = self.get_existing(id); shader.bind(); return use_shader(shader); } // If we reach this line, we didn't have the shader yet let new_length = self.map.len() + 1; // If we would exceed the maximum number of cached shaders, we remove the least recently used half if new_length > self.max_cached_shaders
{ let mut last_used_times: Vec<u64> = self .map .values() .map(|cached_shader| cached_shader.last_used) .collect(); last_used_times.sort(); let median = last_used_times[last_used_times.len() / 2]; // Remove at least half of the cached shaders self.map .retain(|_id, cached_shader| cached_shader.last_used > median); }
conditional_block
TheGiver.py
usersSavedDataObj.keys(): for browserCount in usersSavedDataObj[userName]["previous_browser_counts"]: if browserCount["name"] == browserName: return browserCount def usersBrowserCount(userName, browserName): if userName in usersObj.keys(): for browserCount in usersObj[userName]["previous_browser_counts"]: if browserCount["name"] == browserName: return browserCount def usersPreviousScore(userName): return sum(b["score"] for b in usersObj[userName]["previous_browsers"])
currentLoser = users[0] currentScore = usersPreviousScore(currentLoser["name"]) for user in users: userScore = usersPreviousScore(user["name"]) if userScore > currentScore: currentLoser = user currentScore = userScore return currentLoser # Clean up object for user in users: user["score"] = 0.0 user["last_score"] = 0.0 user["loses"] = 0 user["bails"] = 0 user["previous_browsers"] = [] user["previous_browser_counts"] = [] # Load saved user data into users if user["name"] in usersSavedDataObj.keys(): user["score"] = usersSavedDataObj[user["name"]]["score"] user["loses"] = usersSavedDataObj[user["name"]]["loses"] user["previous_browsers"] = usersSavedDataObj[user["name"]]["previous_browsers"] user["previous_browser_counts"] = usersSavedDataObj[user["name"]]["previous_browser_counts"] user["bails"] = usersSavedDataObj[user["name"]]["bails"] for browser in browsers: browserCount = usersSavedBrowserCount(user["name"], browser["name"]) if browserCount is None: browserCount = {"name":browser["name"], "count": 0} user["previous_browser_counts"].append(browserCount) # Order user by score, highest score more likely to luck out and not get a second browser orderedUsers = sorted(users, key=lambda k: k["score"]) # reset when needed if len(sys.argv) > 1: if sys.argv[1].upper() == "RESET": for user in users: user["score"] = 0.0 user["last_score"] = 0.0 user["loses"] = 0 user["bails"] = 0 user["previous_browsers"] = [] user["previous_browser_counts"] = [] for browser in browsers: browserCount = {"name":browser["name"], "count": 0} user["previous_browser_counts"].append(browserCount) # Check Lose Fairness elif sys.argv[1].upper() == "LOSERS": print("LOSERS:") orderedLosers = sorted(users, key=lambda k: k["loses"], reverse=True) sum = 0 for user in orderedLosers: sum = sum + user["loses"] for user in orderedLosers: perc = 0.0 if sum > perc: perc = float(user["loses"])/float(sum) lineLen = int(gridLen*perc) lineString = '' for j in range(gridLen): if j < lineLen: lineString = lineString + '|' else: lineString = lineString + '.' print(lineString + ' ' + user["name"] + ' (' + str(user["loses"]) + ') : ' + str(int(perc*100)) + '%') # Swap browser testing for previous results elif sys.argv[1].upper() == "SWAP": print("SWAP:") print('\n'.join('[' + str(i) + '] ' + users[i]["name"] + ' (' + str(users[i]["score"]) + ')' for i in range(len(users)))) indexA = int(raw_input('Lucky SOB\'s index: ')) indexB = int(raw_input('Unlucky SOB\' index: ')) if indexA < len(users) and indexB < len(users): loserUser = previousRoundLoser() userA = users[indexA] userB = users[indexB] browsersA = userA["previous_browsers"] browsersB = userB["previous_browsers"] print('') print(userA["name"] + ' can swap the following browsers:') print('\n'.join('[' + str(i) + '] ' + browsersA[i]["name"] + ' (' + str(browsersA[i]["score"]) + ')' for i in range(len(browsersA)))) indexC = int(raw_input('Browser index: ')) if (indexC < len(browsersA)): browserC = browsersA[indexC] confirm = raw_input('Take ' + browserC["name"] + ' from ' + userA["name"] + ' and give it to ' + userB["name"] + ' (y/n)? ') print('') if confirm is 'y': browsersA.pop(indexC) browsersB.append(browserC) # update saved scores userA["score"] = userA["score"] - browserC["score"] userB["score"] = userB["score"] + browserC["score"] # update tested browser counts browserCountA = usersBrowserCount(userA["name"], browserC["name"]) browserCountA["count"] = browserCountA["count"] - 1 browserCountB = usersBrowserCount(userB["name"], browserC["name"]) browserCountB["count"] = browserCountB["count"] + 1 # update last round's user if needed if usersPreviousScore(userB["name"]) > usersPreviousScore(loserUser["name"]): print('Previous Loser: ' + str(usersPreviousScore(loserUser["name"])) + ' ' + loserUser["name"]) print('New Loser: ' + str(usersPreviousScore(userB["name"])) + ' ' + userB["name"]) print('') loserUser["loses"] = loserUser["loses"] - 1 userB["loses"] = userB["loses"] + 1 print(userA["name"] + '\'s browsers:') if (len(browsersA) > 0): print('\n'.join('[' + str(i) + '] ' + browsersA[i]["name"] for i in range(len(browsersA)))) print('') print(userB["name"] + '\'s browsers:') if (len(browsersB) > 0): print('\n'.join('[' + str(i) + '] ' + browsersB[i]["name"] for i in range(len(browsersB)))) # Setup for SAVE writeSave = True else: print('Invalid Browser Index!') else: print('Invalid User Index!') # Check randomness elif len(sys.argv[1]) > 0: for user in orderedUsers: if sys.argv[1].upper() == user["name"].upper(): print(sys.argv[1].upper() + ' CHECK:') browserCounts = [] sum = 0 for browserCount in user["previous_browser_counts"]: browserCounts.append(browserCount) sum = sum + browserCount["count"] browserCounts = sorted(browserCounts, key=lambda k: k["count"], reverse=True) for browserCount in browserCounts: perc = 0.0 if sum > perc: perc = float(browserCount["count"])/float(sum) lineLen = int(gridLen*perc) lineString = '' for j in range(gridLen): if j < lineLen: lineString = lineString + '|' else: lineString = lineString + '.' print(lineString + ' ' + browserCount["name"] + ': ' + str(int(perc*100)) + '%') # Do work else: # init previous browser lists for user in users: user["previous_browsers"] = [] # assign random browsers to users userIndex = 0 usersBrowsers = {} remainingBrowsers = list(browsers) random.shuffle(remainingBrowsers) while len(remainingBrowsers) > 0: user = orderedUsers[userIndex%len(orderedUsers)] browser = remainingBrowsers.pop(random.randrange(100)%len(remainingBrowsers)) user["previous_browsers"].append(browser) userIndex = userIndex + 1 # Identify just_awful double Jeopardy zeroJeopardyUsers = [] doubleJeopardyUsers = [] for user in orderedUsers: ieCount = 0 for browser in user["previous_browsers"]: if browser["just_awful"]: ieCount = ieCount + 1 if ieCount == 0: zeroJeopardyUsers.append(user) elif ieCount == 2: doubleJeopardyUsers.append(user) # Resolve just_awful double Jeopardy for i in range(min(len(zeroJeopardyUsers), len(doubleJeopardyUsers))): tempBrowser = zeroJeopardyUsers[i]["previous_browsers"][0] zeroJeopardyUsers[i]["previous_browsers"][0] = doubleJeopardyUsers[i]["previous_browsers"][0]
def previousRoundLoser():
random_line_split
TheGiver.py
currentLoser # Clean up object for user in users: user["score"] = 0.0 user["last_score"] = 0.0 user["loses"] = 0 user["bails"] = 0 user["previous_browsers"] = [] user["previous_browser_counts"] = [] # Load saved user data into users if user["name"] in usersSavedDataObj.keys(): user["score"] = usersSavedDataObj[user["name"]]["score"] user["loses"] = usersSavedDataObj[user["name"]]["loses"] user["previous_browsers"] = usersSavedDataObj[user["name"]]["previous_browsers"] user["previous_browser_counts"] = usersSavedDataObj[user["name"]]["previous_browser_counts"] user["bails"] = usersSavedDataObj[user["name"]]["bails"] for browser in browsers: browserCount = usersSavedBrowserCount(user["name"], browser["name"]) if browserCount is None: browserCount = {"name":browser["name"], "count": 0} user["previous_browser_counts"].append(browserCount) # Order user by score, highest score more likely to luck out and not get a second browser orderedUsers = sorted(users, key=lambda k: k["score"]) # reset when needed if len(sys.argv) > 1: if sys.argv[1].upper() == "RESET": for user in users: user["score"] = 0.0 user["last_score"] = 0.0 user["loses"] = 0 user["bails"] = 0 user["previous_browsers"] = [] user["previous_browser_counts"] = [] for browser in browsers: browserCount = {"name":browser["name"], "count": 0} user["previous_browser_counts"].append(browserCount) # Check Lose Fairness elif sys.argv[1].upper() == "LOSERS": print("LOSERS:") orderedLosers = sorted(users, key=lambda k: k["loses"], reverse=True) sum = 0 for user in orderedLosers: sum = sum + user["loses"] for user in orderedLosers: perc = 0.0 if sum > perc: perc = float(user["loses"])/float(sum) lineLen = int(gridLen*perc) lineString = '' for j in range(gridLen): if j < lineLen: lineString = lineString + '|' else: lineString = lineString + '.' print(lineString + ' ' + user["name"] + ' (' + str(user["loses"]) + ') : ' + str(int(perc*100)) + '%') # Swap browser testing for previous results elif sys.argv[1].upper() == "SWAP": print("SWAP:") print('\n'.join('[' + str(i) + '] ' + users[i]["name"] + ' (' + str(users[i]["score"]) + ')' for i in range(len(users)))) indexA = int(raw_input('Lucky SOB\'s index: ')) indexB = int(raw_input('Unlucky SOB\' index: ')) if indexA < len(users) and indexB < len(users): loserUser = previousRoundLoser() userA = users[indexA] userB = users[indexB] browsersA = userA["previous_browsers"] browsersB = userB["previous_browsers"] print('') print(userA["name"] + ' can swap the following browsers:') print('\n'.join('[' + str(i) + '] ' + browsersA[i]["name"] + ' (' + str(browsersA[i]["score"]) + ')' for i in range(len(browsersA)))) indexC = int(raw_input('Browser index: ')) if (indexC < len(browsersA)): browserC = browsersA[indexC] confirm = raw_input('Take ' + browserC["name"] + ' from ' + userA["name"] + ' and give it to ' + userB["name"] + ' (y/n)? ') print('') if confirm is 'y': browsersA.pop(indexC) browsersB.append(browserC) # update saved scores userA["score"] = userA["score"] - browserC["score"] userB["score"] = userB["score"] + browserC["score"] # update tested browser counts browserCountA = usersBrowserCount(userA["name"], browserC["name"]) browserCountA["count"] = browserCountA["count"] - 1 browserCountB = usersBrowserCount(userB["name"], browserC["name"]) browserCountB["count"] = browserCountB["count"] + 1 # update last round's user if needed if usersPreviousScore(userB["name"]) > usersPreviousScore(loserUser["name"]): print('Previous Loser: ' + str(usersPreviousScore(loserUser["name"])) + ' ' + loserUser["name"]) print('New Loser: ' + str(usersPreviousScore(userB["name"])) + ' ' + userB["name"]) print('') loserUser["loses"] = loserUser["loses"] - 1 userB["loses"] = userB["loses"] + 1 print(userA["name"] + '\'s browsers:') if (len(browsersA) > 0): print('\n'.join('[' + str(i) + '] ' + browsersA[i]["name"] for i in range(len(browsersA)))) print('') print(userB["name"] + '\'s browsers:') if (len(browsersB) > 0): print('\n'.join('[' + str(i) + '] ' + browsersB[i]["name"] for i in range(len(browsersB)))) # Setup for SAVE writeSave = True else: print('Invalid Browser Index!') else: print('Invalid User Index!') # Check randomness elif len(sys.argv[1]) > 0: for user in orderedUsers: if sys.argv[1].upper() == user["name"].upper(): print(sys.argv[1].upper() + ' CHECK:') browserCounts = [] sum = 0 for browserCount in user["previous_browser_counts"]: browserCounts.append(browserCount) sum = sum + browserCount["count"] browserCounts = sorted(browserCounts, key=lambda k: k["count"], reverse=True) for browserCount in browserCounts: perc = 0.0 if sum > perc: perc = float(browserCount["count"])/float(sum) lineLen = int(gridLen*perc) lineString = '' for j in range(gridLen): if j < lineLen: lineString = lineString + '|' else: lineString = lineString + '.' print(lineString + ' ' + browserCount["name"] + ': ' + str(int(perc*100)) + '%') # Do work else: # init previous browser lists for user in users: user["previous_browsers"] = [] # assign random browsers to users userIndex = 0 usersBrowsers = {} remainingBrowsers = list(browsers) random.shuffle(remainingBrowsers) while len(remainingBrowsers) > 0: user = orderedUsers[userIndex%len(orderedUsers)] browser = remainingBrowsers.pop(random.randrange(100)%len(remainingBrowsers)) user["previous_browsers"].append(browser) userIndex = userIndex + 1 # Identify just_awful double Jeopardy zeroJeopardyUsers = [] doubleJeopardyUsers = [] for user in orderedUsers: ieCount = 0 for browser in user["previous_browsers"]: if browser["just_awful"]: ieCount = ieCount + 1 if ieCount == 0: zeroJeopardyUsers.append(user) elif ieCount == 2: doubleJeopardyUsers.append(user) # Resolve just_awful double Jeopardy for i in range(min(len(zeroJeopardyUsers), len(doubleJeopardyUsers))): tempBrowser = zeroJeopardyUsers[i]["previous_browsers"][0] zeroJeopardyUsers[i]["previous_browsers"][0] = doubleJeopardyUsers[i]["previous_browsers"][0] doubleJeopardyUsers[i]["previous_browsers"][0] = tempBrowser # print results and clean up user objects thisLoser = '' thisLosingScore = 0 biggestLoser = '' biggestLosingScore = 0 scoreLinesForPrint = {} for user in orderedUsers: scoreThisRound = 0 usersBrowsersString = '' for browser in user["previous_browsers"]:
scoreThisRound = scoreThisRound + browser["score"] usersBrowsersString = usersBrowsersString + '[' + browser["name"] + '] ' # update the number of times user has tested this browser browserCount = usersBrowserCount(user["name"], browser["name"]) browserCount["count"] = browserCount["count"] + 1
conditional_block
TheGiver.py
usersSavedDataObj.keys(): for browserCount in usersSavedDataObj[userName]["previous_browser_counts"]: if browserCount["name"] == browserName: return browserCount def usersBrowserCount(userName, browserName): if userName in usersObj.keys(): for browserCount in usersObj[userName]["previous_browser_counts"]: if browserCount["name"] == browserName: return browserCount def usersPreviousScore(userName): return sum(b["score"] for b in usersObj[userName]["previous_browsers"]) def
(): currentLoser = users[0] currentScore = usersPreviousScore(currentLoser["name"]) for user in users: userScore = usersPreviousScore(user["name"]) if userScore > currentScore: currentLoser = user currentScore = userScore return currentLoser # Clean up object for user in users: user["score"] = 0.0 user["last_score"] = 0.0 user["loses"] = 0 user["bails"] = 0 user["previous_browsers"] = [] user["previous_browser_counts"] = [] # Load saved user data into users if user["name"] in usersSavedDataObj.keys(): user["score"] = usersSavedDataObj[user["name"]]["score"] user["loses"] = usersSavedDataObj[user["name"]]["loses"] user["previous_browsers"] = usersSavedDataObj[user["name"]]["previous_browsers"] user["previous_browser_counts"] = usersSavedDataObj[user["name"]]["previous_browser_counts"] user["bails"] = usersSavedDataObj[user["name"]]["bails"] for browser in browsers: browserCount = usersSavedBrowserCount(user["name"], browser["name"]) if browserCount is None: browserCount = {"name":browser["name"], "count": 0} user["previous_browser_counts"].append(browserCount) # Order user by score, highest score more likely to luck out and not get a second browser orderedUsers = sorted(users, key=lambda k: k["score"]) # reset when needed if len(sys.argv) > 1: if sys.argv[1].upper() == "RESET": for user in users: user["score"] = 0.0 user["last_score"] = 0.0 user["loses"] = 0 user["bails"] = 0 user["previous_browsers"] = [] user["previous_browser_counts"] = [] for browser in browsers: browserCount = {"name":browser["name"], "count": 0} user["previous_browser_counts"].append(browserCount) # Check Lose Fairness elif sys.argv[1].upper() == "LOSERS": print("LOSERS:") orderedLosers = sorted(users, key=lambda k: k["loses"], reverse=True) sum = 0 for user in orderedLosers: sum = sum + user["loses"] for user in orderedLosers: perc = 0.0 if sum > perc: perc = float(user["loses"])/float(sum) lineLen = int(gridLen*perc) lineString = '' for j in range(gridLen): if j < lineLen: lineString = lineString + '|' else: lineString = lineString + '.' print(lineString + ' ' + user["name"] + ' (' + str(user["loses"]) + ') : ' + str(int(perc*100)) + '%') # Swap browser testing for previous results elif sys.argv[1].upper() == "SWAP": print("SWAP:") print('\n'.join('[' + str(i) + '] ' + users[i]["name"] + ' (' + str(users[i]["score"]) + ')' for i in range(len(users)))) indexA = int(raw_input('Lucky SOB\'s index: ')) indexB = int(raw_input('Unlucky SOB\' index: ')) if indexA < len(users) and indexB < len(users): loserUser = previousRoundLoser() userA = users[indexA] userB = users[indexB] browsersA = userA["previous_browsers"] browsersB = userB["previous_browsers"] print('') print(userA["name"] + ' can swap the following browsers:') print('\n'.join('[' + str(i) + '] ' + browsersA[i]["name"] + ' (' + str(browsersA[i]["score"]) + ')' for i in range(len(browsersA)))) indexC = int(raw_input('Browser index: ')) if (indexC < len(browsersA)): browserC = browsersA[indexC] confirm = raw_input('Take ' + browserC["name"] + ' from ' + userA["name"] + ' and give it to ' + userB["name"] + ' (y/n)? ') print('') if confirm is 'y': browsersA.pop(indexC) browsersB.append(browserC) # update saved scores userA["score"] = userA["score"] - browserC["score"] userB["score"] = userB["score"] + browserC["score"] # update tested browser counts browserCountA = usersBrowserCount(userA["name"], browserC["name"]) browserCountA["count"] = browserCountA["count"] - 1 browserCountB = usersBrowserCount(userB["name"], browserC["name"]) browserCountB["count"] = browserCountB["count"] + 1 # update last round's user if needed if usersPreviousScore(userB["name"]) > usersPreviousScore(loserUser["name"]): print('Previous Loser: ' + str(usersPreviousScore(loserUser["name"])) + ' ' + loserUser["name"]) print('New Loser: ' + str(usersPreviousScore(userB["name"])) + ' ' + userB["name"]) print('') loserUser["loses"] = loserUser["loses"] - 1 userB["loses"] = userB["loses"] + 1 print(userA["name"] + '\'s browsers:') if (len(browsersA) > 0): print('\n'.join('[' + str(i) + '] ' + browsersA[i]["name"] for i in range(len(browsersA)))) print('') print(userB["name"] + '\'s browsers:') if (len(browsersB) > 0): print('\n'.join('[' + str(i) + '] ' + browsersB[i]["name"] for i in range(len(browsersB)))) # Setup for SAVE writeSave = True else: print('Invalid Browser Index!') else: print('Invalid User Index!') # Check randomness elif len(sys.argv[1]) > 0: for user in orderedUsers: if sys.argv[1].upper() == user["name"].upper(): print(sys.argv[1].upper() + ' CHECK:') browserCounts = [] sum = 0 for browserCount in user["previous_browser_counts"]: browserCounts.append(browserCount) sum = sum + browserCount["count"] browserCounts = sorted(browserCounts, key=lambda k: k["count"], reverse=True) for browserCount in browserCounts: perc = 0.0 if sum > perc: perc = float(browserCount["count"])/float(sum) lineLen = int(gridLen*perc) lineString = '' for j in range(gridLen): if j < lineLen: lineString = lineString + '|' else: lineString = lineString + '.' print(lineString + ' ' + browserCount["name"] + ': ' + str(int(perc*100)) + '%') # Do work else: # init previous browser lists for user in users: user["previous_browsers"] = [] # assign random browsers to users userIndex = 0 usersBrowsers = {} remainingBrowsers = list(browsers) random.shuffle(remainingBrowsers) while len(remainingBrowsers) > 0: user = orderedUsers[userIndex%len(orderedUsers)] browser = remainingBrowsers.pop(random.randrange(100)%len(remainingBrowsers)) user["previous_browsers"].append(browser) userIndex = userIndex + 1 # Identify just_awful double Jeopardy zeroJeopardyUsers = [] doubleJeopardyUsers = [] for user in orderedUsers: ieCount = 0 for browser in user["previous_browsers"]: if browser["just_awful"]: ieCount = ieCount + 1 if ieCount == 0: zeroJeopardyUsers.append(user) elif ieCount == 2: doubleJeopardyUsers.append(user) # Resolve just_awful double Jeopardy for i in range(min(len(zeroJeopardyUsers), len(doubleJeopardyUsers))): tempBrowser = zeroJeopardyUsers[i]["previous_browsers"][0] zeroJeopardyUsers[i]["previous_browsers"][0] = doubleJeopardyUsers[i]["previous_browsers"][0]
previousRoundLoser
identifier_name
TheGiver.py
usersSavedDataObj.keys(): for browserCount in usersSavedDataObj[userName]["previous_browser_counts"]: if browserCount["name"] == browserName: return browserCount def usersBrowserCount(userName, browserName): if userName in usersObj.keys(): for browserCount in usersObj[userName]["previous_browser_counts"]: if browserCount["name"] == browserName: return browserCount def usersPreviousScore(userName):
def previousRoundLoser(): currentLoser = users[0] currentScore = usersPreviousScore(currentLoser["name"]) for user in users: userScore = usersPreviousScore(user["name"]) if userScore > currentScore: currentLoser = user currentScore = userScore return currentLoser # Clean up object for user in users: user["score"] = 0.0 user["last_score"] = 0.0 user["loses"] = 0 user["bails"] = 0 user["previous_browsers"] = [] user["previous_browser_counts"] = [] # Load saved user data into users if user["name"] in usersSavedDataObj.keys(): user["score"] = usersSavedDataObj[user["name"]]["score"] user["loses"] = usersSavedDataObj[user["name"]]["loses"] user["previous_browsers"] = usersSavedDataObj[user["name"]]["previous_browsers"] user["previous_browser_counts"] = usersSavedDataObj[user["name"]]["previous_browser_counts"] user["bails"] = usersSavedDataObj[user["name"]]["bails"] for browser in browsers: browserCount = usersSavedBrowserCount(user["name"], browser["name"]) if browserCount is None: browserCount = {"name":browser["name"], "count": 0} user["previous_browser_counts"].append(browserCount) # Order user by score, highest score more likely to luck out and not get a second browser orderedUsers = sorted(users, key=lambda k: k["score"]) # reset when needed if len(sys.argv) > 1: if sys.argv[1].upper() == "RESET": for user in users: user["score"] = 0.0 user["last_score"] = 0.0 user["loses"] = 0 user["bails"] = 0 user["previous_browsers"] = [] user["previous_browser_counts"] = [] for browser in browsers: browserCount = {"name":browser["name"], "count": 0} user["previous_browser_counts"].append(browserCount) # Check Lose Fairness elif sys.argv[1].upper() == "LOSERS": print("LOSERS:") orderedLosers = sorted(users, key=lambda k: k["loses"], reverse=True) sum = 0 for user in orderedLosers: sum = sum + user["loses"] for user in orderedLosers: perc = 0.0 if sum > perc: perc = float(user["loses"])/float(sum) lineLen = int(gridLen*perc) lineString = '' for j in range(gridLen): if j < lineLen: lineString = lineString + '|' else: lineString = lineString + '.' print(lineString + ' ' + user["name"] + ' (' + str(user["loses"]) + ') : ' + str(int(perc*100)) + '%') # Swap browser testing for previous results elif sys.argv[1].upper() == "SWAP": print("SWAP:") print('\n'.join('[' + str(i) + '] ' + users[i]["name"] + ' (' + str(users[i]["score"]) + ')' for i in range(len(users)))) indexA = int(raw_input('Lucky SOB\'s index: ')) indexB = int(raw_input('Unlucky SOB\' index: ')) if indexA < len(users) and indexB < len(users): loserUser = previousRoundLoser() userA = users[indexA] userB = users[indexB] browsersA = userA["previous_browsers"] browsersB = userB["previous_browsers"] print('') print(userA["name"] + ' can swap the following browsers:') print('\n'.join('[' + str(i) + '] ' + browsersA[i]["name"] + ' (' + str(browsersA[i]["score"]) + ')' for i in range(len(browsersA)))) indexC = int(raw_input('Browser index: ')) if (indexC < len(browsersA)): browserC = browsersA[indexC] confirm = raw_input('Take ' + browserC["name"] + ' from ' + userA["name"] + ' and give it to ' + userB["name"] + ' (y/n)? ') print('') if confirm is 'y': browsersA.pop(indexC) browsersB.append(browserC) # update saved scores userA["score"] = userA["score"] - browserC["score"] userB["score"] = userB["score"] + browserC["score"] # update tested browser counts browserCountA = usersBrowserCount(userA["name"], browserC["name"]) browserCountA["count"] = browserCountA["count"] - 1 browserCountB = usersBrowserCount(userB["name"], browserC["name"]) browserCountB["count"] = browserCountB["count"] + 1 # update last round's user if needed if usersPreviousScore(userB["name"]) > usersPreviousScore(loserUser["name"]): print('Previous Loser: ' + str(usersPreviousScore(loserUser["name"])) + ' ' + loserUser["name"]) print('New Loser: ' + str(usersPreviousScore(userB["name"])) + ' ' + userB["name"]) print('') loserUser["loses"] = loserUser["loses"] - 1 userB["loses"] = userB["loses"] + 1 print(userA["name"] + '\'s browsers:') if (len(browsersA) > 0): print('\n'.join('[' + str(i) + '] ' + browsersA[i]["name"] for i in range(len(browsersA)))) print('') print(userB["name"] + '\'s browsers:') if (len(browsersB) > 0): print('\n'.join('[' + str(i) + '] ' + browsersB[i]["name"] for i in range(len(browsersB)))) # Setup for SAVE writeSave = True else: print('Invalid Browser Index!') else: print('Invalid User Index!') # Check randomness elif len(sys.argv[1]) > 0: for user in orderedUsers: if sys.argv[1].upper() == user["name"].upper(): print(sys.argv[1].upper() + ' CHECK:') browserCounts = [] sum = 0 for browserCount in user["previous_browser_counts"]: browserCounts.append(browserCount) sum = sum + browserCount["count"] browserCounts = sorted(browserCounts, key=lambda k: k["count"], reverse=True) for browserCount in browserCounts: perc = 0.0 if sum > perc: perc = float(browserCount["count"])/float(sum) lineLen = int(gridLen*perc) lineString = '' for j in range(gridLen): if j < lineLen: lineString = lineString + '|' else: lineString = lineString + '.' print(lineString + ' ' + browserCount["name"] + ': ' + str(int(perc*100)) + '%') # Do work else: # init previous browser lists for user in users: user["previous_browsers"] = [] # assign random browsers to users userIndex = 0 usersBrowsers = {} remainingBrowsers = list(browsers) random.shuffle(remainingBrowsers) while len(remainingBrowsers) > 0: user = orderedUsers[userIndex%len(orderedUsers)] browser = remainingBrowsers.pop(random.randrange(100)%len(remainingBrowsers)) user["previous_browsers"].append(browser) userIndex = userIndex + 1 # Identify just_awful double Jeopardy zeroJeopardyUsers = [] doubleJeopardyUsers = [] for user in orderedUsers: ieCount = 0 for browser in user["previous_browsers"]: if browser["just_awful"]: ieCount = ieCount + 1 if ieCount == 0: zeroJeopardyUsers.append(user) elif ieCount == 2: doubleJeopardyUsers.append(user) # Resolve just_awful double Jeopardy for i in range(min(len(zeroJeopardyUsers), len(doubleJeopardyUsers))): tempBrowser = zeroJeopardyUsers[i]["previous_browsers"][0] zeroJeopardyUsers[i]["previous_browsers"][0] = doubleJeopardyUsers[i]["previous_browsers"][0
return sum(b["score"] for b in usersObj[userName]["previous_browsers"])
identifier_body
repr.rs
_profile(&profile); if let Ok(static_credentials) = try_static { break BaseProvider::AccessKey(static_credentials); } } let next_profile = match chain_provider(&profile) { // this provider wasn't a chain provider, reload it as a base provider None => { break base_provider(profile).map_err(|err| { ProfileFileError::InvalidCredentialSource { profile: profile.name().into(), message: format!("could not load source profile: {}", err).into(), } })?; } Some(result) => { let (chain_profile, next) = result?; chain.push(chain_profile); next } }; match next_profile { NextProfile::SelfReference => { // self referential profile, don't go through the loop because it will error // on the infinite loop check. Instead, reload this profile as a base profile // and exit. break base_provider(profile)?; } NextProfile::Named(name) => source_profile_name = name, } }; chain.reverse(); Ok(ProfileChain { base, chain }) } mod role { pub const ROLE_ARN: &str = "role_arn"; pub const EXTERNAL_ID: &str = "external_id"; pub const SESSION_NAME: &str = "role_session_name"; pub const CREDENTIAL_SOURCE: &str = "credential_source"; pub const SOURCE_PROFILE: &str = "source_profile"; } mod web_identity_token { pub const TOKEN_FILE: &str = "web_identity_token_file"; } mod static_credentials { pub const AWS_ACCESS_KEY_ID: &str = "aws_access_key_id"; pub const AWS_SECRET_ACCESS_KEY: &str = "aws_secret_access_key"; pub const AWS_SESSION_TOKEN: &str = "aws_session_token"; } const PROVIDER_NAME: &str = "ProfileFile"; fn base_provider(profile: &Profile) -> Result<BaseProvider, ProfileFileError> { // the profile must define either a `CredentialsSource` or a concrete set of access keys match profile.get(role::CREDENTIAL_SOURCE) { Some(source) => Ok(BaseProvider::NamedSource(source)), None => web_identity_token_from_profile(profile) .unwrap_or_else(|| Ok(BaseProvider::AccessKey(static_creds_from_profile(profile)?))), } } enum NextProfile<'a> { SelfReference, Named(&'a str), } fn chain_provider(profile: &Profile) -> Option<Result<(RoleArn, NextProfile), ProfileFileError>> { let role_provider = role_arn_from_profile(&profile)?; let (source_profile, credential_source) = ( profile.get(role::SOURCE_PROFILE), profile.get(role::CREDENTIAL_SOURCE), ); let profile = match (source_profile, credential_source) { (Some(_), Some(_)) => Err(ProfileFileError::InvalidCredentialSource { profile: profile.name().to_string(), message: "profile contained both source_profile and credential_source. \ Only one or the other can be defined" .into(), }), (None, None) => Err(ProfileFileError::InvalidCredentialSource { profile: profile.name().to_string(), message: "profile must contain `source_profile` or `credential_source` but neither were defined" .into(), }), (Some(source_profile), None) if source_profile == profile.name() => { Ok((role_provider, NextProfile::SelfReference)) } (Some(source_profile), None) => Ok((role_provider, NextProfile::Named(source_profile))), // we want to loop back into this profile and pick up the credential source (None, Some(_credential_source)) => Ok((role_provider, NextProfile::SelfReference)), }; Some(profile) } fn role_arn_from_profile(profile: &Profile) -> Option<RoleArn> { // Web Identity Tokens are root providers, not chained roles if profile.get(web_identity_token::TOKEN_FILE).is_some() { return None; } let role_arn = profile.get(role::ROLE_ARN)?; let session_name = profile.get(role::SESSION_NAME); let external_id = profile.get(role::EXTERNAL_ID); Some(RoleArn { role_arn, external_id, session_name, }) } fn web_identity_token_from_profile( profile: &Profile, ) -> Option<Result<BaseProvider, ProfileFileError>> { let session_name = profile.get(role::SESSION_NAME); match ( profile.get(role::ROLE_ARN), profile.get(web_identity_token::TOKEN_FILE), ) { (Some(role_arn), Some(token_file)) => Some(Ok(BaseProvider::WebIdentityTokenRole { role_arn, web_identity_token_file: token_file, session_name, })), (None, None) => None, (Some(_role_arn), None) => None, (None, Some(_token_file)) => Some(Err(ProfileFileError::InvalidCredentialSource { profile: profile.name().to_string(), message: "`web_identity_token_file` was specified but `role_arn` was missing".into(), })), } } /// Load static credentials from a profile /// /// Example: /// ```ini /// [profile B] /// aws_access_key_id = abc123 /// aws_secret_access_key = def456 /// ``` fn static_creds_from_profile(profile: &Profile) -> Result<Credentials, ProfileFileError> { use static_credentials::*; let access_key = profile.get(AWS_ACCESS_KEY_ID); let secret_key = profile.get(AWS_SECRET_ACCESS_KEY); let session_token = profile.get(AWS_SESSION_TOKEN); if let (None, None, None) = (access_key, secret_key, session_token) { return Err(ProfileFileError::ProfileDidNotContainCredentials { profile: profile.name().to_string(), }); } let access_key = access_key.ok_or_else(|| ProfileFileError::InvalidCredentialSource { profile: profile.name().to_string(), message: "profile missing aws_access_key_id".into(), })?; let secret_key = secret_key.ok_or_else(|| ProfileFileError::InvalidCredentialSource { profile: profile.name().to_string(), message: "profile missing aws_secret_access_key".into(), })?; Ok(Credentials::new( access_key, secret_key, session_token.map(|s| s.to_string()), None, PROVIDER_NAME, )) } #[cfg(test)] mod tests { use crate::profile::credentials::repr::{resolve_chain, BaseProvider, ProfileChain}; use crate::profile::ProfileSet; use serde::Deserialize; use std::collections::HashMap; use std::error::Error; use std::fs; #[test] fn run_test_cases() -> Result<(), Box<dyn Error>> { let test_cases: Vec<TestCase> = serde_json::from_str(&fs::read_to_string("./test-data/assume-role-tests.json")?)?; for test_case in test_cases { print!("checking: {}...", test_case.docs); check(test_case); println!("ok") } Ok(()) } fn check(test_case: TestCase) { let source = ProfileSet::new(test_case.input.profile, test_case.input.selected_profile); let actual = resolve_chain(&source, None); let expected = test_case.output; match (expected, actual) { (TestOutput::Error(s), Err(e)) => assert!( format!("{}", e).contains(&s), "expected {} to contain `{}`", e, s ), (TestOutput::ProfileChain(expected), Ok(actual)) => { assert_eq!(to_test_output(actual), expected) } (expected, actual) => panic!( "error/success mismatch. Expected:\n {:?}\nActual:\n {:?}", &expected, actual ), } } #[derive(Deserialize)] struct TestCase { docs: String, input: TestInput, output: TestOutput, } #[derive(Deserialize)] struct TestInput { profile: HashMap<String, HashMap<String, String>>, selected_profile: String, } fn to_test_output(profile_chain: ProfileChain) -> Vec<Provider> { let mut output = vec![]; match profile_chain.base { BaseProvider::NamedSource(name) => output.push(Provider::NamedSource(name.into())), BaseProvider::AccessKey(creds) => output.push(Provider::AccessKey { access_key_id: creds.access_key_id().into(), secret_access_key: creds.secret_access_key().into(), session_token: creds.session_token().map(|tok| tok.to_string()), }), BaseProvider::WebIdentityTokenRole { role_arn, web_identity_token_file, session_name, } => output.push(Provider::WebIdentityToken { role_arn: role_arn.into(), web_identity_token_file: web_identity_token_file.into(), role_session_name: session_name.map(|sess| sess.to_string()), }), }; for role in profile_chain.chain { output.push(Provider::AssumeRole { role_arn: role.role_arn.into(), external_id: role.external_id.map(ToString::to_string), role_session_name: role.session_name.map(ToString::to_string), }) } output } #[derive(Deserialize, Debug, PartialEq, Eq)] enum TestOutput { ProfileChain(Vec<Provider>), Error(String), } #[derive(Deserialize, Debug, Eq, PartialEq)] enum
Provider
identifier_name
repr.rs
BaseProvider<'a> { /// A profile that specifies a named credential source /// Eg: `credential_source = Ec2InstanceMetadata` /// /// The following profile produces two separate `ProfileProvider` rows: /// 1. `BaseProvider::NamedSource("Ec2InstanceMetadata")` /// 2. `RoleArn { role_arn: "...", ... } /// ```ini /// [profile assume-role] /// role_arn = arn:aws:iam::123456789:role/MyRole /// credential_source = Ec2InstanceMetadata /// ``` NamedSource(&'a str), /// A profile with explicitly configured access keys /// /// Example /// ```ini /// [profile C] /// aws_access_key_id = abc123 /// aws_secret_access_key = def456 /// ``` AccessKey(Credentials), WebIdentityTokenRole { role_arn: &'a str, web_identity_token_file: &'a str, session_name: Option<&'a str>, }, // TODO: add SSO support /* /// An SSO Provider Sso { sso_account_id: &'a str, sso_region: &'a str, sso_role_name: &'a str, sso_start_url: &'a str, }, */ } /// A profile that specifies a role to assume /// /// A RoleArn can only be created from either a profile with `source_profile` /// or one with `credential_source`. #[derive(Debug)] pub struct RoleArn<'a> { /// Role to assume pub role_arn: &'a str, /// external_id parameter to pass to the assume role provider pub external_id: Option<&'a str>, /// session name parameter to pass to the assume role provider pub session_name: Option<&'a str>, } /// Resolve a ProfileChain from a ProfileSet or return an error pub fn resolve_chain<'a>( profile_set: &'a ProfileSet, profile_override: Option<&str>, ) -> Result<ProfileChain<'a>, ProfileFileError> { if profile_set.is_empty() { return Err(ProfileFileError::NoProfilesDefined); } let mut source_profile_name = profile_override.unwrap_or_else(|| profile_set.selected_profile()); let mut visited_profiles = vec![]; let mut chain = vec![]; let base = loop { let profile = profile_set.get_profile(source_profile_name).ok_or( ProfileFileError::MissingProfile { profile: source_profile_name.into(), message: format!( "could not find source profile {} referenced from {}", source_profile_name, visited_profiles.last().unwrap_or(&"the root profile") ) .into(), }, )?; if visited_profiles.contains(&source_profile_name) { return Err(ProfileFileError::CredentialLoop { profiles: visited_profiles .into_iter() .map(|s| s.to_string()) .collect(), next: source_profile_name.to_string(), }); } visited_profiles.push(&source_profile_name); // After the first item in the chain, we will prioritize static credentials if they exist if visited_profiles.len() > 1 { let try_static = static_creds_from_profile(&profile); if let Ok(static_credentials) = try_static { break BaseProvider::AccessKey(static_credentials); } } let next_profile = match chain_provider(&profile) { // this provider wasn't a chain provider, reload it as a base provider None => { break base_provider(profile).map_err(|err| { ProfileFileError::InvalidCredentialSource { profile: profile.name().into(), message: format!("could not load source profile: {}", err).into(), } })?; } Some(result) => { let (chain_profile, next) = result?; chain.push(chain_profile); next } }; match next_profile { NextProfile::SelfReference => { // self referential profile, don't go through the loop because it will error // on the infinite loop check. Instead, reload this profile as a base profile // and exit. break base_provider(profile)?; } NextProfile::Named(name) => source_profile_name = name, } }; chain.reverse(); Ok(ProfileChain { base, chain }) } mod role { pub const ROLE_ARN: &str = "role_arn"; pub const EXTERNAL_ID: &str = "external_id"; pub const SESSION_NAME: &str = "role_session_name"; pub const CREDENTIAL_SOURCE: &str = "credential_source"; pub const SOURCE_PROFILE: &str = "source_profile"; } mod web_identity_token { pub const TOKEN_FILE: &str = "web_identity_token_file"; } mod static_credentials { pub const AWS_ACCESS_KEY_ID: &str = "aws_access_key_id"; pub const AWS_SECRET_ACCESS_KEY: &str = "aws_secret_access_key"; pub const AWS_SESSION_TOKEN: &str = "aws_session_token"; } const PROVIDER_NAME: &str = "ProfileFile"; fn base_provider(profile: &Profile) -> Result<BaseProvider, ProfileFileError> { // the profile must define either a `CredentialsSource` or a concrete set of access keys match profile.get(role::CREDENTIAL_SOURCE) { Some(source) => Ok(BaseProvider::NamedSource(source)), None => web_identity_token_from_profile(profile) .unwrap_or_else(|| Ok(BaseProvider::AccessKey(static_creds_from_profile(profile)?))), } } enum NextProfile<'a> { SelfReference, Named(&'a str), } fn chain_provider(profile: &Profile) -> Option<Result<(RoleArn, NextProfile), ProfileFileError>> { let role_provider = role_arn_from_profile(&profile)?; let (source_profile, credential_source) = ( profile.get(role::SOURCE_PROFILE), profile.get(role::CREDENTIAL_SOURCE), ); let profile = match (source_profile, credential_source) { (Some(_), Some(_)) => Err(ProfileFileError::InvalidCredentialSource { profile: profile.name().to_string(), message: "profile contained both source_profile and credential_source. \ Only one or the other can be defined" .into(), }), (None, None) => Err(ProfileFileError::InvalidCredentialSource { profile: profile.name().to_string(), message: "profile must contain `source_profile` or `credential_source` but neither were defined" .into(), }), (Some(source_profile), None) if source_profile == profile.name() => { Ok((role_provider, NextProfile::SelfReference)) } (Some(source_profile), None) => Ok((role_provider, NextProfile::Named(source_profile))), // we want to loop back into this profile and pick up the credential source (None, Some(_credential_source)) => Ok((role_provider, NextProfile::SelfReference)), }; Some(profile) } fn role_arn_from_profile(profile: &Profile) -> Option<RoleArn> { // Web Identity Tokens are root providers, not chained roles if profile.get(web_identity_token::TOKEN_FILE).is_some() { return None; } let role_arn = profile.get(role::ROLE_ARN)?; let session_name = profile.get(role::SESSION_NAME); let external_id = profile.get(role::EXTERNAL_ID); Some(RoleArn { role_arn, external_id, session_name, }) } fn web_identity_token_from_profile( profile: &Profile, ) -> Option<Result<BaseProvider, ProfileFileError>>
/// Load static credentials from a profile /// /// Example: /// ```ini /// [profile B] /// aws_access_key_id = abc123 /// aws_secret_access_key = def456 /// ``` fn static_creds_from_profile(profile: &Profile) -> Result<Credentials, ProfileFileError> { use static_credentials::*; let access_key = profile.get(AWS_ACCESS_KEY_ID); let secret_key = profile.get(AWS_SECRET_ACCESS_KEY); let session_token = profile.get(AWS_SESSION_TOKEN); if let (None, None, None) = (access_key, secret_key, session_token) { return Err(ProfileFileError::ProfileDidNotContainCredentials { profile: profile.name().to_string(), }); } let access_key = access_key.ok_or_else(|| ProfileFileError::InvalidCredentialSource { profile: profile.name().to_string(), message: "profile missing aws_access_key_id".into(), })?; let secret_key = secret_key.ok_or_else(|| ProfileFileError::InvalidCredentialSource { profile: profile.name().to_string(), message: "profile missing aws_secret
{ let session_name = profile.get(role::SESSION_NAME); match ( profile.get(role::ROLE_ARN), profile.get(web_identity_token::TOKEN_FILE), ) { (Some(role_arn), Some(token_file)) => Some(Ok(BaseProvider::WebIdentityTokenRole { role_arn, web_identity_token_file: token_file, session_name, })), (None, None) => None, (Some(_role_arn), None) => None, (None, Some(_token_file)) => Some(Err(ProfileFileError::InvalidCredentialSource { profile: profile.name().to_string(), message: "`web_identity_token_file` was specified but `role_arn` was missing".into(), })), } }
identifier_body
repr.rs
} pub fn chain(&self) -> &[RoleArn<'a>] { &self.chain.as_slice() } } /// A base member of the profile chain /// /// Base providers do not require input credentials to provide their own credentials, /// eg. IMDS, ECS, Environment variables #[derive(Debug, Clone)] #[non_exhaustive] pub enum BaseProvider<'a> { /// A profile that specifies a named credential source /// Eg: `credential_source = Ec2InstanceMetadata` /// /// The following profile produces two separate `ProfileProvider` rows: /// 1. `BaseProvider::NamedSource("Ec2InstanceMetadata")` /// 2. `RoleArn { role_arn: "...", ... } /// ```ini /// [profile assume-role] /// role_arn = arn:aws:iam::123456789:role/MyRole /// credential_source = Ec2InstanceMetadata /// ``` NamedSource(&'a str), /// A profile with explicitly configured access keys /// /// Example /// ```ini /// [profile C] /// aws_access_key_id = abc123 /// aws_secret_access_key = def456 /// ``` AccessKey(Credentials), WebIdentityTokenRole { role_arn: &'a str, web_identity_token_file: &'a str, session_name: Option<&'a str>, }, // TODO: add SSO support /* /// An SSO Provider Sso { sso_account_id: &'a str, sso_region: &'a str, sso_role_name: &'a str, sso_start_url: &'a str, }, */ } /// A profile that specifies a role to assume /// /// A RoleArn can only be created from either a profile with `source_profile` /// or one with `credential_source`. #[derive(Debug)] pub struct RoleArn<'a> { /// Role to assume pub role_arn: &'a str, /// external_id parameter to pass to the assume role provider pub external_id: Option<&'a str>, /// session name parameter to pass to the assume role provider pub session_name: Option<&'a str>, } /// Resolve a ProfileChain from a ProfileSet or return an error pub fn resolve_chain<'a>( profile_set: &'a ProfileSet, profile_override: Option<&str>, ) -> Result<ProfileChain<'a>, ProfileFileError> { if profile_set.is_empty() { return Err(ProfileFileError::NoProfilesDefined); } let mut source_profile_name = profile_override.unwrap_or_else(|| profile_set.selected_profile()); let mut visited_profiles = vec![]; let mut chain = vec![]; let base = loop { let profile = profile_set.get_profile(source_profile_name).ok_or( ProfileFileError::MissingProfile { profile: source_profile_name.into(), message: format!( "could not find source profile {} referenced from {}", source_profile_name, visited_profiles.last().unwrap_or(&"the root profile") ) .into(), }, )?; if visited_profiles.contains(&source_profile_name) { return Err(ProfileFileError::CredentialLoop { profiles: visited_profiles .into_iter() .map(|s| s.to_string()) .collect(), next: source_profile_name.to_string(), }); } visited_profiles.push(&source_profile_name); // After the first item in the chain, we will prioritize static credentials if they exist if visited_profiles.len() > 1 { let try_static = static_creds_from_profile(&profile); if let Ok(static_credentials) = try_static { break BaseProvider::AccessKey(static_credentials); } } let next_profile = match chain_provider(&profile) { // this provider wasn't a chain provider, reload it as a base provider None => { break base_provider(profile).map_err(|err| { ProfileFileError::InvalidCredentialSource { profile: profile.name().into(), message: format!("could not load source profile: {}", err).into(), } })?; } Some(result) => { let (chain_profile, next) = result?; chain.push(chain_profile); next } }; match next_profile { NextProfile::SelfReference => { // self referential profile, don't go through the loop because it will error // on the infinite loop check. Instead, reload this profile as a base profile // and exit. break base_provider(profile)?; } NextProfile::Named(name) => source_profile_name = name, } }; chain.reverse(); Ok(ProfileChain { base, chain }) } mod role { pub const ROLE_ARN: &str = "role_arn"; pub const EXTERNAL_ID: &str = "external_id"; pub const SESSION_NAME: &str = "role_session_name"; pub const CREDENTIAL_SOURCE: &str = "credential_source"; pub const SOURCE_PROFILE: &str = "source_profile"; } mod web_identity_token { pub const TOKEN_FILE: &str = "web_identity_token_file"; } mod static_credentials { pub const AWS_ACCESS_KEY_ID: &str = "aws_access_key_id"; pub const AWS_SECRET_ACCESS_KEY: &str = "aws_secret_access_key"; pub const AWS_SESSION_TOKEN: &str = "aws_session_token"; } const PROVIDER_NAME: &str = "ProfileFile"; fn base_provider(profile: &Profile) -> Result<BaseProvider, ProfileFileError> { // the profile must define either a `CredentialsSource` or a concrete set of access keys match profile.get(role::CREDENTIAL_SOURCE) { Some(source) => Ok(BaseProvider::NamedSource(source)), None => web_identity_token_from_profile(profile) .unwrap_or_else(|| Ok(BaseProvider::AccessKey(static_creds_from_profile(profile)?))), } } enum NextProfile<'a> { SelfReference, Named(&'a str), } fn chain_provider(profile: &Profile) -> Option<Result<(RoleArn, NextProfile), ProfileFileError>> { let role_provider = role_arn_from_profile(&profile)?; let (source_profile, credential_source) = ( profile.get(role::SOURCE_PROFILE), profile.get(role::CREDENTIAL_SOURCE), ); let profile = match (source_profile, credential_source) { (Some(_), Some(_)) => Err(ProfileFileError::InvalidCredentialSource { profile: profile.name().to_string(), message: "profile contained both source_profile and credential_source. \ Only one or the other can be defined" .into(), }), (None, None) => Err(ProfileFileError::InvalidCredentialSource { profile: profile.name().to_string(), message: "profile must contain `source_profile` or `credential_source` but neither were defined" .into(), }), (Some(source_profile), None) if source_profile == profile.name() => { Ok((role_provider, NextProfile::SelfReference)) } (Some(source_profile), None) => Ok((role_provider, NextProfile::Named(source_profile))), // we want to loop back into this profile and pick up the credential source (None, Some(_credential_source)) => Ok((role_provider, NextProfile::SelfReference)), }; Some(profile) } fn role_arn_from_profile(profile: &Profile) -> Option<RoleArn> { // Web Identity Tokens are root providers, not chained roles if profile.get(web_identity_token::TOKEN_FILE).is_some() { return None; } let role_arn = profile.get(role::ROLE_ARN)?; let session_name = profile.get(role::SESSION_NAME); let external_id = profile.get(role::EXTERNAL_ID); Some(RoleArn { role_arn, external_id, session_name, }) } fn web_identity_token_from_profile( profile: &Profile, ) -> Option<Result<BaseProvider, ProfileFileError>> { let session_name = profile.get(role::SESSION_NAME); match ( profile.get(role::ROLE_ARN), profile.get(web_identity_token::TOKEN_FILE), ) { (Some(role_arn), Some(token_file)) => Some(Ok(BaseProvider::WebIdentityTokenRole { role_arn, web_identity_token_file: token_file, session_name, })), (None, None) => None, (Some(_role_arn), None) => None, (None, Some(_token_file)) => Some(Err(ProfileFileError::InvalidCredentialSource { profile: profile.name().to_string(), message: "`web_identity_token_file` was specified but `role_arn` was missing".into(), })), } } /// Load static credentials from a profile /// /// Example: /// ```ini /// [profile B] /// aws_access_key_id = abc123 /// aws_secret_access_key = def456 /// ``` fn static_creds_from_profile(profile: &Profile) -> Result<Credentials, ProfileFileError> { use static_credentials::*; let access_key = profile.get(AWS_ACCESS_KEY_ID); let secret_key = profile.get(AWS_SECRET_ACCESS_KEY); let session_token = profile.get(AWS_SESSION_TOKEN); if let (None, None, None) = (access_key, secret_key, session_token) { return Err(ProfileFileError::ProfileDidNotContainCredentials { profile: profile.name().to_string(), }); } let access_key = access
random_line_split
neural_network.py
hat = logistic(self.W @ Z).T return Z, Yhat def loss(self, X, V, W, Y): Z, Yhat = self.forward(X, V, W) log_Yhat = log(Yhat) log_Yhat_inv = log(1 - Yhat) log_Yhat[Y == 0] = 0 log_Yhat_inv[Y == 1] = 0 if not (np.isfinite(log_Yhat).all() and np.isfinite(log_Yhat_inv).all()): stderr.write('parameters incompatible with data ' '(log() arising in loss calculation).\n') stderr.flush() log_Yhat[~np.isfinite(log_Yhat)] = log(EPSILON) log_Yhat_inv[~np.isfinite(log_Yhat_inv)] = log(EPSILON) return -(Y * log_Yhat + (1 - Y) * log_Yhat_inv).sum() def fit(self, X, y): """ \grad_{W_k} L = \partiald{L}{\yhat_k} \grad_{W_k} \yhat_k \partiald{L}{\yhat_k} = \frac{y_k - \yhat_k}{\yhat_k (1 - \yhat_k)} \grad_{W_k} \yhat_k = z \yhat_k (1 - \yhat_k) \grad_z L = \sum_k \partiald{L}{\yhat_k} \grad_z \yhat_k \grad_z \yhat_k = W_k \yhat_k (1 - \yhat_k) \grad_{V_h} L = \partiald{L}{z_h} \grad_{V_h} z_h \grad_{V_h} z_h = x(1 - z_h^2) """ assert self.stop_factor or self.n_iterations is not None X, Y = self.prepare_data(X, y) H = self.H K = self.K n, d = X.shape # X has extra offset dimension containing all 1s # The hidden layer z also has a unit whose value is always 1 d -= 1 if self.V is None: self.V = random_normal(0, 0.1, (H + 1, d + 1)) if self.W is None: self.W = random_normal(0, 0.1, (K, H + 1)) V, W = self.V, self.W sample_indices = list(range(n)) shuffle(sample_indices) if self.parallel: pool = Pool() starmap = pool.starmap else: starmap = itertools.starmap it = -self.batch_size while True: it += self.batch_size if it >= self.n_iterations: break if it % 10000 == 0: print('%6d/%-6d %.3f' % (it, self.n_iterations, self.loss(X, V, W, Y))) def args(it): i = sample_indices[it % n] return (X[[i], :], V, W, Y[i, :]) gradients = starmap( self.gradient, map(args, range(it, it + self.batch_size)) ) grad__L__V, grad__L__W = [ reduce(np.add, grads) / len(grads) for grads in zip(*gradients) ] W -= self.learning_rate * grad__L__W V -= self.learning_rate * grad__L__V return self def gradient(self, X, V, W, Y): """ Compute gradient of loss with respect to V and W. """ one, d_plus_one = X.shape K, H_plus_one = W.shape d = d_plus_one - 1 H = H_plus_one - 1 Z, Yhat = self.forward(X, V, W) assert one == 1 x = X y = Y z = Z.ravel() yhat = Yhat.ravel() # Update W # grad__L__yhat = (yhat - y) / np.clip(yhat * (1 - yhat), EPSILON, inf) # grad__L__z[:] = 0.0 # for k in range(K): # grad__yhat_k__W_k = z * yhat[k] * (1 - yhat[k]) # # Last element corresponds to constant offset 1 appended to z # # vector; it does not change / has no derivative. # grad__yhat_k__z = W[k, :-1] * yhat[k] * (1 - yhat[k]) # grad__L__z += grad__L__yhat[k] * grad__yhat_k__z # W[k, :] -= self.learning_rate * grad__L__yhat[k] * grad__yhat_k__W_k grad__L__z = (W.T * (yhat - y)).sum(axis=1) zz = z.reshape((1, H + 1)).repeat(K, 0) grad__L__W = diag(yhat - y) @ zz # Update V # for h in range(H): # grad__z_h__V_h = x * (1 - z[h] ** 2) # grad__L__V_h = grad__L__z[h] * grad__z_h__V_h # V[h, :] -= self.learning_rate * grad__L__V_h xx = x.reshape((1, d + 1)).repeat(H + 1, 0) grad__L__V = diag((1 - z ** 2) * grad__L__z) @ xx return grad__L__V, grad__L__W def estimate_grad__z_h__V_h(self, h, x, V, grad): eps = EPSILON_FINITE_DIFFERENCE def d(j): eps_vec = np.zeros_like(V) eps_vec[h, j] = eps z_plus = tanh((V + eps_vec) @ x) z_minus = tanh((V - eps_vec) @ x) z_plus[-1] = 1 z_minus[-1] = 1 return (z_plus[h] - z_minus[h]) / (2 * eps) return self._do_finite_difference_estimate( d, V[h, :], 'grad__z[%d]__V[%d,:]' % (h, h), grad, ) def estimate_grad__yhat_k__z(self, k, z, W, y, grad): eps = EPSILON_FINITE_DIFFERENCE def d(h): eps_vec = np.zeros_like(z) eps_vec[h] = eps yhat_plus = logistic(W @ (z + eps_vec)) yhat_minus = logistic(W @ (z - eps_vec)) return (yhat_plus[k] - yhat_minus[k]) / (2 * eps) return self._do_finite_difference_estimate( d, z, 'grad__yhat[%d]__z' % k, grad, ) def estimate_grad__yhat_k__W_k(self, k, z, W, y, grad): eps = EPSILON_FINITE_DIFFERENCE def d(h): eps_vec = np.zeros_like(W) eps_vec[k, h] = eps yhat_plus = logistic((W + eps_vec) @ z) yhat_minus = logistic((W - eps_vec) @ z) return (yhat_plus[k] - yhat_minus[k]) / (2 * eps) return self._do_finite_difference_estimate( d, W[k, :], 'grad__yhat[%d]__W[%d,:]' % (k, k), grad, ) def estimate_grad__L__yhat(self, yhat, y, grad): eps = EPSILON_FINITE_DIFFERENCE def d(k): eps_vec = np.zeros_like(yhat) eps_vec[k] = eps L_plus = self.loss(yhat + eps_vec, y) L_minus = self.loss(yhat - eps_vec, y) return (L_plus - L_minus) / (2 * eps) return self._do_finite_difference_estimate( d, yhat, 'grad__L__yhat', grad, ) def estimate_grad__L__z(self, z, W, y, grad):
eps = EPSILON_FINITE_DIFFERENCE def d(h): eps_vec = np.zeros_like(z) eps_vec[h] = eps yhat_plus = logistic(W @ (z + eps_vec)) yhat_minus = logistic(W @ (z - eps_vec)) L_plus = self.loss(yhat_plus, y) L_minus = self.loss(yhat_minus, y) return (L_plus - L_minus) / (2 * eps) return self._do_finite_difference_estimate( d, z, 'grad__L__z', grad, )
identifier_body
neural_network.py
x K). We use stochastic gradient descent, i.e. compute and update gradients for a single input row at a time, so in backpropagation we work with x (d x 1) and y (K x 1). | Input | x | d x 1 | | First weight matrix | V | H x d | | Hidden layer | Z = tanh(Vx) | H x 1 | | Second weight matrix | W | K x H | | Output | yhat | K x 1 | | Loss | L | scalar | The loss function is the cross-entropy -sum_k { y_k log(yhat_k) + (1 - y_k) log(1 - yhat_k) } """ def __init__(self, n_hidden_units, learning_rate, n_iterations=None, batch_size=1, parallel=False, stop_factor=None, stop_window_size=None, outfile=None): if stop_window_size: assert DEBUG self.H = n_hidden_units self.K = None # Determined empirically as distinct training labels self.V = None self.W = None self.learning_rate = learning_rate self.n_iterations = n_iterations self.batch_size = batch_size self.parallel = parallel self.stop_factor = stop_factor self.stop_window_size = stop_window_size self.outfile = outfile def predict(self, X): X = self.prepare_data(X) Z, Yhat = self.forward(X, self.V, self.W) return Yhat def forward(self, X, V, W): Z = tanh(V @ X.T) Z[-1, :] = 1 # The last row of V is unused; z[-1] must always be 1, just as x[-1]. Yhat = logistic(self.W @ Z).T return Z, Yhat def loss(self, X, V, W, Y): Z, Yhat = self.forward(X, V, W) log_Yhat = log(Yhat) log_Yhat_inv = log(1 - Yhat) log_Yhat[Y == 0] = 0 log_Yhat_inv[Y == 1] = 0 if not (np.isfinite(log_Yhat).all() and np.isfinite(log_Yhat_inv).all()): stderr.write('parameters incompatible with data ' '(log() arising in loss calculation).\n') stderr.flush() log_Yhat[~np.isfinite(log_Yhat)] = log(EPSILON) log_Yhat_inv[~np.isfinite(log_Yhat_inv)] = log(EPSILON) return -(Y * log_Yhat + (1 - Y) * log_Yhat_inv).sum() def fit(self, X, y): """ \grad_{W_k} L = \partiald{L}{\yhat_k} \grad_{W_k} \yhat_k \partiald{L}{\yhat_k} = \frac{y_k - \yhat_k}{\yhat_k (1 - \yhat_k)} \grad_{W_k} \yhat_k = z \yhat_k (1 - \yhat_k) \grad_z L = \sum_k \partiald{L}{\yhat_k} \grad_z \yhat_k \grad_z \yhat_k = W_k \yhat_k (1 - \yhat_k) \grad_{V_h} L = \partiald{L}{z_h} \grad_{V_h} z_h \grad_{V_h} z_h = x(1 - z_h^2) """ assert self.stop_factor or self.n_iterations is not None X, Y = self.prepare_data(X, y) H = self.H K = self.K n, d = X.shape # X has extra offset dimension containing all 1s # The hidden layer z also has a unit whose value is always 1 d -= 1 if self.V is None: self.V = random_normal(0, 0.1, (H + 1, d + 1)) if self.W is None: self.W = random_normal(0, 0.1, (K, H + 1)) V, W = self.V, self.W sample_indices = list(range(n)) shuffle(sample_indices) if self.parallel: pool = Pool() starmap = pool.starmap else: starmap = itertools.starmap it = -self.batch_size while True: it += self.batch_size if it >= self.n_iterations: break if it % 10000 == 0: print('%6d/%-6d %.3f' % (it, self.n_iterations, self.loss(X, V, W, Y))) def args(it): i = sample_indices[it % n] return (X[[i], :], V, W, Y[i, :]) gradients = starmap( self.gradient, map(args, range(it, it + self.batch_size)) ) grad__L__V, grad__L__W = [ reduce(np.add, grads) / len(grads) for grads in zip(*gradients) ] W -= self.learning_rate * grad__L__W V -= self.learning_rate * grad__L__V return self def gradient(self, X, V, W, Y): """ Compute gradient of loss with respect to V and W. """ one, d_plus_one = X.shape K, H_plus_one = W.shape d = d_plus_one - 1 H = H_plus_one - 1 Z, Yhat = self.forward(X, V, W) assert one == 1 x = X y = Y z = Z.ravel() yhat = Yhat.ravel() # Update W # grad__L__yhat = (yhat - y) / np.clip(yhat * (1 - yhat), EPSILON, inf) # grad__L__z[:] = 0.0 # for k in range(K): # grad__yhat_k__W_k = z * yhat[k] * (1 - yhat[k]) # # Last element corresponds to constant offset 1 appended to z # # vector; it does not change / has no derivative. # grad__yhat_k__z = W[k, :-1] * yhat[k] * (1 - yhat[k]) # grad__L__z += grad__L__yhat[k] * grad__yhat_k__z # W[k, :] -= self.learning_rate * grad__L__yhat[k] * grad__yhat_k__W_k grad__L__z = (W.T * (yhat - y)).sum(axis=1) zz = z.reshape((1, H + 1)).repeat(K, 0) grad__L__W = diag(yhat - y) @ zz # Update V # for h in range(H):
# grad__z_h__V_h = x * (1 - z[h] ** 2) # grad__L__V_h = grad__L__z[h] * grad__z_h__V_h # V[h, :] -= self.learning_rate * grad__L__V_h xx = x.reshape((1, d + 1)).repeat(H + 1, 0) grad__L__V = diag((1 - z ** 2) * grad__L__z) @ xx return grad__L__V, grad__L__W def estimate_grad__z_h__V_h(self, h, x, V, grad): eps = EPSILON_FINITE_DIFFERENCE def d(j): eps_vec = np.zeros_like(V) eps_vec[h, j] = eps z_plus = tanh((V + eps_vec) @ x) z_minus = tanh((V - eps_vec) @ x) z_plus[-1] = 1 z_minus[-1] = 1 return (z_plus[h] - z_minus[h]) / (2 * eps) return self._do_finite_difference_estimate( d, V[h, :], 'grad__z[%d]__V[%d,:]' % (h, h), grad, ) def estimate_grad__yhat_k__z(self, k, z, W, y, grad): eps = EPSILON_FINITE_DIFFERENCE def d(h): eps_vec = np.zeros_like(z) eps_vec[h] = eps yhat_plus = logistic(W @ (z + eps_vec)) yhat_minus = logistic(W @ (z - eps_vec)) return (yhat_plus[k] - yhat_minus[k]) / (2 * eps) return self._do_finite_difference_estimate( d, z, 'grad__yhat[%d]__z' % k, grad, ) def
random_line_split
neural_network.py
y) H = self.H K = self.K n, d = X.shape # X has extra offset dimension containing all 1s # The hidden layer z also has a unit whose value is always 1 d -= 1 if self.V is None: self.V = random_normal(0, 0.1, (H + 1, d + 1)) if self.W is None: self.W = random_normal(0, 0.1, (K, H + 1)) V, W = self.V, self.W sample_indices = list(range(n)) shuffle(sample_indices) if self.parallel: pool = Pool() starmap = pool.starmap else: starmap = itertools.starmap it = -self.batch_size while True: it += self.batch_size if it >= self.n_iterations: break if it % 10000 == 0: print('%6d/%-6d %.3f' % (it, self.n_iterations, self.loss(X, V, W, Y))) def args(it): i = sample_indices[it % n] return (X[[i], :], V, W, Y[i, :]) gradients = starmap( self.gradient, map(args, range(it, it + self.batch_size)) ) grad__L__V, grad__L__W = [ reduce(np.add, grads) / len(grads) for grads in zip(*gradients) ] W -= self.learning_rate * grad__L__W V -= self.learning_rate * grad__L__V return self def gradient(self, X, V, W, Y): """ Compute gradient of loss with respect to V and W. """ one, d_plus_one = X.shape K, H_plus_one = W.shape d = d_plus_one - 1 H = H_plus_one - 1 Z, Yhat = self.forward(X, V, W) assert one == 1 x = X y = Y z = Z.ravel() yhat = Yhat.ravel() # Update W # grad__L__yhat = (yhat - y) / np.clip(yhat * (1 - yhat), EPSILON, inf) # grad__L__z[:] = 0.0 # for k in range(K): # grad__yhat_k__W_k = z * yhat[k] * (1 - yhat[k]) # # Last element corresponds to constant offset 1 appended to z # # vector; it does not change / has no derivative. # grad__yhat_k__z = W[k, :-1] * yhat[k] * (1 - yhat[k]) # grad__L__z += grad__L__yhat[k] * grad__yhat_k__z # W[k, :] -= self.learning_rate * grad__L__yhat[k] * grad__yhat_k__W_k grad__L__z = (W.T * (yhat - y)).sum(axis=1) zz = z.reshape((1, H + 1)).repeat(K, 0) grad__L__W = diag(yhat - y) @ zz # Update V # for h in range(H): # grad__z_h__V_h = x * (1 - z[h] ** 2) # grad__L__V_h = grad__L__z[h] * grad__z_h__V_h # V[h, :] -= self.learning_rate * grad__L__V_h xx = x.reshape((1, d + 1)).repeat(H + 1, 0) grad__L__V = diag((1 - z ** 2) * grad__L__z) @ xx return grad__L__V, grad__L__W def estimate_grad__z_h__V_h(self, h, x, V, grad): eps = EPSILON_FINITE_DIFFERENCE def d(j): eps_vec = np.zeros_like(V) eps_vec[h, j] = eps z_plus = tanh((V + eps_vec) @ x) z_minus = tanh((V - eps_vec) @ x) z_plus[-1] = 1 z_minus[-1] = 1 return (z_plus[h] - z_minus[h]) / (2 * eps) return self._do_finite_difference_estimate( d, V[h, :], 'grad__z[%d]__V[%d,:]' % (h, h), grad, ) def estimate_grad__yhat_k__z(self, k, z, W, y, grad): eps = EPSILON_FINITE_DIFFERENCE def d(h): eps_vec = np.zeros_like(z) eps_vec[h] = eps yhat_plus = logistic(W @ (z + eps_vec)) yhat_minus = logistic(W @ (z - eps_vec)) return (yhat_plus[k] - yhat_minus[k]) / (2 * eps) return self._do_finite_difference_estimate( d, z, 'grad__yhat[%d]__z' % k, grad, ) def estimate_grad__yhat_k__W_k(self, k, z, W, y, grad): eps = EPSILON_FINITE_DIFFERENCE def d(h): eps_vec = np.zeros_like(W) eps_vec[k, h] = eps yhat_plus = logistic((W + eps_vec) @ z) yhat_minus = logistic((W - eps_vec) @ z) return (yhat_plus[k] - yhat_minus[k]) / (2 * eps) return self._do_finite_difference_estimate( d, W[k, :], 'grad__yhat[%d]__W[%d,:]' % (k, k), grad, ) def estimate_grad__L__yhat(self, yhat, y, grad): eps = EPSILON_FINITE_DIFFERENCE def d(k): eps_vec = np.zeros_like(yhat) eps_vec[k] = eps L_plus = self.loss(yhat + eps_vec, y) L_minus = self.loss(yhat - eps_vec, y) return (L_plus - L_minus) / (2 * eps) return self._do_finite_difference_estimate( d, yhat, 'grad__L__yhat', grad, ) def estimate_grad__L__z(self, z, W, y, grad): eps = EPSILON_FINITE_DIFFERENCE def d(h): eps_vec = np.zeros_like(z) eps_vec[h] = eps yhat_plus = logistic(W @ (z + eps_vec)) yhat_minus = logistic(W @ (z - eps_vec)) L_plus = self.loss(yhat_plus, y) L_minus = self.loss(yhat_minus, y) return (L_plus - L_minus) / (2 * eps) return self._do_finite_difference_estimate( d, z, 'grad__L__z', grad, ) def estimate_grad__L__V_h(self, h, x, V, W, y, grad): eps = EPSILON_FINITE_DIFFERENCE def d(j): eps_vec = np.zeros_like(V) eps_vec[h, j] = eps z_plus = tanh((V + eps_vec) @ x) z_minus = tanh((V - eps_vec) @ x) z_plus[-1] = 1 z_minus[-1] = 1 yhat_plus = logistic(W @ z_plus) yhat_minus = logistic(W @ z_minus) L_plus = self.loss(yhat_plus, y) L_minus = self.loss(yhat_minus, y) return (L_plus - L_minus) / (2 * eps) return self._do_finite_difference_estimate( d, V[h, :], 'grad__L__V_h', grad, ) @staticmethod def _do_finite_difference_estimate(d, wrt, label, grad): grad__n = np.array(list(map(d, range(len(wrt))))) if DEBUG: col = get_colour(re.subn(r'\d+', '%d', label)) print(col('%s = %s' % (label, grad__n))) print(col(', '.join('%.9f' % g for g in describe(grad__n - grad).minmax))) return grad__n def prepare_data(self, X, y=None): n, d = X.shape X = np.hstack([X, np.ones((n, 1))]) if y is None: return X nY, = y.shape assert nY == n K = len(set(y)) # Demand that labels are integers 1...max(y) if not np.issubdtype(y.dtype, np.int):
y_int = np.floor(y).astype(np.int) assert (y_int == y).all() y = y_int
conditional_block
neural_network.py
(self, yhat): """ Map values in output layer to classification predictions. """ raise NotImplementedError class SingleLayerTanhLogisticNeuralNetwork(NeuralNetwork): """ A classification neural net with one hidden layer. The hidden layer uses the tanh activation function. The output layer uses the logistic activation function. Model: The input data are X (n x d) and Y (n x K). We use stochastic gradient descent, i.e. compute and update gradients for a single input row at a time, so in backpropagation we work with x (d x 1) and y (K x 1). | Input | x | d x 1 | | First weight matrix | V | H x d | | Hidden layer | Z = tanh(Vx) | H x 1 | | Second weight matrix | W | K x H | | Output | yhat | K x 1 | | Loss | L | scalar | The loss function is the cross-entropy -sum_k { y_k log(yhat_k) + (1 - y_k) log(1 - yhat_k) } """ def __init__(self, n_hidden_units, learning_rate, n_iterations=None, batch_size=1, parallel=False, stop_factor=None, stop_window_size=None, outfile=None): if stop_window_size: assert DEBUG self.H = n_hidden_units self.K = None # Determined empirically as distinct training labels self.V = None self.W = None self.learning_rate = learning_rate self.n_iterations = n_iterations self.batch_size = batch_size self.parallel = parallel self.stop_factor = stop_factor self.stop_window_size = stop_window_size self.outfile = outfile def predict(self, X): X = self.prepare_data(X) Z, Yhat = self.forward(X, self.V, self.W) return Yhat def forward(self, X, V, W): Z = tanh(V @ X.T) Z[-1, :] = 1 # The last row of V is unused; z[-1] must always be 1, just as x[-1]. Yhat = logistic(self.W @ Z).T return Z, Yhat def loss(self, X, V, W, Y): Z, Yhat = self.forward(X, V, W) log_Yhat = log(Yhat) log_Yhat_inv = log(1 - Yhat) log_Yhat[Y == 0] = 0 log_Yhat_inv[Y == 1] = 0 if not (np.isfinite(log_Yhat).all() and np.isfinite(log_Yhat_inv).all()): stderr.write('parameters incompatible with data ' '(log() arising in loss calculation).\n') stderr.flush() log_Yhat[~np.isfinite(log_Yhat)] = log(EPSILON) log_Yhat_inv[~np.isfinite(log_Yhat_inv)] = log(EPSILON) return -(Y * log_Yhat + (1 - Y) * log_Yhat_inv).sum() def fit(self, X, y): """ \grad_{W_k} L = \partiald{L}{\yhat_k} \grad_{W_k} \yhat_k \partiald{L}{\yhat_k} = \frac{y_k - \yhat_k}{\yhat_k (1 - \yhat_k)} \grad_{W_k} \yhat_k = z \yhat_k (1 - \yhat_k) \grad_z L = \sum_k \partiald{L}{\yhat_k} \grad_z \yhat_k \grad_z \yhat_k = W_k \yhat_k (1 - \yhat_k) \grad_{V_h} L = \partiald{L}{z_h} \grad_{V_h} z_h \grad_{V_h} z_h = x(1 - z_h^2) """ assert self.stop_factor or self.n_iterations is not None X, Y = self.prepare_data(X, y) H = self.H K = self.K n, d = X.shape # X has extra offset dimension containing all 1s # The hidden layer z also has a unit whose value is always 1 d -= 1 if self.V is None: self.V = random_normal(0, 0.1, (H + 1, d + 1)) if self.W is None: self.W = random_normal(0, 0.1, (K, H + 1)) V, W = self.V, self.W sample_indices = list(range(n)) shuffle(sample_indices) if self.parallel: pool = Pool() starmap = pool.starmap else: starmap = itertools.starmap it = -self.batch_size while True: it += self.batch_size if it >= self.n_iterations: break if it % 10000 == 0: print('%6d/%-6d %.3f' % (it, self.n_iterations, self.loss(X, V, W, Y))) def args(it): i = sample_indices[it % n] return (X[[i], :], V, W, Y[i, :]) gradients = starmap( self.gradient, map(args, range(it, it + self.batch_size)) ) grad__L__V, grad__L__W = [ reduce(np.add, grads) / len(grads) for grads in zip(*gradients) ] W -= self.learning_rate * grad__L__W V -= self.learning_rate * grad__L__V return self def gradient(self, X, V, W, Y): """ Compute gradient of loss with respect to V and W. """ one, d_plus_one = X.shape K, H_plus_one = W.shape d = d_plus_one - 1 H = H_plus_one - 1 Z, Yhat = self.forward(X, V, W) assert one == 1 x = X y = Y z = Z.ravel() yhat = Yhat.ravel() # Update W # grad__L__yhat = (yhat - y) / np.clip(yhat * (1 - yhat), EPSILON, inf) # grad__L__z[:] = 0.0 # for k in range(K): # grad__yhat_k__W_k = z * yhat[k] * (1 - yhat[k]) # # Last element corresponds to constant offset 1 appended to z # # vector; it does not change / has no derivative. # grad__yhat_k__z = W[k, :-1] * yhat[k] * (1 - yhat[k]) # grad__L__z += grad__L__yhat[k] * grad__yhat_k__z # W[k, :] -= self.learning_rate * grad__L__yhat[k] * grad__yhat_k__W_k grad__L__z = (W.T * (yhat - y)).sum(axis=1) zz = z.reshape((1, H + 1)).repeat(K, 0) grad__L__W = diag(yhat - y) @ zz # Update V # for h in range(H): # grad__z_h__V_h = x * (1 - z[h] ** 2) # grad__L__V_h = grad__L__z[h] * grad__z_h__V_h # V[h, :] -= self.learning_rate * grad__L__V_h xx = x.reshape((1, d + 1)).repeat(H + 1, 0) grad__L__V = diag((1 - z ** 2) * grad__L__z) @ xx return grad__L__V, grad__L__W def estimate_grad__z_h__V_h(self, h, x, V, grad): eps = EPSILON_FINITE_DIFFERENCE def d(j): eps_vec = np.zeros_like(V) eps_vec[h, j] = eps z_plus = tanh((V + eps_vec) @ x) z_minus = tanh((V - eps_vec) @ x) z_plus[-1] = 1 z_minus[-1] = 1 return (z_plus[h] - z_minus[h]) / (2 * eps) return self._do_finite_difference_estimate( d, V[h, :], 'grad__z[%d]__V[%d,:]' % (h, h), grad, ) def estimate_grad__yhat_k__z(self, k, z, W, y, grad): eps = EPSILON_FINITE_DIFFERENCE def d(h): eps_vec = np.zeros_like(z) eps_vec
prediction_fn
identifier_name