text1
stringlengths 0
536k
| text2
stringlengths 0
536k
| label
int64 0
1
|
---|---|---|
<h2 dir="auto"><g-emoji class="g-emoji" alias="rocket" fallback-src="https://github.githubassets.com/images/icons/emoji/unicode/1f680.png">🚀</g-emoji> Feature</h2>
<p dir="auto">All variants of <code class="notranslate">ReplicationPad</code> should accept 0-dim batch size tensors for both forward and backward autograd. Related to <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="614589851" data-permission-text="Title is private" data-url="https://github.com/pytorch/pytorch/issues/38115" data-hovercard-type="issue" data-hovercard-url="/pytorch/pytorch/issues/38115/hovercard" href="https://github.com/pytorch/pytorch/issues/38115">#38115</a></p>
<p dir="auto">Test code:</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="def test_ReplicationPad1d_Forward():
z = torch.ones(0,3,10)
m = torch.nn.ReplicationPad1d(3)
m(z)
def test_ReplicationPad1d_Backward():
z = torch.ones(0,3,10)
m = torch.nn.ReplicationPad1d(3)
m(z).sum().backward()
def test_ReplicationPad2d_Forward():
z = torch.ones(0,3,10,10)
m = torch.nn.ReplicationPad2d(3)
m(z)
def test_ReplicationPad2d_Backward():
z = torch.ones(0,3,10,10)
m = torch.nn.ReplicationPad2d(3)
m(z).sum().backward()
def test_ReplicationPad3d_Forward():
z = torch.ones(0,3,10,10,10)
m = torch.nn.ReplicationPad3d(3)
m(z)
def test_ReplicationPad3d_Backward():
z = torch.ones(0,3,10,10,10)
m = torch.nn.ReplicationPad3d(3)
m(z).sum().backward()"><pre class="notranslate"><span class="pl-k">def</span> <span class="pl-en">test_ReplicationPad1d_Forward</span>():
<span class="pl-s1">z</span> <span class="pl-c1">=</span> <span class="pl-s1">torch</span>.<span class="pl-en">ones</span>(<span class="pl-c1">0</span>,<span class="pl-c1">3</span>,<span class="pl-c1">10</span>)
<span class="pl-s1">m</span> <span class="pl-c1">=</span> <span class="pl-s1">torch</span>.<span class="pl-s1">nn</span>.<span class="pl-v">ReplicationPad1d</span>(<span class="pl-c1">3</span>)
<span class="pl-en">m</span>(<span class="pl-s1">z</span>)
<span class="pl-k">def</span> <span class="pl-en">test_ReplicationPad1d_Backward</span>():
<span class="pl-s1">z</span> <span class="pl-c1">=</span> <span class="pl-s1">torch</span>.<span class="pl-en">ones</span>(<span class="pl-c1">0</span>,<span class="pl-c1">3</span>,<span class="pl-c1">10</span>)
<span class="pl-s1">m</span> <span class="pl-c1">=</span> <span class="pl-s1">torch</span>.<span class="pl-s1">nn</span>.<span class="pl-v">ReplicationPad1d</span>(<span class="pl-c1">3</span>)
<span class="pl-en">m</span>(<span class="pl-s1">z</span>).<span class="pl-en">sum</span>().<span class="pl-en">backward</span>()
<span class="pl-k">def</span> <span class="pl-en">test_ReplicationPad2d_Forward</span>():
<span class="pl-s1">z</span> <span class="pl-c1">=</span> <span class="pl-s1">torch</span>.<span class="pl-en">ones</span>(<span class="pl-c1">0</span>,<span class="pl-c1">3</span>,<span class="pl-c1">10</span>,<span class="pl-c1">10</span>)
<span class="pl-s1">m</span> <span class="pl-c1">=</span> <span class="pl-s1">torch</span>.<span class="pl-s1">nn</span>.<span class="pl-v">ReplicationPad2d</span>(<span class="pl-c1">3</span>)
<span class="pl-en">m</span>(<span class="pl-s1">z</span>)
<span class="pl-k">def</span> <span class="pl-en">test_ReplicationPad2d_Backward</span>():
<span class="pl-s1">z</span> <span class="pl-c1">=</span> <span class="pl-s1">torch</span>.<span class="pl-en">ones</span>(<span class="pl-c1">0</span>,<span class="pl-c1">3</span>,<span class="pl-c1">10</span>,<span class="pl-c1">10</span>)
<span class="pl-s1">m</span> <span class="pl-c1">=</span> <span class="pl-s1">torch</span>.<span class="pl-s1">nn</span>.<span class="pl-v">ReplicationPad2d</span>(<span class="pl-c1">3</span>)
<span class="pl-en">m</span>(<span class="pl-s1">z</span>).<span class="pl-en">sum</span>().<span class="pl-en">backward</span>()
<span class="pl-k">def</span> <span class="pl-en">test_ReplicationPad3d_Forward</span>():
<span class="pl-s1">z</span> <span class="pl-c1">=</span> <span class="pl-s1">torch</span>.<span class="pl-en">ones</span>(<span class="pl-c1">0</span>,<span class="pl-c1">3</span>,<span class="pl-c1">10</span>,<span class="pl-c1">10</span>,<span class="pl-c1">10</span>)
<span class="pl-s1">m</span> <span class="pl-c1">=</span> <span class="pl-s1">torch</span>.<span class="pl-s1">nn</span>.<span class="pl-v">ReplicationPad3d</span>(<span class="pl-c1">3</span>)
<span class="pl-en">m</span>(<span class="pl-s1">z</span>)
<span class="pl-k">def</span> <span class="pl-en">test_ReplicationPad3d_Backward</span>():
<span class="pl-s1">z</span> <span class="pl-c1">=</span> <span class="pl-s1">torch</span>.<span class="pl-en">ones</span>(<span class="pl-c1">0</span>,<span class="pl-c1">3</span>,<span class="pl-c1">10</span>,<span class="pl-c1">10</span>,<span class="pl-c1">10</span>)
<span class="pl-s1">m</span> <span class="pl-c1">=</span> <span class="pl-s1">torch</span>.<span class="pl-s1">nn</span>.<span class="pl-v">ReplicationPad3d</span>(<span class="pl-c1">3</span>)
<span class="pl-en">m</span>(<span class="pl-s1">z</span>).<span class="pl-en">sum</span>().<span class="pl-en">backward</span>()</pre></div> | <p dir="auto">I've found out that gradient of <code class="notranslate">index_select</code> (i..e <code class="notranslate">index_add</code>) behaves non-deterministically - and wrong - on GPU for some rather random conditions. The following code illustrates the problem:</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="import numpy as np
import torch
from torch.autograd import Variable
def set_seed(seed):
np.random.seed(seed)
torch.manual_seed(seed)
torch.cuda.manual_seed(seed)
def main():
set_seed(1)
cuda = True
n, e, c = 5, 100, 5
input = torch.randn(n,c)
idxn = torch.from_numpy(np.random.randint(n,size=e)) # indices are repeated
if cuda:
input = input.cuda(); idxn = idxn.cuda()
gradsI, gradsS = [], []
N = 4
for i in range(N):
inputv = Variable(input, requires_grad=True)
sel_input = torch.index_select(inputv, 0, Variable(idxn))
sel_input.retain_grad()
# the following computation is one of the "random conditions"
data = [torch.sum(sel_input.narrow(0,0,e//2), 0),
torch.sum(sel_input.narrow(0,e//2,e//2), 0) ]
out = torch.cat(data,0)
out.exp().sum().backward()
gradsI.append(inputv.grad.data.cpu().clone())
gradsS.append(sel_input.grad.data.cpu().clone())
for i in range(N):
for j in range(N):
print(i,j,(gradsI[i]-gradsI[j]).abs().max(), (gradsS[i]-gradsS[j]).abs().max())"><pre class="notranslate"><span class="pl-k">import</span> <span class="pl-s1">numpy</span> <span class="pl-k">as</span> <span class="pl-s1">np</span>
<span class="pl-k">import</span> <span class="pl-s1">torch</span>
<span class="pl-k">from</span> <span class="pl-s1">torch</span>.<span class="pl-s1">autograd</span> <span class="pl-k">import</span> <span class="pl-v">Variable</span>
<span class="pl-k">def</span> <span class="pl-en">set_seed</span>(<span class="pl-s1">seed</span>):
<span class="pl-s1">np</span>.<span class="pl-s1">random</span>.<span class="pl-en">seed</span>(<span class="pl-s1">seed</span>)
<span class="pl-s1">torch</span>.<span class="pl-en">manual_seed</span>(<span class="pl-s1">seed</span>)
<span class="pl-s1">torch</span>.<span class="pl-s1">cuda</span>.<span class="pl-en">manual_seed</span>(<span class="pl-s1">seed</span>)
<span class="pl-k">def</span> <span class="pl-en">main</span>():
<span class="pl-en">set_seed</span>(<span class="pl-c1">1</span>)
<span class="pl-s1">cuda</span> <span class="pl-c1">=</span> <span class="pl-c1">True</span>
<span class="pl-s1">n</span>, <span class="pl-s1">e</span>, <span class="pl-s1">c</span> <span class="pl-c1">=</span> <span class="pl-c1">5</span>, <span class="pl-c1">100</span>, <span class="pl-c1">5</span>
<span class="pl-s1">input</span> <span class="pl-c1">=</span> <span class="pl-s1">torch</span>.<span class="pl-en">randn</span>(<span class="pl-s1">n</span>,<span class="pl-s1">c</span>)
<span class="pl-s1">idxn</span> <span class="pl-c1">=</span> <span class="pl-s1">torch</span>.<span class="pl-en">from_numpy</span>(<span class="pl-s1">np</span>.<span class="pl-s1">random</span>.<span class="pl-en">randint</span>(<span class="pl-s1">n</span>,<span class="pl-s1">size</span><span class="pl-c1">=</span><span class="pl-s1">e</span>)) <span class="pl-c"># indices are repeated</span>
<span class="pl-k">if</span> <span class="pl-s1">cuda</span>:
<span class="pl-s1">input</span> <span class="pl-c1">=</span> <span class="pl-s1">input</span>.<span class="pl-en">cuda</span>(); <span class="pl-s1">idxn</span> <span class="pl-c1">=</span> <span class="pl-s1">idxn</span>.<span class="pl-en">cuda</span>()
<span class="pl-s1">gradsI</span>, <span class="pl-s1">gradsS</span> <span class="pl-c1">=</span> [], []
<span class="pl-v">N</span> <span class="pl-c1">=</span> <span class="pl-c1">4</span>
<span class="pl-k">for</span> <span class="pl-s1">i</span> <span class="pl-c1">in</span> <span class="pl-en">range</span>(<span class="pl-v">N</span>):
<span class="pl-s1">inputv</span> <span class="pl-c1">=</span> <span class="pl-v">Variable</span>(<span class="pl-s1">input</span>, <span class="pl-s1">requires_grad</span><span class="pl-c1">=</span><span class="pl-c1">True</span>)
<span class="pl-s1">sel_input</span> <span class="pl-c1">=</span> <span class="pl-s1">torch</span>.<span class="pl-en">index_select</span>(<span class="pl-s1">inputv</span>, <span class="pl-c1">0</span>, <span class="pl-v">Variable</span>(<span class="pl-s1">idxn</span>))
<span class="pl-s1">sel_input</span>.<span class="pl-en">retain_grad</span>()
<span class="pl-c"># the following computation is one of the "random conditions"</span>
<span class="pl-s1">data</span> <span class="pl-c1">=</span> [<span class="pl-s1">torch</span>.<span class="pl-en">sum</span>(<span class="pl-s1">sel_input</span>.<span class="pl-en">narrow</span>(<span class="pl-c1">0</span>,<span class="pl-c1">0</span>,<span class="pl-s1">e</span><span class="pl-c1">//</span><span class="pl-c1">2</span>), <span class="pl-c1">0</span>),
<span class="pl-s1">torch</span>.<span class="pl-en">sum</span>(<span class="pl-s1">sel_input</span>.<span class="pl-en">narrow</span>(<span class="pl-c1">0</span>,<span class="pl-s1">e</span><span class="pl-c1">//</span><span class="pl-c1">2</span>,<span class="pl-s1">e</span><span class="pl-c1">//</span><span class="pl-c1">2</span>), <span class="pl-c1">0</span>) ]
<span class="pl-s1">out</span> <span class="pl-c1">=</span> <span class="pl-s1">torch</span>.<span class="pl-en">cat</span>(<span class="pl-s1">data</span>,<span class="pl-c1">0</span>)
<span class="pl-s1">out</span>.<span class="pl-en">exp</span>().<span class="pl-en">sum</span>().<span class="pl-en">backward</span>()
<span class="pl-s1">gradsI</span>.<span class="pl-en">append</span>(<span class="pl-s1">inputv</span>.<span class="pl-s1">grad</span>.<span class="pl-s1">data</span>.<span class="pl-en">cpu</span>().<span class="pl-en">clone</span>())
<span class="pl-s1">gradsS</span>.<span class="pl-en">append</span>(<span class="pl-s1">sel_input</span>.<span class="pl-s1">grad</span>.<span class="pl-s1">data</span>.<span class="pl-en">cpu</span>().<span class="pl-en">clone</span>())
<span class="pl-k">for</span> <span class="pl-s1">i</span> <span class="pl-c1">in</span> <span class="pl-en">range</span>(<span class="pl-v">N</span>):
<span class="pl-k">for</span> <span class="pl-s1">j</span> <span class="pl-c1">in</span> <span class="pl-en">range</span>(<span class="pl-v">N</span>):
<span class="pl-en">print</span>(<span class="pl-s1">i</span>,<span class="pl-s1">j</span>,(<span class="pl-s1">gradsI</span>[<span class="pl-s1">i</span>]<span class="pl-c1">-</span><span class="pl-s1">gradsI</span>[<span class="pl-s1">j</span>]).<span class="pl-en">abs</span>().<span class="pl-en">max</span>(), (<span class="pl-s1">gradsS</span>[<span class="pl-s1">i</span>]<span class="pl-c1">-</span><span class="pl-s1">gradsS</span>[<span class="pl-s1">j</span>]).<span class="pl-en">abs</span>().<span class="pl-en">max</span>())</pre></div>
<p dir="auto">If <code class="notranslate">cuda=False</code>, everything works and 0s are printed, i.e. all gradients are equal. However, with cuda on many nonzero differences in <code class="notranslate">gradsI</code> but not in <code class="notranslate">gradsS</code> pop up, sometimes of high magnitude, e.g.:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="0 0 0.0 0.0
0 1 1.4411518807585587e+17 0.0
0 2 1.4411518807585587e+17 0.0
0 3 1.4411518807585587e+17 0.0
1 0 1.4411518807585587e+17 0.0
1 1 0.0 0.0
1 2 3.602879701896397e+16 0.0
1 3 0.015625 0.0
2 0 1.4411518807585587e+17 0.0
2 1 3.602879701896397e+16 0.0
2 2 0.0 0.0
2 3 3.602879701896397e+16 0.0
3 0 1.4411518807585587e+17 0.0
3 1 0.015625 0.0
3 2 3.602879701896397e+16 0.0
3 3 0.0 0.0"><pre class="notranslate"><code class="notranslate">0 0 0.0 0.0
0 1 1.4411518807585587e+17 0.0
0 2 1.4411518807585587e+17 0.0
0 3 1.4411518807585587e+17 0.0
1 0 1.4411518807585587e+17 0.0
1 1 0.0 0.0
1 2 3.602879701896397e+16 0.0
1 3 0.015625 0.0
2 0 1.4411518807585587e+17 0.0
2 1 3.602879701896397e+16 0.0
2 2 0.0 0.0
2 3 3.602879701896397e+16 0.0
3 0 1.4411518807585587e+17 0.0
3 1 0.015625 0.0
3 2 3.602879701896397e+16 0.0
3 3 0.0 0.0
</code></pre></div>
<p dir="auto">This indicates that there is something wrong with backward of <code class="notranslate">torch.index_select</code>. Setting <code class="notranslate">torch.set_default_tensor_type('torch.DoubleTensor')</code> doesn't help. Can anybody shed light onto this?</p> | 0 |
<p dir="auto">This feature request pertains to the following stack overflow question. <a href="http://stackoverflow.com/questions/32185411/break-in-x-axis-of-matplotlib" rel="nofollow">http://stackoverflow.com/questions/32185411/break-in-x-axis-of-matplotlib</a></p>
<p dir="auto">The answer for making a break works, however how the method works for a user should probably be abstracted away since it is not only unintuitive, but also quite lengthy for a simple break. I wish to create set_axis_break_x and set_axis_break_y functions that would take a tuple and create the break between the 2 data points without need for those long lines of code.</p> | <p dir="auto">So often times it is necessary to trim out irrelevant in information from a plot to show regions of interest, which lie far apart from each other, in higher detail, in a way that they can be easily compared. The solution to this problem is to put breaks in the axis lines (spines) to indicate a jump in magnitude.</p>
<p dir="auto">Example (from second link below):<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/20605205/42841848-a8436958-8a0b-11e8-84ba-b8b1b04f5b21.png"><img src="https://user-images.githubusercontent.com/20605205/42841848-a8436958-8a0b-11e8-84ba-b8b1b04f5b21.png" alt="image" style="max-width: 100%;"></a></p>
<p dir="auto">I recently had to make such plots and was disappointed by the solutions. The most predominant one seems to be <a href="https://stackoverflow.com/questions/44731152/matplotlib-create-broken-axis-in-subplot" rel="nofollow">making additional subplots, erasing the spines, and plotting fake lines to achieve the axis breaks. </a> This is the solution I used to get the plots made right now.</p>
<p dir="auto">However, I was thinking of solutions and did see <a href="https://stackoverflow.com/questions/5656798/python-matplotlib-is-there-a-way-to-make-a-discontinuous-axis" rel="nofollow">that involved creating a custom Axis subclass with it's own Transform subclass.</a> I liked this approach and decided to look a little further into it. However, this solution does not introduce the "breaks" in the axis (spine) which are critical to this display method.</p>
<p dir="auto">As such, I played with creating a subclass for the Spines to see if there was a solution for that there as well.</p>
<p dir="auto">I would like to share my code so far, just so people can see it, try it out, and offer suggestions. I'm not going directly for a PR because I'm not 100% sure where something like this should go into the code, and I would like it to be a little more polished and cohesive before starting a PR to track development. For the time being, the code only applies to the top or bottom axis, but porting to left and right is not a significant issue. I would just rather post it that let it end up forgotten in my home folder.</p>
<p dir="auto">Currently there are significant issues in the logic and drawing of the spine if the axis limits are adjusted such that the axis breaks reach the extreme end of the axes.</p>
<p dir="auto"><strong>Header</strong></p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="%matplotlib notebook
from matplotlib import pyplot as plt
from matplotlib import scale as mscale
from matplotlib import transforms as mtransforms
from matplotlib import ticker
from matplotlib import patches
from matplotlib.path import Path
from matplotlib.spines import Spine
import matplotlib.path as mpath
from matplotlib import rcParams
import numpy as np
from numpy import ma
plt.ion()"><pre class="notranslate"><span class="pl-c1">%</span><span class="pl-s1">matplotlib</span> <span class="pl-s1">notebook</span>
<span class="pl-k">from</span> <span class="pl-s1">matplotlib</span> <span class="pl-k">import</span> <span class="pl-s1">pyplot</span> <span class="pl-k">as</span> <span class="pl-s1">plt</span>
<span class="pl-k">from</span> <span class="pl-s1">matplotlib</span> <span class="pl-k">import</span> <span class="pl-s1">scale</span> <span class="pl-k">as</span> <span class="pl-s1">mscale</span>
<span class="pl-k">from</span> <span class="pl-s1">matplotlib</span> <span class="pl-k">import</span> <span class="pl-s1">transforms</span> <span class="pl-k">as</span> <span class="pl-s1">mtransforms</span>
<span class="pl-k">from</span> <span class="pl-s1">matplotlib</span> <span class="pl-k">import</span> <span class="pl-s1">ticker</span>
<span class="pl-k">from</span> <span class="pl-s1">matplotlib</span> <span class="pl-k">import</span> <span class="pl-s1">patches</span>
<span class="pl-k">from</span> <span class="pl-s1">matplotlib</span>.<span class="pl-s1">path</span> <span class="pl-k">import</span> <span class="pl-v">Path</span>
<span class="pl-k">from</span> <span class="pl-s1">matplotlib</span>.<span class="pl-s1">spines</span> <span class="pl-k">import</span> <span class="pl-v">Spine</span>
<span class="pl-k">import</span> <span class="pl-s1">matplotlib</span>.<span class="pl-s1">path</span> <span class="pl-k">as</span> <span class="pl-s1">mpath</span>
<span class="pl-k">from</span> <span class="pl-s1">matplotlib</span> <span class="pl-k">import</span> <span class="pl-s1">rcParams</span>
<span class="pl-k">import</span> <span class="pl-s1">numpy</span> <span class="pl-k">as</span> <span class="pl-s1">np</span>
<span class="pl-k">from</span> <span class="pl-s1">numpy</span> <span class="pl-k">import</span> <span class="pl-s1">ma</span>
<span class="pl-s1">plt</span>.<span class="pl-en">ion</span>()</pre></div>
<p dir="auto"><strong>Axis Scale and included Transform subclass</strong></p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="class BrokenScale(mscale.ScaleBase):
name = 'broken'
def __init__(self, axis, **kwargs):
mscale.ScaleBase.__init__(self)
self.thresh = None #thresh
self.breaks = kwargs.pop('breaks', [])
def get_transform(self):
return self.BrokenTransform(self.breaks, self.thresh)
def set_default_locators_and_formatters(self, axis):
#axis.set_major_locator(BrokenScale.BrokenLocator(self.breaks))
pass
class BrokenTransform(mtransforms.Transform):
input_dims = 1
output_dims = 1
is_separable = True
def __init__(self, breaks, thresh):
mtransforms.Transform.__init__(self)
self.thresh = thresh
self.breaks = breaks
def transform_non_affine(self, a):
mask = [any([low<i<high for (low,high) in self.breaks]) for i in a]
aa = a.copy()
tot_removed = 0
for _range in self.breaks[::-1]:
aa[a>(_range[0])] = aa[a>(_range[0])]-(_range[1]-_range[0]) + 2 # The 2 here is an offset factor so
# that they points before and after the
# break don't share the same x position.
# This will need to become a kwarg.
mma = ma.masked_array(aa, mask)
return mma
def inverted(self):
return BrokenScale.InvertedBrokenTransform(self.breaks, self.thresh)
class InvertedBrokenTransform(mtransforms.Transform):
input_dims = 1
output_dims = 1
is_separable = True
def __init__(self, breaks, thresh):
mtransforms.Transform.__init__(self)
self.thresh = thresh
self.breaks = breaks
def transform_non_affine(self, a):
aa = a.copy()
for _range in self.breaks:
aa[a>_range[0]] = a[a>_range[0]]+(_range[1]-_range[0])
return aa
def inverted(self):
return BrokenScale.BrokenTransform(self.breaks, self.thresh)
class BrokenLocator(ticker.Locator):
# To override
def __init__(self):
# To override
pass
def tick_values(self, vmin, vmax):
# To override
pass
def __call__(self):
# To override
# note: some locators return data limits, other return view limits,
# hence there is no *one* interface to call self.tick_values.
pass"><pre class="notranslate"><span class="pl-k">class</span> <span class="pl-v">BrokenScale</span>(<span class="pl-s1">mscale</span>.<span class="pl-v">ScaleBase</span>):
<span class="pl-s1">name</span> <span class="pl-c1">=</span> <span class="pl-s">'broken'</span>
<span class="pl-k">def</span> <span class="pl-en">__init__</span>(<span class="pl-s1">self</span>, <span class="pl-s1">axis</span>, <span class="pl-c1">**</span><span class="pl-s1">kwargs</span>):
<span class="pl-s1">mscale</span>.<span class="pl-v">ScaleBase</span>.<span class="pl-en">__init__</span>(<span class="pl-s1">self</span>)
<span class="pl-s1">self</span>.<span class="pl-s1">thresh</span> <span class="pl-c1">=</span> <span class="pl-c1">None</span> <span class="pl-c">#thresh</span>
<span class="pl-s1">self</span>.<span class="pl-s1">breaks</span> <span class="pl-c1">=</span> <span class="pl-s1">kwargs</span>.<span class="pl-en">pop</span>(<span class="pl-s">'breaks'</span>, [])
<span class="pl-k">def</span> <span class="pl-en">get_transform</span>(<span class="pl-s1">self</span>):
<span class="pl-k">return</span> <span class="pl-s1">self</span>.<span class="pl-v">BrokenTransform</span>(<span class="pl-s1">self</span>.<span class="pl-s1">breaks</span>, <span class="pl-s1">self</span>.<span class="pl-s1">thresh</span>)
<span class="pl-k">def</span> <span class="pl-en">set_default_locators_and_formatters</span>(<span class="pl-s1">self</span>, <span class="pl-s1">axis</span>):
<span class="pl-c">#axis.set_major_locator(BrokenScale.BrokenLocator(self.breaks))</span>
<span class="pl-k">pass</span>
<span class="pl-k">class</span> <span class="pl-v">BrokenTransform</span>(<span class="pl-s1">mtransforms</span>.<span class="pl-v">Transform</span>):
<span class="pl-s1">input_dims</span> <span class="pl-c1">=</span> <span class="pl-c1">1</span>
<span class="pl-s1">output_dims</span> <span class="pl-c1">=</span> <span class="pl-c1">1</span>
<span class="pl-s1">is_separable</span> <span class="pl-c1">=</span> <span class="pl-c1">True</span>
<span class="pl-k">def</span> <span class="pl-en">__init__</span>(<span class="pl-s1">self</span>, <span class="pl-s1">breaks</span>, <span class="pl-s1">thresh</span>):
<span class="pl-s1">mtransforms</span>.<span class="pl-v">Transform</span>.<span class="pl-en">__init__</span>(<span class="pl-s1">self</span>)
<span class="pl-s1">self</span>.<span class="pl-s1">thresh</span> <span class="pl-c1">=</span> <span class="pl-s1">thresh</span>
<span class="pl-s1">self</span>.<span class="pl-s1">breaks</span> <span class="pl-c1">=</span> <span class="pl-s1">breaks</span>
<span class="pl-k">def</span> <span class="pl-en">transform_non_affine</span>(<span class="pl-s1">self</span>, <span class="pl-s1">a</span>):
<span class="pl-s1">mask</span> <span class="pl-c1">=</span> [<span class="pl-en">any</span>([<span class="pl-s1">low</span><span class="pl-c1"><</span><span class="pl-s1">i</span><span class="pl-c1"><</span><span class="pl-s1">high</span> <span class="pl-k">for</span> (<span class="pl-s1">low</span>,<span class="pl-s1">high</span>) <span class="pl-c1">in</span> <span class="pl-s1">self</span>.<span class="pl-s1">breaks</span>]) <span class="pl-k">for</span> <span class="pl-s1">i</span> <span class="pl-c1">in</span> <span class="pl-s1">a</span>]
<span class="pl-s1">aa</span> <span class="pl-c1">=</span> <span class="pl-s1">a</span>.<span class="pl-en">copy</span>()
<span class="pl-s1">tot_removed</span> <span class="pl-c1">=</span> <span class="pl-c1">0</span>
<span class="pl-k">for</span> <span class="pl-s1">_range</span> <span class="pl-c1">in</span> <span class="pl-s1">self</span>.<span class="pl-s1">breaks</span>[::<span class="pl-c1">-</span><span class="pl-c1">1</span>]:
<span class="pl-s1">aa</span>[<span class="pl-s1">a</span><span class="pl-c1">></span>(<span class="pl-s1">_range</span>[<span class="pl-c1">0</span>])] <span class="pl-c1">=</span> <span class="pl-s1">aa</span>[<span class="pl-s1">a</span><span class="pl-c1">></span>(<span class="pl-s1">_range</span>[<span class="pl-c1">0</span>])]<span class="pl-c1">-</span>(<span class="pl-s1">_range</span>[<span class="pl-c1">1</span>]<span class="pl-c1">-</span><span class="pl-s1">_range</span>[<span class="pl-c1">0</span>]) <span class="pl-c1">+</span> <span class="pl-c1">2</span> <span class="pl-c"># The 2 here is an offset factor so </span>
<span class="pl-c"># that they points before and after the </span>
<span class="pl-c"># break don't share the same x position.</span>
<span class="pl-c"># This will need to become a kwarg.</span>
<span class="pl-s1">mma</span> <span class="pl-c1">=</span> <span class="pl-s1">ma</span>.<span class="pl-en">masked_array</span>(<span class="pl-s1">aa</span>, <span class="pl-s1">mask</span>)
<span class="pl-k">return</span> <span class="pl-s1">mma</span>
<span class="pl-k">def</span> <span class="pl-en">inverted</span>(<span class="pl-s1">self</span>):
<span class="pl-k">return</span> <span class="pl-v">BrokenScale</span>.<span class="pl-v">InvertedBrokenTransform</span>(<span class="pl-s1">self</span>.<span class="pl-s1">breaks</span>, <span class="pl-s1">self</span>.<span class="pl-s1">thresh</span>)
<span class="pl-k">class</span> <span class="pl-v">InvertedBrokenTransform</span>(<span class="pl-s1">mtransforms</span>.<span class="pl-v">Transform</span>):
<span class="pl-s1">input_dims</span> <span class="pl-c1">=</span> <span class="pl-c1">1</span>
<span class="pl-s1">output_dims</span> <span class="pl-c1">=</span> <span class="pl-c1">1</span>
<span class="pl-s1">is_separable</span> <span class="pl-c1">=</span> <span class="pl-c1">True</span>
<span class="pl-k">def</span> <span class="pl-en">__init__</span>(<span class="pl-s1">self</span>, <span class="pl-s1">breaks</span>, <span class="pl-s1">thresh</span>):
<span class="pl-s1">mtransforms</span>.<span class="pl-v">Transform</span>.<span class="pl-en">__init__</span>(<span class="pl-s1">self</span>)
<span class="pl-s1">self</span>.<span class="pl-s1">thresh</span> <span class="pl-c1">=</span> <span class="pl-s1">thresh</span>
<span class="pl-s1">self</span>.<span class="pl-s1">breaks</span> <span class="pl-c1">=</span> <span class="pl-s1">breaks</span>
<span class="pl-k">def</span> <span class="pl-en">transform_non_affine</span>(<span class="pl-s1">self</span>, <span class="pl-s1">a</span>):
<span class="pl-s1">aa</span> <span class="pl-c1">=</span> <span class="pl-s1">a</span>.<span class="pl-en">copy</span>()
<span class="pl-k">for</span> <span class="pl-s1">_range</span> <span class="pl-c1">in</span> <span class="pl-s1">self</span>.<span class="pl-s1">breaks</span>:
<span class="pl-s1">aa</span>[<span class="pl-s1">a</span><span class="pl-c1">></span><span class="pl-s1">_range</span>[<span class="pl-c1">0</span>]] <span class="pl-c1">=</span> <span class="pl-s1">a</span>[<span class="pl-s1">a</span><span class="pl-c1">></span><span class="pl-s1">_range</span>[<span class="pl-c1">0</span>]]<span class="pl-c1">+</span>(<span class="pl-s1">_range</span>[<span class="pl-c1">1</span>]<span class="pl-c1">-</span><span class="pl-s1">_range</span>[<span class="pl-c1">0</span>])
<span class="pl-k">return</span> <span class="pl-s1">aa</span>
<span class="pl-k">def</span> <span class="pl-en">inverted</span>(<span class="pl-s1">self</span>):
<span class="pl-k">return</span> <span class="pl-v">BrokenScale</span>.<span class="pl-v">BrokenTransform</span>(<span class="pl-s1">self</span>.<span class="pl-s1">breaks</span>, <span class="pl-s1">self</span>.<span class="pl-s1">thresh</span>)
<span class="pl-k">class</span> <span class="pl-v">BrokenLocator</span>(<span class="pl-s1">ticker</span>.<span class="pl-v">Locator</span>):
<span class="pl-c"># To override</span>
<span class="pl-k">def</span> <span class="pl-en">__init__</span>(<span class="pl-s1">self</span>):
<span class="pl-c"># To override</span>
<span class="pl-k">pass</span>
<span class="pl-k">def</span> <span class="pl-en">tick_values</span>(<span class="pl-s1">self</span>, <span class="pl-s1">vmin</span>, <span class="pl-s1">vmax</span>):
<span class="pl-c"># To override</span>
<span class="pl-k">pass</span>
<span class="pl-k">def</span> <span class="pl-en">__call__</span>(<span class="pl-s1">self</span>):
<span class="pl-c"># To override</span>
<span class="pl-c"># note: some locators return data limits, other return view limits,</span>
<span class="pl-c"># hence there is no *one* interface to call self.tick_values.</span>
<span class="pl-k">pass</span></pre></div>
<p dir="auto"><strong>Spine subclass</strong></p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="class BrokenSpine(Spine):
def __init__(self, axes, spine_type, path, **kwargs):
self.breaks = kwargs.pop('breaks', [])
super().__init__(axes, spine_type, path, **kwargs)
def _adjust_location(self):
"""automatically set spine bounds to the view interval"""
if self._bounds is None:
if self.spine_type in ('left', 'right'):
low, high = self.axes.viewLim.intervaly
elif self.spine_type in ('top', 'bottom'):
low, high = self.axes.viewLim.intervalx
else:
raise ValueError('unknown spine spine_type: %s' %
self.spine_type)
if self._smart_bounds:
# attempt to set bounds in sophisticated way
# handle inverted limits
viewlim_low, viewlim_high = sorted([low, high])
if self.spine_type in ('left', 'right'):
datalim_low, datalim_high = self.axes.dataLim.intervaly
ticks = self.axes.get_yticks()
elif self.spine_type in ('top', 'bottom'):
datalim_low, datalim_high = self.axes.dataLim.intervalx
ticks = self.axes.get_xticks()
# handle inverted limits
ticks = np.sort(ticks)
datalim_low, datalim_high = sorted([datalim_low, datalim_high])
if datalim_low < viewlim_low:
# Data extends past view. Clip line to view.
low = viewlim_low
else:
# Data ends before view ends.
cond = (ticks <= datalim_low) & (ticks >= viewlim_low)
tickvals = ticks[cond]
if len(tickvals):
# A tick is less than or equal to lowest data point.
low = tickvals[-1]
else:
# No tick is available
low = datalim_low
low = max(low, viewlim_low)
if datalim_high > viewlim_high:
# Data extends past view. Clip line to view.
high = viewlim_high
else:
# Data ends before view ends.
cond = (ticks >= datalim_high) & (ticks <= viewlim_high)
tickvals = ticks[cond]
if len(tickvals):
# A tick is greater than or equal to highest data
# point.
high = tickvals[0]
else:
# No tick is available
high = datalim_high
high = min(high, viewlim_high)
else:
low, high = self._bounds
v1 = self._path.vertices
c1 = self._path.codes
# The below line is commented out from the copy of the parent method
#assert v1.shape == (2, 2), 'unexpected vertices shape'
if self.spine_type in ['left', 'right']:
v1[0, 1] = low
v1[1, 1] = high
elif self.spine_type in ['bottom', 'top']:
y = v1[0,1]
v1 = []
c1 = []
v1.append([low, y])
c1.append(Path.MOVETO)
for i, (lower, upper) in enumerate(self.breaks):
if lower<low or upper>high or lower>high or upper<low:
continue
c1.append(Path.LINETO)
v1.append([lower, y])
c1.append(Path.MOVETO)
v1.append([upper, y])
c1.append(Path.LINETO)
v1.append([high, y])
self._path = Path(v1,c1) # Changing the number of vertices like the
# original code doesn't seem to work.
# Must create new path.
else:
raise ValueError('unable to set bounds for spine "%s"' %
self.spine_type)
@classmethod
def linear_spine(cls, axes, spine_type, **kwargs):
"""
(staticmethod) Returns a linear :class:`Spine`.
"""
breaks = kwargs.pop('breaks')
# all values of 0.999 get replaced upon call to set_bounds()
if spine_type == 'left':
path = Path([(0.0, 0.999), (0.0, 0.999)])
elif spine_type == 'right':
path = Path([(1.0, 0.999), (1.0, 0.999)])
elif spine_type == 'top':
path = Path([(0.999, 1.0), (0.999, 1.0)])
elif spine_type == 'bottom':
verts = []
codes = []
low, high = axes.viewLim.intervalx
codes.append(Path.MOVETO)
verts.append([low, 0.0])
for (lower, upper) in breaks:
codes.append(Path.LINETO)
verts.append([lower, 0.0])
codes.append(Path.MOVETO)
verts.append([upper, 0.0])
codes.append(Path.LINETO)
verts.append([high, 0.0])
path = Path(verts,codes)
print(path)
else:
raise ValueError('unable to make path for spine "%s"' % spine_type)
result = cls(axes, spine_type, path, breaks=breaks, **kwargs)
result.set_visible(rcParams['axes.spines.{0}'.format(spine_type)])
return result"><pre class="notranslate"><span class="pl-k">class</span> <span class="pl-v">BrokenSpine</span>(<span class="pl-v">Spine</span>):
<span class="pl-k">def</span> <span class="pl-en">__init__</span>(<span class="pl-s1">self</span>, <span class="pl-s1">axes</span>, <span class="pl-s1">spine_type</span>, <span class="pl-s1">path</span>, <span class="pl-c1">**</span><span class="pl-s1">kwargs</span>):
<span class="pl-s1">self</span>.<span class="pl-s1">breaks</span> <span class="pl-c1">=</span> <span class="pl-s1">kwargs</span>.<span class="pl-en">pop</span>(<span class="pl-s">'breaks'</span>, [])
<span class="pl-en">super</span>().<span class="pl-en">__init__</span>(<span class="pl-s1">axes</span>, <span class="pl-s1">spine_type</span>, <span class="pl-s1">path</span>, <span class="pl-c1">**</span><span class="pl-s1">kwargs</span>)
<span class="pl-k">def</span> <span class="pl-en">_adjust_location</span>(<span class="pl-s1">self</span>):
<span class="pl-s">"""automatically set spine bounds to the view interval"""</span>
<span class="pl-k">if</span> <span class="pl-s1">self</span>.<span class="pl-s1">_bounds</span> <span class="pl-c1">is</span> <span class="pl-c1">None</span>:
<span class="pl-k">if</span> <span class="pl-s1">self</span>.<span class="pl-s1">spine_type</span> <span class="pl-c1">in</span> (<span class="pl-s">'left'</span>, <span class="pl-s">'right'</span>):
<span class="pl-s1">low</span>, <span class="pl-s1">high</span> <span class="pl-c1">=</span> <span class="pl-s1">self</span>.<span class="pl-s1">axes</span>.<span class="pl-s1">viewLim</span>.<span class="pl-s1">intervaly</span>
<span class="pl-k">elif</span> <span class="pl-s1">self</span>.<span class="pl-s1">spine_type</span> <span class="pl-c1">in</span> (<span class="pl-s">'top'</span>, <span class="pl-s">'bottom'</span>):
<span class="pl-s1">low</span>, <span class="pl-s1">high</span> <span class="pl-c1">=</span> <span class="pl-s1">self</span>.<span class="pl-s1">axes</span>.<span class="pl-s1">viewLim</span>.<span class="pl-s1">intervalx</span>
<span class="pl-k">else</span>:
<span class="pl-k">raise</span> <span class="pl-v">ValueError</span>(<span class="pl-s">'unknown spine spine_type: %s'</span> <span class="pl-c1">%</span>
<span class="pl-s1">self</span>.<span class="pl-s1">spine_type</span>)
<span class="pl-k">if</span> <span class="pl-s1">self</span>.<span class="pl-s1">_smart_bounds</span>:
<span class="pl-c"># attempt to set bounds in sophisticated way</span>
<span class="pl-c"># handle inverted limits</span>
<span class="pl-s1">viewlim_low</span>, <span class="pl-s1">viewlim_high</span> <span class="pl-c1">=</span> <span class="pl-en">sorted</span>([<span class="pl-s1">low</span>, <span class="pl-s1">high</span>])
<span class="pl-k">if</span> <span class="pl-s1">self</span>.<span class="pl-s1">spine_type</span> <span class="pl-c1">in</span> (<span class="pl-s">'left'</span>, <span class="pl-s">'right'</span>):
<span class="pl-s1">datalim_low</span>, <span class="pl-s1">datalim_high</span> <span class="pl-c1">=</span> <span class="pl-s1">self</span>.<span class="pl-s1">axes</span>.<span class="pl-s1">dataLim</span>.<span class="pl-s1">intervaly</span>
<span class="pl-s1">ticks</span> <span class="pl-c1">=</span> <span class="pl-s1">self</span>.<span class="pl-s1">axes</span>.<span class="pl-en">get_yticks</span>()
<span class="pl-k">elif</span> <span class="pl-s1">self</span>.<span class="pl-s1">spine_type</span> <span class="pl-c1">in</span> (<span class="pl-s">'top'</span>, <span class="pl-s">'bottom'</span>):
<span class="pl-s1">datalim_low</span>, <span class="pl-s1">datalim_high</span> <span class="pl-c1">=</span> <span class="pl-s1">self</span>.<span class="pl-s1">axes</span>.<span class="pl-s1">dataLim</span>.<span class="pl-s1">intervalx</span>
<span class="pl-s1">ticks</span> <span class="pl-c1">=</span> <span class="pl-s1">self</span>.<span class="pl-s1">axes</span>.<span class="pl-en">get_xticks</span>()
<span class="pl-c"># handle inverted limits</span>
<span class="pl-s1">ticks</span> <span class="pl-c1">=</span> <span class="pl-s1">np</span>.<span class="pl-en">sort</span>(<span class="pl-s1">ticks</span>)
<span class="pl-s1">datalim_low</span>, <span class="pl-s1">datalim_high</span> <span class="pl-c1">=</span> <span class="pl-en">sorted</span>([<span class="pl-s1">datalim_low</span>, <span class="pl-s1">datalim_high</span>])
<span class="pl-k">if</span> <span class="pl-s1">datalim_low</span> <span class="pl-c1"><</span> <span class="pl-s1">viewlim_low</span>:
<span class="pl-c"># Data extends past view. Clip line to view.</span>
<span class="pl-s1">low</span> <span class="pl-c1">=</span> <span class="pl-s1">viewlim_low</span>
<span class="pl-k">else</span>:
<span class="pl-c"># Data ends before view ends.</span>
<span class="pl-s1">cond</span> <span class="pl-c1">=</span> (<span class="pl-s1">ticks</span> <span class="pl-c1"><=</span> <span class="pl-s1">datalim_low</span>) <span class="pl-c1">&</span> (<span class="pl-s1">ticks</span> <span class="pl-c1">>=</span> <span class="pl-s1">viewlim_low</span>)
<span class="pl-s1">tickvals</span> <span class="pl-c1">=</span> <span class="pl-s1">ticks</span>[<span class="pl-s1">cond</span>]
<span class="pl-k">if</span> <span class="pl-en">len</span>(<span class="pl-s1">tickvals</span>):
<span class="pl-c"># A tick is less than or equal to lowest data point.</span>
<span class="pl-s1">low</span> <span class="pl-c1">=</span> <span class="pl-s1">tickvals</span>[<span class="pl-c1">-</span><span class="pl-c1">1</span>]
<span class="pl-k">else</span>:
<span class="pl-c"># No tick is available</span>
<span class="pl-s1">low</span> <span class="pl-c1">=</span> <span class="pl-s1">datalim_low</span>
<span class="pl-s1">low</span> <span class="pl-c1">=</span> <span class="pl-en">max</span>(<span class="pl-s1">low</span>, <span class="pl-s1">viewlim_low</span>)
<span class="pl-k">if</span> <span class="pl-s1">datalim_high</span> <span class="pl-c1">></span> <span class="pl-s1">viewlim_high</span>:
<span class="pl-c"># Data extends past view. Clip line to view.</span>
<span class="pl-s1">high</span> <span class="pl-c1">=</span> <span class="pl-s1">viewlim_high</span>
<span class="pl-k">else</span>:
<span class="pl-c"># Data ends before view ends.</span>
<span class="pl-s1">cond</span> <span class="pl-c1">=</span> (<span class="pl-s1">ticks</span> <span class="pl-c1">>=</span> <span class="pl-s1">datalim_high</span>) <span class="pl-c1">&</span> (<span class="pl-s1">ticks</span> <span class="pl-c1"><=</span> <span class="pl-s1">viewlim_high</span>)
<span class="pl-s1">tickvals</span> <span class="pl-c1">=</span> <span class="pl-s1">ticks</span>[<span class="pl-s1">cond</span>]
<span class="pl-k">if</span> <span class="pl-en">len</span>(<span class="pl-s1">tickvals</span>):
<span class="pl-c"># A tick is greater than or equal to highest data</span>
<span class="pl-c"># point.</span>
<span class="pl-s1">high</span> <span class="pl-c1">=</span> <span class="pl-s1">tickvals</span>[<span class="pl-c1">0</span>]
<span class="pl-k">else</span>:
<span class="pl-c"># No tick is available</span>
<span class="pl-s1">high</span> <span class="pl-c1">=</span> <span class="pl-s1">datalim_high</span>
<span class="pl-s1">high</span> <span class="pl-c1">=</span> <span class="pl-en">min</span>(<span class="pl-s1">high</span>, <span class="pl-s1">viewlim_high</span>)
<span class="pl-k">else</span>:
<span class="pl-s1">low</span>, <span class="pl-s1">high</span> <span class="pl-c1">=</span> <span class="pl-s1">self</span>.<span class="pl-s1">_bounds</span>
<span class="pl-s1">v1</span> <span class="pl-c1">=</span> <span class="pl-s1">self</span>.<span class="pl-s1">_path</span>.<span class="pl-s1">vertices</span>
<span class="pl-s1">c1</span> <span class="pl-c1">=</span> <span class="pl-s1">self</span>.<span class="pl-s1">_path</span>.<span class="pl-s1">codes</span>
<span class="pl-c"># The below line is commented out from the copy of the parent method</span>
<span class="pl-c">#assert v1.shape == (2, 2), 'unexpected vertices shape'</span>
<span class="pl-k">if</span> <span class="pl-s1">self</span>.<span class="pl-s1">spine_type</span> <span class="pl-c1">in</span> [<span class="pl-s">'left'</span>, <span class="pl-s">'right'</span>]:
<span class="pl-s1">v1</span>[<span class="pl-c1">0</span>, <span class="pl-c1">1</span>] <span class="pl-c1">=</span> <span class="pl-s1">low</span>
<span class="pl-s1">v1</span>[<span class="pl-c1">1</span>, <span class="pl-c1">1</span>] <span class="pl-c1">=</span> <span class="pl-s1">high</span>
<span class="pl-k">elif</span> <span class="pl-s1">self</span>.<span class="pl-s1">spine_type</span> <span class="pl-c1">in</span> [<span class="pl-s">'bottom'</span>, <span class="pl-s">'top'</span>]:
<span class="pl-s1">y</span> <span class="pl-c1">=</span> <span class="pl-s1">v1</span>[<span class="pl-c1">0</span>,<span class="pl-c1">1</span>]
<span class="pl-s1">v1</span> <span class="pl-c1">=</span> []
<span class="pl-s1">c1</span> <span class="pl-c1">=</span> []
<span class="pl-s1">v1</span>.<span class="pl-en">append</span>([<span class="pl-s1">low</span>, <span class="pl-s1">y</span>])
<span class="pl-s1">c1</span>.<span class="pl-en">append</span>(<span class="pl-v">Path</span>.<span class="pl-v">MOVETO</span>)
<span class="pl-k">for</span> <span class="pl-s1">i</span>, (<span class="pl-s1">lower</span>, <span class="pl-s1">upper</span>) <span class="pl-c1">in</span> <span class="pl-en">enumerate</span>(<span class="pl-s1">self</span>.<span class="pl-s1">breaks</span>):
<span class="pl-k">if</span> <span class="pl-s1">lower</span><span class="pl-c1"><</span><span class="pl-s1">low</span> <span class="pl-c1">or</span> <span class="pl-s1">upper</span><span class="pl-c1">></span><span class="pl-s1">high</span> <span class="pl-c1">or</span> <span class="pl-s1">lower</span><span class="pl-c1">></span><span class="pl-s1">high</span> <span class="pl-c1">or</span> <span class="pl-s1">upper</span><span class="pl-c1"><</span><span class="pl-s1">low</span>:
<span class="pl-k">continue</span>
<span class="pl-s1">c1</span>.<span class="pl-en">append</span>(<span class="pl-v">Path</span>.<span class="pl-v">LINETO</span>)
<span class="pl-s1">v1</span>.<span class="pl-en">append</span>([<span class="pl-s1">lower</span>, <span class="pl-s1">y</span>])
<span class="pl-s1">c1</span>.<span class="pl-en">append</span>(<span class="pl-v">Path</span>.<span class="pl-v">MOVETO</span>)
<span class="pl-s1">v1</span>.<span class="pl-en">append</span>([<span class="pl-s1">upper</span>, <span class="pl-s1">y</span>])
<span class="pl-s1">c1</span>.<span class="pl-en">append</span>(<span class="pl-v">Path</span>.<span class="pl-v">LINETO</span>)
<span class="pl-s1">v1</span>.<span class="pl-en">append</span>([<span class="pl-s1">high</span>, <span class="pl-s1">y</span>])
<span class="pl-s1">self</span>.<span class="pl-s1">_path</span> <span class="pl-c1">=</span> <span class="pl-v">Path</span>(<span class="pl-s1">v1</span>,<span class="pl-s1">c1</span>) <span class="pl-c"># Changing the number of vertices like the </span>
<span class="pl-c"># original code doesn't seem to work. </span>
<span class="pl-c"># Must create new path.</span>
<span class="pl-k">else</span>:
<span class="pl-k">raise</span> <span class="pl-v">ValueError</span>(<span class="pl-s">'unable to set bounds for spine "%s"'</span> <span class="pl-c1">%</span>
<span class="pl-s1">self</span>.<span class="pl-s1">spine_type</span>)
<span class="pl-en">@<span class="pl-s1">classmethod</span></span>
<span class="pl-k">def</span> <span class="pl-en">linear_spine</span>(<span class="pl-s1">cls</span>, <span class="pl-s1">axes</span>, <span class="pl-s1">spine_type</span>, <span class="pl-c1">**</span><span class="pl-s1">kwargs</span>):
<span class="pl-s">"""</span>
<span class="pl-s"> (staticmethod) Returns a linear :class:`Spine`.</span>
<span class="pl-s"> """</span>
<span class="pl-s1">breaks</span> <span class="pl-c1">=</span> <span class="pl-s1">kwargs</span>.<span class="pl-en">pop</span>(<span class="pl-s">'breaks'</span>)
<span class="pl-c"># all values of 0.999 get replaced upon call to set_bounds()</span>
<span class="pl-k">if</span> <span class="pl-s1">spine_type</span> <span class="pl-c1">==</span> <span class="pl-s">'left'</span>:
<span class="pl-s1">path</span> <span class="pl-c1">=</span> <span class="pl-v">Path</span>([(<span class="pl-c1">0.0</span>, <span class="pl-c1">0.999</span>), (<span class="pl-c1">0.0</span>, <span class="pl-c1">0.999</span>)])
<span class="pl-k">elif</span> <span class="pl-s1">spine_type</span> <span class="pl-c1">==</span> <span class="pl-s">'right'</span>:
<span class="pl-s1">path</span> <span class="pl-c1">=</span> <span class="pl-v">Path</span>([(<span class="pl-c1">1.0</span>, <span class="pl-c1">0.999</span>), (<span class="pl-c1">1.0</span>, <span class="pl-c1">0.999</span>)])
<span class="pl-k">elif</span> <span class="pl-s1">spine_type</span> <span class="pl-c1">==</span> <span class="pl-s">'top'</span>:
<span class="pl-s1">path</span> <span class="pl-c1">=</span> <span class="pl-v">Path</span>([(<span class="pl-c1">0.999</span>, <span class="pl-c1">1.0</span>), (<span class="pl-c1">0.999</span>, <span class="pl-c1">1.0</span>)])
<span class="pl-k">elif</span> <span class="pl-s1">spine_type</span> <span class="pl-c1">==</span> <span class="pl-s">'bottom'</span>:
<span class="pl-s1">verts</span> <span class="pl-c1">=</span> []
<span class="pl-s1">codes</span> <span class="pl-c1">=</span> []
<span class="pl-s1">low</span>, <span class="pl-s1">high</span> <span class="pl-c1">=</span> <span class="pl-s1">axes</span>.<span class="pl-s1">viewLim</span>.<span class="pl-s1">intervalx</span>
<span class="pl-s1">codes</span>.<span class="pl-en">append</span>(<span class="pl-v">Path</span>.<span class="pl-v">MOVETO</span>)
<span class="pl-s1">verts</span>.<span class="pl-en">append</span>([<span class="pl-s1">low</span>, <span class="pl-c1">0.0</span>])
<span class="pl-k">for</span> (<span class="pl-s1">lower</span>, <span class="pl-s1">upper</span>) <span class="pl-c1">in</span> <span class="pl-s1">breaks</span>:
<span class="pl-s1">codes</span>.<span class="pl-en">append</span>(<span class="pl-v">Path</span>.<span class="pl-v">LINETO</span>)
<span class="pl-s1">verts</span>.<span class="pl-en">append</span>([<span class="pl-s1">lower</span>, <span class="pl-c1">0.0</span>])
<span class="pl-s1">codes</span>.<span class="pl-en">append</span>(<span class="pl-v">Path</span>.<span class="pl-v">MOVETO</span>)
<span class="pl-s1">verts</span>.<span class="pl-en">append</span>([<span class="pl-s1">upper</span>, <span class="pl-c1">0.0</span>])
<span class="pl-s1">codes</span>.<span class="pl-en">append</span>(<span class="pl-v">Path</span>.<span class="pl-v">LINETO</span>)
<span class="pl-s1">verts</span>.<span class="pl-en">append</span>([<span class="pl-s1">high</span>, <span class="pl-c1">0.0</span>])
<span class="pl-s1">path</span> <span class="pl-c1">=</span> <span class="pl-v">Path</span>(<span class="pl-s1">verts</span>,<span class="pl-s1">codes</span>)
<span class="pl-en">print</span>(<span class="pl-s1">path</span>)
<span class="pl-k">else</span>:
<span class="pl-k">raise</span> <span class="pl-v">ValueError</span>(<span class="pl-s">'unable to make path for spine "%s"'</span> <span class="pl-c1">%</span> <span class="pl-s1">spine_type</span>)
<span class="pl-s1">result</span> <span class="pl-c1">=</span> <span class="pl-en">cls</span>(<span class="pl-s1">axes</span>, <span class="pl-s1">spine_type</span>, <span class="pl-s1">path</span>, <span class="pl-s1">breaks</span><span class="pl-c1">=</span><span class="pl-s1">breaks</span>, <span class="pl-c1">**</span><span class="pl-s1">kwargs</span>)
<span class="pl-s1">result</span>.<span class="pl-en">set_visible</span>(<span class="pl-s1">rcParams</span>[<span class="pl-s">'axes.spines.{0}'</span>.<span class="pl-en">format</span>(<span class="pl-s1">spine_type</span>)])
<span class="pl-k">return</span> <span class="pl-s1">result</span></pre></div>
<p dir="auto"><strong>Example</strong></p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="mscale.register_scale(BrokenScale)
plt.figure()
x = np.concatenate((np.linspace(0,2*np.pi,100), np.linspace(10*np.pi,12*np.pi,100)))
y = np.sin(x)
plt.plot(x, y,'.')
plt.xscale('broken', breaks=[[2*np.pi, 10*np.pi]])
plt.show()
ax = plt.gca()
ax.spines['bottom'] = BrokenSpine.linear_spine(ax, 'bottom', breaks=[[2*np.pi, 10*np.pi]])"><pre class="notranslate"><span class="pl-s1">mscale</span>.<span class="pl-en">register_scale</span>(<span class="pl-v">BrokenScale</span>)
<span class="pl-s1">plt</span>.<span class="pl-en">figure</span>()
<span class="pl-s1">x</span> <span class="pl-c1">=</span> <span class="pl-s1">np</span>.<span class="pl-en">concatenate</span>((<span class="pl-s1">np</span>.<span class="pl-en">linspace</span>(<span class="pl-c1">0</span>,<span class="pl-c1">2</span><span class="pl-c1">*</span><span class="pl-s1">np</span>.<span class="pl-s1">pi</span>,<span class="pl-c1">100</span>), <span class="pl-s1">np</span>.<span class="pl-en">linspace</span>(<span class="pl-c1">10</span><span class="pl-c1">*</span><span class="pl-s1">np</span>.<span class="pl-s1">pi</span>,<span class="pl-c1">12</span><span class="pl-c1">*</span><span class="pl-s1">np</span>.<span class="pl-s1">pi</span>,<span class="pl-c1">100</span>)))
<span class="pl-s1">y</span> <span class="pl-c1">=</span> <span class="pl-s1">np</span>.<span class="pl-en">sin</span>(<span class="pl-s1">x</span>)
<span class="pl-s1">plt</span>.<span class="pl-en">plot</span>(<span class="pl-s1">x</span>, <span class="pl-s1">y</span>,<span class="pl-s">'.'</span>)
<span class="pl-s1">plt</span>.<span class="pl-en">xscale</span>(<span class="pl-s">'broken'</span>, <span class="pl-s1">breaks</span><span class="pl-c1">=</span>[[<span class="pl-c1">2</span><span class="pl-c1">*</span><span class="pl-s1">np</span>.<span class="pl-s1">pi</span>, <span class="pl-c1">10</span><span class="pl-c1">*</span><span class="pl-s1">np</span>.<span class="pl-s1">pi</span>]])
<span class="pl-s1">plt</span>.<span class="pl-en">show</span>()
<span class="pl-s1">ax</span> <span class="pl-c1">=</span> <span class="pl-s1">plt</span>.<span class="pl-en">gca</span>()
<span class="pl-s1">ax</span>.<span class="pl-s1">spines</span>[<span class="pl-s">'bottom'</span>] <span class="pl-c1">=</span> <span class="pl-v">BrokenSpine</span>.<span class="pl-en">linear_spine</span>(<span class="pl-s1">ax</span>, <span class="pl-s">'bottom'</span>, <span class="pl-s1">breaks</span><span class="pl-c1">=</span>[[<span class="pl-c1">2</span><span class="pl-c1">*</span><span class="pl-s1">np</span>.<span class="pl-s1">pi</span>, <span class="pl-c1">10</span><span class="pl-c1">*</span><span class="pl-s1">np</span>.<span class="pl-s1">pi</span>]])</pre></div> | 1 |
<p dir="auto"><strong>I'm submitting a ...</strong> (check one with "x")</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="[x] bug report => search github for a similar issue or PR before submitting
[ ] feature request
[ ] support request => Please do not submit support request here, instead see https://github.com/angular/angular/blob/master/CONTRIBUTING.md#question"><pre class="notranslate"><code class="notranslate">[x] bug report => search github for a similar issue or PR before submitting
[ ] feature request
[ ] support request => Please do not submit support request here, instead see https://github.com/angular/angular/blob/master/CONTRIBUTING.md#question
</code></pre></div>
<p dir="auto"><strong>Current behavior</strong></p>
<p dir="auto">canDeactivate guard is not called for child components when params change in parent route.</p>
<p dir="auto">Given that I have a general layout for a page if I have a user param, When I change the user while maintaining the subroute for the UserForm, canDeactivate is not called.</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="{ path: ':id', component: UserLayout,
children: [
{ path: 'edit', component: UserForm, canDeactivate: [CanDeactivateUserForm] }
{ path: 'favorites', ... },
{ path: 'commments', ... }
]
}"><pre class="notranslate"><code class="notranslate">{ path: ':id', component: UserLayout,
children: [
{ path: 'edit', component: UserForm, canDeactivate: [CanDeactivateUserForm] }
{ path: 'favorites', ... },
{ path: 'commments', ... }
]
}
</code></pre></div>
<p dir="auto">If I put the <code class="notranslate">canDeactivate</code> on the <code class="notranslate">:id</code> path, the method is called, however the <code class="notranslate">component</code> I receive as the first parameter in the <code class="notranslate">canDeactivate</code> method is not the <code class="notranslate">UserForm</code> component, but the <code class="notranslate">UserLayout</code>.</p>
<p dir="auto"><strong>Expected behavior</strong></p>
<p dir="auto">Expected behavior would be either to implement a <code class="notranslate">canDeactivateChild</code> similar to <code class="notranslate">canActivateChild</code> that is called on the subroutes if the parent route changes or have the children's <code class="notranslate">canDeactivate</code> calls trigger when a parent route changes.</p>
<p dir="auto">Another alternative would be to make its usage similar to the way it worked before RC.5, where <code class="notranslate">canDeactivate</code> can be directly implemented on the <code class="notranslate">UserForm</code> class.</p>
<p dir="auto"><strong>Minimal reproduction of the problem with instructions</strong></p>
<p dir="auto"><a href="https://plnkr.co/edit/zxFODbsGl8pL36K8CYUy?p=preview" rel="nofollow">Here is a minimal plunker demonstrating the issue</a>. When you navigate to the a user, the <code class="notranslate">EditInfo</code> form is currently the only subroute of <code class="notranslate">UserDetail</code> (for brevety), but you could imagine other routes. The <code class="notranslate">canDeactivate</code> is attached to this route but when you click on another user, we want to still maintain the editing page, however, <code class="notranslate">canDeactivate</code> is not called.</p>
<p dir="auto"><strong>What is the motivation / use case for changing the behavior?</strong></p>
<p dir="auto">So we can check with the user before navigating away from an unfinished form.</p>
<p dir="auto"><strong>Please tell us about your environment:</strong></p>
<p dir="auto">MacOS 10.10/Linux Ubuntu, VSCode, npm, rails server</p>
<ul dir="auto">
<li><strong>Angular version:</strong> 2.0.0</li>
<li><strong>Browser:</strong> all</li>
<li><strong>Language:</strong> all</li>
</ul> | <p dir="auto"><strong>I'm submitting a ...</strong> (check one with "x")</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="[ ] bug report => search github for a similar issue or PR before submitting
[x] feature request
[ ] support request => Please do not submit support request here, instead see https://github.com/angular/angular/blob/master/CONTRIBUTING.md#question"><pre class="notranslate"><code class="notranslate">[ ] bug report => search github for a similar issue or PR before submitting
[x] feature request
[ ] support request => Please do not submit support request here, instead see https://github.com/angular/angular/blob/master/CONTRIBUTING.md#question
</code></pre></div>
<p dir="auto"><strong>What is the motivation / use case for changing the behavior?</strong><br>
There is already a canActivateChild guard. Having the symmetric guard is equally important. Not sure why this wasn't added at the same time. We have many scenarios where we need to protect child routes from navigating away if they have pending unsaved changes. That's currently only possible by putting a canDeactivate guard on each child route.</p>
<ul dir="auto">
<li><strong>Angular version:</strong> Latest</li>
<li><strong>Browser:</strong> all</li>
<li><strong>Language:</strong> all</li>
</ul> | 1 |
<p dir="auto">xref comment in <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="296639568" data-permission-text="Title is private" data-url="https://github.com/pandas-dev/pandas/issues/19673" data-hovercard-type="pull_request" data-hovercard-url="/pandas-dev/pandas/pull/19673/hovercard" href="https://github.com/pandas-dev/pandas/pull/19673">#19673</a>, we don't have any ASV benchmarks in <code class="notranslate">groupby.GroupByMethods</code> to measure performance on datetimelike types (timestamp / timedelta)</p> | <p dir="auto">Hello,</p>
<p dir="auto">Series doesn't have is a <code class="notranslate">is_unique</code> attribute</p>
<p dir="auto">Index have a <code class="notranslate">is_unique</code> attribute but not (values of a) Series</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content=">>> df.index.is_unique
True
>>> df['Column1'].is_unique
---------------------------------------------------------------------------
AttributeError Traceback (most recent call last)
<ipython-input-417-c450d4c4ceda> in <module>()
----> 1 df['Date'].is_unique
//anaconda/envs/py27/lib/python2.7/site-packages/pandas/core/generic.pyc in __getattr__(self, name)
2358 return self[name]
2359 raise AttributeError("'%s' object has no attribute '%s'" %
-> 2360 (type(self).__name__, name))
2361
2362 def __setattr__(self, name, value):
AttributeError: 'Series' object has no attribute 'is_unique'"><pre class="notranslate"><span class="pl-c1">>></span><span class="pl-c1">></span> <span class="pl-s1">df</span>.<span class="pl-s1">index</span>.<span class="pl-s1">is_unique</span>
<span class="pl-c1">True</span>
<span class="pl-c1">>></span><span class="pl-c1">></span> <span class="pl-s1">df</span>[<span class="pl-s">'Column1'</span>].<span class="pl-s1">is_unique</span>
<span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span>
<span class="pl-v">AttributeError</span> <span class="pl-v">Traceback</span> (<span class="pl-s1">most</span> <span class="pl-s1">recent</span> <span class="pl-s1">call</span> <span class="pl-s1">last</span>)
<span class="pl-c1"><</span><span class="pl-s1">ipython</span><span class="pl-c1">-</span><span class="pl-s1">input</span><span class="pl-c1">-</span><span class="pl-c1">417</span><span class="pl-c1">-</span><span class="pl-s1">c450d4c4ceda</span><span class="pl-c1">></span> <span class="pl-s1">in</span> <span class="pl-c1"><</span><span class="pl-s1">module</span><span class="pl-c1">></span>()
<span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">-</span><span class="pl-c1">></span> <span class="pl-c1">1</span> <span class="pl-s1">df</span>[<span class="pl-s">'Date'</span>].<span class="pl-s1">is_unique</span>
<span class="pl-c1">//</span><span class="pl-s1">anaconda</span><span class="pl-c1">/</span><span class="pl-s1">envs</span><span class="pl-c1">/</span><span class="pl-s1">py27</span><span class="pl-c1">/</span><span class="pl-s1">lib</span><span class="pl-c1">/</span><span class="pl-s1">python2</span>.<span class="pl-c1">7</span><span class="pl-c1">/</span><span class="pl-s1">site</span><span class="pl-c1">-</span><span class="pl-s1">packages</span><span class="pl-c1">/</span><span class="pl-s1">pandas</span><span class="pl-c1">/</span><span class="pl-s1">core</span><span class="pl-c1">/</span><span class="pl-s1">generic</span>.<span class="pl-s1">pyc</span> <span class="pl-c1">in</span> <span class="pl-en">__getattr__</span>(<span class="pl-s1">self</span>, <span class="pl-s1">name</span>)
<span class="pl-c1">2358</span> <span class="pl-k">return</span> <span class="pl-s1">self</span>[<span class="pl-s1">name</span>]
<span class="pl-c1">2359</span> <span class="pl-k">raise</span> <span class="pl-v">AttributeError</span>(<span class="pl-s">"'%s' object has no attribute '%s'"</span> <span class="pl-c1">%</span>
<span class="pl-c1">-</span><span class="pl-c1">></span> <span class="pl-c1">2360</span> (<span class="pl-en">type</span>(<span class="pl-s1">self</span>).<span class="pl-s1">__name__</span>, <span class="pl-s1">name</span>))
<span class="pl-c1">2361</span>
<span class="pl-c1">2362</span> <span class="pl-k">def</span> <span class="pl-en">__setattr__</span>(<span class="pl-s1">self</span>, <span class="pl-s1">name</span>, <span class="pl-s1">value</span>):
<span class="pl-v">AttributeError</span>: <span class="pl-s">'Series'</span> <span class="pl-s1">object</span> <span class="pl-s1">has</span> <span class="pl-s1">no</span> <span class="pl-s1">attribute</span> <span class="pl-s">'is_unique'</span></pre></div>
<p dir="auto">I'm doing this</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="def is_unique(ser):
return len(np.unique(ser.values)) == len(ser.values)"><pre class="notranslate"><span class="pl-k">def</span> <span class="pl-en">is_unique</span>(<span class="pl-s1">ser</span>):
<span class="pl-k">return</span> <span class="pl-en">len</span>(<span class="pl-s1">np</span>.<span class="pl-en">unique</span>(<span class="pl-s1">ser</span>.<span class="pl-s1">values</span>)) <span class="pl-c1">==</span> <span class="pl-en">len</span>(<span class="pl-s1">ser</span>.<span class="pl-s1">values</span>)</pre></div>
<p dir="auto">but this kind of attribute could be useful when you want to set a column as index and ensure that there is no duplicate keys.</p>
<p dir="auto">PS: I know there is</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content=" df.set_index('Column1', verify_integrity=True)"><pre class="notranslate"> <span class="pl-s1">df</span>.<span class="pl-en">set_index</span>(<span class="pl-s">'Column1'</span>, <span class="pl-s1">verify_integrity</span><span class="pl-c1">=</span><span class="pl-c1">True</span>)</pre></div> | 0 |
<p dir="auto">Looking at the docs, many fields and structs are under-specified. At some point (soon) we need to do an audit/scrub of API per-field and per-struct comments (which generate docs).</p>
<ul dir="auto">
<li>What is this field?</li>
<li>Is this field required? (omitempty is sort of a bad way IMO)</li>
<li>What happens if this field is not specified?</li>
<li>What are valid values or bounds for this field?</li>
<li>What happens if you specify an out-of-bounds value?</li>
<li>Does this field have a literal default value (e.g. 42), a derived default value (e.g. same as another field), or no default value (the system will choose a value)?</li>
<li>Is this field updateable? What happens if you update it?</li>
</ul> | <ul dir="auto">
<li>
<h1 dir="auto">10933 removes the description tag in types.go, the description of fields are merged with other comments as the new description. Some of these comments don't read as a sentence after the merge. Because our api-reference <a href="http://kubernetes.io/v1.0/docs/api-reference/definitions.html" rel="nofollow">docs</a> use these descriptions, we need to fix them before <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="93842482" data-permission-text="Title is private" data-url="https://github.com/kubernetes/kubernetes/issues/10933" data-hovercard-type="pull_request" data-hovercard-url="/kubernetes/kubernetes/pull/10933/hovercard" href="https://github.com/kubernetes/kubernetes/pull/10933">#10933</a>.</h1>
</li>
<li>
<h1 dir="auto">10933 detects these <a href="https://github.com/GoogleCloudPlatform/kubernetes/pull/10933#issuecomment-125270576" data-hovercard-type="pull_request" data-hovercard-url="/kubernetes/kubernetes/pull/10933/hovercard">structs</a> don't have any description. We need to add description for them.</h1>
</li>
</ul>
<p dir="auto"><a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/andronat/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/andronat">@andronat</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/nikhiljindal/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/nikhiljindal">@nikhiljindal</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/bgrant0607/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/bgrant0607">@bgrant0607</a></p> | 1 |
<p dir="auto">We should add support to save files that require admin access. The way Sublime Text does it is when you hit save, there is a dialog prompt for your password.</p>
<p dir="auto">Currently this is the error thrown:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="Uncaught Error: EACCES, permission denied '/private/etc/apache2/extra/httpd-vhosts.conf' - fs.js"><pre class="notranslate"><code class="notranslate">Uncaught Error: EACCES, permission denied '/private/etc/apache2/extra/httpd-vhosts.conf' - fs.js
</code></pre></div> | <p dir="auto">I edit my <code class="notranslate">/etc/hosts</code> file a lot to work with Enterprise VMs. My typical workflow for that is to run <code class="notranslate">subl /etc/hosts</code>, edit the file, <code class="notranslate">cmd-s</code>, and then I get prompted for my password since it's trying to save and my user doesn't have permission to.</p>
<p dir="auto">In Atom, it opens the file just fine, but when I try saving the dev tools pane pops up and I see this in the javascript console:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="Uncaught Error: You don’t have permission to save the file “hosts” in the folder “etc”."><pre class="notranslate"><code class="notranslate">Uncaught Error: You don’t have permission to save the file “hosts” in the folder “etc”.
</code></pre></div>
<p dir="auto">It would be cool if we had support for nicely handling cases where you try saving files that require privilege escalation to write.</p> | 1 |
<p dir="auto">In some hosting environments/configurations, the network traffic between node and master may traverse the public Internet. As a result, we'd like to secure the communication between the node components (e.g. kubelet and proxy) and master. To avoid the complexity of securing the kubelet API, we'd like to secure the node -> master communication, but not the reverse. This simplification has a downside; it means all communication between kubelet and master would have to be initiated by the kubelet. For example, we'd have to change health checks to be initiated by the kubelet, which in turn raises a question of how to do flow control (master apply backpressure when it becomes overloaded).</p> | <p dir="auto"><strong>Is this a request for help?</strong> (If yes, you should use our troubleshooting guide and community support channels, see <a href="http://kubernetes.io/docs/troubleshooting/.):NO" rel="nofollow">http://kubernetes.io/docs/troubleshooting/.):NO</a></p>
<p dir="auto"><strong>What keywords did you search in Kubernetes issues before filing this one?</strong> (If you have found any duplicates, you should instead reply there.): <code class="notranslate">is:issue is:open zsh completion</code></p>
<hr>
<p dir="auto"><strong>Is this a BUG REPORT or FEATURE REQUEST?</strong> (choose one):</p>
<p dir="auto"><strong>Kubernetes version</strong> (use <code class="notranslate">kubectl version</code>):</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="% kubectl version
Client Version: version.Info{Major:"1", Minor:"3", GitVersion:"v1.3.7+a2cba278", GitCommit:"a2cba278cba1f6881bb0a7704d9cac6fca6ed435", GitTreeState:"not a git tree", BuildDate:"2016-09-23T02:19:56Z", GoVersion:"go1.7.1", Compiler:"gc", Platform:"darwin/amd64"}"><pre class="notranslate"><code class="notranslate">% kubectl version
Client Version: version.Info{Major:"1", Minor:"3", GitVersion:"v1.3.7+a2cba278", GitCommit:"a2cba278cba1f6881bb0a7704d9cac6fca6ed435", GitTreeState:"not a git tree", BuildDate:"2016-09-23T02:19:56Z", GoVersion:"go1.7.1", Compiler:"gc", Platform:"darwin/amd64"}
</code></pre></div>
<p dir="auto"><strong>Environment</strong>:</p>
<ul dir="auto">
<li><strong>Cloud provider or hardware configuration</strong>: MacBook Pro</li>
<li><strong>OS</strong> (e.g. from /etc/os-release): macOS</li>
<li><strong>Kernel</strong> (e.g. <code class="notranslate">uname -a</code>): Sierra</li>
<li><strong>Install tools</strong>: homebrew</li>
<li><strong>Others</strong>:</li>
</ul>
<p dir="auto"><strong>What happened</strong>:<br>
If you source the kubectl complete for zsh like so:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="$ source <(kubectl completion zsh)"><pre class="notranslate"><code class="notranslate">$ source <(kubectl completion zsh)
</code></pre></div>
<p dir="auto">This breaks zsh vi history search.</p>
<p dir="auto"><strong>What you expected to happen</strong>:<br>
ZSH vi history to work as normal</p>
<p dir="auto"><strong>How to reproduce it</strong> (as minimally and precisely as possible):</p>
<p dir="auto">In <code class="notranslate">.zshrc</code>, enable the <code class="notranslate">vi-mode</code> plugin from <em>oh-my-zsh</em><br>
This should allow to hit the <code class="notranslate">ESC-/</code> to perform a history search.</p>
<p dir="auto">Without source the <code class="notranslate">kubectl</code> completion, the command works normally.<br>
After sourcing the <code class="notranslate">kubectl</code> completion, zsh prompts:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="zsh: do you wish to see all 16016 possibilities (16144 lines)?"><pre class="notranslate"><code class="notranslate">zsh: do you wish to see all 16016 possibilities (16144 lines)?
</code></pre></div>
<p dir="auto">Select 'n' for NO and drops back to the prompt.</p>
<p dir="auto"><strong>Anything else do we need to know</strong>:<br>
It seems like <code class="notranslate">kubectl</code> completion is conflicting with <code class="notranslate">vi-mode</code> plugin potentially overwriting the previous configuration.</p>
<p dir="auto">If I source <code class="notranslate">vi-mode</code> after <code class="notranslate">kubectl</code>, everything works as expected.</p> | 0 |
<p dir="auto">Hi!</p>
<p dir="auto">Is there any reason why running plugins inside atom-shell won't work, on principle?</p>
<p dir="auto">I know it will never be a primary goal of atom-shell, but I'm trying to run a silverlight application (sigh) inside atom-shell.</p>
<p dir="auto">I had to enable plugins on WebPreferences, and I'm still running into some issues, but I wanted to know if there was an architectural reason, because of how things are structured, that just made it impossible to run plugins successfully over atom-shell.</p>
<p dir="auto">I'm well aware of the end of support (in the very near future) for plugins on chromium, but even so, we are already running using CEF. It requires us to build different shells for OSX and Win, which I hope atom-shell might free us by letting us write all the OS dependent stuff on JS instead of objC/WPF (and provide us a way to start migrating code incrementally away from SL).</p>
<p dir="auto">Thanks for any input!</p> | 0 |
|
<h3 dir="auto">Description</h3>
<p dir="auto"><code class="notranslate">apache-airflow-providers-google</code> has a lot of dependencies, some pinned to pretty old versions compared to <a href="https://github.com/googleapis/google-cloud-python#libraries">current release versions</a>. In particular, many of these older versions requires <code class="notranslate">protobuf<4.0.0dev</code> whereas if you already have newer versions (say google-cloud-bigquery==3.7.0) it requires newer versions of protobuf and causes incompatibility</p>
<h3 dir="auto">Use case/motivation</h3>
<p dir="auto">Installing <code class="notranslate">dbt-bigquery</code> then <code class="notranslate">apache-airflow-providers-google</code> takes forever to resolve dependencies.</p>
<p dir="auto">Installing <code class="notranslate">apache-airflow-providers-google</code> then <code class="notranslate">dbt-bigquery</code> leads to dependency errors:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
google-cloud-aiplatform 1.16.1 requires google-cloud-bigquery<3.0.0dev,>=1.15.0, but you have google-cloud-bigquery 3.7.0 which is incompatible.
google-ads 18.0.0 requires protobuf!=3.18.*,!=3.19.*,<=3.20.0,>=3.12.0, but you have protobuf 3.20.3 which is incompatible."><pre class="notranslate"><code class="notranslate">ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
google-cloud-aiplatform 1.16.1 requires google-cloud-bigquery<3.0.0dev,>=1.15.0, but you have google-cloud-bigquery 3.7.0 which is incompatible.
google-ads 18.0.0 requires protobuf!=3.18.*,!=3.19.*,<=3.20.0,>=3.12.0, but you have protobuf 3.20.3 which is incompatible.
</code></pre></div>
<h3 dir="auto">Related issues</h3>
<p dir="auto"><em>No response</em></p>
<h3 dir="auto">Are you willing to submit a PR?</h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> Yes I am willing to submit a PR!</li>
</ul>
<h3 dir="auto">Code of Conduct</h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I agree to follow this project's <a href="https://github.com/apache/airflow/blob/main/CODE_OF_CONDUCT.md">Code of Conduct</a></li>
</ul> | <h3 dir="auto">Apache Airflow Provider(s)</h3>
<p dir="auto">amazon</p>
<h3 dir="auto">Versions of Apache Airflow Providers</h3>
<p dir="auto">apache-airflow-providers-amazon==7.2.0</p>
<h3 dir="auto">Apache Airflow version</h3>
<p dir="auto">2.4.3</p>
<h3 dir="auto">Operating System</h3>
<p dir="auto">Arch Linux</p>
<h3 dir="auto">Deployment</h3>
<p dir="auto">Docker-Compose</p>
<h3 dir="auto">Deployment details</h3>
<p dir="auto">the same for MWAA (aws-managed airflow)</p>
<h3 dir="auto">What happened</h3>
<p dir="auto">exception is raised:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="[2023-02-20, 14:32:02 UTC] {subprocess.py:92} INFO - [�[34m2023-02-20, 14:32:02 UTC�[0m] {�[34mconnectionpool.py:�[0m475} WARNING�[0m - Failed to parse headers url=[https://BUCKET.s3.us-west-2.amazonaws.com:443/object-key.json:[NoBoundaryInMultipartDefect()], unparsed data: ''�[0m
[2023-02-20, 14:32:02 UTC] {subprocess.py:92} INFO - Traceback (most recent call last):
[2023-02-20, 14:32:02 UTC] {subprocess.py:92} INFO - File "/home/***/.local/lib/python3.7/site-packages/urllib3/connectionpool.py", line 469, in _make_request
[2023-02-20, 14:32:02 UTC] {subprocess.py:92} INFO - assert_header_parsing(httplib_response.msg)
[2023-02-20, 14:32:02 UTC] {subprocess.py:92} INFO - File "/home/***/.local/lib/python3.7/site-packages/urllib3/util/response.py", line 91, in assert_header_parsing
[2023-02-20, 14:32:02 UTC] {subprocess.py:92} INFO - raise HeaderParsingError(defects=defects, unparsed_data=unparsed_data)
[2023-02-20, 14:32:02 UTC] {subprocess.py:92} INFO - urllib3.exceptions.HeaderParsingError: [NoBoundaryInMultipartDefect()], unparsed data: ''"><pre class="notranslate"><code class="notranslate">[2023-02-20, 14:32:02 UTC] {subprocess.py:92} INFO - [�[34m2023-02-20, 14:32:02 UTC�[0m] {�[34mconnectionpool.py:�[0m475} WARNING�[0m - Failed to parse headers url=[https://BUCKET.s3.us-west-2.amazonaws.com:443/object-key.json:[NoBoundaryInMultipartDefect()], unparsed data: ''�[0m
[2023-02-20, 14:32:02 UTC] {subprocess.py:92} INFO - Traceback (most recent call last):
[2023-02-20, 14:32:02 UTC] {subprocess.py:92} INFO - File "/home/***/.local/lib/python3.7/site-packages/urllib3/connectionpool.py", line 469, in _make_request
[2023-02-20, 14:32:02 UTC] {subprocess.py:92} INFO - assert_header_parsing(httplib_response.msg)
[2023-02-20, 14:32:02 UTC] {subprocess.py:92} INFO - File "/home/***/.local/lib/python3.7/site-packages/urllib3/util/response.py", line 91, in assert_header_parsing
[2023-02-20, 14:32:02 UTC] {subprocess.py:92} INFO - raise HeaderParsingError(defects=defects, unparsed_data=unparsed_data)
[2023-02-20, 14:32:02 UTC] {subprocess.py:92} INFO - urllib3.exceptions.HeaderParsingError: [NoBoundaryInMultipartDefect()], unparsed data: ''
</code></pre></div>
<h3 dir="auto">What you think should happen instead</h3>
<p dir="auto">shouldn't be such an exception :)</p>
<h3 dir="auto">How to reproduce</h3>
<p dir="auto">the code that downloads data is simple:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="
def download_from_s3(key: str, bucket_name: str, local_path: str) -> str:
boto3.set_stream_logger('boto3.resources', logging.DEBUG)
hook = S3Hook(aws_conn_id='s3_conn')
file_name = hook.download_file(key=key, bucket_name=bucket_name, preserve_file_name=True)
return file_name
"><pre class="notranslate"><code class="notranslate">
def download_from_s3(key: str, bucket_name: str, local_path: str) -> str:
boto3.set_stream_logger('boto3.resources', logging.DEBUG)
hook = S3Hook(aws_conn_id='s3_conn')
file_name = hook.download_file(key=key, bucket_name=bucket_name, preserve_file_name=True)
return file_name
</code></pre></div>
<h3 dir="auto">Anything else</h3>
<p dir="auto">anyway, file is downldaed and looks valid.</p>
<p dir="auto">some logs:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="[�[34m2023-02-20, 15:18:38 UTC�[0m] {�[34mconnection_wrapper.py:�[0m337} INFO�[0m - AWS Connection (conn_id='s3_conn', conn_type='aws') credentials retrieved from login and password.�[0m
2023-02-20, 15:18:38 UTC boto3.resources.factory [DEBUG] Loading s3:s3
[�[34m2023-02-20, 15:18:38 UTC�[0m] {�[34mfactory.py:�[0m66} DEBUG�[0m - Loading s3:s3�[0m
2023-02-20, 15:18:38 UTC boto3.resources.factory [DEBUG] Loading s3:Object
[�[34m2023-02-20, 15:18:38 UTC�[0m] {�[34mfactory.py:�[0m66} DEBUG�[0m - Loading s3:Object�[0m
2023-02-20, 15:18:38 UTC boto3.resources.action [DEBUG] Calling s3:head_object with {'Bucket': 'BUCKET', 'Key': 'object_key.json'}
[�[34m2023-02-20, 15:18:38 UTC�[0m] {�[34maction.py:�[0m85} DEBUG�[0m - Calling s3:head_object with {'Bucket': 'BUCKET', 'Key': 'object_key.json'}�[0m
[�[34m2023-02-20, 15:18:40 UTC�[0m] {�[34mconnectionpool.py:�[0m475} WARNING�[0m - Failed to parse headers (url=https://BUCKET.s3.us-west-2.amazonaws.com:443/object_key.json): [NoBoundaryInMultipartDefect()], unparsed data: ''�[0m
Traceback (most recent call last):
File "/home/***/.local/lib/python3.7/site-packages/urllib3/connectionpool.py", line 469, in _make_request
assert_header_parsing(httplib_response.msg)
File "/home/***/.local/lib/python3.7/site-packages/urllib3/util/response.py", line 91, in assert_header_parsing
raise HeaderParsingError(defects=defects, unparsed_data=unparsed_data)
urllib3.exceptions.HeaderParsingError: [NoBoundaryInMultipartDefect()], unparsed data: ''
2023-02-20, 15:18:40 UTC boto3.resources.action [DEBUG] Response: {'ResponseMetadata': {'RequestId': 'W3J4VRW3WQVV8AV7', 'HostId': 'uRLn/mC6mUAPtgAZRcPbdIlkzWNQ8/AKuPn5HuHjJK1CLNAxfES3DXQsnF7HYSia4guuylFLItY=', 'HTTPStatusCode': 200, 'HTTPHeaders': {'x-amz-id-2': 'uRLn/mC6mUAPtgAZRcPbdIlkzWNQ8/AKuPn5HuHjJK1CLNAxfES3DXQsnF7HYSia4guuylFLItY=', 'x-amz-request-id': 'W3J4VRW3WQVV8AV7', 'date': 'Mon, 20 Feb 2023 15:18:40 GMT', 'last-modified': 'Thu, 09 Feb 2023 10:34:28 GMT', 'etag': '"e7d2a315e24716624b1085cfa7f31ad8"', 'x-amz-server-side-encryption': 'AES256', 'accept-ranges': 'bytes', 'content-type': 'multipart/form-data', 'server': 'AmazonS3', 'content-length': '7004'}, 'RetryAttempts': 0}, 'AcceptRanges': 'bytes', 'LastModified': datetime.datetime(2023, 2, 9, 10, 34, 28, tzinfo=tzutc()), 'ContentLength': 7004, 'ETag': '"e7d2a315e24716624b1085cfa7f31ad8"', 'ContentType': 'multipart/form-data', 'ServerSideEncryption': 'AES256', 'Metadata': {}}
[�[34m2023-02-20, 15:18:40 UTC�[0m] {�[34maction.py:�[0m90} DEBUG�[0m - Response: {'ResponseMetadata': {'RequestId': 'W3J4VRW3WQVV8AV7', 'HostId': 'uRLn/mC6mUAPtgAZRcPbdIlkzWNQ8/AKuPn5HuHjJK1CLNAxfES3DXQsnF7HYSia4guuylFLItY=', 'HTTPStatusCode': 200, 'HTTPHeaders': {'x-amz-id-2': 'uRLn/mC6mUAPtgAZRcPbdIlkzWNQ8/AKuPn5HuHjJK1CLNAxfES3DXQsnF7HYSia4guuylFLItY=', 'x-amz-request-id': 'W3J4VRW3WQVV8AV7', 'date': 'Mon, 20 Feb 2023 15:18:40 GMT', 'last-modified': 'Thu, 09 Feb 2023 10:34:28 GMT', 'etag': '"e7d2a315e24716624b1085cfa7f31ad8"', 'x-amz-server-side-encryption': 'AES256', 'accept-ranges': 'bytes', 'content-type': 'multipart/form-data', 'server': 'AmazonS3', 'content-length': '7004'}, 'RetryAttempts': 0}, 'AcceptRanges': 'bytes', 'LastModified': datetime.datetime(2023, 2, 9, 10, 34, 28, tzinfo=tzutc()), 'ContentLength': 7004, 'ETag': '"e7d2a315e24716624b1085cfa7f31ad8"', 'ContentType': 'multipart/form-data', 'ServerSideEncryption': 'AES256', 'Metadata': {}}�[0m
[�[34m2023-02-20, 15:18:40 UTC�[0m] {�[34mconnectionpool.py:�[0m475} WARNING�[0m - Failed to parse headers (url=https://BUCKET.s3.us-west-2.amazonaws.com:443/object_key.json): [NoBoundaryInMultipartDefect()], unparsed data: ''�[0m
Traceback (most recent call last):
File "/home/***/.local/lib/python3.7/site-packages/urllib3/connectionpool.py", line 469, in _make_request
assert_header_parsing(httplib_response.msg)
File "/home/***/.local/lib/python3.7/site-packages/urllib3/util/response.py", line 91, in assert_header_parsing
raise HeaderParsingError(defects=defects, unparsed_data=unparsed_data)
urllib3.exceptions.HeaderParsingError: [NoBoundaryInMultipartDefect()], unparsed data: ''
[�[34m2023-02-20, 15:18:40 UTC�[0m] {�[34mconnectionpool.py:�[0m475} WARNING�[0m - Failed to parse headers (url=https://BUCKET.s3.us-west-2.amazonaws.com:443/object_key.json): [NoBoundaryInMultipartDefect()], unparsed data: ''�[0m
Traceback (most recent call last):
File "/home/***/.local/lib/python3.7/site-packages/urllib3/connectionpool.py", line 469, in _make_request
assert_header_parsing(httplib_response.msg)
File "/home/***/.local/lib/python3.7/site-packages/urllib3/util/response.py", line 91, in assert_header_parsing
raise HeaderParsingError(defects=defects, unparsed_data=unparsed_data)
urllib3.exceptions.HeaderParsingError: [NoBoundaryInMultipartDefect()], unparsed data: ''"><pre class="notranslate"><code class="notranslate">[�[34m2023-02-20, 15:18:38 UTC�[0m] {�[34mconnection_wrapper.py:�[0m337} INFO�[0m - AWS Connection (conn_id='s3_conn', conn_type='aws') credentials retrieved from login and password.�[0m
2023-02-20, 15:18:38 UTC boto3.resources.factory [DEBUG] Loading s3:s3
[�[34m2023-02-20, 15:18:38 UTC�[0m] {�[34mfactory.py:�[0m66} DEBUG�[0m - Loading s3:s3�[0m
2023-02-20, 15:18:38 UTC boto3.resources.factory [DEBUG] Loading s3:Object
[�[34m2023-02-20, 15:18:38 UTC�[0m] {�[34mfactory.py:�[0m66} DEBUG�[0m - Loading s3:Object�[0m
2023-02-20, 15:18:38 UTC boto3.resources.action [DEBUG] Calling s3:head_object with {'Bucket': 'BUCKET', 'Key': 'object_key.json'}
[�[34m2023-02-20, 15:18:38 UTC�[0m] {�[34maction.py:�[0m85} DEBUG�[0m - Calling s3:head_object with {'Bucket': 'BUCKET', 'Key': 'object_key.json'}�[0m
[�[34m2023-02-20, 15:18:40 UTC�[0m] {�[34mconnectionpool.py:�[0m475} WARNING�[0m - Failed to parse headers (url=https://BUCKET.s3.us-west-2.amazonaws.com:443/object_key.json): [NoBoundaryInMultipartDefect()], unparsed data: ''�[0m
Traceback (most recent call last):
File "/home/***/.local/lib/python3.7/site-packages/urllib3/connectionpool.py", line 469, in _make_request
assert_header_parsing(httplib_response.msg)
File "/home/***/.local/lib/python3.7/site-packages/urllib3/util/response.py", line 91, in assert_header_parsing
raise HeaderParsingError(defects=defects, unparsed_data=unparsed_data)
urllib3.exceptions.HeaderParsingError: [NoBoundaryInMultipartDefect()], unparsed data: ''
2023-02-20, 15:18:40 UTC boto3.resources.action [DEBUG] Response: {'ResponseMetadata': {'RequestId': 'W3J4VRW3WQVV8AV7', 'HostId': 'uRLn/mC6mUAPtgAZRcPbdIlkzWNQ8/AKuPn5HuHjJK1CLNAxfES3DXQsnF7HYSia4guuylFLItY=', 'HTTPStatusCode': 200, 'HTTPHeaders': {'x-amz-id-2': 'uRLn/mC6mUAPtgAZRcPbdIlkzWNQ8/AKuPn5HuHjJK1CLNAxfES3DXQsnF7HYSia4guuylFLItY=', 'x-amz-request-id': 'W3J4VRW3WQVV8AV7', 'date': 'Mon, 20 Feb 2023 15:18:40 GMT', 'last-modified': 'Thu, 09 Feb 2023 10:34:28 GMT', 'etag': '"e7d2a315e24716624b1085cfa7f31ad8"', 'x-amz-server-side-encryption': 'AES256', 'accept-ranges': 'bytes', 'content-type': 'multipart/form-data', 'server': 'AmazonS3', 'content-length': '7004'}, 'RetryAttempts': 0}, 'AcceptRanges': 'bytes', 'LastModified': datetime.datetime(2023, 2, 9, 10, 34, 28, tzinfo=tzutc()), 'ContentLength': 7004, 'ETag': '"e7d2a315e24716624b1085cfa7f31ad8"', 'ContentType': 'multipart/form-data', 'ServerSideEncryption': 'AES256', 'Metadata': {}}
[�[34m2023-02-20, 15:18:40 UTC�[0m] {�[34maction.py:�[0m90} DEBUG�[0m - Response: {'ResponseMetadata': {'RequestId': 'W3J4VRW3WQVV8AV7', 'HostId': 'uRLn/mC6mUAPtgAZRcPbdIlkzWNQ8/AKuPn5HuHjJK1CLNAxfES3DXQsnF7HYSia4guuylFLItY=', 'HTTPStatusCode': 200, 'HTTPHeaders': {'x-amz-id-2': 'uRLn/mC6mUAPtgAZRcPbdIlkzWNQ8/AKuPn5HuHjJK1CLNAxfES3DXQsnF7HYSia4guuylFLItY=', 'x-amz-request-id': 'W3J4VRW3WQVV8AV7', 'date': 'Mon, 20 Feb 2023 15:18:40 GMT', 'last-modified': 'Thu, 09 Feb 2023 10:34:28 GMT', 'etag': '"e7d2a315e24716624b1085cfa7f31ad8"', 'x-amz-server-side-encryption': 'AES256', 'accept-ranges': 'bytes', 'content-type': 'multipart/form-data', 'server': 'AmazonS3', 'content-length': '7004'}, 'RetryAttempts': 0}, 'AcceptRanges': 'bytes', 'LastModified': datetime.datetime(2023, 2, 9, 10, 34, 28, tzinfo=tzutc()), 'ContentLength': 7004, 'ETag': '"e7d2a315e24716624b1085cfa7f31ad8"', 'ContentType': 'multipart/form-data', 'ServerSideEncryption': 'AES256', 'Metadata': {}}�[0m
[�[34m2023-02-20, 15:18:40 UTC�[0m] {�[34mconnectionpool.py:�[0m475} WARNING�[0m - Failed to parse headers (url=https://BUCKET.s3.us-west-2.amazonaws.com:443/object_key.json): [NoBoundaryInMultipartDefect()], unparsed data: ''�[0m
Traceback (most recent call last):
File "/home/***/.local/lib/python3.7/site-packages/urllib3/connectionpool.py", line 469, in _make_request
assert_header_parsing(httplib_response.msg)
File "/home/***/.local/lib/python3.7/site-packages/urllib3/util/response.py", line 91, in assert_header_parsing
raise HeaderParsingError(defects=defects, unparsed_data=unparsed_data)
urllib3.exceptions.HeaderParsingError: [NoBoundaryInMultipartDefect()], unparsed data: ''
[�[34m2023-02-20, 15:18:40 UTC�[0m] {�[34mconnectionpool.py:�[0m475} WARNING�[0m - Failed to parse headers (url=https://BUCKET.s3.us-west-2.amazonaws.com:443/object_key.json): [NoBoundaryInMultipartDefect()], unparsed data: ''�[0m
Traceback (most recent call last):
File "/home/***/.local/lib/python3.7/site-packages/urllib3/connectionpool.py", line 469, in _make_request
assert_header_parsing(httplib_response.msg)
File "/home/***/.local/lib/python3.7/site-packages/urllib3/util/response.py", line 91, in assert_header_parsing
raise HeaderParsingError(defects=defects, unparsed_data=unparsed_data)
urllib3.exceptions.HeaderParsingError: [NoBoundaryInMultipartDefect()], unparsed data: ''
</code></pre></div>
<h3 dir="auto">Are you willing to submit PR?</h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> Yes I am willing to submit a PR!</li>
</ul>
<h3 dir="auto">Code of Conduct</h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I agree to follow this project's <a href="https://github.com/apache/airflow/blob/main/CODE_OF_CONDUCT.md">Code of Conduct</a></li>
</ul> | 0 |
<p dir="auto">Hi,</p>
<p dir="auto">The example <a href="https://github.com/fchollet/keras/blob/master/examples/cifar10_cnn.py">here </a> claims to go to 0.65 loss in 25 epochs. I tried out the code with both the theano and tensorflow backend and with the default hyper-parameters as is but it does not seem to go below 2.30 validation (and 0.1 accuracy) in 25 epochs.</p>
<p dir="auto">Is there anything that I am doing wrong and can anyone else reproduce this? (At least a friend of mine could)</p> | <p dir="auto">The goal is to implement data-parallel multi-GPU training with gradient-averaging properly in Keras (at least explicitly for TensorFlow backend).</p>
<p dir="auto">In this issue I'd like to discuss a particular approach which tries to fix problems of current solutions. Since Keras seems not to be designed for data-parallel SGD, I'm trying to find ways how to modify or adapt the Keras code, while keeping the API philosophy. Since this problem is quite important for many people, including our team at <a class="user-mention notranslate" data-hovercard-type="organization" data-hovercard-url="/orgs/rossumai/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/rossumai">@rossumai</a>, I'd like to ask for advice. Any feedback is really welcome.</p>
<h2 dir="auto">Quick outline of the data-parallel SGD algorithm</h2>
<p dir="auto">We use N identical model replicas (towers) to train on slices on a mini-batch. Model parameters are placed on a parameter server device (PS), CPU or one of the GPUs, computations are made on N worker devices. A minibatch of inputs is split into N sub-batches, distributed to each worker which computes the forward and backwards pass, the resulting gradients are sent to the PS device, averaged and used to update the weights, which are then copies back to the workers.</p>
<h2 dir="auto">Previous experiments</h2>
<p dir="auto">As a baseline I checked the <a href="https://www.tensorflow.org/tutorials/deep_cnn#training_a_model_using_multiple_gpu_cards" rel="nofollow">TensorFlow CIFAR 10 multi-GPU tutorial</a>. It worked as expected for 1-4 GPUs (TensorFlow 1.2.1, CUDA 8.0, GTX 1070).</p>
<p dir="auto">I tried the approach of <a href="https://github.com/kuza55/keras-extras">kuza55/keras-extras</a>, discussed earlier in other issues (<a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="149949934" data-permission-text="Title is private" data-url="https://github.com/keras-team/keras/issues/2436" data-hovercard-type="issue" data-hovercard-url="/keras-team/keras/issues/2436/hovercard" href="https://github.com/keras-team/keras/issues/2436">#2436</a>) and blog post <a href="https://medium.com/@kuza55/transparent-multi-gpu-training-on-tensorflow-with-keras-8b0016fd9012" rel="nofollow">Transparent Multi-GPU Training on TensorFlow with Keras</a>, adapting MNIST MLP and CIFAR10 Keras examples (Keras 2.0.6, TensorFlow 1.2.1, CUDA 8.0). In practice using more than one GPU lead to decrease of performance. Between 2 and 4 GPUs there was a 2x speedup, however.</p>
<p dir="auto"><a href="https://gist.github.com/bzamecnik/390d3802b31ce766fbf6fd6c9fd682d3">https://gist.github.com/bzamecnik/390d3802b31ce766fbf6fd6c9fd682d3</a></p>
<h2 dir="auto">Problems with kuza55/keras-extras</h2>
<p dir="auto">After examining the graph in TensorBoard I discovered a problem in this approach: gradients are not computed in parallel on each device, but in whole on the parameter service device. Indeed each worker computes only the predictions which are distributed to PS and concatenated. The loss + gradients are computed there. Another potential problem is that the whole mini-batch is fed to each device which only takes it's slice. We waste our precious IO bandwidth.</p>
<h2 dir="auto">Proposed fixes</h2>
<ul dir="auto">
<li>gradients should be computed in each tower separately, then averaged on PS device</li>
<li>only tower sub-batch slices of input/labels transferred to each tower (not full batch)</li>
<li>should we use queue for providing inputs asynchronously?</li>
<li>are the model parameters properly placed on the PS device and shared?</li>
<li>in addition for a correct parallel SGD implementation we should incorporate corrections outlined in the <a href="https://arxiv.org/abs/1706.02677" rel="nofollow">Accurate, Large Minibatch SGD: Training ImageNet in 1 Hour</a> paper:
<ul dir="auto">
<li>scaled learning rate with warm-up</li>
<li>momentum correction</li>
<li>batch normalization correction</li>
<li>gradient aggregation
<ul dir="auto">
<li>for averaging we can put all the normalization terms inside the loss and then reduce by sum</li>
<li>"Normalize the per-worker loss by total minibatch size kn, not per-worker size n."</li>
</ul>
</li>
<li>random shuffling: "Use a single random shuffling of the training data (per epoch) that is divided amongst all k workers."</li>
<li>gradients aggregatin should be done in parallel with back-propagation
<ul dir="auto">
<li>"As soon as the gradient for a layer is computed, it is aggregated across workers, while gradient computation for the next layer continues."</li>
</ul>
</li>
</ul>
</li>
</ul>
<h2 dir="auto">Proposed implementation</h2>
<h3 dir="auto">Parallel losses and gradients - DataParallelOptimizer</h3>
<p dir="auto">Since the Keras API is at present not directly suitable for data-parallel SGD computation, in the first step of making a working prototype we can make different implementations of Optimizer and Model, let's say DataParallelOptimizer and DataParallelModel.</p>
<p dir="auto">We need to compute loss and gradients in parallel. Tensor for loss is created by Keras within <code class="notranslate">Model.compile()</code> and stored as <code class="notranslate">Model.total_loss</code>. Gradients are computed in <code class="notranslate">Optimizer.get_gradients()</code> which is called in lazily created functions <code class="notranslate">Model.{train,test,predict}_function()</code> (called from <code class="notranslate">fit()</code>, etc.). This function accepts single loss tensor. Various optimizers then compute updates based on the gradients. The problem is a single loss tensor (which can be placed on one device) passed to <code class="notranslate">Optimizer.get_updates()</code>.</p>
<p dir="auto">So far the only way I see it so change <code class="notranslate">Model.total_loss</code> from a single tensor into a list of tensors, each of them able to be placed on a different device.<br>
<code class="notranslate">DataParallelOptimizer</code> wrapper class can derive from <code class="notranslate">Optimizer</code> and override <code class="notranslate">get_gradients()</code> to accept loss as a list of tensors and average them. This would be the place of synchronization of the GPU workers. The <code class="notranslate">get_updates()</code> function (implemented any of the wrapped <code class="notranslate">Optimizer</code>) just calls <code class="notranslate">get_gradients()</code>. Note that thanks to the option <code class="notranslate">collocate_gradients_with_ops=True</code> in TF implementation of <code class="notranslate">K.gradients()</code> the gradient ops would automatically be placed on the same device as the loss ops even though it's called outside compile() and the device scope. (TODO: issue link)</p>
<h3 dir="auto">Model replication and feeding data - DataParallelModel</h3>
<p dir="auto">We need a Model which contains the replicas (towers) and provides the list of losses to the <code class="notranslate">DataParallelOptimizer</code>. We would adapt the code in the <code class="notranslate">make_parallel()</code> function from kuza55/keras-extras. The <code class="notranslate">DataParallelModel</code> would take via the constructor an instance of the basic model. Similar like in <code class="notranslate">make_parallel()</code> it would make N replicas of this model placed on different devices. We could try to set TF variable reuse after the first replica. Also we make the merged model, which concatenates the outputs, and use it for the actual training. Better then slicing the duplicate inputs we can pass the sub-batch inputs separately and route them to each replicate directly.</p>
<p dir="auto">Then it would override <code class="notranslate">compile()</code> which has to call <code class="notranslate">compile()</code> on each replica (and the merged model) - in order to place losses and gradients - and gather <code class="notranslate">total_loss</code> operation from all replicas. In <code class="notranslate">compile()</code> we also wrap the provided optimizer with <code class="notranslate">DataParallelOptimizer</code> and inject both the total_loss list and the DataParallelOptimizer instance to the merged model. The rest of the methods in <code class="notranslate">DataParallelModel</code> will be proxied to the merged model.</p>
<p dir="auto">In case we want to avoid slicing the inputs we could change the inputs in {train,test,predict}_function() and perform the slice in <code class="notranslate">*_on_batch()</code> functions.</p>
<h2 dir="auto">Code</h2>
<p dir="auto"><del><a href="https://gist.github.com/bzamecnik/92607207af912ae53dd2aa557631b977">https://gist.github.com/bzamecnik/92607207af912ae53dd2aa557631b977</a></del></p>
<p dir="auto"><a href="https://github.com/rossumai/keras-multi-gpu">https://github.com/rossumai/keras-multi-gpu</a></p>
<p dir="auto">I have prepared an implementation of DataParallelOptimizer and I'm working on DataParallelModel. The mechanism of the latter is not as clear at the moment. In the first stage I'd like to make a working prototype, then make experiments to show that the model produces correct results and that we obtain benefit from scaling to multiple GPUs. Next I wish to make the API cleaner. So far I think the code might be separate from Keras, since it will depend on TensorFlow explicitly and I'm not sure about Theano support.</p>
<p dir="auto">If you read this rather longer I'd like to kindly ask for advice if you think this approach is feasible or you see any problems with that. Any feedback is really welcome. Thanks!</p>
<p dir="auto">/cc <a class="user-mention notranslate" data-hovercard-type="organization" data-hovercard-url="/orgs/rossumai/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/rossumai">@rossumai</a></p> | 0 |
<p dir="auto">Could anyone help me with the following please?</p>
<p dir="auto">I'm unable to light custom geometries, even just a single triangle.</p>
<p dir="auto">I made the following "as-simple-as-possible" demo: <a href="http://gnitsuk.leadhoster.com/ThreeTest/Test.html" rel="nofollow">http://gnitsuk.leadhoster.com/ThreeTest/Test.html</a></p>
<p dir="auto">The scene shows a cube and a triangle. Notice that only the cube is responding to the PointLIght, the triangle is dark. It makes no difference if I change the winding of the traingle in the call triangleGeometry.faces.push.</p>
<p dir="auto">If you left click on the page the pointlight is alternately removed and added to the scene. The source code is viewable from the browser.</p>
<p dir="auto">Can anyone see what I've done wrongly that prevents the triangle from responding to the lighting?</p>
<p dir="auto">Thanks,</p>
<p dir="auto">Mitch.</p> | <p dir="auto">When trying to add object loaded with ObjectLoader that has a BufferGeometry and materials I'm getting an error (r65):</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="TypeError: material is undefined
if ( material.transparent ) {
three.js (line 23305)"><pre class="notranslate"><code class="notranslate">TypeError: material is undefined
if ( material.transparent ) {
three.js (line 23305)
</code></pre></div>
<p dir="auto">I'm using the ObjectExporter to save a file with BufferGeometry and materials.<br>
Object file:<br>
<a href="https://drive.google.com/file/d/0B12AhxvnYHrAUlpBSFhyUUNTLUE/edit?usp=sharing" rel="nofollow">https://drive.google.com/file/d/0B12AhxvnYHrAUlpBSFhyUUNTLUE/edit?usp=sharing</a></p>
<p dir="auto">JSON model:<br>
<a href="https://drive.google.com/file/d/0B12AhxvnYHrAU0ZxT0M0NzYtd28/edit?usp=sharing" rel="nofollow">https://drive.google.com/file/d/0B12AhxvnYHrAU0ZxT0M0NzYtd28/edit?usp=sharing</a></p>
<p dir="auto">Original collada model:<br>
<a href="https://github.com/KhronosGroup/glTF/blob/master/model/duck/duck.dae">https://github.com/KhronosGroup/glTF/blob/master/model/duck/duck.dae</a></p>
<p dir="auto">Exporter:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="var loader = new THREE.JSONLoader();
loader.load('model/model.js', function(geometry, materials) {
var bufferGeometry = THREE.BufferGeometryUtils.fromGeometry(geometry);
var mesh = new THREE.Mesh(bufferGeometry, new THREE.MeshFaceMaterial(materials));
var exporter = new THREE.ObjectExporter();
var output = JSON.stringify( exporter.parse(mesh), null, '\t' );
output = output.replace( /[\n\t]+([\d\.e\-\[\]]+)/g, '$1' );
var blob = new Blob( [ output ], { type: 'text/plain' } );
//window.navigator.msSaveBlob(blob, 'model.js');
var a = document.querySelector("#a");
a.href = window.URL.createObjectURL(blob);
a.download = "model.js";
});"><pre class="notranslate"><code class="notranslate">var loader = new THREE.JSONLoader();
loader.load('model/model.js', function(geometry, materials) {
var bufferGeometry = THREE.BufferGeometryUtils.fromGeometry(geometry);
var mesh = new THREE.Mesh(bufferGeometry, new THREE.MeshFaceMaterial(materials));
var exporter = new THREE.ObjectExporter();
var output = JSON.stringify( exporter.parse(mesh), null, '\t' );
output = output.replace( /[\n\t]+([\d\.e\-\[\]]+)/g, '$1' );
var blob = new Blob( [ output ], { type: 'text/plain' } );
//window.navigator.msSaveBlob(blob, 'model.js');
var a = document.querySelector("#a");
a.href = window.URL.createObjectURL(blob);
a.download = "model.js";
});
</code></pre></div>
<p dir="auto">Trying to load the obkect:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="var loader = new THREE.ObjectLoader();
loader.load('model.js', function(object) {
scene.add(object);
});"><pre class="notranslate"><code class="notranslate">var loader = new THREE.ObjectLoader();
loader.load('model.js', function(object) {
scene.add(object);
});
</code></pre></div>
<p dir="auto">By the way, using a with a download attribute might be nicer than opening a new browser window in the editor.<br>
<a href="https://github.com/mrdoob/three.js/blob/master/editor/js/Menubar.File.js#L168">https://github.com/mrdoob/three.js/blob/master/editor/js/Menubar.File.js#L168</a></p>
<p dir="auto">Thanks</p> | 0 |
<p dir="auto">When using nn.ParameterList and nn.DataParallel at the same time, the length of ParameterList will be 0 after DataParallel.</p>
<p dir="auto">See and run this code (with pytorch=1.6+cu101 installed by pip):</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="import torch
class Model(torch.nn.Module):
def __init__(self):
super(Model, self).__init__()
self.param = torch.nn.ParameterList([torch.nn.Parameter(torch.zeros(3,3)) for _ in range(16)])
print('Before DataParallel: ' + str(len(self.param)))
def forward(self, x):
print('After DataParallel: ' + str(len(self.param)))
return x
model = Model()
model = model.to(torch.device('cuda'))
model = torch.nn.DataParallel(model)
x = model(torch.zeros(3,3))"><pre class="notranslate"><code class="notranslate">import torch
class Model(torch.nn.Module):
def __init__(self):
super(Model, self).__init__()
self.param = torch.nn.ParameterList([torch.nn.Parameter(torch.zeros(3,3)) for _ in range(16)])
print('Before DataParallel: ' + str(len(self.param)))
def forward(self, x):
print('After DataParallel: ' + str(len(self.param)))
return x
model = Model()
model = model.to(torch.device('cuda'))
model = torch.nn.DataParallel(model)
x = model(torch.zeros(3,3))
</code></pre></div>
<p dir="auto">And the results are:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="Before DataParallel: 16
After DataParallel: 0
After DataParallel: 0
After DataParallel: 0
After DataParallel: 0"><pre class="notranslate"><code class="notranslate">Before DataParallel: 16
After DataParallel: 0
After DataParallel: 0
After DataParallel: 0
After DataParallel: 0
</code></pre></div>
<p dir="auto">cc <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/pietern/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/pietern">@pietern</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/mrshenli/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/mrshenli">@mrshenli</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/pritamdamania87/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/pritamdamania87">@pritamdamania87</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/zhaojuanmao/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/zhaojuanmao">@zhaojuanmao</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/satgera/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/satgera">@satgera</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/rohan-varma/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/rohan-varma">@rohan-varma</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/gqchen/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/gqchen">@gqchen</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/aazzolini/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/aazzolini">@aazzolini</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/xush6528/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/xush6528">@xush6528</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/osalpekar/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/osalpekar">@osalpekar</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/jiayisuse/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/jiayisuse">@jiayisuse</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/agolynski/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/agolynski">@agolynski</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/albanD/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/albanD">@albanD</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/mruberry/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/mruberry">@mruberry</a></p> | <h2 dir="auto">🐛 Bug</h2>
<p dir="auto">When I use nn.DataParallel to wrap an nn.Module X, nn.Parameter in X is not copied to gpus in the forward pass. I think nn.Parameter can be considered as a part of module parameters, so it should be treated like other nn.Module parameters in X as well. Is it an intentional design?</p>
<h2 dir="auto">To Reproduce</h2>
<p dir="auto">test.py:</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="import sys
import torch
import torch.nn as nn
import torch.nn.functional as F
gpus = list(map(int, sys.argv[1].split(',')))
class Net(nn.Module):
def __init__(self):
super().__init__()
self.alpha = nn.ParameterList()
for i in range(4):
self.alpha.append(nn.Parameter(1e-3*torch.randn(i+2, 5)))
self.cnn = nn.Conv2d(1, 1, 1, 1, 1)
def forward(self, x):
print(self.alpha)
print(self.cnn)
return x
if __name__ == '__main__':
net = Net().cuda()
if len(gpus) > 1:
net = nn.DataParallel(net, device_ids=gpus)
net(torch.rand(4, 5))"><pre class="notranslate"><span class="pl-k">import</span> <span class="pl-s1">sys</span>
<span class="pl-k">import</span> <span class="pl-s1">torch</span>
<span class="pl-k">import</span> <span class="pl-s1">torch</span>.<span class="pl-s1">nn</span> <span class="pl-k">as</span> <span class="pl-s1">nn</span>
<span class="pl-k">import</span> <span class="pl-s1">torch</span>.<span class="pl-s1">nn</span>.<span class="pl-s1">functional</span> <span class="pl-k">as</span> <span class="pl-v">F</span>
<span class="pl-s1">gpus</span> <span class="pl-c1">=</span> <span class="pl-en">list</span>(<span class="pl-en">map</span>(<span class="pl-s1">int</span>, <span class="pl-s1">sys</span>.<span class="pl-s1">argv</span>[<span class="pl-c1">1</span>].<span class="pl-en">split</span>(<span class="pl-s">','</span>)))
<span class="pl-k">class</span> <span class="pl-v">Net</span>(<span class="pl-s1">nn</span>.<span class="pl-v">Module</span>):
<span class="pl-k">def</span> <span class="pl-en">__init__</span>(<span class="pl-s1">self</span>):
<span class="pl-en">super</span>().<span class="pl-en">__init__</span>()
<span class="pl-s1">self</span>.<span class="pl-s1">alpha</span> <span class="pl-c1">=</span> <span class="pl-s1">nn</span>.<span class="pl-v">ParameterList</span>()
<span class="pl-k">for</span> <span class="pl-s1">i</span> <span class="pl-c1">in</span> <span class="pl-en">range</span>(<span class="pl-c1">4</span>):
<span class="pl-s1">self</span>.<span class="pl-s1">alpha</span>.<span class="pl-en">append</span>(<span class="pl-s1">nn</span>.<span class="pl-v">Parameter</span>(<span class="pl-c1">1e-3</span><span class="pl-c1">*</span><span class="pl-s1">torch</span>.<span class="pl-en">randn</span>(<span class="pl-s1">i</span><span class="pl-c1">+</span><span class="pl-c1">2</span>, <span class="pl-c1">5</span>)))
<span class="pl-s1">self</span>.<span class="pl-s1">cnn</span> <span class="pl-c1">=</span> <span class="pl-s1">nn</span>.<span class="pl-v">Conv2d</span>(<span class="pl-c1">1</span>, <span class="pl-c1">1</span>, <span class="pl-c1">1</span>, <span class="pl-c1">1</span>, <span class="pl-c1">1</span>)
<span class="pl-k">def</span> <span class="pl-en">forward</span>(<span class="pl-s1">self</span>, <span class="pl-s1">x</span>):
<span class="pl-en">print</span>(<span class="pl-s1">self</span>.<span class="pl-s1">alpha</span>)
<span class="pl-en">print</span>(<span class="pl-s1">self</span>.<span class="pl-s1">cnn</span>)
<span class="pl-k">return</span> <span class="pl-s1">x</span>
<span class="pl-k">if</span> <span class="pl-s1">__name__</span> <span class="pl-c1">==</span> <span class="pl-s">'__main__'</span>:
<span class="pl-s1">net</span> <span class="pl-c1">=</span> <span class="pl-v">Net</span>().<span class="pl-en">cuda</span>()
<span class="pl-k">if</span> <span class="pl-en">len</span>(<span class="pl-s1">gpus</span>) <span class="pl-c1">></span> <span class="pl-c1">1</span>:
<span class="pl-s1">net</span> <span class="pl-c1">=</span> <span class="pl-s1">nn</span>.<span class="pl-v">DataParallel</span>(<span class="pl-s1">net</span>, <span class="pl-s1">device_ids</span><span class="pl-c1">=</span><span class="pl-s1">gpus</span>)
<span class="pl-en">net</span>(<span class="pl-s1">torch</span>.<span class="pl-en">rand</span>(<span class="pl-c1">4</span>, <span class="pl-c1">5</span>))</pre></div>
<p dir="auto">When I run <code class="notranslate">python3 test.py 0</code> (which means device_id = [0]), the output is</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="ParameterList(
(0): Parameter containing: [torch.cuda.FloatTensor of size 2x5 (GPU 0)]
(1): Parameter containing: [torch.cuda.FloatTensor of size 3x5 (GPU 0)]
(2): Parameter containing: [torch.cuda.FloatTensor of size 4x5 (GPU 0)]
(3): Parameter containing: [torch.cuda.FloatTensor of size 5x5 (GPU 0)]
)
Conv2d(1, 1, kernel_size=(1, 1), stride=(1, 1), padding=(1, 1))"><pre class="notranslate"><code class="notranslate">ParameterList(
(0): Parameter containing: [torch.cuda.FloatTensor of size 2x5 (GPU 0)]
(1): Parameter containing: [torch.cuda.FloatTensor of size 3x5 (GPU 0)]
(2): Parameter containing: [torch.cuda.FloatTensor of size 4x5 (GPU 0)]
(3): Parameter containing: [torch.cuda.FloatTensor of size 5x5 (GPU 0)]
)
Conv2d(1, 1, kernel_size=(1, 1), stride=(1, 1), padding=(1, 1))
</code></pre></div>
<p dir="auto">However, when I run <code class="notranslate">python3 test.py 0,1</code> (which means device_id = [0, 1]), the output is</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="ParameterList()
Conv2d(1, 1, kernel_size=(1, 1), stride=(1, 1), padding=(1, 1))
ParameterList()
Conv2d(1, 1, kernel_size=(1, 1), stride=(1, 1), padding=(1, 1))"><pre class="notranslate"><code class="notranslate">ParameterList()
Conv2d(1, 1, kernel_size=(1, 1), stride=(1, 1), padding=(1, 1))
ParameterList()
Conv2d(1, 1, kernel_size=(1, 1), stride=(1, 1), padding=(1, 1))
</code></pre></div>
<p dir="auto">Only nn.Module is copied to gpus in forward pass.<br>
How can I use and train nn.Parameter just like nn.Module with nn.DataParallel?</p>
<h2 dir="auto">Expected behavior</h2>
<p dir="auto">When the nn.Module X is wrapped with nn.DataParallel, both nn.Module and nn.Parameter in X should be copied to gpus.</p>
<h2 dir="auto">Environment</h2>
<p dir="auto">PyTorch version: 1.6.0.dev20200401+cu101<br>
Is debug build: No<br>
CUDA used to build PyTorch: 10.1</p>
<p dir="auto">OS: Arch Linux<br>
GCC version: (Arch Linux 9.3.0-1) 9.3.0<br>
CMake version: version 3.17.0</p>
<p dir="auto">Python version: 3.8<br>
Is CUDA available: Yes<br>
CUDA runtime version: 10.2.89<br>
GPU models and configuration:<br>
GPU 0: GeForce GTX TITAN X<br>
GPU 1: GeForce GTX 1060 6GB</p>
<p dir="auto">Nvidia driver version: 440.64<br>
cuDNN version: /usr/lib/libcudnn.so.7.6.5</p>
<p dir="auto">Versions of relevant libraries:<br>
[pip3] numpy==1.18.2<br>
[pip3] torch==1.6.0.dev20200401+cu101<br>
[pip3] torchexp==0.1.0<br>
[pip3] torchvision==0.6.0.dev20200401+cu101<br>
[conda] Could not collect</p>
<p dir="auto">cc <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/ezyang/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/ezyang">@ezyang</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/gchanan/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/gchanan">@gchanan</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/zou3519/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/zou3519">@zou3519</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/albanD/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/albanD">@albanD</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/mruberry/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/mruberry">@mruberry</a></p> | 1 |
<p dir="auto">checkout with-amp expanse, install deps with yarn, yarn run build produces an error:</p>
<p dir="auto"><code class="notranslate">> Failed to build on /var/folders/96/_4p69_wn5jj32w5tl1prsd640000gn/T/1c702ef9-d328-4aad-a4be-713d02183e83 { Error: ./pages/_document.js?entry Module not found: Error: Can't resolve 'react-dom/lib/ReactInjection' in '/Users/philipheinser/Downloads/next.js-master/examples/with-amp/pages' resolve 'react-dom/lib/ReactInjection' in '/Users/philipheinser/Downloads/next.js-master/examples/with-amp/pages' Parsed request is a module using description file: /Users/philipheinser/Downloads/next.js-master/examples/with-amp/package.json (relative path: ./pages) Field 'browser' doesn't contain a valid alias configuration aliased with mapping 'react-dom': '/Users/philipheinser/Downloads/next.js-master/examples/with-amp/node_modules/react-dom/dist/react-dom.min.js' to '/Users/philipheinser/Downloads/next.js-master/examples/with-amp/node_modules/react-dom/dist/react-dom.min.js/lib/ReactInjection'</code></p> | <p dir="auto">The prototype I built following the tutorial on learnnextjs.com needs a lot of optimization on its final build. When run past <a href="https://developers.google.com/speed/pagespeed/insights" rel="nofollow">Google PageSpeed Insights</a>, it gets scored down on the following parameters:</p>
<ul dir="auto">
<li>
<p dir="auto">Enable compression</p>
</li>
<li>
<p dir="auto">Reduce server response time</p>
</li>
</ul>
<p dir="auto">According to Google, the following scripts are uncompressed:</p>
<ul dir="auto">
<li>
<p dir="auto">/_next/7edb944ce4757e3fa4d6f778f42d60f1/app.js</p>
</li>
<li>
<p dir="auto">/_next/97e680aa-2946-4f9e-ad5b-8f938979a15c/page/_error</p>
</li>
<li>
<p dir="auto">/_next/97e680aa-2946-4f9e-ad5b-8f938979a15c/page/(index)</p>
</li>
<li>
<p dir="auto">/(index)</p>
</li>
</ul>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I have searched the <a href="https://github.com/zeit/next.js/issues">issues</a> of this repository and believe that this is not a duplicate.</li>
</ul>
<h2 dir="auto">Expected Behavior</h2>
<p dir="auto">All bundled files served to the client should be minified and compressed.</p>
<h2 dir="auto">Current Behavior</h2>
<p dir="auto">Bundled files being served to the client are not being minified or compressed.</p> | 0 |
<p dir="auto">I have a kind request for all the contributors to the latest provider packages release.<br>
Could you help us to test the RC versions of the providers and let us know in the comment,<br>
if the issue is addressed there.</p>
<h2 dir="auto">Providers that need testing</h2>
<p dir="auto">Those are providers that require testing as there were some substantial changes introduced:</p>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-airbyte/2.0.0rc2" rel="nofollow">airbyte: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16236" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16236/hovercard">Add test_connection method to Airbyte hook (#16236)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/msumit/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/msumit">@msumit</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16109" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16109/hovercard">Fix hooks extended from http hook (#16109)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/msumit/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/msumit">@msumit</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-amazon/2.0.0rc2" rel="nofollow">amazon: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/15173" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15173/hovercard">read timestamp from Cloudwatch events (#15173)</a>: @codenamestif</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16150" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16150/hovercard">remove retry for now (#16150)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/zachliu/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/zachliu">@zachliu</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/15609" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15609/hovercard">Remove the <code class="notranslate">not-allow-trailing-slash</code> rule on S3_hook (#15609)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/Isaacwhyuenac/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/Isaacwhyuenac">@Isaacwhyuenac</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/15848" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15848/hovercard">Add support of capacity provider strategy for ECSOperator (#15848)</a>: @codenamestif</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16241" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16241/hovercard">Update copy command for s3 to redshift (#16241)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/sunki-hong/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/sunki-hong">@sunki-hong</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/16327" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16327/hovercard">Make job name check optional in SageMakerTrainingOperator (#16327)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/olivermeyer/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/olivermeyer">@olivermeyer</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15850" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15850/hovercard">Add AWS DMS replication task operators (#15850)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/Visya/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/Visya">@Visya</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/16189" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16189/hovercard">Fix S3 Select payload join (#16189)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/TAKEDA-Takashi/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/TAKEDA-Takashi">@TAKEDA-Takashi</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15839" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15839/hovercard">Fix spacing in AwsBatchWaitersHook docstring (#15839)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/jlaneve/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/jlaneve">@jlaneve</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/15680" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15680/hovercard">MongoToS3Operator failed when running with a single query (not aggregate pipeline) (#15680)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/amatellanes/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/amatellanes">@amatellanes</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16216" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16216/hovercard">fix: AwsGlueJobOperator change order of args for load_file (#16216)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/avocadomaster/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/avocadomaster">@avocadomaster</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/13796" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/13796/hovercard">Fix S3ToFTPOperator (#13796)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/JavierLopezT/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/JavierLopezT">@JavierLopezT</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-apache-spark/2.0.0rc2" rel="nofollow">apache.spark: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15794" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15794/hovercard">Make SparkSqlHook use Connection (#15794)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/uranusjr/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/uranusjr">@uranusjr</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-cncf-kubernetes/2.0.0rc2" rel="nofollow">cncf.kubernetes: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15942" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15942/hovercard">Add KPO pod-template-file jinja template support. (#15942)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/Dr-Denzy/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/Dr-Denzy">@Dr-Denzy</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15755" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15755/hovercard">Save pod name to xcom for KubernetesPodOperator (#15755)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/junnplus/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/junnplus">@junnplus</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15787" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15787/hovercard">Bug Fix Pod-Template Affinity Ignored due to empty Affinity K8S Object (#15787)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/jpyen/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/jpyen">@jpyen</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16095" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16095/hovercard">Bug Pod Template File Values Ignored (#16095)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/jpyen/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/jpyen">@jpyen</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15638" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15638/hovercard">Fix issue with parsing error logs in the KPO (#15638)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/dimberman/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/dimberman">@dimberman</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/15490" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15490/hovercard">Fix unsuccessful KubernetesPod final_state call when <code class="notranslate">is_delete_operator_pod=True</code> (#15490)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/MatthewRBruce/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/MatthewRBruce">@MatthewRBruce</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-dingding/2.0.0rc2" rel="nofollow">dingding: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16109" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16109/hovercard">Fix hooks extended from http hook (#16109)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/msumit/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/msumit">@msumit</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-discord/2.0.0rc2" rel="nofollow">discord: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16109" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16109/hovercard">Fix hooks extended from http hook (#16109)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/msumit/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/msumit">@msumit</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-docker/2.0.0rc2" rel="nofollow">docker: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15843" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15843/hovercard">Replace DockerOperator's 'volumes' arg for 'mounts' (#15843)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/uranusjr/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/uranusjr">@uranusjr</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-elasticsearch/2.0.1rc1" rel="nofollow">elasticsearch: 2.0.1rc1</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/16465" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16465/hovercard">Remove support jinja templated log_id in elasticsearch (#16465)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/jedcunningham/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/jedcunningham">@jedcunningham</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/14625" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/14625/hovercard">Support remote logging in elasticsearch with filebeat 7 (#14625)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/jedcunningham/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/jedcunningham">@jedcunningham</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/16489" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16489/hovercard">Support non-https elasticsearch external links (#16489)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/jedcunningham/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/jedcunningham">@jedcunningham</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/16357" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16357/hovercard">Fix external elasticsearch logs link (#16357)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/jedcunningham/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/jedcunningham">@jedcunningham</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/16467" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16467/hovercard">Fix elasticsearch external log link with json_format (#16467)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/jedcunningham/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/jedcunningham">@jedcunningham</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-google/4.0.0rc2" rel="nofollow">google: 4.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15812" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15812/hovercard">Move plyvel to google provider extra (#15812)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/dstandish/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/dstandish">@dstandish</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/16388" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16388/hovercard">Fixes AzureFileShare connection extras (#16388)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/potiuk/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/potiuk">@potiuk</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/10343" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/10343/hovercard">Add extra links for google dataproc (#10343)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/yesemsanthoshkumar/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/yesemsanthoshkumar">@yesemsanthoshkumar</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/15632" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15632/hovercard">Add link to Oracle Connection Docs (#15632)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/sunkickr/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/sunkickr">@sunkickr</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/15541" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15541/hovercard">pass wait_for_done parameter down to _DataflowJobsController (#15541)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/dejii/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/dejii">@dejii</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15266" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15266/hovercard">Update Google Ads hook (#15266)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/jacobhjkim/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/jacobhjkim">@jacobhjkim</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15367" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15367/hovercard">Implement BigQuery Table Schema Update Operator (#15367)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/thejens/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/thejens">@thejens</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15422" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15422/hovercard">Add BigQueryToMsSqlOperator (#15422)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/subkanthi/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/subkanthi">@subkanthi</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16160" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16160/hovercard">Fix: GCS To BigQuery source_object (#16160)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/tegardp/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/tegardp">@tegardp</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16171" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16171/hovercard">FIX: unnecessary downloads in GCSToLocalFilesystemOperator (#16171)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/p-kachalov/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/p-kachalov">@p-kachalov</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16027" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16027/hovercard">Fix bigquery type error when export format is parquet (#16027)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/sunki-hong/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/sunki-hong">@sunki-hong</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15738" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15738/hovercard">Fix argument ordering and type of bucket and object (#15738)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/sjvanrossum/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/sjvanrossum">@sjvanrossum</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/15730" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15730/hovercard">Fix sql_to_gcs docstring lint error (#15730)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/natanweinberger/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/natanweinberger">@natanweinberger</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/15026" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15026/hovercard">Ensure <code class="notranslate">mysql_to_gcs</code> fully compatible with MySQL and BigQuery for <code class="notranslate">datetime</code>-related values (#15026)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/tianjianjiang/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/tianjianjiang">@tianjianjiang</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16403" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16403/hovercard">Fix deprecation warnings location in google provider (#16403)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/ashb/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/ashb">@ashb</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-hashicorp/2.0.0rc2" rel="nofollow">hashicorp: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16407" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16407/hovercard">Sanitize end of line character when loading token from a file (vault) (#16407)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/mmenarguezpear/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/mmenarguezpear">@mmenarguezpear</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-http/2.0.0rc2" rel="nofollow">http: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/15605" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15605/hovercard">Update SimpleHttpOperator to take auth type object (#15605)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/fredthomsen/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/fredthomsen">@fredthomsen</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/14701" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/14701/hovercard">HttpHook. Use request factory and respect defaults (#14701)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/ngaranko/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/ngaranko">@ngaranko</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-microsoft-azure/3.0.0rc2" rel="nofollow">microsoft.azure: 3.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/16388" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16388/hovercard">Fixes AzureFileShare connection extras (#16388)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/potiuk/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/potiuk">@potiuk</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15632" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15632/hovercard">Add link to Oracle Connection Docs (#15632)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/sunkickr/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/sunkickr">@sunkickr</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/15637" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15637/hovercard">Fix WasbHook.delete_file broken when using prefix (#15637)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/monti-python/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/monti-python">@monti-python</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15841" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15841/hovercard">Fix colon spacing in AzureDataExplorerHook docstring (#15841)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/jlaneve/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/jlaneve">@jlaneve</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/16280" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16280/hovercard">fix wasb remote logging when blob already exists (#16280)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/flolas/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/flolas">@flolas</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-odbc/2.0.0rc2" rel="nofollow">odbc: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/15510" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15510/hovercard">OdbcHook returns None. Related to #15016 issue. (#15510)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/Goodkat/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/Goodkat">@Goodkat</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15772" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15772/hovercard">Fix OdbcHook handling of port (#15772)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/dstandish/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/dstandish">@dstandish</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-opsgenie/2.0.0rc2" rel="nofollow">opsgenie: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16109" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16109/hovercard">Fix hooks extended from http hook (#16109)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/msumit/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/msumit">@msumit</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-oracle/2.0.0rc2" rel="nofollow">oracle: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15581" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15581/hovercard">Add optional result handler to database hooks (#15581)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/malthe/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/malthe">@malthe</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15589" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15589/hovercard">[Oracle] Add port to DSN (#15589)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/malthe/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/malthe">@malthe</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15632" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15632/hovercard">Add link to Oracle Connection Docs (#15632)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/sunkickr/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/sunkickr">@sunkickr</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-papermill/2.0.0rc2" rel="nofollow">papermill: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15302" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15302/hovercard">Emit error on duplicated DAG ID (#15302)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/uranusjr/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/uranusjr">@uranusjr</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-plexus/2.0.0rc2" rel="nofollow">plexus: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/16026" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16026/hovercard">Removes arrow higher limits for plexus provider (#16026)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/potiuk/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/potiuk">@potiuk</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-postgres/2.0.0rc2" rel="nofollow">postgres: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/15412" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15412/hovercard">PostgresHook: deepcopy connection to avoid mutating connection obj (#15412)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/zhzhang/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/zhzhang">@zhzhang</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16100" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16100/hovercard">Avoid passing <code class="notranslate">aws_conn_id</code> as conn_args for <code class="notranslate">psycopg2.connect</code> (#16100)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/gabrielsyapse/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/gabrielsyapse">@gabrielsyapse</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-samba/2.0.0rc2" rel="nofollow">samba: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16115" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16115/hovercard">Add support for extra parameters to samba client (#16115)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/malthe/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/malthe">@malthe</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-sftp/2.0.0rc2" rel="nofollow">sftp: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/14028" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/14028/hovercard">Depreciate private_key_pass extra param and rename to private_key_passphrase (#14028)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/pgillet/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/pgillet">@pgillet</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-slack/4.0.0rc2" rel="nofollow">slack: 4.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16109" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16109/hovercard">Fix hooks extended from http hook (#16109)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/msumit/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/msumit">@msumit</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-snowflake/2.0.0rc2" rel="nofollow">snowflake: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15926" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15926/hovercard">Add <code class="notranslate">template_fields</code> to <code class="notranslate">S3ToSnowflake</code> operator (#15926)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/nlecoy/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/nlecoy">@nlecoy</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15817" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15817/hovercard">Allow S3ToSnowflakeOperator to omit schema (#15817)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/uranusjr/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/uranusjr">@uranusjr</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16420" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16420/hovercard">Added ability for Snowflake to attribute usage to Airflow by adding an application parameter (#16420)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/sfc-gh-madkins/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/sfc-gh-madkins">@sfc-gh-madkins</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/16102" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/16102/hovercard">fix: restore parameters support when sql passed to SnowflakeHook as str (#16102)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/grassten/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/grassten">@grassten</a></li>
</ul>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-ssh/2.0.0rc2" rel="nofollow">ssh: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> <a href="https://github.com/apache/airflow/pull/15212" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15212/hovercard">Fixed #9963: Display explicit error in case UID has no actual username (#15212)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/andrewgodwin/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/andrewgodwin">@andrewgodwin</a></li>
</ul>
<h2 dir="auto">Providers for which previous RC was fully tested (thanks <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/levyitay/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/levyitay">@levyitay</a> !)</h2>
<h3 dir="auto">Provider <a href="https://pypi.org/project/apache-airflow-providers-qubole/2.0.0rc2" rel="nofollow">qubole: 2.0.0rc2</a></h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/issues/15598" data-hovercard-type="issue" data-hovercard-url="/apache/airflow/issues/15598/hovercard">Qubole Hook Does Not Support 'include_headers' (#15598)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/levyitay/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/levyitay">@levyitay</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/15683" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15683/hovercard">Feature qubole hook support headers (#15683)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/levyitay/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/levyitay">@levyitay</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> <a href="https://github.com/apache/airflow/pull/15615" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/15615/hovercard">Feature qubole hook support headers (#15615)</a>: <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/levyitay/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/levyitay">@levyitay</a></li>
</ul>
<h2 dir="auto">Providers that do not need testing</h2>
<p dir="auto">Those are providers that were either doc-only or had changes that do not require testing.</p>
<ul dir="auto">
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-apache-beam/3.0.0rc2" rel="nofollow">apache.beam: 3.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-apache-cassandra/2.0.0rc2" rel="nofollow">apache.cassandra: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-apache-druid/2.0.0rc2" rel="nofollow">apache.druid: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-apache-hdfs/2.0.0rc2" rel="nofollow">apache.hdfs: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-apache-hive/2.0.0rc2" rel="nofollow">apache.hive: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-apache-kylin/2.0.0rc2" rel="nofollow">apache.kylin: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-apache-livy/2.0.0rc2" rel="nofollow">apache.livy: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-apache-pig/2.0.0rc2" rel="nofollow">apache.pig: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-apache-pinot/2.0.0rc2" rel="nofollow">apache.pinot: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-apache-sqoop/2.0.0rc2" rel="nofollow">apache.sqoop: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-asana/1.0.0rc2" rel="nofollow">asana: 1.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-celery/2.0.0rc2" rel="nofollow">celery: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-cloudant/2.0.0rc2" rel="nofollow">cloudant: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-databricks/2.0.0rc2" rel="nofollow">databricks: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-datadog/2.0.0rc2" rel="nofollow">datadog: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-exasol/2.0.0rc2" rel="nofollow">exasol: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-facebook/2.0.0rc2" rel="nofollow">facebook: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-ftp/2.0.0rc2" rel="nofollow">ftp: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-grpc/2.0.0rc2" rel="nofollow">grpc: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-imap/2.0.0rc2" rel="nofollow">imap: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-jdbc/2.0.0rc2" rel="nofollow">jdbc: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-jenkins/2.0.0rc2" rel="nofollow">jenkins: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-jira/2.0.0rc2" rel="nofollow">jira: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-microsoft-mssql/2.0.0rc2" rel="nofollow">microsoft.mssql: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-microsoft-winrm/2.0.0rc2" rel="nofollow">microsoft.winrm: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-mongo/2.0.0rc2" rel="nofollow">mongo: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-mysql/2.0.0rc2" rel="nofollow">mysql: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-neo4j/2.0.0rc2" rel="nofollow">neo4j: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-openfaas/2.0.0rc2" rel="nofollow">openfaas: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-pagerduty/2.0.0rc2" rel="nofollow">pagerduty: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-presto/2.0.0rc2" rel="nofollow">presto: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-redis/2.0.0rc2" rel="nofollow">redis: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-salesforce/3.0.0rc2" rel="nofollow">salesforce: 3.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-segment/2.0.0rc2" rel="nofollow">segment: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-sendgrid/2.0.0rc2" rel="nofollow">sendgrid: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-singularity/2.0.0rc2" rel="nofollow">singularity: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-sqlite/2.0.0rc2" rel="nofollow">sqlite: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-tableau/2.0.0rc2" rel="nofollow">tableau: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-telegram/2.0.0rc2" rel="nofollow">telegram: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-trino/2.0.0rc2" rel="nofollow">trino: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-vertica/2.0.0rc2" rel="nofollow">vertica: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-yandex/2.0.0rc2" rel="nofollow">yandex: 2.0.0rc2</a></li>
<li>Provider <a href="https://pypi.org/project/apache-airflow-providers-zendesk/2.0.0rc2" rel="nofollow">zendesk: 2.0.0rc2</a></li>
</ul> | <h3 dir="auto">What do you see as an issue?</h3>
<p dir="auto">In <code class="notranslate">CONTRIBUTORS_QUICK_START.rst</code>, the links in the table of contents that direct users to parts of the guide that are hidden by the drop down don't work if the drop down isn't expanded. For example, clicking "<a href="https://github.com/apache/airflow/blob/main/CONTRIBUTORS_QUICK_START.rst#setup-airflow-with-breeze">Setup Airflow with Breeze</a>" does nothing until you open the appropriate drop down <code class="notranslate">Setup and develop using <PyCharm, Visual Studio Code, Gitpod></code></p>
<h3 dir="auto">Solving the problem</h3>
<p dir="auto">Instead of having the entire documentation blocks under <code class="notranslate">Setup and develop using {method}</code> dropdowns, there could be drop downs under each section so that the guide remains concise without sacrificing the functionality of the table of contents.</p>
<h3 dir="auto">Anything else</h3>
<p dir="auto">I'm happy to submit a PR eventually, but I might not be able to get around to it for a bit if anyone else wants to handle it real quick.</p>
<h3 dir="auto">Are you willing to submit PR?</h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> Yes I am willing to submit a PR!</li>
</ul>
<h3 dir="auto">Code of Conduct</h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I agree to follow this project's <a href="https://github.com/apache/airflow/blob/main/CODE_OF_CONDUCT.md">Code of Conduct</a></li>
</ul> | 0 |
<h5 dir="auto">Description of the problem</h5>
<p dir="auto">BLEND mode when loading a gltf is broken in three.js if a material has both texture + emission texture. Interestingly, a broken material will also break all other transparent materials (e.g. adding a 3rd object with a broken material will break totally separate materials on objects 1 + 2).</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/2693840/75697694-bd1b4500-5cad-11ea-9cc3-27a2c2c608e6.png"><img src="https://user-images.githubusercontent.com/2693840/75697694-bd1b4500-5cad-11ea-9cc3-27a2c2c608e6.png" alt="image" style="max-width: 100%;"></a></p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/2693840/75697716-c60c1680-5cad-11ea-9688-d95dc84338d5.png"><img src="https://user-images.githubusercontent.com/2693840/75697716-c60c1680-5cad-11ea-9688-d95dc84338d5.png" alt="image" style="max-width: 100%;"></a></p>
<p dir="auto">A workaround, for whatever reason, is to use KHR_Texture_Transform - if TexTransform is present on a material, it will not break but instead display texture + emission texture correctly.</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/2693840/75697736-cdcbbb00-5cad-11ea-8b03-ea36fb5e339c.png"><img src="https://user-images.githubusercontent.com/2693840/75697736-cdcbbb00-5cad-11ea-8b03-ea36fb5e339c.png" alt="image" style="max-width: 100%;"></a></p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/2693840/75697764-d6bc8c80-5cad-11ea-9135-04f1efa54971.png"><img src="https://user-images.githubusercontent.com/2693840/75697764-d6bc8c80-5cad-11ea-9135-04f1efa54971.png" alt="image" style="max-width: 100%;"></a></p>
<p dir="auto">Live reproduction:<br>
<a href="https://glitch.com/edit/#!/spice-water-dormouse" rel="nofollow">https://glitch.com/edit/#!/spice-water-dormouse</a></p>
<h5 dir="auto">Three.js version</h5>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> r114</li>
</ul>
<p dir="auto">Note that this breaks in all of the following:</p>
<ul dir="auto">
<li>Visual Studio Code glTF extension, only in three.js mode, Cesium and Babylon work fine</li>
<li>model-viewer (e.g. your <a href="https://model-viewer-tester.glitch.me/" rel="nofollow">https://model-viewer-tester.glitch.me/</a>)</li>
<li><a href="https://gltf-viewer.donmccurdy.com/" rel="nofollow">https://gltf-viewer.donmccurdy.com/</a></li>
</ul>
<h5 dir="auto">Browser</h5>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> All of them</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> Chrome</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> Firefox</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> Internet Explorer</li>
</ul>
<h5 dir="auto">OS</h5>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> All of them</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> Windows</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> macOS</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> Linux</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> Android</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> iOS</li>
</ul> | <h5 dir="auto"><code class="notranslate">Box3</code>'s <code class="notranslate">expandByObject</code> fails for buffer geometry whose <code class="notranslate">itemSize !== 3</code>.</h5>
<p dir="auto">If you call <code class="notranslate">expandByObject(obj)</code> where <code class="notranslate">obj</code> has a child with (for example) a <code class="notranslate">TextGeometry</code>, this line will attempt to read past the end of an array, then call <code class="notranslate">applyMatrix4</code> so that <code class="notranslate">v1</code> is <code class="notranslate">{x:NaN, y:NaN, z:NaN}</code>:<br>
<a href="https://github.com/mrdoob/three.js/blob/master/src/math/Box3.js#L255">https://github.com/mrdoob/three.js/blob/master/src/math/Box3.js#L255</a><br>
This means that the box's <code class="notranslate">min</code> and <code class="notranslate">max</code> both become <code class="notranslate">NaN, NaN, NaN</code>.</p>
<p dir="auto">One (obviously not optimized) solution is to replace that line with:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="if (attribute.itemSize === 2) {
var v2 = new Vector2();
v2.fromBufferAttribute(attribute, i);
v1.x = v2.x;
v1.y = v2.y;
v1.z = 0;
} else {
v1.fromBufferAttribute(attribute, i);
}
v1.applyMatrix4(node.matrixWorld);"><pre class="notranslate"><code class="notranslate">if (attribute.itemSize === 2) {
var v2 = new Vector2();
v2.fromBufferAttribute(attribute, i);
v1.x = v2.x;
v1.y = v2.y;
v1.z = 0;
} else {
v1.fromBufferAttribute(attribute, i);
}
v1.applyMatrix4(node.matrixWorld);
</code></pre></div>
<p dir="auto">Of course, this would fail if <code class="notranslate">attribute.itemSize === 4</code>. (I'm not sure ever happens in practice.) You can fix that by checking if <code class="notranslate">itemSize ===3</code> before reading from it with a <code class="notranslate">Vector3</code>.</p>
<p dir="auto">This is not a problem with <code class="notranslate">setFromBufferAttribute</code> because that does not rely on <code class="notranslate">Math.min</code> or <code class="notranslate">Math.max</code>, which favor <code class="notranslate">NaN</code> over any numbers. (Instead it uses <code class="notranslate"><</code> and <code class="notranslate">></code> which favor numbers over <code class="notranslate">NaN</code>.)</p>
<h5 dir="auto">Three.js version</h5>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> Dev</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> r93</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> ...</li>
</ul>
<h5 dir="auto">Browser</h5>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> All of them</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> Chrome</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> Firefox</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> Internet Explorer</li>
</ul>
<h5 dir="auto">OS</h5>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> All of them</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> Windows</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> macOS</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> Linux</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> Android</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> iOS</li>
</ul>
<h5 dir="auto">Hardware Requirements (graphics card, VR Device, ...)</h5> | 0 |
<p dir="auto"><strong>I'm submitting a ...</strong> (check one with "x")</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="[x] bug report => search github for a similar issue or PR before submitting
[ ] feature request
[ ] support request => Please do not submit support request here, instead see https://github.com/angular/angular/blob/master/CONTRIBUTING.md#question"><pre class="notranslate"><code class="notranslate">[x] bug report => search github for a similar issue or PR before submitting
[ ] feature request
[ ] support request => Please do not submit support request here, instead see https://github.com/angular/angular/blob/master/CONTRIBUTING.md#question
</code></pre></div>
<p dir="auto"><strong>Current behavior</strong></p>
<p dir="auto">I have an angular app where am using a webcomponent based on custom elements V0 specification.</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="<custom-page>
<p>This is child element inside my custom page</p>
<step>
<custom-input-element></custom-input-element>
</step>
<step></step>
</custom-page>"><pre class="notranslate"><code class="notranslate"><custom-page>
<p>This is child element inside my custom page</p>
<step>
<custom-input-element></custom-input-element>
</step>
<step></step>
</custom-page>
</code></pre></div>
<p dir="auto">Unable to get access to child elements of custom element "custom-page" in attachedCallBack (lifecylce callback of webcomponents custom element) in the webcomponent.</p>
<p dir="auto"><strong>Expected behavior</strong></p>
<p dir="auto">Angular needs to provide a way for the webcomponent to access its child elements.</p>
<p dir="auto"><strong>Minimal reproduction of the problem with instructions</strong></p>
<p dir="auto"><a href="https://plnkr.co/edit/SGndyIXp8alqaSP4Av5F?p=preview" rel="nofollow">https://plnkr.co/edit/SGndyIXp8alqaSP4Av5F?p=preview</a></p>
<p dir="auto">Please look into the console message after - "Accessing child Nodes".<br>
On the first preview, it might look like the node references are resolved, please stop and re-run the plunker preview to reproduce the issue</p>
<ul dir="auto">
<li><strong>Angular version:</strong> 2.0.X</li>
</ul> | <p dir="auto"><strong>I'm submitting a ...</strong> (check one with "x")</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="[x] bug report
[ ] feature request
[ ] support request"><pre class="notranslate"><code class="notranslate">[x] bug report
[ ] feature request
[ ] support request
</code></pre></div>
<p dir="auto"><strong>Current behavior</strong><br>
I'm trying to integrate ChartIQ library which has HTML Custom Elements to my angular 2 application.<br>
Some of these elements have logic in createdCallback, but when it is firing innerHTML of Custom element is empty.</p>
<p dir="auto"><strong>Expected behavior</strong><br>
If I use Custom Elements without Angular, createdCallback is firing when the full element (with all childrens added into it) inserted into DOM.</p>
<p dir="auto"><strong>Minimal reproduction of the problem with instructions</strong><br>
<a href="http://plnkr.co/edit/OW500u9fr27aVteo3zJh" rel="nofollow">http://plnkr.co/edit/OW500u9fr27aVteo3zJh</a></p>
<p dir="auto"><strong>What is the motivation / use case for changing the behavior?</strong><br>
For now Angular 2 breaks some part of expected behavior of Custom Elements. It needs to be fixed to allow using Custom Elements in Angular apps.</p>
<p dir="auto"><strong>Please tell us about your environment:</strong><br>
Windows 8, WebStorm, npm</p>
<ul dir="auto">
<li>
<p dir="auto"><strong>Angular version:</strong><br>
2.2.1</p>
</li>
<li>
<p dir="auto"><strong>Browser:</strong><br>
Chrome 55</p>
</li>
<li>
<p dir="auto"><strong>Language:</strong><br>
all</p>
</li>
<li>
<p dir="auto"><strong>Node (for AoT issues):</strong> <code class="notranslate">node --version</code> =<br>
6.7.0</p>
</li>
</ul> | 1 |
<p dir="auto">Under <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="273440900" data-permission-text="Title is private" data-url="https://github.com/pandas-dev/pandas/issues/18262" data-hovercard-type="issue" data-hovercard-url="/pandas-dev/pandas/issues/18262/hovercard" href="https://github.com/pandas-dev/pandas/issues/18262">#18262</a> a <code class="notranslate">FutureWarning</code> was added suggesting that existing code like this:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="pd.DataFrame.from_items(x)"><pre class="notranslate"><code class="notranslate">pd.DataFrame.from_items(x)
</code></pre></div>
<p dir="auto">Should be changed to this:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="import collections
pd.DataFrame.from_dict(collections.OrderedDict(x))"><pre class="notranslate"><code class="notranslate">import collections
pd.DataFrame.from_dict(collections.OrderedDict(x))
</code></pre></div>
<p dir="auto">The fact that <code class="notranslate">from_items()</code> appeared only 6 times (now 8 times) in a Stack Overflow search was used as partial justification for removing it. But if you search on GitHub, <a href="https://github.com/search?q=pd.DataFrame.from_items&type=Code"><code class="notranslate">pd.DataFrame.from_items()</code></a> appears more than 15,000 times in Python--almost half as many as <code class="notranslate">from_records()</code>!</p>
<p dir="auto">We should celebrate the fact that this function doesn't cause enough confusion to appear often on Stack Overflow. But it does occur (a lot!) in real code, and deprecating it is a mistake.</p>
<p dir="auto">If constructing a temporary <code class="notranslate">OrderedDict</code> around items is the best way to construct a DataFrame, Pandas should implement that as a short function called <code class="notranslate">DataFrame.from_items()</code>, rather than asking thousands of people to busy themselves to accommodate this unnecessary API change.</p>
<p dir="auto">I recommend removing the FutureWarning, and retaining this widely-used, longstanding function.</p>
<p dir="auto">For reference, the <code class="notranslate">FutureWarning</code> starts in 0.23 and looks like this:</p>
<blockquote>
<p dir="auto">FutureWarning: from_items is deprecated. Please use DataFrame.from_dict(dict(items), ...) instead. DataFrame.from_dict(OrderedDict(items)) may be used to preserve the key order.</p>
</blockquote> | <p dir="auto">I want to slice log files to extract only values within a certain time range. This works well with all but one log file. <em>"Must be the log file then!"</em>. However, the log file is well-formed like all the others and I don't see the reason my code fails:</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="import pandas as pan
import datetime as dt
time_f = '%Y-%m-%d %H:%M:%S.%f'
def get_sublog_by_index(log, t_stime, t_etime):
s_idx, e_idx = log.index.slice_locs(t_stime, t_etime)
return log.ix[s_idx: e_idx + 1]
if __name__ == '__main__':
#http://pastebin.com/tQ6MFhgN
df_good = pan.read_csv('dstat_good', header=True, parse_dates=True, index_col=0)
#http://pastebin.com/M9r59Seq
df_bad = pan.read_csv('dstat_bad', header=True, parse_dates=True, index_col=0)
#manual construction just to be shure; why does pan.tslib.Timestamp.strptime fail?
s_t_g = pan.tslib.Timestamp(dt.datetime.strptime('2015-04-23 03:46:37.030256', time_f))
s_e_g = pan.tslib.Timestamp(dt.datetime.strptime('2015-04-23 04:46:40.560409', time_f))
s_t_b = pan.tslib.Timestamp(dt.datetime.strptime('2015-04-23 02:44:49.310807', time_f))
s_e_b = pan.tslib.Timestamp(dt.datetime.strptime('2015-04-23 03:44:52.922636', time_f))
sublog_g = get_sublog_by_index(df_good, s_t_g, s_e_g)
sublog_b = get_sublog_by_index(df_bad, s_t_b, s_e_b)"><pre class="notranslate"><span class="pl-k">import</span> <span class="pl-s1">pandas</span> <span class="pl-k">as</span> <span class="pl-s1">pan</span>
<span class="pl-k">import</span> <span class="pl-s1">datetime</span> <span class="pl-k">as</span> <span class="pl-s1">dt</span>
<span class="pl-s1">time_f</span> <span class="pl-c1">=</span> <span class="pl-s">'%Y-%m-%d %H:%M:%S.%f'</span>
<span class="pl-k">def</span> <span class="pl-en">get_sublog_by_index</span>(<span class="pl-s1">log</span>, <span class="pl-s1">t_stime</span>, <span class="pl-s1">t_etime</span>):
<span class="pl-s1">s_idx</span>, <span class="pl-s1">e_idx</span> <span class="pl-c1">=</span> <span class="pl-s1">log</span>.<span class="pl-s1">index</span>.<span class="pl-en">slice_locs</span>(<span class="pl-s1">t_stime</span>, <span class="pl-s1">t_etime</span>)
<span class="pl-k">return</span> <span class="pl-s1">log</span>.<span class="pl-s1">ix</span>[<span class="pl-s1">s_idx</span>: <span class="pl-s1">e_idx</span> <span class="pl-c1">+</span> <span class="pl-c1">1</span>]
<span class="pl-k">if</span> <span class="pl-s1">__name__</span> <span class="pl-c1">==</span> <span class="pl-s">'__main__'</span>:
<span class="pl-c">#http://pastebin.com/tQ6MFhgN</span>
<span class="pl-s1">df_good</span> <span class="pl-c1">=</span> <span class="pl-s1">pan</span>.<span class="pl-en">read_csv</span>(<span class="pl-s">'dstat_good'</span>, <span class="pl-s1">header</span><span class="pl-c1">=</span><span class="pl-c1">True</span>, <span class="pl-s1">parse_dates</span><span class="pl-c1">=</span><span class="pl-c1">True</span>, <span class="pl-s1">index_col</span><span class="pl-c1">=</span><span class="pl-c1">0</span>)
<span class="pl-c">#http://pastebin.com/M9r59Seq</span>
<span class="pl-s1">df_bad</span> <span class="pl-c1">=</span> <span class="pl-s1">pan</span>.<span class="pl-en">read_csv</span>(<span class="pl-s">'dstat_bad'</span>, <span class="pl-s1">header</span><span class="pl-c1">=</span><span class="pl-c1">True</span>, <span class="pl-s1">parse_dates</span><span class="pl-c1">=</span><span class="pl-c1">True</span>, <span class="pl-s1">index_col</span><span class="pl-c1">=</span><span class="pl-c1">0</span>)
<span class="pl-c">#manual construction just to be shure; why does pan.tslib.Timestamp.strptime fail?</span>
<span class="pl-s1">s_t_g</span> <span class="pl-c1">=</span> <span class="pl-s1">pan</span>.<span class="pl-s1">tslib</span>.<span class="pl-v">Timestamp</span>(<span class="pl-s1">dt</span>.<span class="pl-s1">datetime</span>.<span class="pl-en">strptime</span>(<span class="pl-s">'2015-04-23 03:46:37.030256'</span>, <span class="pl-s1">time_f</span>))
<span class="pl-s1">s_e_g</span> <span class="pl-c1">=</span> <span class="pl-s1">pan</span>.<span class="pl-s1">tslib</span>.<span class="pl-v">Timestamp</span>(<span class="pl-s1">dt</span>.<span class="pl-s1">datetime</span>.<span class="pl-en">strptime</span>(<span class="pl-s">'2015-04-23 04:46:40.560409'</span>, <span class="pl-s1">time_f</span>))
<span class="pl-s1">s_t_b</span> <span class="pl-c1">=</span> <span class="pl-s1">pan</span>.<span class="pl-s1">tslib</span>.<span class="pl-v">Timestamp</span>(<span class="pl-s1">dt</span>.<span class="pl-s1">datetime</span>.<span class="pl-en">strptime</span>(<span class="pl-s">'2015-04-23 02:44:49.310807'</span>, <span class="pl-s1">time_f</span>))
<span class="pl-s1">s_e_b</span> <span class="pl-c1">=</span> <span class="pl-s1">pan</span>.<span class="pl-s1">tslib</span>.<span class="pl-v">Timestamp</span>(<span class="pl-s1">dt</span>.<span class="pl-s1">datetime</span>.<span class="pl-en">strptime</span>(<span class="pl-s">'2015-04-23 03:44:52.922636'</span>, <span class="pl-s1">time_f</span>))
<span class="pl-s1">sublog_g</span> <span class="pl-c1">=</span> <span class="pl-en">get_sublog_by_index</span>(<span class="pl-s1">df_good</span>, <span class="pl-s1">s_t_g</span>, <span class="pl-s1">s_e_g</span>)
<span class="pl-s1">sublog_b</span> <span class="pl-c1">=</span> <span class="pl-en">get_sublog_by_index</span>(<span class="pl-s1">df_bad</span>, <span class="pl-s1">s_t_b</span>, <span class="pl-s1">s_e_b</span>)</pre></div>
<p dir="auto">I have two log files to test, one working, the other not:<br>
<a href="http://pastebin.com/tQ6MFhgN" rel="nofollow">http://pastebin.com/tQ6MFhgN</a> (working)<br>
<a href="http://pastebin.com/M9r59Seq" rel="nofollow">http://pastebin.com/M9r59Seq</a> (not working)</p>
<p dir="auto">At closer inspection you will notice that the start and end times (<code class="notranslate">s_t*</code> and <code class="notranslate">s_e*</code>) are within the range of the logs' timestamps. You will also notice that the <em>bad</em> log has no odd data or anything else that gives a hint why the code is not working in that case. This is truly odd and seems to origin from the deep. Or I'm struck with utter blindness?</p>
<p dir="auto">For completeness, this is what I get if I try to slice the <em>bad</em> log file:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content=" ...
File "/usr/local/bin/dstatlogsplit", line 54, in get_sublog_by_index
s_idx, e_idx = log.index.slice_locs(t_stime, t_etime)
File "/usr/local/lib/python2.7/dist-packages/pandas/core/index.py", line 2484, in slice_locs
start_slice = self.get_slice_bound(start, 'left', kind)
File "/usr/local/lib/python2.7/dist-packages/pandas/core/index.py", line 2432, in get_slice_bound
raise err
KeyError: 1429757089310807000L"><pre class="notranslate"><code class="notranslate"> ...
File "/usr/local/bin/dstatlogsplit", line 54, in get_sublog_by_index
s_idx, e_idx = log.index.slice_locs(t_stime, t_etime)
File "/usr/local/lib/python2.7/dist-packages/pandas/core/index.py", line 2484, in slice_locs
start_slice = self.get_slice_bound(start, 'left', kind)
File "/usr/local/lib/python2.7/dist-packages/pandas/core/index.py", line 2432, in get_slice_bound
raise err
KeyError: 1429757089310807000L
</code></pre></div> | 0 |
<p dir="auto">I found an odd discrepancy when creating a custom dtype in numpy. The code below:</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="import numpy as np
print "VERSION", np.__version__
def f(data, size):
return np.frombuffer(data, np.dtype([('d', 'i', size)]))
d1 = np.arange(1)
d2 = np.arange(2)
a1 = f(d1, 1)
a2 = f(d2, 2)
b1 = f(d1, [1])
b2 = f(d2, [2])
print "a1['d'].shape: ", a1['d'].shape, " b1['d'].shape: ", b1['d'].shape
print "a2['d'].shape: ", a2['d'].shape, " b2['d'].shape: ", b2['d'].shape"><pre class="notranslate"><span class="pl-k">import</span> <span class="pl-s1">numpy</span> <span class="pl-k">as</span> <span class="pl-s1">np</span>
<span class="pl-k">print</span> <span class="pl-s">"VERSION"</span>, <span class="pl-s1">np</span>.<span class="pl-s1">__version__</span>
<span class="pl-k">def</span> <span class="pl-en">f</span>(<span class="pl-s1">data</span>, <span class="pl-s1">size</span>):
<span class="pl-k">return</span> <span class="pl-s1">np</span>.<span class="pl-en">frombuffer</span>(<span class="pl-s1">data</span>, <span class="pl-s1">np</span>.<span class="pl-en">dtype</span>([(<span class="pl-s">'d'</span>, <span class="pl-s">'i'</span>, <span class="pl-s1">size</span>)]))
<span class="pl-s1">d1</span> <span class="pl-c1">=</span> <span class="pl-s1">np</span>.<span class="pl-en">arange</span>(<span class="pl-c1">1</span>)
<span class="pl-s1">d2</span> <span class="pl-c1">=</span> <span class="pl-s1">np</span>.<span class="pl-en">arange</span>(<span class="pl-c1">2</span>)
<span class="pl-s1">a1</span> <span class="pl-c1">=</span> <span class="pl-en">f</span>(<span class="pl-s1">d1</span>, <span class="pl-c1">1</span>)
<span class="pl-s1">a2</span> <span class="pl-c1">=</span> <span class="pl-en">f</span>(<span class="pl-s1">d2</span>, <span class="pl-c1">2</span>)
<span class="pl-s1">b1</span> <span class="pl-c1">=</span> <span class="pl-en">f</span>(<span class="pl-s1">d1</span>, [<span class="pl-c1">1</span>])
<span class="pl-s1">b2</span> <span class="pl-c1">=</span> <span class="pl-en">f</span>(<span class="pl-s1">d2</span>, [<span class="pl-c1">2</span>])
<span class="pl-k">print</span> <span class="pl-s">"a1['d'].shape: "</span>, <span class="pl-s1">a1</span>[<span class="pl-s">'d'</span>].<span class="pl-s1">shape</span>, <span class="pl-s">" b1['d'].shape: "</span>, <span class="pl-s1">b1</span>[<span class="pl-s">'d'</span>].<span class="pl-s1">shape</span>
<span class="pl-k">print</span> <span class="pl-s">"a2['d'].shape: "</span>, <span class="pl-s1">a2</span>[<span class="pl-s">'d'</span>].<span class="pl-s1">shape</span>, <span class="pl-s">" b2['d'].shape: "</span>, <span class="pl-s1">b2</span>[<span class="pl-s">'d'</span>].<span class="pl-s1">shape</span></pre></div>
<p dir="auto">produces this:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="VERSION 1.12.1
a1['d'].shape: (1L,) b1['d'].shape: (1L, 1L)
a2['d'].shape: (1L, 2L) b2['d'].shape: (1L, 2L)"><pre class="notranslate"><code class="notranslate">VERSION 1.12.1
a1['d'].shape: (1L,) b1['d'].shape: (1L, 1L)
a2['d'].shape: (1L, 2L) b2['d'].shape: (1L, 2L)
</code></pre></div>
<p dir="auto">where <code class="notranslate">a1['d'].shape</code> is the odd one out. This inconsistency is annoying... but at least it seems like always defining the size with a list produces consistent behavior.</p>
<p dir="auto">I did a quick search and I didn't see a duplicate, but I might have missed it.</p> | <p dir="auto">So for context this question comes from this post on stackoverflow: <a href="https://stackoverflow.com/questions/50047174/array-comparison-not-matching-elementwise-comparison-in-numpy#50047327" rel="nofollow">https://stackoverflow.com/questions/50047174/array-comparison-not-matching-elementwise-comparison-in-numpy#50047327</a></p>
<blockquote>
<p dir="auto">I have a numpy array arr. It's a numpy.ndarray, size is (5553110,), dtype=float32.</p>
<p dir="auto">When I do:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="(arr > np.pi )[3154950]
False
(arr[3154950] > np.pi )
True"><pre class="notranslate"><code class="notranslate">(arr > np.pi )[3154950]
False
(arr[3154950] > np.pi )
True
</code></pre></div>
<p dir="auto">Why is the first comparison getting it wrong? And how can I fix it?</p>
<p dir="auto">The values:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="arr[3154950]= 3.1415927
np.pi= 3.141592653589793"><pre class="notranslate"><code class="notranslate">arr[3154950]= 3.1415927
np.pi= 3.141592653589793
</code></pre></div>
<p dir="auto">Is the problem with precision?</p>
</blockquote>
<p dir="auto">Would someone be able to explain the difference in the way numpy approaches the problem? It looks like using np.float64 explicitly fixes the problem, but I'm curious to know if numpy does a different casting in the two indexing methods.</p> | 0 |
<p dir="auto">This may be trivial problem, but there seems to be inconsistency between variance and covariance</p>
<p dir="auto">import numpy as np<br>
x = [1,2,3,4,5,6,7]<br>
np.var(x)</p>
<p dir="auto">produces 4.0, but then</p>
<p dir="auto">np.cov(x,x)</p>
<p dir="auto">produces</p>
<p dir="auto">array([[4.66666667, 4.66666667],<br>
[4.66666667, 4.66666667]])</p>
<p dir="auto">Of course, mathematically var(x) = cov(x,x). The difference here is clearly because in np.var the sum seems to be divided by n while in np.cov the sum seems to be divided by n - 1. For a naive user without paying attention to the detail, this inconsistency may cause some problem.</p> | <p dir="auto">cov() uses a bias of 1 by default.<br>
var() uses a bias of 0 by default.</p>
<p dir="auto">Such that</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="import numpy as np
x = np.random.rand(100)
if np.isclose(np.cov([x,x])[0,0], np.var(x)):
print("Consistent by default.")
if np.isclose(np.cov([x,x],ddof=0)[0,0], np.var(x,ddof=0))
print("Consistent.")"><pre class="notranslate"><span class="pl-k">import</span> <span class="pl-s1">numpy</span> <span class="pl-k">as</span> <span class="pl-s1">np</span>
<span class="pl-s1">x</span> <span class="pl-c1">=</span> <span class="pl-s1">np</span>.<span class="pl-s1">random</span>.<span class="pl-en">rand</span>(<span class="pl-c1">100</span>)
<span class="pl-k">if</span> <span class="pl-s1">np</span>.<span class="pl-en">isclose</span>(<span class="pl-s1">np</span>.<span class="pl-en">cov</span>([<span class="pl-s1">x</span>,<span class="pl-s1">x</span>])[<span class="pl-c1">0</span>,<span class="pl-c1">0</span>], <span class="pl-s1">np</span>.<span class="pl-en">var</span>(<span class="pl-s1">x</span>)):
<span class="pl-en">print</span>(<span class="pl-s">"Consistent by default."</span>)
<span class="pl-k">if</span> <span class="pl-s1">np</span>.<span class="pl-en">isclose</span>(<span class="pl-s1">np</span>.<span class="pl-en">cov</span>([<span class="pl-s1">x</span>,<span class="pl-s1">x</span>],<span class="pl-s1">ddof</span><span class="pl-c1">=</span><span class="pl-c1">0</span>)[<span class="pl-c1">0</span>,<span class="pl-c1">0</span>], <span class="pl-s1">np</span>.<span class="pl-en">var</span>(<span class="pl-s1">x</span>,<span class="pl-s1">ddof</span><span class="pl-c1">=</span><span class="pl-c1">0</span>))
<span class="pl-s1">print</span>(<span class="pl-s">"Consistent."</span>)</pre></div>
<p dir="auto">will only print the second line.</p> | 1 |
<h4 dir="auto">Code Sample, a copy-pastable example if possible</h4>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="import pandas as pd
import numpy as np
df = pd.DataFrame(
np.ones([6, 4], dtype=int),
columns=pd.MultiIndex.from_product([['A', 'B'], [1, 2]])
)
(
df
.groupby(level=0, axis=1)
.apply(
lambda df: pd.concat(
[df.xs(df.name, axis=1), df.sum(axis=1).to_frame('Total')],
axis=1
)
)
)
"><pre class="notranslate"><span class="pl-k">import</span> <span class="pl-s1">pandas</span> <span class="pl-k">as</span> <span class="pl-s1">pd</span>
<span class="pl-k">import</span> <span class="pl-s1">numpy</span> <span class="pl-k">as</span> <span class="pl-s1">np</span>
<span class="pl-s1">df</span> <span class="pl-c1">=</span> <span class="pl-s1">pd</span>.<span class="pl-v">DataFrame</span>(
<span class="pl-s1">np</span>.<span class="pl-en">ones</span>([<span class="pl-c1">6</span>, <span class="pl-c1">4</span>], <span class="pl-s1">dtype</span><span class="pl-c1">=</span><span class="pl-s1">int</span>),
<span class="pl-s1">columns</span><span class="pl-c1">=</span><span class="pl-s1">pd</span>.<span class="pl-v">MultiIndex</span>.<span class="pl-en">from_product</span>([[<span class="pl-s">'A'</span>, <span class="pl-s">'B'</span>], [<span class="pl-c1">1</span>, <span class="pl-c1">2</span>]])
)
(
<span class="pl-s1">df</span>
.<span class="pl-en">groupby</span>(<span class="pl-s1">level</span><span class="pl-c1">=</span><span class="pl-c1">0</span>, <span class="pl-s1">axis</span><span class="pl-c1">=</span><span class="pl-c1">1</span>)
.<span class="pl-en">apply</span>(
<span class="pl-k">lambda</span> <span class="pl-s1">df</span>: <span class="pl-s1">pd</span>.<span class="pl-en">concat</span>(
[<span class="pl-s1">df</span>.<span class="pl-en">xs</span>(<span class="pl-s1">df</span>.<span class="pl-s1">name</span>, <span class="pl-s1">axis</span><span class="pl-c1">=</span><span class="pl-c1">1</span>), <span class="pl-s1">df</span>.<span class="pl-en">sum</span>(<span class="pl-s1">axis</span><span class="pl-c1">=</span><span class="pl-c1">1</span>).<span class="pl-en">to_frame</span>(<span class="pl-s">'Total'</span>)],
<span class="pl-s1">axis</span><span class="pl-c1">=</span><span class="pl-c1">1</span>
)
)
)</pre></div>
<h4 dir="auto">Problem description</h4>
<p dir="auto">The code above produces the error:</p>
<blockquote>
<p dir="auto">cannot reindex from a duplicate axis</p>
</blockquote>
<p dir="auto">I believe this is a bug because, as described below, essentially the same operations can be successfully run along the other axis.</p>
<h4 dir="auto">Expected Output</h4>
<p dir="auto">It should, as far as I can tell, produce the following output:</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/15697539/26962477-b92ed344-4d29-11e7-9cd8-ca5406eee097.png"><img src="https://user-images.githubusercontent.com/15697539/26962477-b92ed344-4d29-11e7-9cd8-ca5406eee097.png" alt="image" style="max-width: 100%;"></a></p>
<p dir="auto">The desired output can be obtained when working on the transposed DataFrame along the <code class="notranslate">index</code> rather than the <code class="notranslate">columns</code>:</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="(
df
.T
.groupby(level=0)
.apply(
lambda df: pd.concat(
[df.xs(df.name), df.sum().to_frame('Total').T]
)
)
.T
)"><pre class="notranslate">(
<span class="pl-s1">df</span>
.<span class="pl-v">T</span>
.<span class="pl-en">groupby</span>(<span class="pl-s1">level</span><span class="pl-c1">=</span><span class="pl-c1">0</span>)
.<span class="pl-en">apply</span>(
<span class="pl-k">lambda</span> <span class="pl-s1">df</span>: <span class="pl-s1">pd</span>.<span class="pl-en">concat</span>(
[<span class="pl-s1">df</span>.<span class="pl-en">xs</span>(<span class="pl-s1">df</span>.<span class="pl-s1">name</span>), <span class="pl-s1">df</span>.<span class="pl-en">sum</span>().<span class="pl-en">to_frame</span>(<span class="pl-s">'Total'</span>).<span class="pl-v">T</span>]
)
)
.<span class="pl-v">T</span>
)</pre></div>
<h4 dir="auto">Output of <code class="notranslate">pd.show_versions()</code></h4>
<details>
<p dir="auto">pd.show_versions()<br>
pd.show_versions()</p>
<h2 dir="auto">INSTALLED VERSIONS</h2>
<p dir="auto">commit: None<br>
python: 2.7.9.final.0<br>
python-bits: 64<br>
OS: Windows<br>
OS-release: 8<br>
machine: AMD64<br>
processor: Intel64 Family 6 Model 69 Stepping 1, GenuineIntel<br>
byteorder: little<br>
LC_ALL: None<br>
LANG: None<br>
LOCALE: None.None</p>
<p dir="auto">pandas: 0.20.2<br>
pytest: None<br>
pip: 9.0.1<br>
setuptools: 34.3.3<br>
Cython: None<br>
numpy: 1.13.0<br>
scipy: 0.19.0<br>
xarray: None<br>
IPython: 5.3.0<br>
sphinx: None<br>
patsy: None<br>
dateutil: 2.6.0<br>
pytz: 2017.2<br>
blosc: None<br>
bottleneck: None<br>
tables: None<br>
numexpr: None<br>
feather: None<br>
matplotlib: 2.0.0<br>
openpyxl: None<br>
xlrd: 1.0.0<br>
xlwt: None<br>
xlsxwriter: None<br>
lxml: None<br>
bs4: None<br>
html5lib: 0.999<br>
sqlalchemy: None<br>
pymysql: None<br>
psycopg2: 2.7.1 (dt dec pq3 ext lo64)<br>
jinja2: 2.9.5<br>
s3fs: None<br>
pandas_gbq: None<br>
pandas_datareader: None</p>
</details> | <h4 dir="auto">Code Sample, a copy-pastable example if possible</h4>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="import pandas as pd
df = pd.DataFrame([['2018-01-01','MSFT','short'],
['2018-01-01','MSFT','long'],
['2018-01-01','MSFT',pd.np.NaN],
['2018-01-01','AAPL','short'],
['2018-01-01','AAPL',pd.np.NaN],
['2018-01-02','AAPL','short'],
['2018-01-01','TSLA','neutral'],
['2018-01-02','TSLA',pd.np.NaN]],
columns=['date','symbol','status'])
df['date'] = pd.to_datetime(df['date'])
df = df.set_index(['date','symbol'])
df.groupby('symbol')['status'].ffill()"><pre class="notranslate"><span class="pl-k">import</span> <span class="pl-s1">pandas</span> <span class="pl-k">as</span> <span class="pl-s1">pd</span>
<span class="pl-s1">df</span> <span class="pl-c1">=</span> <span class="pl-s1">pd</span>.<span class="pl-v">DataFrame</span>([[<span class="pl-s">'2018-01-01'</span>,<span class="pl-s">'MSFT'</span>,<span class="pl-s">'short'</span>],
[<span class="pl-s">'2018-01-01'</span>,<span class="pl-s">'MSFT'</span>,<span class="pl-s">'long'</span>],
[<span class="pl-s">'2018-01-01'</span>,<span class="pl-s">'MSFT'</span>,<span class="pl-s1">pd</span>.<span class="pl-s1">np</span>.<span class="pl-v">NaN</span>],
[<span class="pl-s">'2018-01-01'</span>,<span class="pl-s">'AAPL'</span>,<span class="pl-s">'short'</span>],
[<span class="pl-s">'2018-01-01'</span>,<span class="pl-s">'AAPL'</span>,<span class="pl-s1">pd</span>.<span class="pl-s1">np</span>.<span class="pl-v">NaN</span>],
[<span class="pl-s">'2018-01-02'</span>,<span class="pl-s">'AAPL'</span>,<span class="pl-s">'short'</span>],
[<span class="pl-s">'2018-01-01'</span>,<span class="pl-s">'TSLA'</span>,<span class="pl-s">'neutral'</span>],
[<span class="pl-s">'2018-01-02'</span>,<span class="pl-s">'TSLA'</span>,<span class="pl-s1">pd</span>.<span class="pl-s1">np</span>.<span class="pl-v">NaN</span>]],
<span class="pl-s1">columns</span><span class="pl-c1">=</span>[<span class="pl-s">'date'</span>,<span class="pl-s">'symbol'</span>,<span class="pl-s">'status'</span>])
<span class="pl-s1">df</span>[<span class="pl-s">'date'</span>] <span class="pl-c1">=</span> <span class="pl-s1">pd</span>.<span class="pl-en">to_datetime</span>(<span class="pl-s1">df</span>[<span class="pl-s">'date'</span>])
<span class="pl-s1">df</span> <span class="pl-c1">=</span> <span class="pl-s1">df</span>.<span class="pl-en">set_index</span>([<span class="pl-s">'date'</span>,<span class="pl-s">'symbol'</span>])
<span class="pl-s1">df</span>.<span class="pl-en">groupby</span>(<span class="pl-s">'symbol'</span>)[<span class="pl-s">'status'</span>].<span class="pl-en">ffill</span>()</pre></div>
<h4 dir="auto">Raised exception</h4>
<details>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="---------------------------------------------------------------------------
Exception Traceback (most recent call last)
<ipython-input-58-2fe0989150cf> in <module>()
13 df['date'] = pd.to_datetime(df['date'])
14 df = df.set_index(['date','symbol'])
---> 15 df.groupby('symbol')['status'].ffill()
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in pad(self, limit)
1362 DataFrame.fillna
1363 """
-> 1364 return self.apply(lambda x: x.ffill(limit=limit))
1365 ffill = pad
1366
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in apply(self, func, *args, **kwargs)
3110 examples=_apply_docs['series_examples']))
3111 def apply(self, func, *args, **kwargs):
-> 3112 return super(SeriesGroupBy, self).apply(func, *args, **kwargs)
3113
3114 @Appender(_agg_doc)
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in apply(self, func, *args, **kwargs)
803 # ignore SettingWithCopy here in case the user mutates
804 with option_context('mode.chained_assignment', None):
--> 805 return self._python_apply_general(f)
806
807 def _python_apply_general(self, f):
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _python_apply_general(self, f)
812 keys,
813 values,
--> 814 not_indexed_same=mutated or self.mutated)
815
816 def _iterate_slices(self):
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _wrap_applied_output(self, keys, values, not_indexed_same)
3250 if isinstance(values[0], (Series, dict)):
3251 return self._concat_objects(keys, values,
-> 3252 not_indexed_same=not_indexed_same)
3253 elif isinstance(values[0], DataFrame):
3254 # possible that Series -> DataFrame by applied function
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _concat_objects(self, keys, values, not_indexed_same)
978
979 if isinstance(result, Series):
--> 980 result = result.reindex(ax)
981 else:
982
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/series.pyc in reindex(self, index, **kwargs)
2679 @Appender(generic._shared_docs['reindex'] % _shared_doc_kwargs)
2680 def reindex(self, index=None, **kwargs):
-> 2681 return super(Series, self).reindex(index=index, **kwargs)
2682
2683 @Appender(generic._shared_docs['fillna'] % _shared_doc_kwargs)
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/generic.pyc in reindex(self, *args, **kwargs)
3021 # perform the reindex on the axes
3022 return self._reindex_axes(axes, level, limit, tolerance, method,
-> 3023 fill_value, copy).__finalize__(self)
3024
3025 def _reindex_axes(self, axes, level, limit, tolerance, method, fill_value,
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/generic.pyc in _reindex_axes(self, axes, level, limit, tolerance, method, fill_value, copy)
3034 ax = self._get_axis(a)
3035 new_index, indexer = ax.reindex(labels, level=level, limit=limit,
-> 3036 tolerance=tolerance, method=method)
3037
3038 axis = self._get_axis_number(a)
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/indexes/multi.pyc in reindex(self, target, method, level, limit, tolerance)
1901 tolerance=tolerance)
1902 else:
-> 1903 raise Exception("cannot handle a non-unique multi-index!")
1904
1905 if not isinstance(target, MultiIndex):
Exception: cannot handle a non-unique multi-index!"><pre class="notranslate"><code class="notranslate">---------------------------------------------------------------------------
Exception Traceback (most recent call last)
<ipython-input-58-2fe0989150cf> in <module>()
13 df['date'] = pd.to_datetime(df['date'])
14 df = df.set_index(['date','symbol'])
---> 15 df.groupby('symbol')['status'].ffill()
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in pad(self, limit)
1362 DataFrame.fillna
1363 """
-> 1364 return self.apply(lambda x: x.ffill(limit=limit))
1365 ffill = pad
1366
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in apply(self, func, *args, **kwargs)
3110 examples=_apply_docs['series_examples']))
3111 def apply(self, func, *args, **kwargs):
-> 3112 return super(SeriesGroupBy, self).apply(func, *args, **kwargs)
3113
3114 @Appender(_agg_doc)
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in apply(self, func, *args, **kwargs)
803 # ignore SettingWithCopy here in case the user mutates
804 with option_context('mode.chained_assignment', None):
--> 805 return self._python_apply_general(f)
806
807 def _python_apply_general(self, f):
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _python_apply_general(self, f)
812 keys,
813 values,
--> 814 not_indexed_same=mutated or self.mutated)
815
816 def _iterate_slices(self):
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _wrap_applied_output(self, keys, values, not_indexed_same)
3250 if isinstance(values[0], (Series, dict)):
3251 return self._concat_objects(keys, values,
-> 3252 not_indexed_same=not_indexed_same)
3253 elif isinstance(values[0], DataFrame):
3254 # possible that Series -> DataFrame by applied function
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _concat_objects(self, keys, values, not_indexed_same)
978
979 if isinstance(result, Series):
--> 980 result = result.reindex(ax)
981 else:
982
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/series.pyc in reindex(self, index, **kwargs)
2679 @Appender(generic._shared_docs['reindex'] % _shared_doc_kwargs)
2680 def reindex(self, index=None, **kwargs):
-> 2681 return super(Series, self).reindex(index=index, **kwargs)
2682
2683 @Appender(generic._shared_docs['fillna'] % _shared_doc_kwargs)
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/generic.pyc in reindex(self, *args, **kwargs)
3021 # perform the reindex on the axes
3022 return self._reindex_axes(axes, level, limit, tolerance, method,
-> 3023 fill_value, copy).__finalize__(self)
3024
3025 def _reindex_axes(self, axes, level, limit, tolerance, method, fill_value,
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/generic.pyc in _reindex_axes(self, axes, level, limit, tolerance, method, fill_value, copy)
3034 ax = self._get_axis(a)
3035 new_index, indexer = ax.reindex(labels, level=level, limit=limit,
-> 3036 tolerance=tolerance, method=method)
3037
3038 axis = self._get_axis_number(a)
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/indexes/multi.pyc in reindex(self, target, method, level, limit, tolerance)
1901 tolerance=tolerance)
1902 else:
-> 1903 raise Exception("cannot handle a non-unique multi-index!")
1904
1905 if not isinstance(target, MultiIndex):
Exception: cannot handle a non-unique multi-index!
</code></pre></div>
</details>
<h4 dir="auto">Problem description</h4>
<p dir="auto">Performing <code class="notranslate">ffill()</code> does not forward fill values as intended. The index has duplicate values, the expectation is that the values will be filled according to the order the rows are in already.</p>
<p dir="auto">Duplicate index values are possible when using high frequency time intervals.</p>
<h4 dir="auto">Expected Output</h4>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="df = pd.DataFrame([['2018-01-01','MSFT','short'],
['2018-01-01','MSFT','long'],
['2018-01-01','MSFT','long'],
['2018-01-01','AAPL','short'],
['2018-01-01','AAPL','short'],
['2018-01-02','AAPL','short'],
['2018-01-01','TSLA','neutral'],
['2018-01-02','TSLA','neutral']],
columns=['date','symbol','status'])
df['date'] = pd.to_datetime(df['date'])"><pre class="notranslate"><span class="pl-s1">df</span> <span class="pl-c1">=</span> <span class="pl-s1">pd</span>.<span class="pl-v">DataFrame</span>([[<span class="pl-s">'2018-01-01'</span>,<span class="pl-s">'MSFT'</span>,<span class="pl-s">'short'</span>],
[<span class="pl-s">'2018-01-01'</span>,<span class="pl-s">'MSFT'</span>,<span class="pl-s">'long'</span>],
[<span class="pl-s">'2018-01-01'</span>,<span class="pl-s">'MSFT'</span>,<span class="pl-s">'long'</span>],
[<span class="pl-s">'2018-01-01'</span>,<span class="pl-s">'AAPL'</span>,<span class="pl-s">'short'</span>],
[<span class="pl-s">'2018-01-01'</span>,<span class="pl-s">'AAPL'</span>,<span class="pl-s">'short'</span>],
[<span class="pl-s">'2018-01-02'</span>,<span class="pl-s">'AAPL'</span>,<span class="pl-s">'short'</span>],
[<span class="pl-s">'2018-01-01'</span>,<span class="pl-s">'TSLA'</span>,<span class="pl-s">'neutral'</span>],
[<span class="pl-s">'2018-01-02'</span>,<span class="pl-s">'TSLA'</span>,<span class="pl-s">'neutral'</span>]],
<span class="pl-s1">columns</span><span class="pl-c1">=</span>[<span class="pl-s">'date'</span>,<span class="pl-s">'symbol'</span>,<span class="pl-s">'status'</span>])
<span class="pl-s1">df</span>[<span class="pl-s">'date'</span>] <span class="pl-c1">=</span> <span class="pl-s1">pd</span>.<span class="pl-en">to_datetime</span>(<span class="pl-s1">df</span>[<span class="pl-s">'date'</span>])</pre></div>
<h4 dir="auto">Output of <code class="notranslate">pd.show_versions()</code></h4>
<details>
<h2 dir="auto">INSTALLED VERSIONS</h2>
<p dir="auto">commit: None<br>
python: 2.7.14.final.0<br>
python-bits: 64<br>
OS: Linux<br>
OS-release: 4.13.0-31-generic<br>
machine: x86_64<br>
processor: x86_64<br>
byteorder: little<br>
LC_ALL: None<br>
LANG: en_AU.UTF-8<br>
LOCALE: None.None</p>
<p dir="auto">pandas: 0.22.0<br>
pytest: 3.2.1<br>
pip: 9.0.1<br>
setuptools: 36.5.0.post20170921<br>
Cython: 0.26.1<br>
numpy: 1.14.0<br>
scipy: 0.19.1<br>
pyarrow: None<br>
xarray: None<br>
IPython: 5.4.1<br>
sphinx: 1.6.3<br>
patsy: 0.4.1<br>
dateutil: 2.6.1<br>
pytz: 2017.3<br>
blosc: None<br>
bottleneck: 1.2.1<br>
tables: 3.4.2<br>
numexpr: 2.6.2<br>
feather: None<br>
matplotlib: 2.1.0<br>
openpyxl: 2.4.8<br>
xlrd: 1.1.0<br>
xlwt: 1.3.0<br>
xlsxwriter: 1.0.2<br>
lxml: 4.1.0<br>
bs4: 4.6.0<br>
html5lib: 0.999999999<br>
sqlalchemy: 1.1.13<br>
pymysql: None<br>
psycopg2: 2.7.3.2 (dt dec pq3 ext lo64)<br>
jinja2: 2.9.6<br>
s3fs: None<br>
fastparquet: None<br>
pandas_gbq: None<br>
pandas_datareader: None</p>
</details>
<p dir="auto">This also happens with <code class="notranslate">axis=1</code>.</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="import pandas as pd
import numpy as np
df = pd.DataFrame(
np.ones([6, 4], dtype=int),
columns=pd.MultiIndex.from_product([['A', 'B'], [1, 2]])
)
(
df
.groupby(level=0, axis=1)
.apply(
lambda df: 2*df.xs(df.name, axis=1)
)
)"><pre class="notranslate"><span class="pl-k">import</span> <span class="pl-s1">pandas</span> <span class="pl-k">as</span> <span class="pl-s1">pd</span>
<span class="pl-k">import</span> <span class="pl-s1">numpy</span> <span class="pl-k">as</span> <span class="pl-s1">np</span>
<span class="pl-s1">df</span> <span class="pl-c1">=</span> <span class="pl-s1">pd</span>.<span class="pl-v">DataFrame</span>(
<span class="pl-s1">np</span>.<span class="pl-en">ones</span>([<span class="pl-c1">6</span>, <span class="pl-c1">4</span>], <span class="pl-s1">dtype</span><span class="pl-c1">=</span><span class="pl-s1">int</span>),
<span class="pl-s1">columns</span><span class="pl-c1">=</span><span class="pl-s1">pd</span>.<span class="pl-v">MultiIndex</span>.<span class="pl-en">from_product</span>([[<span class="pl-s">'A'</span>, <span class="pl-s">'B'</span>], [<span class="pl-c1">1</span>, <span class="pl-c1">2</span>]])
)
(
<span class="pl-s1">df</span>
.<span class="pl-en">groupby</span>(<span class="pl-s1">level</span><span class="pl-c1">=</span><span class="pl-c1">0</span>, <span class="pl-s1">axis</span><span class="pl-c1">=</span><span class="pl-c1">1</span>)
.<span class="pl-en">apply</span>(
<span class="pl-k">lambda</span> <span class="pl-s1">df</span>: <span class="pl-c1">2</span><span class="pl-c1">*</span><span class="pl-s1">df</span>.<span class="pl-en">xs</span>(<span class="pl-s1">df</span>.<span class="pl-s1">name</span>, <span class="pl-s1">axis</span><span class="pl-c1">=</span><span class="pl-c1">1</span>)
)
)</pre></div>
<details>
<div class="highlight highlight-text-python-traceback notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-12-e7c022399f89> in <module>()
3 .groupby(level=0, axis=1)
4 .apply(
----> 5 lambda df: 2*df.xs(df.name, axis=1)
6 )
7 )
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in apply(self, func, *args, **kwargs)
882 # ignore SettingWithCopy here in case the user mutates
883 with option_context('mode.chained_assignment', None):
--> 884 return self._python_apply_general(f)
885
886 def _python_apply_general(self, f):
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _python_apply_general(self, f)
891 keys,
892 values,
--> 893 not_indexed_same=mutated or self.mutated)
894
895 def _iterate_slices(self):
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _wrap_applied_output(self, keys, values, not_indexed_same)
3920 elif isinstance(v, DataFrame):
3921 return self._concat_objects(keys, values,
-> 3922 not_indexed_same=not_indexed_same)
3923 elif self.grouper.groupings is not None:
3924 if len(self.grouper.groupings) > 1:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _concat_objects(self, keys, values, not_indexed_same)
1085 result = result.take(indexer, axis=self.axis)
1086 else:
-> 1087 result = result.reindex(ax, axis=self.axis)
1088
1089 elif self.group_keys:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/util/_decorators.py in wrapper(*args, **kwargs)
145 @wraps(func)
146 def wrapper(*args, **kwargs):
--> 147 return func(*args, **kwargs)
148
149 if not PY2:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in reindex(self, *args, **kwargs)
2979 kwargs.pop('axis', None)
2980 kwargs.pop('labels', None)
-> 2981 return super(DataFrame, self).reindex(**kwargs)
2982
2983 @Appender(_shared_docs['reindex_axis'] % _shared_doc_kwargs)
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/generic.py in reindex(self, *args, **kwargs)
3356 # perform the reindex on the axes
3357 return self._reindex_axes(axes, level, limit, tolerance, method,
-> 3358 fill_value, copy).__finalize__(self)
3359
3360 def _reindex_axes(self, axes, level, limit, tolerance, method, fill_value,
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in _reindex_axes(self, axes, level, limit, tolerance, method, fill_value, copy)
2909 if columns is not None:
2910 frame = frame._reindex_columns(columns, method, copy, level,
-> 2911 fill_value, limit, tolerance)
2912
2913 index = axes['index']
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in _reindex_columns(self, new_columns, method, copy, level, fill_value, limit, tolerance)
2934 return self._reindex_with_indexers({1: [new_columns, indexer]},
2935 copy=copy, fill_value=fill_value,
-> 2936 allow_dups=False)
2937
2938 def _reindex_multi(self, axes, copy, fill_value):
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/generic.py in _reindex_with_indexers(self, reindexers, fill_value, copy, allow_dups)
3477 fill_value=fill_value,
3478 allow_dups=allow_dups,
-> 3479 copy=copy)
3480
3481 if copy and new_data is self._data:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/internals.py in reindex_indexer(self, new_axis, indexer, axis, fill_value, allow_dups, copy)
4125 # some axes don't allow reindexing with dups
4126 if not allow_dups:
-> 4127 self.axes[axis]._can_reindex(indexer)
4128
4129 if axis >= self.ndim:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/indexes/base.py in _can_reindex(self, indexer)
2940 # trying to reindex on an axis with duplicates
2941 if not self.is_unique and len(indexer):
-> 2942 raise ValueError("cannot reindex from a duplicate axis")
2943
2944 def reindex(self, target, method=None, level=None, limit=None,
ValueError: cannot reindex from a duplicate axis"><pre class="notranslate">---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-12-e7c022399f89> in <module>()
<span class="pl-c1">3</span> .groupby(<span class="pl-v">level</span><span class="pl-k">=</span><span class="pl-c1">0</span>, <span class="pl-v">axis</span><span class="pl-k">=</span><span class="pl-c1">1</span>)
<span class="pl-c1">4</span> .apply(
----> 5 lambda df: 2*df.xs(df.name, axis=1)
<span class="pl-c1">6</span> )
<span class="pl-c1">7</span> )
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in apply(self, func, *args, **kwargs)
<span class="pl-c1">882</span> <span class="pl-c"><span class="pl-c">#</span> ignore SettingWithCopy here in case the user mutates</span>
<span class="pl-c1">883</span> <span class="pl-k">with</span> option_context(<span class="pl-s"><span class="pl-pds">'</span>mode.chained_assignment<span class="pl-pds">'</span></span>, <span class="pl-c1">None</span>):
--> 884 return self._python_apply_general(f)
<span class="pl-c1">885</span>
<span class="pl-c1">886</span> <span class="pl-k">def</span> <span class="pl-en">_python_apply_general</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">f</span>):
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _python_apply_general(self, f)
<span class="pl-c1">891</span> keys,
<span class="pl-c1">892</span> values,
--> 893 not_indexed_same=mutated or self.mutated)
<span class="pl-c1">894</span>
<span class="pl-c1">895</span> <span class="pl-k">def</span> <span class="pl-en">_iterate_slices</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _wrap_applied_output(self, keys, values, not_indexed_same)
3920 elif isinstance(v, DataFrame):
3921 return self._concat_objects(keys, values,
-> 3922 not_indexed_same=not_indexed_same)
3923 elif self.grouper.groupings is not None:
3924 if len(self.grouper.groupings) > 1:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _concat_objects(self, keys, values, not_indexed_same)
1085 result = result.take(indexer, axis=self.axis)
1086 else:
-> 1087 result = result.reindex(ax, axis=self.axis)
1088
1089 elif self.group_keys:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/util/_decorators.py in wrapper(*args, **kwargs)
<span class="pl-c1">145</span> <span class="pl-k">@</span>wraps(func)
<span class="pl-c1">146</span> <span class="pl-k">def</span> <span class="pl-en">wrapper</span>(<span class="pl-k">*</span><span class="pl-smi">args</span>, <span class="pl-k">**</span><span class="pl-smi">kwargs</span>):
--> 147 return func(*args, **kwargs)
<span class="pl-c1">148</span>
<span class="pl-c1">149</span> <span class="pl-k">if</span> <span class="pl-k">not</span> <span class="pl-c1">PY2</span>:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in reindex(self, *args, **kwargs)
2979 kwargs.pop('axis', None)
2980 kwargs.pop('labels', None)
-> 2981 return super(DataFrame, self).reindex(**kwargs)
2982
2983 @Appender(_shared_docs['reindex_axis'] % _shared_doc_kwargs)
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/generic.py in reindex(self, *args, **kwargs)
3356 # perform the reindex on the axes
3357 return self._reindex_axes(axes, level, limit, tolerance, method,
-> 3358 fill_value, copy).__finalize__(self)
3359
3360 def _reindex_axes(self, axes, level, limit, tolerance, method, fill_value,
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in _reindex_axes(self, axes, level, limit, tolerance, method, fill_value, copy)
2909 if columns is not None:
2910 frame = frame._reindex_columns(columns, method, copy, level,
-> 2911 fill_value, limit, tolerance)
2912
2913 index = axes['index']
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in _reindex_columns(self, new_columns, method, copy, level, fill_value, limit, tolerance)
2934 return self._reindex_with_indexers({1: [new_columns, indexer]},
2935 copy=copy, fill_value=fill_value,
-> 2936 allow_dups=False)
2937
2938 def _reindex_multi(self, axes, copy, fill_value):
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/generic.py in _reindex_with_indexers(self, reindexers, fill_value, copy, allow_dups)
3477 fill_value=fill_value,
3478 allow_dups=allow_dups,
-> 3479 copy=copy)
3480
3481 if copy and new_data is self._data:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/internals.py in reindex_indexer(self, new_axis, indexer, axis, fill_value, allow_dups, copy)
4125 # some axes don't allow reindexing with dups
4126 if not allow_dups:
-> 4127 self.axes[axis]._can_reindex(indexer)
4128
4129 if axis >= self.ndim:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/indexes/base.py in _can_reindex(self, indexer)
2940 # trying to reindex on an axis with duplicates
2941 if not self.is_unique and len(indexer):
-> 2942 raise ValueError("cannot reindex from a duplicate axis")
2943
2944 def reindex(self, target, method=None, level=None, limit=None,
<span class="pl-en">ValueError</span>: <span class="pl-s">cannot reindex from a duplicate axis</span></pre></div>
</details> | 1 |
<h4 dir="auto">Challenge Name</h4>
<p dir="auto"><a href="url">https://www.freecodecamp.com/challenges/show-the-local-weather</a></p>
<h4 dir="auto">Issue Description</h4>
<p dir="auto">For the weather challenge it is recommend to use the open weather api but when using the service on google Chrome the content will not be served on codepen.io due to new security changes in Chrome. So I suggest that we look for alternatives.</p>
<h4 dir="auto">Browser Information</h4>
<p dir="auto">Browser Name, Version:<br>
Google Chrome 55.0.2883.95 (Official Build) (64-bit)</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://cloud.githubusercontent.com/assets/19998186/21945197/d870155e-d9d1-11e6-8958-a240a7863ef5.jpeg"><img src="https://cloud.githubusercontent.com/assets/19998186/21945197/d870155e-d9d1-11e6-8958-a240a7863ef5.jpeg" alt="security" style="max-width: 100%;"></a></p> | <h4 dir="auto">Show the Local Weather</h4>
<p dir="auto"><a href="https://www.freecodecamp.com/challenges/show-the-local-weather" rel="nofollow">https://www.freecodecamp.com/challenges/show-the-local-weather</a></p>
<h4 dir="auto">Weather App Location issue</h4>
<p dir="auto">I am currently using Google chrome and I have been trying to figure out why my location could not be found when I copied the exact code from FCC and placed it in Codepen.io. After a week of fighting I asked for help and I learn that I had to use https:// so I can find my location. The second issue is that when I tried to pass this location to the Weather API it would not work, so I asked the community again and I learn that the FCC given locator will not work when we pass the location to the weather API. Could you please add a side note to the other codes as a heads up so that they don't end up beating their computer with a baseball bat?</p>
<h4 dir="auto">Browser Information</h4>
<ul dir="auto">
<li>Browser Name, Version: Google Chrome Version 51.0.2704.84 m</li>
<li>Operating System: Windows 10</li>
<li>Mobile, Desktop, or Tablet: Desktop</li>
</ul>
<h4 dir="auto">Your Code</h4>
<div class="highlight highlight-source-js notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="// If relevant, paste all of your challenge code in here
"><pre class="notranslate"><span class="pl-c">// If relevant, paste all of your challenge code in here</span></pre></div>
<h4 dir="auto">Screenshot</h4> | 1 |
<p dir="auto">Minimal example:</p>
<div class="highlight highlight-source-rust notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="trait Foo {
type T;
}
#[derive(Copy)]
struct PairFoo<F: Foo>(pub F::T, pub F::T);
fn main() {}"><pre class="notranslate"><span class="pl-k">trait</span> <span class="pl-smi">Foo</span> <span class="pl-kos">{</span>
<span class="pl-k">type</span> <span class="pl-smi">T</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span>
<span class="pl-c1">#<span class="pl-kos">[</span>derive<span class="pl-kos">(</span><span class="pl-v">Copy</span><span class="pl-kos">)</span><span class="pl-kos">]</span></span>
<span class="pl-k">struct</span> <span class="pl-smi">PairFoo</span><span class="pl-kos"><</span><span class="pl-smi">F</span><span class="pl-kos">:</span> <span class="pl-smi">Foo</span><span class="pl-kos">></span><span class="pl-kos">(</span><span class="pl-k">pub</span> <span class="pl-smi">F</span><span class="pl-kos">::</span><span class="pl-smi">T</span><span class="pl-kos">,</span> <span class="pl-k">pub</span> <span class="pl-smi">F</span><span class="pl-kos">::</span><span class="pl-smi">T</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-k">fn</span> <span class="pl-en">main</span><span class="pl-kos">(</span><span class="pl-kos">)</span> <span class="pl-kos">{</span><span class="pl-kos">}</span></pre></div>
<p dir="auto"><code class="notranslate">rustc -Z unstable-options --pretty=expanded</code> yields:</p>
<div class="highlight highlight-source-rust notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="#![feature(no_std)]
#![no_std]
#[prelude_import]
use std::prelude::v1::*;
#[macro_use]
extern crate "std" as std;
trait Foo {
type
T;
}
struct PairFoo<F: Foo>(pub F::T, pub F::T);
#[automatically_derived]
impl <F: ::std::marker::Copy + Foo> ::std::marker::Copy for PairFoo<F> { }
fn main() { }"><pre class="notranslate"><span class="pl-c1">#!<span class="pl-kos">[</span>feature<span class="pl-kos">(</span>no_std<span class="pl-kos">)</span><span class="pl-kos">]</span></span>
<span class="pl-c1">#!<span class="pl-kos">[</span>no_std<span class="pl-kos">]</span></span>
<span class="pl-c1">#<span class="pl-kos">[</span>prelude_import<span class="pl-kos">]</span></span>
<span class="pl-k">use</span> std<span class="pl-kos">::</span>prelude<span class="pl-kos">::</span>v1<span class="pl-kos">::</span><span class="pl-c1">*</span><span class="pl-kos">;</span>
<span class="pl-c1">#<span class="pl-kos">[</span>macro_use<span class="pl-kos">]</span></span>
<span class="pl-k">extern</span> <span class="pl-k">crate</span> <span class="pl-s">"std"</span> as std<span class="pl-kos">;</span>
<span class="pl-k">trait</span> <span class="pl-smi">Foo</span> <span class="pl-kos">{</span>
<span class="pl-k">type</span>
<span class="pl-smi">T</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span>
<span class="pl-k">struct</span> <span class="pl-smi">PairFoo</span><span class="pl-kos"><</span><span class="pl-smi">F</span><span class="pl-kos">:</span> <span class="pl-smi">Foo</span><span class="pl-kos">></span><span class="pl-kos">(</span><span class="pl-k">pub</span> <span class="pl-smi">F</span><span class="pl-kos">::</span><span class="pl-smi">T</span><span class="pl-kos">,</span> <span class="pl-k">pub</span> <span class="pl-smi">F</span><span class="pl-kos">::</span><span class="pl-smi">T</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-c1">#<span class="pl-kos">[</span>automatically_derived<span class="pl-kos">]</span></span>
<span class="pl-k">impl</span> <span class="pl-kos"><</span><span class="pl-smi">F</span><span class="pl-kos">:</span> <span class="pl-kos">::</span>std<span class="pl-kos">::</span>marker<span class="pl-kos">::</span><span class="pl-smi">Copy</span> + <span class="pl-smi">Foo</span><span class="pl-kos">></span> <span class="pl-kos">::</span>std<span class="pl-kos">::</span>marker<span class="pl-kos">::</span><span class="pl-smi">Copy</span> <span class="pl-k">for</span> <span class="pl-smi">PairFoo</span><span class="pl-kos"><</span><span class="pl-smi">F</span><span class="pl-kos">></span> <span class="pl-kos">{</span> <span class="pl-kos">}</span>
<span class="pl-k">fn</span> <span class="pl-en">main</span><span class="pl-kos">(</span><span class="pl-kos">)</span> <span class="pl-kos">{</span> <span class="pl-kos">}</span></pre></div>
<p dir="auto">This is wrong, because the bound shouldn't be <code class="notranslate">F: Copy</code> but <code class="notranslate"><F as Foo>::T: Copy</code> instead.</p> | <p dir="auto">In the following snippet, <code class="notranslate">PairFoo<F></code>'s members are all <code class="notranslate">Copy</code> (resp. <code class="notranslate">Clone</code>) whenever <code class="notranslate">F::T</code> is <code class="notranslate">Copy</code> (resp. <code class="notranslate">Clone</code>). In spite of this, neither does <code class="notranslate">#[derive(Copy)] #[derive(Clone)]</code> work, nor can I provide my own <code class="notranslate">Copy</code> impl.</p>
<div class="highlight highlight-source-rust notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="use std::marker::Copy;
use std::clone::Clone;
#[derive(Copy)]
#[derive(Clone)]
struct Pair<T>(pub T, pub T);
trait Foo { type T; }
// #[derive(Copy)]
// #[derive(Clone)]
struct PairFoo<F: Foo>(pub F::T, pub F::T);
// this one doesn't work
impl<F, T> Copy for PairFoo<F>
where F: Foo<T = T>,
T: Copy
{
}
// this one does work
impl<F, T> Clone for PairFoo<F>
where F: Foo<T = T>,
T: Clone
{
fn clone(&self) -> PairFoo<F>
{
PairFoo(self.0.clone(), self.1.clone())
}
}"><pre class="notranslate"><span class="pl-k">use</span> std<span class="pl-kos">::</span>marker<span class="pl-kos">::</span><span class="pl-v">Copy</span><span class="pl-kos">;</span>
<span class="pl-k">use</span> std<span class="pl-kos">::</span>clone<span class="pl-kos">::</span><span class="pl-v">Clone</span><span class="pl-kos">;</span>
<span class="pl-c1">#<span class="pl-kos">[</span>derive<span class="pl-kos">(</span><span class="pl-v">Copy</span><span class="pl-kos">)</span><span class="pl-kos">]</span></span>
<span class="pl-c1">#<span class="pl-kos">[</span>derive<span class="pl-kos">(</span><span class="pl-v">Clone</span><span class="pl-kos">)</span><span class="pl-kos">]</span></span>
<span class="pl-k">struct</span> <span class="pl-smi">Pair</span><span class="pl-kos"><</span><span class="pl-smi">T</span><span class="pl-kos">></span><span class="pl-kos">(</span><span class="pl-k">pub</span> <span class="pl-smi">T</span><span class="pl-kos">,</span> <span class="pl-k">pub</span> <span class="pl-smi">T</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-k">trait</span> <span class="pl-smi">Foo</span> <span class="pl-kos">{</span> <span class="pl-k">type</span> <span class="pl-smi">T</span><span class="pl-kos">;</span> <span class="pl-kos">}</span>
<span class="pl-c">// #[derive(Copy)]</span>
<span class="pl-c">// #[derive(Clone)]</span>
<span class="pl-k">struct</span> <span class="pl-smi">PairFoo</span><span class="pl-kos"><</span><span class="pl-smi">F</span><span class="pl-kos">:</span> <span class="pl-smi">Foo</span><span class="pl-kos">></span><span class="pl-kos">(</span><span class="pl-k">pub</span> <span class="pl-smi">F</span><span class="pl-kos">::</span><span class="pl-smi">T</span><span class="pl-kos">,</span> <span class="pl-k">pub</span> <span class="pl-smi">F</span><span class="pl-kos">::</span><span class="pl-smi">T</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-c">// this one doesn't work</span>
<span class="pl-k">impl</span><span class="pl-kos"><</span><span class="pl-smi">F</span><span class="pl-kos">,</span> <span class="pl-smi">T</span><span class="pl-kos">></span> <span class="pl-smi">Copy</span> <span class="pl-k">for</span> <span class="pl-smi">PairFoo</span><span class="pl-kos"><</span><span class="pl-smi">F</span><span class="pl-kos">></span>
<span class="pl-k">where</span> <span class="pl-smi">F</span><span class="pl-kos">:</span> <span class="pl-smi">Foo</span><span class="pl-kos"><</span><span class="pl-smi">T</span> = <span class="pl-smi">T</span><span class="pl-kos">></span><span class="pl-kos">,</span>
<span class="pl-smi">T</span><span class="pl-kos">:</span> <span class="pl-smi">Copy</span>
<span class="pl-kos">{</span>
<span class="pl-kos">}</span>
<span class="pl-c">// this one does work</span>
<span class="pl-k">impl</span><span class="pl-kos"><</span><span class="pl-smi">F</span><span class="pl-kos">,</span> <span class="pl-smi">T</span><span class="pl-kos">></span> <span class="pl-smi">Clone</span> <span class="pl-k">for</span> <span class="pl-smi">PairFoo</span><span class="pl-kos"><</span><span class="pl-smi">F</span><span class="pl-kos">></span>
<span class="pl-k">where</span> <span class="pl-smi">F</span><span class="pl-kos">:</span> <span class="pl-smi">Foo</span><span class="pl-kos"><</span><span class="pl-smi">T</span> = <span class="pl-smi">T</span><span class="pl-kos">></span><span class="pl-kos">,</span>
<span class="pl-smi">T</span><span class="pl-kos">:</span> <span class="pl-smi">Clone</span>
<span class="pl-kos">{</span>
<span class="pl-k">fn</span> <span class="pl-en">clone</span><span class="pl-kos">(</span><span class="pl-c1">&</span><span class="pl-smi">self</span><span class="pl-kos">)</span> -> <span class="pl-smi">PairFoo</span><span class="pl-kos"><</span><span class="pl-smi">F</span><span class="pl-kos">></span>
<span class="pl-kos">{</span>
<span class="pl-v">PairFoo</span><span class="pl-kos">(</span><span class="pl-smi">self</span><span class="pl-kos">.</span><span class="pl-c1">0</span><span class="pl-kos">.</span><span class="pl-en">clone</span><span class="pl-kos">(</span><span class="pl-kos">)</span><span class="pl-kos">,</span> <span class="pl-smi">self</span><span class="pl-kos">.</span><span class="pl-c1">1</span><span class="pl-kos">.</span><span class="pl-en">clone</span><span class="pl-kos">(</span><span class="pl-kos">)</span><span class="pl-kos">)</span>
<span class="pl-kos">}</span>
<span class="pl-kos">}</span></pre></div>
<p dir="auto">rustc's output:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="<anon>:12:1: 16:2 error: the trait `Copy` may not be implemented for this type; field `<unnamed_field>` does not implement `Copy` [E0204]
<anon>:12 impl<F, T> Copy for PairFoo<F>
<anon>:13 where F: Foo<T = T>,
<anon>:14 T: Copy
<anon>:15 {
<anon>:16 }
error: aborting due to previous error
playpen: application terminated with error code 101"><pre class="notranslate"><code class="notranslate"><anon>:12:1: 16:2 error: the trait `Copy` may not be implemented for this type; field `<unnamed_field>` does not implement `Copy` [E0204]
<anon>:12 impl<F, T> Copy for PairFoo<F>
<anon>:13 where F: Foo<T = T>,
<anon>:14 T: Copy
<anon>:15 {
<anon>:16 }
error: aborting due to previous error
playpen: application terminated with error code 101
</code></pre></div> | 1 |
<h5 dir="auto">System information (version)</h5>
<ul dir="auto">
<li>OpenCV => 4.2</li>
<li>Operating System / Platform => Windows 10</li>
<li>Compiler => Visual C++ 2019</li>
</ul>
<h5 dir="auto">Detailed description</h5>
<p dir="auto">This is not a duplicate of similar looking issues reported earlier because it explains how those unreleased blocks accumulate in certain cases and this situation cannot be resolved by suppressing the report as suggested here <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="556985994" data-permission-text="Title is private" data-url="https://github.com/opencv/opencv/issues/16464" data-hovercard-type="issue" data-hovercard-url="/opencv/opencv/issues/16464/hovercard?comment_id=579852547&comment_type=issue_comment" href="https://github.com/opencv/opencv/issues/16464#issuecomment-579852547">#16464 (comment)</a></p>
<p dir="auto">Suppose there's <code class="notranslate">ProcessingModule.dll</code> made by a user and it links to <code class="notranslate">opencv_core420d.dll</code> via a .lib file - such that when anyone tries to load <code class="notranslate">ProcessingModule.dll</code> then Windows automatically loads <code class="notranslate">opencv_core420d.dll</code> too.</p>
<p dir="auto">Each time <code class="notranslate">ProcessingModule.dll</code> is loaded using <code class="notranslate">LoadLibrary()</code> Windows loads <code class="notranslate">opencv_core420d.dll</code> and this causes OpenCV to constructs some global objects some of which allocate dynamic memory. When <code class="notranslate">ProcessingModule.dll</code> is unloaded with <code class="notranslate">FreeLibrary()</code> Windows also unloads <code class="notranslate">opencv_core420d.dll</code> and dynamic memory from those global objects remains unfreed. If <code class="notranslate">LoadLibrary()</code>/<code class="notranslate">FreeLibrary()</code> are called in a loop those unreleased blocks accumulate and form a leak which gets worse over time because those global objects are created anew every time <code class="notranslate">opencv_core420d.dll</code> is loaded and those global objects are never released.</p>
<p dir="auto">So the code to reproduce the problem would be something like this:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="int main()
{
_CrtSetDbgFlag(_CrtSetDbgFlag(_CRTDBG_REPORT_FLAG ) | _CRTDBG_LEAK_CHECK_DF);
for (int c = 0; c < 50; c++)
{
HMODULE h = LoadLibraryW(L"ProcessingModule.dll");
FreeLibrary(h);
}
}"><pre class="notranslate"><code class="notranslate">int main()
{
_CrtSetDbgFlag(_CrtSetDbgFlag(_CRTDBG_REPORT_FLAG ) | _CRTDBG_LEAK_CHECK_DF);
for (int c = 0; c < 50; c++)
{
HMODULE h = LoadLibraryW(L"ProcessingModule.dll");
FreeLibrary(h);
}
}
</code></pre></div>
<p dir="auto">and after this runs and the program exits there's a huge number of leaks reported in Visual Studio debugger output - about ten leaks per each iteration of the loop and all of them belong to global objects when OpenCV created for its own needs and never released. When upper limit is increased from 50 to a larger number then the number of reported unreleased blocks increases.</p>
<p dir="auto">You may say "don't write stupid programs" but the scenario above is perfectly natural when using in-proc COM servers. Here's how:</p>
<ul dir="auto">
<li>user calls <code class="notranslate">CoCreateInstance()</code> to instantiate an object which is served by <code class="notranslate">ProcessingModule.dll</code></li>
<li>Windows loads <code class="notranslate">ProcessingModule.dll</code> and also <code class="notranslate">opencv_core420d.dll</code>, those global objects are initialized</li>
<li>user releases the object he requested earlier</li>
<li>some time passes and Windows calls <code class="notranslate">DllCanUnloadNow()</code> in <code class="notranslate">ProcessingModule.dll</code> to find whether it's still serving any objects</li>
<li><code class="notranslate">ProcessingModule.dll</code> returns "no objects, you may unload me" and Windows unloads <code class="notranslate">ProcessingModule.dll</code> and also <code class="notranslate">opencv_core420d.dll</code></li>
<li>some more time passes and user again calls <code class="notranslate">CoCreateInstance()</code> to instantiate an object which is served by <code class="notranslate">ProcessingModule.dll</code></li>
<li>the cycle continues</li>
</ul>
<p dir="auto">This is effectively the same as a "stupid for loop" above and it may occur naturally when using in-proc COM servers on Windows.</p>
<p dir="auto">This cannot be resolved by ignoring those unreleased blocks using valgrind suppressions or their equivalent.</p>
<h5 dir="auto">Steps to reproduce</h5>
<h5 dir="auto">Issue submission checklist</h5>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I report the issue, it's not a question
</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I checked the problem with documentation, FAQ, open issues,<br>
forum.opencv.org, Stack Overflow, etc and have not found solution
</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> I updated to latest OpenCV version and the issue is still there
</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> There is reproducer code and related data files: videos, images, onnx, etc
</li>
</ul> | <h5 dir="auto">System information (version)</h5>
<ul dir="auto">
<li>OpenCV => 4.1.0</li>
<li>Operating System / Platform =>Ubuntu 18.04</li>
</ul>
<h5 dir="auto">Detailed description</h5>
<p dir="auto">I want to test out the pipeline for stereo camera calibration and depth map estimation. I used the OpenCV sample (samples/data/left01.jpg – left14.jpg and right ...) for calibration. Is there any pair of images on which I could test depth map estimation after the stereo set has been calibrated with the <strong>given example</strong>?</p>
<h5 dir="auto">Relevant Documentation</h5>
<p dir="auto">The <a href="https://docs.opencv.org/4.1.0/dd/d53/tutorial_py_depthmap.html" rel="nofollow">camera calibration tutorial</a> uses the same sample (samples/data/left01.jpg – left14.jpg and right ...) for calibration as I did. Next, it is unclear whether the pair of images from <a href="https://docs.opencv.org/4.1.0/dd/d53/tutorial_py_depthmap.html" rel="nofollow">depth map from stereo images tutorial</a> is actually the pair of images taken with the same stereo camera that was calibrated in the previous tutorial. Also, this tutorial only shows how to compute disparity but not actually estimate the depth.</p>
<p dir="auto">It'd be great if the depth map tutorial had a stereo image pair, so that depth from this pair could be estimated using camera parameters from the calibration tutorial.</p> | 0 |
<p dir="auto">This is difficult to explain, so I've written an isolated test case:</p>
<div class="highlight highlight-source-js notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="let instances = new Map();
class Resource{
static get(id){
let list = instances.get(this) || {};
return list[id];
}
static add(id, instance){
let list = instances.get(this);
if(!list)
instances.set(this, list = {});
list[id] = instance;
}
constructor(id){
let type = this.constructor;
let byId = type.get(id);
// Already created. Return reference to existing instance.
if(byId){
return byId;
}
this.id = id;
type.add(id, this);
}
}
class SubtypeResource extends Resource{
}
let resourceA = new SubtypeResource(200);
let resourceB = new SubtypeResource(200);
console.log(resourceA === resourceB); // "true" if natively supported; "false" if transpiled"><pre class="notranslate"><span class="pl-k">let</span> <span class="pl-s1">instances</span> <span class="pl-c1">=</span> <span class="pl-k">new</span> <span class="pl-v">Map</span><span class="pl-kos">(</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-k">class</span> <span class="pl-v">Resource</span><span class="pl-kos">{</span>
<span class="pl-k">static</span> <span class="pl-en">get</span><span class="pl-kos">(</span><span class="pl-s1">id</span><span class="pl-kos">)</span><span class="pl-kos">{</span>
<span class="pl-k">let</span> <span class="pl-s1">list</span> <span class="pl-c1">=</span> <span class="pl-s1">instances</span><span class="pl-kos">.</span><span class="pl-en">get</span><span class="pl-kos">(</span><span class="pl-smi">this</span><span class="pl-kos">)</span> <span class="pl-c1">||</span> <span class="pl-kos">{</span><span class="pl-kos">}</span><span class="pl-kos">;</span>
<span class="pl-k">return</span> <span class="pl-s1">list</span><span class="pl-kos">[</span><span class="pl-s1">id</span><span class="pl-kos">]</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span>
<span class="pl-k">static</span> <span class="pl-en">add</span><span class="pl-kos">(</span><span class="pl-s1">id</span><span class="pl-kos">,</span> <span class="pl-s1">instance</span><span class="pl-kos">)</span><span class="pl-kos">{</span>
<span class="pl-k">let</span> <span class="pl-s1">list</span> <span class="pl-c1">=</span> <span class="pl-s1">instances</span><span class="pl-kos">.</span><span class="pl-en">get</span><span class="pl-kos">(</span><span class="pl-smi">this</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-k">if</span><span class="pl-kos">(</span><span class="pl-c1">!</span><span class="pl-s1">list</span><span class="pl-kos">)</span>
<span class="pl-s1">instances</span><span class="pl-kos">.</span><span class="pl-en">set</span><span class="pl-kos">(</span><span class="pl-smi">this</span><span class="pl-kos">,</span> <span class="pl-s1">list</span> <span class="pl-c1">=</span> <span class="pl-kos">{</span><span class="pl-kos">}</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-s1">list</span><span class="pl-kos">[</span><span class="pl-s1">id</span><span class="pl-kos">]</span> <span class="pl-c1">=</span> <span class="pl-s1">instance</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span>
<span class="pl-en">constructor</span><span class="pl-kos">(</span><span class="pl-s1">id</span><span class="pl-kos">)</span><span class="pl-kos">{</span>
<span class="pl-k">let</span> <span class="pl-s1">type</span> <span class="pl-c1">=</span> <span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-c1">constructor</span><span class="pl-kos">;</span>
<span class="pl-k">let</span> <span class="pl-s1">byId</span> <span class="pl-c1">=</span> <span class="pl-s1">type</span><span class="pl-kos">.</span><span class="pl-en">get</span><span class="pl-kos">(</span><span class="pl-s1">id</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-c">// Already created. Return reference to existing instance.</span>
<span class="pl-k">if</span><span class="pl-kos">(</span><span class="pl-s1">byId</span><span class="pl-kos">)</span><span class="pl-kos">{</span>
<span class="pl-k">return</span> <span class="pl-s1">byId</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span>
<span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-c1">id</span> <span class="pl-c1">=</span> <span class="pl-s1">id</span><span class="pl-kos">;</span>
<span class="pl-s1">type</span><span class="pl-kos">.</span><span class="pl-en">add</span><span class="pl-kos">(</span><span class="pl-s1">id</span><span class="pl-kos">,</span> <span class="pl-smi">this</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span>
<span class="pl-kos">}</span>
<span class="pl-k">class</span> <span class="pl-v">SubtypeResource</span> <span class="pl-k">extends</span> <span class="pl-v">Resource</span><span class="pl-kos">{</span>
<span class="pl-kos">}</span>
<span class="pl-k">let</span> <span class="pl-s1">resourceA</span> <span class="pl-c1">=</span> <span class="pl-k">new</span> <span class="pl-v">SubtypeResource</span><span class="pl-kos">(</span><span class="pl-c1">200</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-k">let</span> <span class="pl-s1">resourceB</span> <span class="pl-c1">=</span> <span class="pl-k">new</span> <span class="pl-v">SubtypeResource</span><span class="pl-kos">(</span><span class="pl-c1">200</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-smi">console</span><span class="pl-kos">.</span><span class="pl-en">log</span><span class="pl-kos">(</span><span class="pl-s1">resourceA</span> <span class="pl-c1">===</span> <span class="pl-s1">resourceB</span><span class="pl-kos">)</span><span class="pl-kos">;</span> <span class="pl-c">// "true" if natively supported; "false" if transpiled</span></pre></div>
<p dir="auto">Blacklisting <code class="notranslate">es6.classes</code> solves the problem, but that naturally requires an interpreter to support it natively. I'm running the latest version of Node (4.2.1), so it's no big deal. But it'd be different if it were for client-side delivery.</p>
<p dir="auto">This only happens with subclasses, so it might have something to do with the way the transformer maps references to constructor functions. E.g., the following code runs fine either natively or through Babel:</p>
<div class="highlight highlight-source-js notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="console.log(new Resource(200) === new Resource(200)); // true"><pre class="notranslate"><span class="pl-smi">console</span><span class="pl-kos">.</span><span class="pl-en">log</span><span class="pl-kos">(</span><span class="pl-k">new</span> <span class="pl-v">Resource</span><span class="pl-kos">(</span><span class="pl-c1">200</span><span class="pl-kos">)</span> <span class="pl-c1">===</span> <span class="pl-k">new</span> <span class="pl-v">Resource</span><span class="pl-kos">(</span><span class="pl-c1">200</span><span class="pl-kos">)</span><span class="pl-kos">)</span><span class="pl-kos">;</span> <span class="pl-c">// true</span></pre></div>
<hr>
<p dir="auto">To summarise what my code's doing: I'm writing a Node program to scrape data from a site that lacks a formal API. Resources are loaded once and done so asynchronously, collating patches of data from numerous different sources. I thought an elegant approach would be to return references to existing instances instead of manually checking if they've been created already:</p>
<div class="highlight highlight-source-js notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="let instance;
if(SubtypeResource.get(id)){
instance = SubtypeResource.get(id);
}
else{
instance = new SubtypeResource(id);
SubtypeResource.add(id, instance);
}"><pre class="notranslate"><span class="pl-k">let</span> <span class="pl-s1">instance</span><span class="pl-kos">;</span>
<span class="pl-k">if</span><span class="pl-kos">(</span><span class="pl-v">SubtypeResource</span><span class="pl-kos">.</span><span class="pl-en">get</span><span class="pl-kos">(</span><span class="pl-s1">id</span><span class="pl-kos">)</span><span class="pl-kos">)</span><span class="pl-kos">{</span>
<span class="pl-s1">instance</span> <span class="pl-c1">=</span> <span class="pl-v">SubtypeResource</span><span class="pl-kos">.</span><span class="pl-en">get</span><span class="pl-kos">(</span><span class="pl-s1">id</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span>
<span class="pl-k">else</span><span class="pl-kos">{</span>
<span class="pl-s1">instance</span> <span class="pl-c1">=</span> <span class="pl-k">new</span> <span class="pl-v">SubtypeResource</span><span class="pl-kos">(</span><span class="pl-s1">id</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-v">SubtypeResource</span><span class="pl-kos">.</span><span class="pl-en">add</span><span class="pl-kos">(</span><span class="pl-s1">id</span><span class="pl-kos">,</span> <span class="pl-s1">instance</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span></pre></div> | <p dir="auto">The current implementation doesn't handle parent classes that return specific objects properly. The value returned from <code class="notranslate">[[Construct]]</code> should be used as the <code class="notranslate">this</code> binding value in the child constructor.</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="class Other {
method(){
console.log('other');
}
}
class Base {
constructor(){
return new Other();
}
method(){
console.log('base');
}
}
class Sub extends Base {
constructor(){
super();
this.method();
}
}
var s = new Sub();
console.log('expected true, got ' + (s instanceof Other));"><pre class="notranslate"><code class="notranslate">class Other {
method(){
console.log('other');
}
}
class Base {
constructor(){
return new Other();
}
method(){
console.log('base');
}
}
class Sub extends Base {
constructor(){
super();
this.method();
}
}
var s = new Sub();
console.log('expected true, got ' + (s instanceof Other));
</code></pre></div>
<p dir="auto">Seems like</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="var Sub = (function (_Base) {
_inherits(Sub, _Base);
function Sub() {
_classCallCheck(this, Sub);
_get(Object.getPrototypeOf(Sub.prototype), 'constructor', this).call(this);
this.method();
}
return Sub;
})(Base);"><pre class="notranslate"><code class="notranslate">var Sub = (function (_Base) {
_inherits(Sub, _Base);
function Sub() {
_classCallCheck(this, Sub);
_get(Object.getPrototypeOf(Sub.prototype), 'constructor', this).call(this);
this.method();
}
return Sub;
})(Base);
</code></pre></div>
<p dir="auto">should be more like</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="var Sub = (function (_Base) {
_inherits(Sub, _Base);
function Sub() {
_classCallCheck(this, Sub);
var _this = _get(Object.getPrototypeOf(Sub.prototype), 'constructor', this).call(this);
if (_this === undefined) _this = this;
_this.method();
return _this;
}
return Sub;
})(Base);"><pre class="notranslate"><code class="notranslate">var Sub = (function (_Base) {
_inherits(Sub, _Base);
function Sub() {
_classCallCheck(this, Sub);
var _this = _get(Object.getPrototypeOf(Sub.prototype), 'constructor', this).call(this);
if (_this === undefined) _this = this;
_this.method();
return _this;
}
return Sub;
})(Base);
</code></pre></div> | 1 |
<p dir="auto">There's a link in the Deployments section of the <a href="http://kubernetes.io/docs/user-guide/walkthrough/k8s201/" rel="nofollow">201 walkthrough</a> to a longer description that's broken. It links to <code class="notranslate">http://kubernetes.io/docs/user-guide/deployment</code> but should link to <code class="notranslate">http://kubernetes.io/docs/user-guide/deployments</code> (missing an <code class="notranslate">s</code>).</p> | <p dir="auto">I wish /docs/user-guide/deployment/ existed. It is linked from <a href="http://kubernetes.io/docs/user-guide/walkthrough/k8s201/" rel="nofollow">http://kubernetes.io/docs/user-guide/walkthrough/k8s201/</a> under the <code class="notranslate">Deployments</code> section</p> | 1 |
<p dir="auto">ERROR: /private/var/tmp/_bazel_chenxu/eac512cc1715441e7965427cfb136e20/external/protobuf_archive/BUILD:665:1: C++ compilation of rule '@protobuf_archive//:python/google/protobuf/pyext/_message.so' failed (Exit 1)<br>
external/protobuf_archive/python/google/protobuf/pyext/descriptor_containers.cc:172:13: error: <em>assigning to 'char *' from incompatible type 'const char *'</em><br>
if (PyString_AsStringAndSize(key, &name, &name_size) < 0) {<br>
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~<br>
external/protobuf_archive/python/google/protobuf/pyext/descriptor_containers.cc:69:22: note: expanded from macro 'PyString_AsStringAndSize'<br>
((<em>(charpp) = PyUnicode_AsUTF8AndSize(ob, (sizep))) == NULL? -1: 0): <br>
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~<br>
external/protobuf_archive/python/google/protobuf/pyext/descriptor_containers.cc:189:13: error: <em>assigning to 'char *' from incompatible type 'const char *'</em><br>
if (PyString_AsStringAndSize(key, &camelcase_name, &name_size) < 0) {<br>
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~<br>
external/protobuf_archive/python/google/protobuf/pyext/descriptor_containers.cc:69:22: note: expanded from macro 'PyString_AsStringAndSize'<br>
((</em>(charpp) = PyUnicode_AsUTF8AndSize(ob, (sizep))) == NULL? -1: 0): <br>
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~</p>
<hr>
<h3 dir="auto">System information</h3>
<ul dir="auto">
<li><strong>OS Platform and Distribution (e.g., Linux Ubuntu 16.04)</strong>:<br>
MACOS high sierra</li>
<li><strong>TensorFlow installed from (source or binary)</strong>:<br>
source</li>
<li><strong>TensorFlow version (use command below)</strong>:<br>
lastest</li>
<li><strong>Python version</strong>:<br>
3.7</li>
<li><strong>Bazel version (if compiling from source)</strong>:<br>
bazel version<br>
homebrew build<br>
Build label: 0.15.2-homebrew</li>
<li><strong>GCC/Compiler version (if compiling from source)</strong>:<br>
mac default gcc( Apple LLVM version 9.1.0 (clang-902.0.39.2))</li>
</ul> | <p dir="auto">I'm sure developers are working hard to catch up with Python 3.7.<br>
Is there any timeline?</p>
<p dir="auto">pip3 install tensorflow - apparently does not work, building from source:</p>
<p dir="auto">OS Platform and Distribution: Mac OS X 10.13.5<br>
Python: Python 3.7.0 (Homebrew)<br>
TensorFlow installed from: source (<a href="https://github.com/tensorflow/tensorflow.git">https://github.com/tensorflow/tensorflow.git</a>)<br>
TensorFlow version: TensorFlow 1.9.0-rc2<br>
Bazel version:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="Build label: 0.15.0-homebrew
Build target: bazel-out/darwin-opt/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer_deploy.jar
Build time: Tue Jun 26 12:42:27 2018 (1530016947)
Build timestamp: 1530016947
Build timestamp as int: 1530016947"><pre class="notranslate"><code class="notranslate">Build label: 0.15.0-homebrew
Build target: bazel-out/darwin-opt/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer_deploy.jar
Build time: Tue Jun 26 12:42:27 2018 (1530016947)
Build timestamp: 1530016947
Build timestamp as int: 1530016947
</code></pre></div>
<p dir="auto">CUDA/cuDNN version: None<br>
GPU model and memory: None<br>
Exact command to reproduce:<br>
<code class="notranslate">bazel build --config=opt //tensorflow/tools/pip_package:build_pip_package</code></p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="Starting local Bazel server and connecting to it...
...........................
WARNING: /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/BUILD:1992:1: in srcs attribute of cc_library rule @grpc//:grpc_nanopb: please do not import '@grpc//third_party/nanopb:pb_common.c' directly. You should either move the file to this package or depend on an appropriate rule there. Since this rule was created by the macro 'grpc_generate_one_off_targets', the error might have been caused by the macro implementation in /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/bazel/grpc_build_system.bzl:172:12
WARNING: /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/BUILD:1992:1: in srcs attribute of cc_library rule @grpc//:grpc_nanopb: please do not import '@grpc//third_party/nanopb:pb_decode.c' directly. You should either move the file to this package or depend on an appropriate rule there. Since this rule was created by the macro 'grpc_generate_one_off_targets', the error might have been caused by the macro implementation in /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/bazel/grpc_build_system.bzl:172:12
WARNING: /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/BUILD:1992:1: in srcs attribute of cc_library rule @grpc//:grpc_nanopb: please do not import '@grpc//third_party/nanopb:pb_encode.c' directly. You should either move the file to this package or depend on an appropriate rule there. Since this rule was created by the macro 'grpc_generate_one_off_targets', the error might have been caused by the macro implementation in /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/bazel/grpc_build_system.bzl:172:12
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/learn/BUILD:17:1: in py_library rule //tensorflow/contrib/learn:learn: target '//tensorflow/contrib/learn:learn' depends on deprecated target '//tensorflow/contrib/session_bundle:exporter': No longer supported. Switch to SavedModel immediately.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/learn/BUILD:17:1: in py_library rule //tensorflow/contrib/learn:learn: target '//tensorflow/contrib/learn:learn' depends on deprecated target '//tensorflow/contrib/session_bundle:gc': No longer supported. Switch to SavedModel immediately.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/timeseries/python/timeseries/BUILD:356:1: in py_library rule //tensorflow/contrib/timeseries/python/timeseries:ar_model: target '//tensorflow/contrib/timeseries/python/timeseries:ar_model' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/timeseries/python/timeseries/state_space_models/BUILD:73:1: in py_library rule //tensorflow/contrib/timeseries/python/timeseries/state_space_models:kalman_filter: target '//tensorflow/contrib/timeseries/python/timeseries/state_space_models:kalman_filter' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/timeseries/python/timeseries/state_space_models/BUILD:230:1: in py_library rule //tensorflow/contrib/timeseries/python/timeseries/state_space_models:filtering_postprocessor: target '//tensorflow/contrib/timeseries/python/timeseries/state_space_models:filtering_postprocessor' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/bayesflow/BUILD:17:1: in py_library rule //tensorflow/contrib/bayesflow:bayesflow_py: target '//tensorflow/contrib/bayesflow:bayesflow_py' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/seq2seq/BUILD:23:1: in py_library rule //tensorflow/contrib/seq2seq:seq2seq_py: target '//tensorflow/contrib/seq2seq:seq2seq_py' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/kfac/python/ops/BUILD:80:1: in py_library rule //tensorflow/contrib/kfac/python/ops:loss_functions: target '//tensorflow/contrib/kfac/python/ops:loss_functions' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/BUILD:14:1: in py_library rule //tensorflow/contrib:contrib_py: target '//tensorflow/contrib:contrib_py' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
INFO: Analysed target //tensorflow/tools/pip_package:build_pip_package (303 packages loaded).
INFO: Found 1 target...
INFO: From Linking external/grpc/libgrpc_base_c.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(endpoint_pair_uv.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(endpoint_pair_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(ev_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(fork_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(gethostname_fallback.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(gethostname_host_name_max.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(iocp_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(iomgr_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(pollset_set_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(pollset_uv.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(pollset_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(resolve_address_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(socket_utils_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(socket_utils_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(socket_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_client_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_server_utils_posix_noifaddrs.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_server_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_uv.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(timer_uv.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(unix_sockets_posix_noop.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(wakeup_fd_eventfd.o) has no symbols
INFO: From Linking external/grpc/libalts_util.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libalts_util.a(check_gcp_environment_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libalts_util.a(check_gcp_environment_windows.o) has no symbols
INFO: From Linking external/grpc/libtsi.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libtsi.a(ssl_session_openssl.o) has no symbols
INFO: From Linking external/grpc/libgrpc++_base.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc++_base.a(rpc_method.o) has no symbols
INFO: From Linking external/grpc/libgpr_base.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(cpu_iphone.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(cpu_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(cpu_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(env_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(env_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(log_android.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(log_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(log_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(string_util_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(string_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(sync_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(time_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(tls_pthread.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(tmpfile_msys.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(tmpfile_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(wrap_memcpy.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(thd_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(stap_timers.o) has no symbols
INFO: From Linking external/grpc/third_party/address_sorting/libaddress_sorting.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/third_party/address_sorting/libaddress_sorting.a(address_sorting_windows.o) has no symbols
ERROR: /Users/zardoz/Projects/tensorflow/tensorflow/python/BUILD:5315:1: Executing genrule //tensorflow/python:framework/fast_tensor_util.pyx_cython_translation failed (Exit 1)
Traceback (most recent call last):
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/execroot/org_tensorflow/bazel-out/host/bin/external/cython/cython_binary.runfiles/cython/cython.py", line 17, in <module>
main(command_line = 1)
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 720, in main
result = compile(sources, options)
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 695, in compile
return compile_multiple(source, options)
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 666, in compile_multiple
context = options.create_context()
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 590, in create_context
self.cplus, self.language_level, options=self)
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 75, in __init__
from . import Builtin, CythonScope
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/CythonScope.py", line 5, in <module>
from .UtilityCode import CythonUtilityCode
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/UtilityCode.py", line 3, in <module>
from .TreeFragment import parse_from_strings, StringParseContext
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/TreeFragment.py", line 17, in <module>
from .Visitor import VisitorTransform
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Visitor.py", line 15, in <module>
from . import ExprNodes
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/ExprNodes.py", line 2875
await = None
^
SyntaxError: invalid syntax
Target //tensorflow/tools/pip_package:build_pip_package failed to build
Use --verbose_failures to see the command lines of failed build steps.
INFO: Elapsed time: 179.318s, Critical Path: 6.38s
INFO: 413 processes: 413 local.
FAILED: Build did NOT complete successfully"><pre class="notranslate"><code class="notranslate">Starting local Bazel server and connecting to it...
...........................
WARNING: /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/BUILD:1992:1: in srcs attribute of cc_library rule @grpc//:grpc_nanopb: please do not import '@grpc//third_party/nanopb:pb_common.c' directly. You should either move the file to this package or depend on an appropriate rule there. Since this rule was created by the macro 'grpc_generate_one_off_targets', the error might have been caused by the macro implementation in /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/bazel/grpc_build_system.bzl:172:12
WARNING: /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/BUILD:1992:1: in srcs attribute of cc_library rule @grpc//:grpc_nanopb: please do not import '@grpc//third_party/nanopb:pb_decode.c' directly. You should either move the file to this package or depend on an appropriate rule there. Since this rule was created by the macro 'grpc_generate_one_off_targets', the error might have been caused by the macro implementation in /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/bazel/grpc_build_system.bzl:172:12
WARNING: /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/BUILD:1992:1: in srcs attribute of cc_library rule @grpc//:grpc_nanopb: please do not import '@grpc//third_party/nanopb:pb_encode.c' directly. You should either move the file to this package or depend on an appropriate rule there. Since this rule was created by the macro 'grpc_generate_one_off_targets', the error might have been caused by the macro implementation in /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/bazel/grpc_build_system.bzl:172:12
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/learn/BUILD:17:1: in py_library rule //tensorflow/contrib/learn:learn: target '//tensorflow/contrib/learn:learn' depends on deprecated target '//tensorflow/contrib/session_bundle:exporter': No longer supported. Switch to SavedModel immediately.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/learn/BUILD:17:1: in py_library rule //tensorflow/contrib/learn:learn: target '//tensorflow/contrib/learn:learn' depends on deprecated target '//tensorflow/contrib/session_bundle:gc': No longer supported. Switch to SavedModel immediately.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/timeseries/python/timeseries/BUILD:356:1: in py_library rule //tensorflow/contrib/timeseries/python/timeseries:ar_model: target '//tensorflow/contrib/timeseries/python/timeseries:ar_model' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/timeseries/python/timeseries/state_space_models/BUILD:73:1: in py_library rule //tensorflow/contrib/timeseries/python/timeseries/state_space_models:kalman_filter: target '//tensorflow/contrib/timeseries/python/timeseries/state_space_models:kalman_filter' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/timeseries/python/timeseries/state_space_models/BUILD:230:1: in py_library rule //tensorflow/contrib/timeseries/python/timeseries/state_space_models:filtering_postprocessor: target '//tensorflow/contrib/timeseries/python/timeseries/state_space_models:filtering_postprocessor' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/bayesflow/BUILD:17:1: in py_library rule //tensorflow/contrib/bayesflow:bayesflow_py: target '//tensorflow/contrib/bayesflow:bayesflow_py' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/seq2seq/BUILD:23:1: in py_library rule //tensorflow/contrib/seq2seq:seq2seq_py: target '//tensorflow/contrib/seq2seq:seq2seq_py' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/kfac/python/ops/BUILD:80:1: in py_library rule //tensorflow/contrib/kfac/python/ops:loss_functions: target '//tensorflow/contrib/kfac/python/ops:loss_functions' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/BUILD:14:1: in py_library rule //tensorflow/contrib:contrib_py: target '//tensorflow/contrib:contrib_py' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
INFO: Analysed target //tensorflow/tools/pip_package:build_pip_package (303 packages loaded).
INFO: Found 1 target...
INFO: From Linking external/grpc/libgrpc_base_c.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(endpoint_pair_uv.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(endpoint_pair_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(ev_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(fork_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(gethostname_fallback.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(gethostname_host_name_max.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(iocp_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(iomgr_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(pollset_set_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(pollset_uv.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(pollset_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(resolve_address_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(socket_utils_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(socket_utils_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(socket_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_client_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_server_utils_posix_noifaddrs.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_server_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_uv.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(timer_uv.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(unix_sockets_posix_noop.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(wakeup_fd_eventfd.o) has no symbols
INFO: From Linking external/grpc/libalts_util.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libalts_util.a(check_gcp_environment_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libalts_util.a(check_gcp_environment_windows.o) has no symbols
INFO: From Linking external/grpc/libtsi.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libtsi.a(ssl_session_openssl.o) has no symbols
INFO: From Linking external/grpc/libgrpc++_base.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc++_base.a(rpc_method.o) has no symbols
INFO: From Linking external/grpc/libgpr_base.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(cpu_iphone.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(cpu_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(cpu_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(env_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(env_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(log_android.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(log_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(log_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(string_util_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(string_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(sync_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(time_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(tls_pthread.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(tmpfile_msys.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(tmpfile_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(wrap_memcpy.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(thd_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(stap_timers.o) has no symbols
INFO: From Linking external/grpc/third_party/address_sorting/libaddress_sorting.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/third_party/address_sorting/libaddress_sorting.a(address_sorting_windows.o) has no symbols
ERROR: /Users/zardoz/Projects/tensorflow/tensorflow/python/BUILD:5315:1: Executing genrule //tensorflow/python:framework/fast_tensor_util.pyx_cython_translation failed (Exit 1)
Traceback (most recent call last):
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/execroot/org_tensorflow/bazel-out/host/bin/external/cython/cython_binary.runfiles/cython/cython.py", line 17, in <module>
main(command_line = 1)
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 720, in main
result = compile(sources, options)
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 695, in compile
return compile_multiple(source, options)
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 666, in compile_multiple
context = options.create_context()
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 590, in create_context
self.cplus, self.language_level, options=self)
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 75, in __init__
from . import Builtin, CythonScope
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/CythonScope.py", line 5, in <module>
from .UtilityCode import CythonUtilityCode
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/UtilityCode.py", line 3, in <module>
from .TreeFragment import parse_from_strings, StringParseContext
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/TreeFragment.py", line 17, in <module>
from .Visitor import VisitorTransform
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Visitor.py", line 15, in <module>
from . import ExprNodes
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/ExprNodes.py", line 2875
await = None
^
SyntaxError: invalid syntax
Target //tensorflow/tools/pip_package:build_pip_package failed to build
Use --verbose_failures to see the command lines of failed build steps.
INFO: Elapsed time: 179.318s, Critical Path: 6.38s
INFO: 413 processes: 413 local.
FAILED: Build did NOT complete successfully
</code></pre></div> | 1 |
<p dir="auto">Multi-line Text Input is a common use case (Email, Messages...).</p>
<p dir="auto">It would also be useful to be able track "cursor location" in a "input-area" so that type-ahead UIs can be correct positioned.</p> | <p dir="auto"><a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/abarth/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/abarth">@abarth</a></p> | 1 |
<p dir="auto"><strong>Apache Airflow version</strong>: 2.1.1</p>
<p dir="auto"><strong>Kubernetes version (if you are using kubernetes)</strong> (use <code class="notranslate">kubectl version</code>): AWS EKS, AWS ECS</p>
<p dir="auto"><strong>Environment</strong>:</p>
<ul dir="auto">
<li><strong>Cloud provider or hardware configuration</strong>: AWS</li>
<li><strong>OS</strong> (e.g. from /etc/os-release): Debian GNU/Linux 10</li>
<li><strong>Kernel</strong> (e.g. <code class="notranslate">uname -a</code>):</li>
<li><strong>Install tools</strong>:</li>
<li><strong>Others</strong>:</li>
</ul>
<p dir="auto"><strong>What happened</strong>:</p>
<p dir="auto">I have enabled RBAC and using FAB OAUTH to authenticate users using AZURE AD. Users are able to sign-in using Azure SSO. I have referenced my webserver config file from <a href="https://github.com/dpgaspar/Flask-AppBuilder/blob/master/examples/oauth/config.py">https://github.com/dpgaspar/Flask-AppBuilder/blob/master/examples/oauth/config.py</a> with OAUTH_PROVIDER as Azure.</p>
<p dir="auto">Now all our users are acquiring role ADMIN and I am not able to edit their role to other available roles as there is no edit option coming in the List USER page. My user to have an ADMIN role.</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/19628707/125004469-926e9880-e027-11eb-8c9c-2d370f61b2be.png"><img src="https://user-images.githubusercontent.com/19628707/125004469-926e9880-e027-11eb-8c9c-2d370f61b2be.png" alt="Screen Shot 2021-07-08 at 10 07 05 AM" style="max-width: 100%;"></a></p>
<p dir="auto"><strong>What you expected to happen</strong>:</p>
<p dir="auto">As ADMIN is having <code class="notranslate">can edit on UserOAuthModelView</code> permission I should be able to edit the roles for different users.</p>
<p dir="auto"><strong>How to reproduce it</strong>:</p>
<p dir="auto"><strong>Anything else we need to know</strong>:</p> | <p dir="auto"><strong>Apache Airflow version</strong>: 2.1.0</p>
<p dir="auto"><strong>Browsers</strong>: Chrome and Firefox</p>
<p dir="auto"><strong>What happened</strong>:</p>
<p dir="auto">Before upgrading to 2.1.0</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/14293802/120359517-c1ca1100-c2d5-11eb-95ba-58ccc0a3ac37.png"><img src="https://user-images.githubusercontent.com/14293802/120359517-c1ca1100-c2d5-11eb-95ba-58ccc0a3ac37.png" alt="before" style="max-width: 100%;"></a></p>
<p dir="auto">After upgrading to 2.1.0</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/14293802/120359528-c4c50180-c2d5-11eb-8e04-f34846ea2736.png"><img src="https://user-images.githubusercontent.com/14293802/120359528-c4c50180-c2d5-11eb-8e04-f34846ea2736.png" alt="after" style="max-width: 100%;"></a></p>
<p dir="auto"><strong>What you expected to happen</strong>:</p>
<p dir="auto">Show/Edit/Delete under Security -> Users are available</p>
<p dir="auto"><strong>How to reproduce it</strong>:</p>
<p dir="auto">Go to Security -> Users (as an admin of course)</p> | 1 |
<h2 dir="auto">Environment info</h2>
<ul dir="auto">
<li><code class="notranslate">transformers</code> version:4.3.3</li>
<li>Platform:Pytorch</li>
<li>Python version:3.7.0</li>
<li>PyTorch version (GPU?):GPU</li>
<li>Tensorflow version (GPU?):</li>
<li>Using GPU in script?:</li>
<li>Using distributed or parallel set-up in script?:NO</li>
</ul>
<h3 dir="auto">Who can help</h3>
<h2 dir="auto">Information</h2>
<p dir="auto">Model I am using (Bert, XLNet ...):</p>
<p dir="auto">The problem arises when using:</p>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> the official example scripts: (give details below)<br>
<a href="https://github.com/huggingface/transformers/blob/master/examples/language-modeling/run_mlm.py">https://github.com/huggingface/transformers/blob/master/examples/language-modeling/run_mlm.py</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> my own modified scripts: (give details below)</li>
</ul>
<p dir="auto">The tasks I am working on is:</p>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> an official GLUE/SQUaD task: (give the name)</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> my own task or dataset: (give details below)</li>
</ul>
<h2 dir="auto">To reproduce</h2>
<p dir="auto">Steps to reproduce the behavior:</p>
<p dir="auto">1.python run_mlm.py --model_name_or_path release_model/ --dataset_name wikitext --dataset_config_name wikitext-2-raw-v1 --do_train --do_eval --output_dir test-mlm --max_seq_length 128<br>
2.Got an error:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="[INFO|trainer.py:1408] 2021-03-15 11:10:32,884 >> Saving model checkpoint to test-mlm
[INFO|configuration_utils.py:304] 2021-03-15 11:10:32,886 >> Configuration saved in test-mlm/config.json
[INFO|modeling_utils.py:817] 2021-03-15 11:10:33,863 >> Model weights saved in test-mlm/pytorch_model.bin
Traceback (most recent call last):
File "run_mlm.py", line 475, in <module>
main()
File "run_mlm.py", line 450, in main
trainer.log_metrics("train", metrics)
AttributeError: 'Trainer' object has no attribute 'log_metrics'"><pre class="notranslate"><code class="notranslate">[INFO|trainer.py:1408] 2021-03-15 11:10:32,884 >> Saving model checkpoint to test-mlm
[INFO|configuration_utils.py:304] 2021-03-15 11:10:32,886 >> Configuration saved in test-mlm/config.json
[INFO|modeling_utils.py:817] 2021-03-15 11:10:33,863 >> Model weights saved in test-mlm/pytorch_model.bin
Traceback (most recent call last):
File "run_mlm.py", line 475, in <module>
main()
File "run_mlm.py", line 450, in main
trainer.log_metrics("train", metrics)
AttributeError: 'Trainer' object has no attribute 'log_metrics'
</code></pre></div>
<h2 dir="auto">Expected behavior</h2>
<p dir="auto">The script finish without error.</p> | <p dir="auto">I've been trying to work out why I keep getting a CUDA assert in a specific mini-batch when training RoBERTa from scratch. I finally tracked it down after switching to CPU.</p>
<p dir="auto">I don't understand why <code class="notranslate">padding_idx</code> is added to <code class="notranslate">incremental_indices</code> below? - <em>edit: I do in that the embedding needs a padding mask but it I'm not sure it's the correct way to do it.</em></p>
<p dir="auto">In my case padding_idx=3. And one of my input_ids rows was truncated. Say I have input_ids = [[4,5,6],[4,3,3]], this results in mask=[[1,1,1],[1,0,0]] and incremental index=[[1,2,3],[1,0,0]]. Adding padding_idx then produces [[4,5,6],[4,3,3]].</p>
<p dir="auto">The issue is <code class="notranslate">self.position_embeddings = nn.Embedding(config.max_position_embeddings, config.hidden_size)</code> so for any sequences which are truncated adding anything to the indices results in an index which is greater than the embedding dim.</p>
<p dir="auto">Perhaps you can argue that max_position_embeddings is supposed to be larger than the largest possible sequence so this doesn't happen? There is a check in <code class="notranslate">run_mlm.py</code> that <code class="notranslate">data_args.max_seq_length > tokenizer.model_max_length</code> but it seems that in actual fact to avoid a very hard to track down error you must have tokenizer.model_max_length < data_args.max_seq_length</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="def create_position_ids_from_input_ids(input_ids, padding_idx):
"""
Replace non-padding symbols with their position numbers. Position numbers begin at padding_idx+1. Padding symbols
are ignored. This is modified from fairseq's `utils.make_positions`.
Args:
x: torch.Tensor x:
Returns: torch.Tensor
"""
# The series of casts and type-conversions here are carefully balanced to both work with ONNX export and XLA.
mask = input_ids.ne(padding_idx).int()
incremental_indices = torch.cumsum(mask, dim=1).type_as(mask) * mask
return incremental_indices.long() + padding_idx"><pre class="notranslate"><code class="notranslate">def create_position_ids_from_input_ids(input_ids, padding_idx):
"""
Replace non-padding symbols with their position numbers. Position numbers begin at padding_idx+1. Padding symbols
are ignored. This is modified from fairseq's `utils.make_positions`.
Args:
x: torch.Tensor x:
Returns: torch.Tensor
"""
# The series of casts and type-conversions here are carefully balanced to both work with ONNX export and XLA.
mask = input_ids.ne(padding_idx).int()
incremental_indices = torch.cumsum(mask, dim=1).type_as(mask) * mask
return incremental_indices.long() + padding_idx
</code></pre></div> | 0 |
<p dir="auto">by <strong>j.christoff25</strong>:</p>
<pre class="notranslate">Before filing a bug, please check whether it has been fixed since
the latest release: run "hg pull -u" and retry what you did to
reproduce the problem. Thanks.
What steps will reproduce the problem?
1.6g demo.go
6l demo.6
2.
3.
What is the expected output?
a working demo
What do you see instead?
too many loads
Which compiler are you using (5g, 6g, 8g, gccgo)?
I use the "make install", followed by "make example"
Which operating system are you using?
Darwin AMD64 on a Imac 10.6
Which revision are you using? (hg identify)
117785e448a6+ tip
Please provide any additional information below.
I clone the file with git clone <a href="https://github.com/mattn/go-gtk.git">https://github.com/mattn/go-gtk.git</a>
This program will compile & link using 8g & 8l. I have tried to link this demo
with several versions of GO in the past. All yield the same error,"too many
loads". What does "too many loads" mean?
The following is the message I get when using "make install', followed by
"make example"
jchristoffs-iMac:go-gtk jim$ make install
cd glib && gomake install
CGOPKGPATH= cgo -- `pkg-config --cflags glib-2.0 gobject-2.0` glib.go
touch _cgo_run
6g -o _go_.6 glib.cgo1.go _cgo_gotypes.go
6c -FVw -I/Users/jim/go/pkg/darwin_amd64 _cgo_defun.c
gcc -m64 -g -fPIC -O2 -o _cgo_main.o -c `pkg-config --cflags glib-2.0 gobject-2.0`
_cgo_main.c
gcc -m64 -g -fPIC -O2 -o glib.cgo2.o -c `pkg-config --cflags glib-2.0 gobject-2.0`
glib.cgo2.c
gcc -m64 -g -fPIC -O2 -o _cgo_export.o -c `pkg-config --cflags glib-2.0 gobject-2.0`
_cgo_export.c
gcc -m64 -g -fPIC -O2 -o _cgo1_.o _cgo_main.o glib.cgo2.o _cgo_export.o `pkg-config
--libs glib-2.0 gobject-2.0`
cgo -dynimport _cgo1_.o >__cgo_import.c && mv -f __cgo_import.c _cgo_import.c
6c -FVw _cgo_import.c
rm -f _obj/glib.a
gopack grc _obj/glib.a _go_.6 _cgo_defun.6 _cgo_import.6 glib.cgo2.o _cgo_export.o
cp _obj/glib.a "/Users/jim/go/pkg/darwin_amd64/glib.a"
cd gdk && gomake install
CGOPKGPATH= cgo -- `pkg-config --cflags gdk-2.0` gdk.go
touch _cgo_run
6g -o _go_.6 gdk.cgo1.go _cgo_gotypes.go
6c -FVw -I/Users/jim/go/pkg/darwin_amd64 _cgo_defun.c
gcc -m64 -g -fPIC -O2 -o _cgo_main.o -c `pkg-config --cflags gdk-2.0` _cgo_main.c
gcc -m64 -g -fPIC -O2 -o gdk.cgo2.o -c `pkg-config --cflags gdk-2.0` gdk.cgo2.c
gcc -m64 -g -fPIC -O2 -o _cgo_export.o -c `pkg-config --cflags gdk-2.0` _cgo_export.c
gcc -m64 -g -fPIC -O2 -o _cgo1_.o _cgo_main.o gdk.cgo2.o _cgo_export.o `pkg-config
--libs gdk-2.0 gthread-2.0`
cgo -dynimport _cgo1_.o >__cgo_import.c && mv -f __cgo_import.c _cgo_import.c
6c -FVw _cgo_import.c
rm -f _obj/gdk.a
gopack grc _obj/gdk.a _go_.6 _cgo_defun.6 _cgo_import.6 gdk.cgo2.o _cgo_export.o
cp _obj/gdk.a "/Users/jim/go/pkg/darwin_amd64/gdk.a"
cd gdkpixbuf && gomake install
CGOPKGPATH= cgo -- `pkg-config --cflags gdk-pixbuf-2.0` gdkpixbuf.go
touch _cgo_run
6g -I ../glib/_obj -o _go_.6 gdkpixbuf.cgo1.go _cgo_gotypes.go
6c -FVw -I/Users/jim/go/pkg/darwin_amd64 _cgo_defun.c
gcc -m64 -g -fPIC -O2 -o _cgo_main.o -c `pkg-config --cflags gdk-pixbuf-2.0` _cgo_main.c
gcc -m64 -g -fPIC -O2 -o gdkpixbuf.cgo2.o -c `pkg-config --cflags gdk-pixbuf-2.0`
gdkpixbuf.cgo2.c
gcc -m64 -g -fPIC -O2 -o _cgo_export.o -c `pkg-config --cflags gdk-pixbuf-2.0`
_cgo_export.c
gcc -m64 -g -fPIC -O2 -o _cgo1_.o _cgo_main.o gdkpixbuf.cgo2.o _cgo_export.o `pkg-config
--libs gdk-pixbuf-2.0`
cgo -dynimport _cgo1_.o >__cgo_import.c && mv -f __cgo_import.c _cgo_import.c
6c -FVw _cgo_import.c
rm -f _obj/gdkpixbuf.a
gopack grc _obj/gdkpixbuf.a _go_.6 _cgo_defun.6 _cgo_import.6 gdkpixbuf.cgo2.o
_cgo_export.o
cp _obj/gdkpixbuf.a "/Users/jim/go/pkg/darwin_amd64/gdkpixbuf.a"
cd gtk && gomake install
CGOPKGPATH= cgo -- `pkg-config --cflags gtk+-2.0 gtksourceview-2.0` gtk.go
touch _cgo_run
6g -I ../glib/_obj -I ../gdk/_obj -I ../gdkpixbuf/_obj -o _go_.6 gtk.cgo1.go
_cgo_gotypes.go
6c -FVw -I/Users/jim/go/pkg/darwin_amd64 _cgo_defun.c
gcc -m64 -g -fPIC -O2 -o _cgo_main.o -c `pkg-config --cflags gtk+-2.0 gtksourceview-2.0`
_cgo_main.c
gcc -m64 -g -fPIC -O2 -o gtk.cgo2.o -c `pkg-config --cflags gtk+-2.0 gtksourceview-2.0`
gtk.cgo2.c
gtk.go: In function '_cgo_17e138e95a5d_Cfunc_gtk_about_dialog_set_artists':
gtk.go:1655: warning: passing argument 2 of 'gtk_about_dialog_set_artists' from
incompatible pointer type
gtk.go: In function '_cgo_17e138e95a5d_Cfunc_gtk_about_dialog_get_documenters':
gtk.go:2377: warning: assignment discards qualifiers from pointer target type
gtk.go: In function '_cgo_17e138e95a5d_Cfunc_gtk_about_dialog_set_documenters':
gtk.go:5071: warning: passing argument 2 of 'gtk_about_dialog_set_documenters' from
incompatible pointer type
gtk.go: In function '_cgo_17e138e95a5d_Cfunc_make_strings':
gtk.go:6935: warning: assignment from incompatible pointer type
gtk.go: In function '_cgo_17e138e95a5d_Cfunc_gtk_about_dialog_get_authors':
gtk.go:8005: warning: assignment discards qualifiers from pointer target type
gtk.go: In function '_cgo_17e138e95a5d_Cfunc_gtk_about_dialog_set_authors':
gtk.go:9161: warning: passing argument 2 of 'gtk_about_dialog_set_authors' from
incompatible pointer type
gtk.go: In function '_cgo_17e138e95a5d_Cfunc_next_gcharptr':
gtk.go:9430: warning: assignment from incompatible pointer type
gtk.go: In function '_cgo_17e138e95a5d_Cfunc_gtk_about_dialog_get_artists':
gtk.go:9704: warning: assignment discards qualifiers from pointer target type
gcc -m64 -g -fPIC -O2 -o _cgo_export.o -c `pkg-config --cflags gtk+-2.0
gtksourceview-2.0` _cgo_export.c
gcc -m64 -g -fPIC -O2 -o _cgo1_.o _cgo_main.o gtk.cgo2.o _cgo_export.o -lpthread
`pkg-config --libs gtk+-2.0 gtksourceview-2.0`
cgo -dynimport _cgo1_.o >__cgo_import.c && mv -f __cgo_import.c _cgo_import.c
6c -FVw _cgo_import.c
rm -f _obj/gtk.a
gopack grc _obj/gtk.a _go_.6 _cgo_defun.6 _cgo_import.6 gtk.cgo2.o _cgo_export.o
cp _obj/gtk.a "/Users/jim/go/pkg/darwin_amd64/gtk.a"
jchristoffs-iMac:go-gtk jim$ make example
cd glib && gomake install
cd gdk && gomake install
cd gdkpixbuf && gomake install
cd gtk && gomake install
cd example && gomake
cd demo && gomake
6g -o _go_.6 demo.go
6l -o demo _go_.6
too many loads
make[2]: *** [demo] Error 1
make[1]: *** [all] Error 2
make: *** [example] Error 2
jchristoffs-iMac:go-gtk jim$</pre>
<p dir="auto">Attachments:</p>
<ol dir="auto">
<li><a href="https://storage.googleapis.com/go-attachment/1570/0/demo.go" rel="nofollow">demo.go</a> (9763 bytes)</li>
</ol> | <pre class="notranslate">I have installed lots of third pkg.
The local godoc's /pkg page looks very messy.
I think show the gopath pkg in a separate page is better.</pre> | 0 |
<table role="table">
<thead>
<tr>
<th>Q</th>
<th>A</th>
</tr>
</thead>
<tbody>
<tr>
<td>Bug report?</td>
<td>yes (documentation)</td>
</tr>
<tr>
<td>Feature request?</td>
<td>no</td>
</tr>
<tr>
<td>BC Break report?</td>
<td>no</td>
</tr>
<tr>
<td>RFC?</td>
<td>no</td>
</tr>
<tr>
<td>Symfony version</td>
<td>all</td>
</tr>
</tbody>
</table>
<p dir="auto">Hello,<br>
I'm digging through Symfony's Security components trying to implement a custom security bundle, and I really miss one thing: The Big Picture. There is a lot written about individual components and even source code is quite understandable, but there is nothing about how these components work together.</p>
<p dir="auto">The workflow diagrams in <a href="http://symfony.com/doc/current/components/http_kernel.html" rel="nofollow">HttpKernel documentation</a> are extremely helpful. There should be similar diagrams included in <a href="https://symfony.com/doc/current/components/security.html" rel="nofollow">Security component documentation</a>. How all these components interact with each other? What happens during ordinary HTTP request? What happens during login? What happens during subsequent HTTP requests? …? …?</p>
<p dir="auto">The documentation says there are some components (Guards missing, btw) and that's it. No relation between them is provided. No context. Nothing about basic principles nor ideas.</p>
<p dir="auto">Security is important and must be well understood to make it right. A cookbook without context is not enough — it helps when one understands, but does not make him understand. Symfony Security component is powerful and complex, so it is hard to get into it.</p>
<p dir="auto">Someone who has deep understanding of the Security component, please, draw few diagrams how it works. It will help many other developers.</p> | <p dir="auto">If write following config:</p>
<div class="highlight highlight-text-xml notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content=" <service id="service" alias="service_alias"/>
<service id="sevice_alias" alias="service"/>"><pre class="notranslate"> <<span class="pl-ent">service</span> <span class="pl-e">id</span>=<span class="pl-s"><span class="pl-pds">"</span>service<span class="pl-pds">"</span></span> <span class="pl-e">alias</span>=<span class="pl-s"><span class="pl-pds">"</span>service_alias<span class="pl-pds">"</span></span>/>
<<span class="pl-ent">service</span> <span class="pl-e">id</span>=<span class="pl-s"><span class="pl-pds">"</span>sevice_alias<span class="pl-pds">"</span></span> <span class="pl-e">alias</span>=<span class="pl-s"><span class="pl-pds">"</span>service<span class="pl-pds">"</span></span>/></pre></div>
<p dir="auto">Infinite loop will occur in the ResolveReferencesToAliasesPass class. Think such circular references should be catched and an exception should occur instead of falling into infinite loop.</p> | 0 |
<p dir="auto">TypeScript version: 1.7.5<br>
For example:</p>
<p dir="auto">hoge.ts</p>
<div class="highlight highlight-source-ts notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="export default class Hoge {}"><pre class="notranslate"><span class="pl-k">export</span> <span class="pl-k">default</span> <span class="pl-k">class</span> <span class="pl-smi">Hoge</span> <span class="pl-kos">{</span><span class="pl-kos">}</span></pre></div>
<p dir="auto">fuga.ts</p>
<div class="highlight highlight-source-ts notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="import Hoge from "./hoge"
new Hoge();"><pre class="notranslate"><span class="pl-k">import</span> <span class="pl-smi">Hoge</span> <span class="pl-k">from</span> <span class="pl-s">"./hoge"</span>
<span class="pl-k">new</span> <span class="pl-smi">Hoge</span><span class="pl-kos">(</span><span class="pl-kos">)</span><span class="pl-kos">;</span></pre></div>
<p dir="auto"><code class="notranslate">tsc hoge.ts fuga.ts -t es6 -m commonjs</code><br>
compile result:<br>
hoge.js</p>
<div class="highlight highlight-source-js notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="class Hoge {
}
exports.Hoge = Hoge;"><pre class="notranslate"><span class="pl-k">class</span> <span class="pl-v">Hoge</span> <span class="pl-kos">{</span>
<span class="pl-kos">}</span>
<span class="pl-s1">exports</span><span class="pl-kos">.</span><span class="pl-c1">Hoge</span> <span class="pl-c1">=</span> <span class="pl-v">Hoge</span><span class="pl-kos">;</span></pre></div>
<p dir="auto">fuga.js</p>
<div class="highlight highlight-source-js notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="var hoge_1 = require("./hoge");
new hoge_1.default();"><pre class="notranslate"><span class="pl-k">var</span> <span class="pl-s1">hoge_1</span> <span class="pl-c1">=</span> <span class="pl-en">require</span><span class="pl-kos">(</span><span class="pl-s">"./hoge"</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-k">new</span> <span class="pl-s1">hoge_1</span><span class="pl-kos">.</span><span class="pl-c1">default</span><span class="pl-kos">(</span><span class="pl-kos">)</span><span class="pl-kos">;</span></pre></div>
<p dir="auto">Compiled code doesn't work because <code class="notranslate">require("./hoge")</code> returns <code class="notranslate">{ Hoge: Hoge }</code> and <code class="notranslate">hoge_1.default</code> is undefined.</p> | <div class="highlight highlight-source-ts notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="export default class {
}"><pre class="notranslate"><span class="pl-k">export</span> <span class="pl-k">default</span> <span class="pl-k">class</span> <span class="pl-kos">{</span>
<span class="pl-kos">}</span></pre></div>
<p dir="auto">compiled code is:</p>
<div class="highlight highlight-source-js notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="class {
}
exports. = default_1;
"><pre class="notranslate"><span class="pl-k">class</span> <span class="pl-kos">{</span>
<span class="pl-kos">}</span>
<span class="pl-s1">exports</span><span class="pl-kos">.</span> <span class="pl-c1">=</span> <span class="pl-c1">default_1</span><span class="pl-kos">;</span></pre></div>
<p dir="auto">compile command is <code class="notranslate">tsc --module commonjs --target ES6 default.ts</code>, TypeScript version is 1.7.5</p> | 1 |
<h3 dir="auto">Describe the issue:</h3>
<p dir="auto">We have the following code:</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="I_wb = I_wb.astype(np.uint16)"><pre class="notranslate"><span class="pl-v">I_wb</span> <span class="pl-c1">=</span> <span class="pl-v">I_wb</span>.<span class="pl-en">astype</span>(<span class="pl-s1">np</span>.<span class="pl-s1">uint16</span>)</pre></div>
<p dir="auto">I_wb was float32 before the above conversion.<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/326807/148121640-5a047b35-ef45-456c-90d2-407030bc51c4.png"><img width="811" alt="Screen Shot 2022-01-04 at 12 31 16 PM" src="https://user-images.githubusercontent.com/326807/148121640-5a047b35-ef45-456c-90d2-407030bc51c4.png" style="max-width: 100%;"></a></p>
<p dir="auto">We noticed that when overflow happens, astype could behave differently on different platforms.</p>
<p dir="auto">as shown in the above image, when running the same code on mac, the value 68469.13 will be truncated to 65535, whereas on linux, the value is applied with a modulo operation . 68469.13 % 65535</p>
<p dir="auto">on both sides, the code uses numpy 1.21.2</p>
<p dir="auto">also tried 1.22.0, it's the same.</p>
<p dir="auto">note: The shape of the data matters,</p>
<p dir="auto">for example:</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="I2 = np.zeros((2448, 2048, 3), dtype="float32")
I2[1700,970] = [68469.13, 65472., 65535.]"><pre class="notranslate"><span class="pl-v">I2</span> <span class="pl-c1">=</span> <span class="pl-s1">np</span>.<span class="pl-en">zeros</span>((<span class="pl-c1">2448</span>, <span class="pl-c1">2048</span>, <span class="pl-c1">3</span>), <span class="pl-s1">dtype</span><span class="pl-c1">=</span><span class="pl-s">"float32"</span>)
<span class="pl-v">I2</span>[<span class="pl-c1">1700</span>,<span class="pl-c1">970</span>] <span class="pl-c1">=</span> [<span class="pl-c1">68469.13</span>, <span class="pl-c1">65472.</span>, <span class="pl-c1">65535.</span>]</pre></div>
<p dir="auto">The above can reproduce, the value 68469.13 gets truncated on certain platforms, but applied with modulo on other platforms.</p>
<p dir="auto">Whereas the following won't reproduce:</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="I = np.array([68469.13, 65472., 65535.], dtype="float32")"><pre class="notranslate"><span class="pl-v">I</span> <span class="pl-c1">=</span> <span class="pl-s1">np</span>.<span class="pl-en">array</span>([<span class="pl-c1">68469.13</span>, <span class="pl-c1">65472.</span>, <span class="pl-c1">65535.</span>], <span class="pl-s1">dtype</span><span class="pl-c1">=</span><span class="pl-s">"float32"</span>)</pre></div>
<p dir="auto">on both platforms I tried, it behaves the same ( modulo)</p>
<p dir="auto">could this be due to the underlining cpp compiler and optimization level (clang vs. gcc) ?</p>
<h3 dir="auto">Reproduce the code example:</h3>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="import numpy as np
import sys
print(np.__version__ , sys.version)
I2 = np.zeros((2448, 2048, 3), dtype="float32")
I2[1700,970] = [68469.13, 65472., 65535.]
print("pixel color: ", I2[1700,970])
I2_wb = I2.astype(np.uint16)
print(I2_wb[1700,970])"><pre class="notranslate"><span class="pl-k">import</span> <span class="pl-s1">numpy</span> <span class="pl-k">as</span> <span class="pl-s1">np</span>
<span class="pl-k">import</span> <span class="pl-s1">sys</span>
<span class="pl-en">print</span>(<span class="pl-s1">np</span>.<span class="pl-s1">__version__</span> , <span class="pl-s1">sys</span>.<span class="pl-s1">version</span>)
<span class="pl-v">I2</span> <span class="pl-c1">=</span> <span class="pl-s1">np</span>.<span class="pl-en">zeros</span>((<span class="pl-c1">2448</span>, <span class="pl-c1">2048</span>, <span class="pl-c1">3</span>), <span class="pl-s1">dtype</span><span class="pl-c1">=</span><span class="pl-s">"float32"</span>)
<span class="pl-v">I2</span>[<span class="pl-c1">1700</span>,<span class="pl-c1">970</span>] <span class="pl-c1">=</span> [<span class="pl-c1">68469.13</span>, <span class="pl-c1">65472.</span>, <span class="pl-c1">65535.</span>]
<span class="pl-en">print</span>(<span class="pl-s">"pixel color: "</span>, <span class="pl-v">I2</span>[<span class="pl-c1">1700</span>,<span class="pl-c1">970</span>])
<span class="pl-v">I2_wb</span> <span class="pl-c1">=</span> <span class="pl-v">I2</span>.<span class="pl-en">astype</span>(<span class="pl-s1">np</span>.<span class="pl-s1">uint16</span>)
<span class="pl-en">print</span>(<span class="pl-v">I2_wb</span>[<span class="pl-c1">1700</span>,<span class="pl-c1">970</span>])</pre></div>
<h3 dir="auto">Error message:</h3>
<div class="highlight highlight-source-shell notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="compare the reproduce code on different platforms, notice output difference even with matched numpy version."><pre class="notranslate">compare the reproduce code on different platforms, notice output difference even with matched numpy version.</pre></div>
<h3 dir="auto">NumPy/Python version information:</h3>
<p dir="auto">1.22.0<br>
1.21.2</p>
<p dir="auto">sample output on my mac (value truncated):</p>
<div class="highlight highlight-source-shell notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="1.21.2 3.9.1 (default, Oct 15 2021, 13:26:04)
[Clang 13.0.0 (clang-1300.0.29.3)]
pixel color: [68469.13 65472. 65535. ]
[65535 65472 65535]"><pre class="notranslate">1.21.2 3.9.1 (default, Oct 15 2021, 13:26:04)
[Clang 13.0.0 (clang-1300.0.29.3)]
pixel color: [68469.13 65472. 65535. ]
[65535 65472 65535]</pre></div>
<p dir="auto">sample output on my linux (modulo)</p>
<div class="highlight highlight-source-shell notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="1.21.2 3.9.7 (default, Sep 10 2021, 14:59:43)
[GCC 11.2.0]
pixel color: [68469.13 65472. 65535. ]
[ 2933 65472 65535]
"><pre class="notranslate">1.21.2 3.9.7 (default, Sep 10 2021, 14:59:43)
[GCC 11.2.0]
pixel color: [68469.13 65472. 65535. ]
[ 2933 65472 65535]
</pre></div> | <h3 dir="auto">Describe the issue:</h3>
<p dir="auto">We have the following code:</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="I_wb = I_wb.astype(np.uint16)"><pre class="notranslate"><span class="pl-v">I_wb</span> <span class="pl-c1">=</span> <span class="pl-v">I_wb</span>.<span class="pl-en">astype</span>(<span class="pl-s1">np</span>.<span class="pl-s1">uint16</span>)</pre></div>
<p dir="auto">I_wb was float32 before the above conversion.<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/326807/148121640-5a047b35-ef45-456c-90d2-407030bc51c4.png"><img width="811" alt="Screen Shot 2022-01-04 at 12 31 16 PM" src="https://user-images.githubusercontent.com/326807/148121640-5a047b35-ef45-456c-90d2-407030bc51c4.png" style="max-width: 100%;"></a></p>
<p dir="auto">We noticed that when overflow happens, astype could behave differently on different platforms.</p>
<p dir="auto">as shown in the above image, when running the same code on mac, the value 68469.13 will be truncated to 65535, whereas on linux, the value is applied with a modulo operation . 68469.13 % 65535</p>
<p dir="auto">on both sides, the code uses numpy 1.21.2</p>
<p dir="auto">also tried 1.22.0, it's the same.</p>
<p dir="auto">note: The shape of the data matters,</p>
<p dir="auto">for example:</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="I2 = np.zeros((2448, 2048, 3), dtype="float32")
I2[1700,970] = [68469.13, 65472., 65535.]"><pre class="notranslate"><span class="pl-v">I2</span> <span class="pl-c1">=</span> <span class="pl-s1">np</span>.<span class="pl-en">zeros</span>((<span class="pl-c1">2448</span>, <span class="pl-c1">2048</span>, <span class="pl-c1">3</span>), <span class="pl-s1">dtype</span><span class="pl-c1">=</span><span class="pl-s">"float32"</span>)
<span class="pl-v">I2</span>[<span class="pl-c1">1700</span>,<span class="pl-c1">970</span>] <span class="pl-c1">=</span> [<span class="pl-c1">68469.13</span>, <span class="pl-c1">65472.</span>, <span class="pl-c1">65535.</span>]</pre></div>
<p dir="auto">The above can reproduce, the value 68469.13 gets truncated on certain platforms, but applied with modulo on other platforms.</p>
<p dir="auto">Whereas the following won't reproduce:</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="I = np.array([68469.13, 65472., 65535.], dtype="float32")"><pre class="notranslate"><span class="pl-v">I</span> <span class="pl-c1">=</span> <span class="pl-s1">np</span>.<span class="pl-en">array</span>([<span class="pl-c1">68469.13</span>, <span class="pl-c1">65472.</span>, <span class="pl-c1">65535.</span>], <span class="pl-s1">dtype</span><span class="pl-c1">=</span><span class="pl-s">"float32"</span>)</pre></div>
<p dir="auto">on both platforms I tried, it behaves the same ( modulo)</p>
<p dir="auto">could this be due to the underlining cpp compiler and optimization level (clang vs. gcc) ?</p>
<h3 dir="auto">Reproduce the code example:</h3>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="import numpy as np
import sys
print(np.__version__ , sys.version)
I2 = np.zeros((2448, 2048, 3), dtype="float32")
I2[1700,970] = [68469.13, 65472., 65535.]
print("pixel color: ", I2[1700,970])
I2_wb = I2.astype(np.uint16)
print(I2_wb[1700,970])"><pre class="notranslate"><span class="pl-k">import</span> <span class="pl-s1">numpy</span> <span class="pl-k">as</span> <span class="pl-s1">np</span>
<span class="pl-k">import</span> <span class="pl-s1">sys</span>
<span class="pl-en">print</span>(<span class="pl-s1">np</span>.<span class="pl-s1">__version__</span> , <span class="pl-s1">sys</span>.<span class="pl-s1">version</span>)
<span class="pl-v">I2</span> <span class="pl-c1">=</span> <span class="pl-s1">np</span>.<span class="pl-en">zeros</span>((<span class="pl-c1">2448</span>, <span class="pl-c1">2048</span>, <span class="pl-c1">3</span>), <span class="pl-s1">dtype</span><span class="pl-c1">=</span><span class="pl-s">"float32"</span>)
<span class="pl-v">I2</span>[<span class="pl-c1">1700</span>,<span class="pl-c1">970</span>] <span class="pl-c1">=</span> [<span class="pl-c1">68469.13</span>, <span class="pl-c1">65472.</span>, <span class="pl-c1">65535.</span>]
<span class="pl-en">print</span>(<span class="pl-s">"pixel color: "</span>, <span class="pl-v">I2</span>[<span class="pl-c1">1700</span>,<span class="pl-c1">970</span>])
<span class="pl-v">I2_wb</span> <span class="pl-c1">=</span> <span class="pl-v">I2</span>.<span class="pl-en">astype</span>(<span class="pl-s1">np</span>.<span class="pl-s1">uint16</span>)
<span class="pl-en">print</span>(<span class="pl-v">I2_wb</span>[<span class="pl-c1">1700</span>,<span class="pl-c1">970</span>])</pre></div>
<h3 dir="auto">Error message:</h3>
<div class="highlight highlight-source-shell notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="compare the reproduce code on different platforms, notice output difference even with matched numpy version."><pre class="notranslate">compare the reproduce code on different platforms, notice output difference even with matched numpy version.</pre></div>
<h3 dir="auto">NumPy/Python version information:</h3>
<p dir="auto">1.22.0<br>
1.21.2</p>
<p dir="auto">sample output on my mac (value truncated):</p>
<div class="highlight highlight-source-shell notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="1.21.2 3.9.1 (default, Oct 15 2021, 13:26:04)
[Clang 13.0.0 (clang-1300.0.29.3)]
pixel color: [68469.13 65472. 65535. ]
[65535 65472 65535]"><pre class="notranslate">1.21.2 3.9.1 (default, Oct 15 2021, 13:26:04)
[Clang 13.0.0 (clang-1300.0.29.3)]
pixel color: [68469.13 65472. 65535. ]
[65535 65472 65535]</pre></div>
<p dir="auto">sample output on my linux (modulo)</p>
<div class="highlight highlight-source-shell notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="1.21.2 3.9.7 (default, Sep 10 2021, 14:59:43)
[GCC 11.2.0]
pixel color: [68469.13 65472. 65535. ]
[ 2933 65472 65535]
"><pre class="notranslate">1.21.2 3.9.7 (default, Sep 10 2021, 14:59:43)
[GCC 11.2.0]
pixel color: [68469.13 65472. 65535. ]
[ 2933 65472 65535]
</pre></div> | 1 |
<h1 dir="auto">Feature request</h1>
<h2 dir="auto">Is your feature request related to a problem? Please describe.</h2>
<p dir="auto">url structures are still visible on devices, so its a UI-element and often is bound to editorial decisions, yet in nextjs, url routes always reflect the filesystem of the project</p>
<p dir="auto">Having the filename of a page be the route path segment might be ok, if you develop for an english target group, but leads to problems in other markets. E.g. in switzerland we often need to have an app in multiple languages and customers often wish that the url paths are also translated.</p>
<p dir="auto">Also, there is a common practice, that source code is always in english. So this leads to the problem, that when you decide to have e.g. german url paths, you need to have german-named files in your app, which just feels wrong.</p>
<p dir="auto">Nextjs should therefore provide a firstclass support to translate url path segments.</p>
<h2 dir="auto">Describe the solution you'd like</h2>
<p dir="auto">There should be a way to provide aliases to pages in code. So a page would be delivered on all available aliases. It should be possible to use translation services to provide these aliases. They should work both on server and on client.</p>
<p dir="auto">If you use a <code class="notranslate"><Link /></code>� to a page, you can provide some context, so that nextjs would use the right alias to make the route. E.g. you could provide the locale.</p>
<p dir="auto">I don't know nextjs good enough to suggest an api for it yet, but suggestions are very welcome!</p>
<h2 dir="auto">Describe alternatives you've considered</h2>
<p dir="auto">Doing it with dynamic routes (next 9 feature), is nearly impossible, because every page and folder would basically look like <code class="notranslate">[slug].js</code>, so you would need to deliver the right page yourself, defeating the purpose of the whole file-system-routing.</p>
<p dir="auto">It's only possible with a custom server and some boilerplate, so that your <code class="notranslate"><Link /></code> behave correctly. But this is needs careful setup, so that you don't have duplicated route declarations and it is therefore error-prone.</p> | <h2 dir="auto">Problem</h2>
<p dir="auto">I created a project via create-<a class="issue-link js-issue-link notranslate" rel="noopener noreferrer nofollow" href="https://linear.app/vercel/issue/NEXT-app">next-app</a>, but when I tried to run dev server I got an error below:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="[ error ] ./node_modules/next/dist/client/dev/amp-dev.js
Error: [BABEL] C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\client\dev\amp-dev.js: Cannot find module './src/data'
Require stack:
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\core-js-compat\helpers.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\core-js-compat\get-modules-list-for-target-version.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\@babel\preset-env\lib\polyfills\corejs3\usage-plugin.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\@babel\preset-env\lib\index.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\build\babel\preset.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\build\webpack\loaders\next-babel-loader.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\loader-runner\lib\loadLoader.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\loader-runner\lib\LoaderRunner.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\webpack\lib\NormalModule.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\webpack\lib\NormalModuleFactory.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\webpack\lib\Compiler.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\webpack\lib\webpack.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\server\hot-reloader.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\server\next-dev-server.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\server\next.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\server\lib\start-server.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\cli\next-dev.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\bin\next (While processing: "programmatic item")"><pre class="notranslate"><code class="notranslate">[ error ] ./node_modules/next/dist/client/dev/amp-dev.js
Error: [BABEL] C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\client\dev\amp-dev.js: Cannot find module './src/data'
Require stack:
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\core-js-compat\helpers.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\core-js-compat\get-modules-list-for-target-version.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\@babel\preset-env\lib\polyfills\corejs3\usage-plugin.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\@babel\preset-env\lib\index.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\build\babel\preset.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\build\webpack\loaders\next-babel-loader.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\loader-runner\lib\loadLoader.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\loader-runner\lib\LoaderRunner.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\webpack\lib\NormalModule.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\webpack\lib\NormalModuleFactory.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\webpack\lib\Compiler.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\webpack\lib\webpack.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\server\hot-reloader.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\server\next-dev-server.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\server\next.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\server\lib\start-server.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\cli\next-dev.js
- C:\Users\serge\Coding\web\playground\nextjs-tests\bitzprice\node_modules\next\dist\bin\next (While processing: "programmatic item")
</code></pre></div>
<p dir="auto">Packages information:</p>
<ul dir="auto">
<li>next: 9.1.4,</li>
<li>react: 16.12.0</li>
<li>react-dom: 16.12.0</li>
</ul>
<p dir="auto">My dependency manager is Yarn (1.19.2)</p>
<h2 dir="auto">Expected Behavior</h2>
<p dir="auto">Working dev server</p>
<h2 dir="auto">System Information</h2>
<ul dir="auto">
<li>OS: Windows 10 (1903)</li>
</ul> | 0 |
<h4 dir="auto">Challenge Name</h4>
<p dir="auto"><a href="http://beta.freecodecamp.com/en/challenges/css-flexbox/using-the-flexdirection-property-to-make-a-row" rel="nofollow">http://beta.freecodecamp.com/en/challenges/css-flexbox/using-the-flexdirection-property-to-make-a-row</a></p>
<h4 dir="auto">Issue Description</h4>
<p dir="auto">I guess since the default value for <code class="notranslate">flex-direction</code> is row, this test can pass without doing anything at all. Maybe a reg-ex is in order to make sure the user enters code?</p> | <p dir="auto">CHALLENGE:<br>
<a href="http://beta.freecodecamp.com/en/challenges/css-flexbox/using-the-flexdirection-property-to-make-a-row" rel="nofollow">http://beta.freecodecamp.com/en/challenges/css-flexbox/using-the-flexdirection-property-to-make-a-row</a></p>
<p dir="auto">ISSUE:<br>
Clicking 'Run tests' button passes the tests without the test conditions being met. I did not add the CSS property flex-direction to the #box-container element, and give it a value of row.</p>
<p dir="auto">BROWSER INFO:</p>
<ul dir="auto">
<li>Chrome 55.0.2883.87 m</li>
<li>Windows 7</li>
<li>Desktop</li>
</ul>
<p dir="auto">SCREENSHOT:<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://cloud.githubusercontent.com/assets/16689321/22271736/02667804-e24b-11e6-803c-c60f00637c91.png"><img src="https://cloud.githubusercontent.com/assets/16689321/22271736/02667804-e24b-11e6-803c-c60f00637c91.png" alt="image" style="max-width: 100%;"></a></p> | 1 |
<p dir="auto">Splitting off from: <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="215152737" data-permission-text="Title is private" data-url="https://github.com/flutter/flutter/issues/8876" data-hovercard-type="issue" data-hovercard-url="/flutter/flutter/issues/8876/hovercard" href="https://github.com/flutter/flutter/issues/8876">#8876</a></p> | <p dir="auto">Just helped track down an issue an internal customer was having whereby multiple <code class="notranslate">adb</code> instances were fighting each other:</p>
<ul dir="auto">
<li><code class="notranslate">/usr/bin/adb</code> existed as part of the <code class="notranslate">adb</code> Debian package that was installed on the user's system. It seemed to be restarting its server automatically when killed because of an entry in <code class="notranslate">init.d</code></li>
<li><code class="notranslate">$ANDROID_HOME/platform-tools/adb</code> was the user's preferred <code class="notranslate">adb</code> instance and was also the instance detected by flutter tools since it lived in a proper SDK folder structure.</li>
</ul>
<p dir="auto">There was initially an issue whereby one <code class="notranslate">adb</code> client process would detect a mismatch between the client and server versions and restart the appropriate server, and in so doing, it'd echo a message to the terminal that flutter tools mistook for a device line (fixed in <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="225344131" data-permission-text="Title is private" data-url="https://github.com/flutter/flutter/issues/9693" data-hovercard-type="pull_request" data-hovercard-url="/flutter/flutter/pull/9693/hovercard" href="https://github.com/flutter/flutter/pull/9693">#9693</a>).</p>
<p dir="auto">Once that issue was fixed, the two <code class="notranslate">adb</code> servers still seemed to be fighting each other, causing the preferred <code class="notranslate">adb</code> to show no connected devices even though there were devices connected. The solution we eventually found for this was uninstalling <code class="notranslate">/usr/bin/adb</code> (via <code class="notranslate">sudo apt-get remove adb</code>)...</p>
<p dir="auto">This experience, along with reports within the Flutter team of multiple <code class="notranslate">adb</code> instances not playing nicely with each other and with Flutter tools, means we probably want to add a doctor validator that will warn the user if multiple adb binaries are detected.</p> | 1 |
<h2 dir="auto"><g-emoji class="g-emoji" alias="bug" fallback-src="https://github.githubassets.com/images/icons/emoji/unicode/1f41b.png">🐛</g-emoji> Bug</h2>
<p dir="auto">I think the Kl divergence is summing over the batch dimensions when it shouldn't, at least for the Gaussian case</p>
<p dir="auto">Here's the transformed distribution function:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="def _kl_transformed_transformed(p, q):
if p.transforms != q.transforms:
raise NotImplementedError
if p.event_shape != q.event_shape:
raise NotImplementedError
# extra_event_dim = len(p.event_shape) - len(p.base_dist.event_shape)
extra_event_dim = len(p.event_shape)
base_kl_divergence = kl_divergence(p.base_dist, q.base_dist) #call to indep_indep below
#this will again sum over kl_divergence for each entry in batch
return _sum_rightmost(base_kl_divergence, extra_event_dim)"><pre class="notranslate"><code class="notranslate">def _kl_transformed_transformed(p, q):
if p.transforms != q.transforms:
raise NotImplementedError
if p.event_shape != q.event_shape:
raise NotImplementedError
# extra_event_dim = len(p.event_shape) - len(p.base_dist.event_shape)
extra_event_dim = len(p.event_shape)
base_kl_divergence = kl_divergence(p.base_dist, q.base_dist) #call to indep_indep below
#this will again sum over kl_divergence for each entry in batch
return _sum_rightmost(base_kl_divergence, extra_event_dim)
</code></pre></div>
<p dir="auto">Here's independent_independent KL</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="@register_kl(Independent, Independent)
def _kl_independent_independent(p, q):
shared_ndims = min(p.reinterpreted_batch_ndims, q.reinterpreted_batch_ndims)
p_ndims = p.reinterpreted_batch_ndims - shared_ndims
q_ndims = q.reinterpreted_batch_ndims - shared_ndims
p = Independent(p.base_dist, p_ndims) if p_ndims else p.base_dist
q = Independent(q.base_dist, q_ndims) if q_ndims else q.base_dist
kl = kl_divergence(p, q)
if shared_ndims:
#this line gets called when base_dist is Gaussian
kl = sum_rightmost(kl, shared_ndims)
return kl"><pre class="notranslate"><code class="notranslate">@register_kl(Independent, Independent)
def _kl_independent_independent(p, q):
shared_ndims = min(p.reinterpreted_batch_ndims, q.reinterpreted_batch_ndims)
p_ndims = p.reinterpreted_batch_ndims - shared_ndims
q_ndims = q.reinterpreted_batch_ndims - shared_ndims
p = Independent(p.base_dist, p_ndims) if p_ndims else p.base_dist
q = Independent(q.base_dist, q_ndims) if q_ndims else q.base_dist
kl = kl_divergence(p, q)
if shared_ndims:
#this line gets called when base_dist is Gaussian
kl = sum_rightmost(kl, shared_ndims)
return kl
</code></pre></div>
<h2 dir="auto">To Reproduce</h2>
<p dir="auto">I think something like this will do it:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="p = Gaussian(torch.zeros(3, 10), torch.ones(3, 10)).independent(1)
q = Gaussian(torch.ones(3,10) *2, torch.ones(3,10).independent(1)
kl = kl_divergence(p, q)
print(kl.shape) #should be something like [3, 1] but will instead output []"><pre class="notranslate"><code class="notranslate">p = Gaussian(torch.zeros(3, 10), torch.ones(3, 10)).independent(1)
q = Gaussian(torch.ones(3,10) *2, torch.ones(3,10).independent(1)
kl = kl_divergence(p, q)
print(kl.shape) #should be something like [3, 1] but will instead output []
</code></pre></div>
<h2 dir="auto">expected behavior</h2>
<p dir="auto">Maybe I misunderstand kl_divergence function, but I don't think it should be summing over batch.</p>
<ul dir="auto">
<li>PyTorch Version (e.g., 1.0): 1.2</li>
<li>OS (e.g., Linux): Ubuntu 16.04</li>
<li>How you installed PyTorch (<code class="notranslate">conda</code>, <code class="notranslate">pip</code>, source): conda/ pip</li>
<li>Build command you used (if compiling from source):</li>
<li>Python version: 3.6.9</li>
<li>CUDA/cuDNN version: 9.0</li>
<li>GPU models and configuration:</li>
<li>Any other relevant information:<br>
I originally found this using the MeanFieldTrace(...) class in the Pyro probabilistic programming language with normalizing flows. The Kl divergence function is used underneath the hood, and checks that the kl_divergence dimensions are the same as the q.batch_shape which in my case they were not.</li>
</ul>
<p dir="auto">cc <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/vincentqb/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/vincentqb">@vincentqb</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/fritzo/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/fritzo">@fritzo</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/neerajprad/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/neerajprad">@neerajprad</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/alicanb/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/alicanb">@alicanb</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/vishwakftw/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/vishwakftw">@vishwakftw</a></p> | <p dir="auto">This is a copy of an issue I posted for pytorch: <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="521900193" data-permission-text="Title is private" data-url="https://github.com/pytorch/pytorch/issues/29698" data-hovercard-type="issue" data-hovercard-url="/pytorch/pytorch/issues/29698/hovercard" href="https://github.com/pytorch/pytorch/issues/29698">pytorch/pytorch#29698</a></p>
<p dir="auto">I think this is an issue with the _kl_independent_independent class which is summing over batches when it shouldn't...</p>
<h3 dir="auto">Guidelines</h3>
<p dir="auto">I think the Kl divergence is summing over the batch dimensions when it shouldn't, at least for the Gaussian case</p>
<p dir="auto">Here's the transformed distribution function:</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="def _kl_transformed_transformed(p, q):
if p.transforms != q.transforms:
raise NotImplementedError
if p.event_shape != q.event_shape:
raise NotImplementedError
# extra_event_dim = len(p.event_shape) - len(p.base_dist.event_shape)
extra_event_dim = len(p.event_shape)
base_kl_divergence = kl_divergence(p.base_dist, q.base_dist) #call to indep_indep below
#this will again sum over kl_divergence for each entry in batch
return _sum_rightmost(base_kl_divergence, extra_event_dim)"><pre class="notranslate"><span class="pl-k">def</span> <span class="pl-en">_kl_transformed_transformed</span>(<span class="pl-s1">p</span>, <span class="pl-s1">q</span>):
<span class="pl-k">if</span> <span class="pl-s1">p</span>.<span class="pl-s1">transforms</span> <span class="pl-c1">!=</span> <span class="pl-s1">q</span>.<span class="pl-s1">transforms</span>:
<span class="pl-k">raise</span> <span class="pl-v">NotImplementedError</span>
<span class="pl-k">if</span> <span class="pl-s1">p</span>.<span class="pl-s1">event_shape</span> <span class="pl-c1">!=</span> <span class="pl-s1">q</span>.<span class="pl-s1">event_shape</span>:
<span class="pl-k">raise</span> <span class="pl-v">NotImplementedError</span>
<span class="pl-c"># extra_event_dim = len(p.event_shape) - len(p.base_dist.event_shape)</span>
<span class="pl-s1">extra_event_dim</span> <span class="pl-c1">=</span> <span class="pl-en">len</span>(<span class="pl-s1">p</span>.<span class="pl-s1">event_shape</span>)
<span class="pl-s1">base_kl_divergence</span> <span class="pl-c1">=</span> <span class="pl-en">kl_divergence</span>(<span class="pl-s1">p</span>.<span class="pl-s1">base_dist</span>, <span class="pl-s1">q</span>.<span class="pl-s1">base_dist</span>) <span class="pl-c">#call to indep_indep below</span>
<span class="pl-c">#this will again sum over kl_divergence for each entry in batch</span>
<span class="pl-k">return</span> <span class="pl-en">_sum_rightmost</span>(<span class="pl-s1">base_kl_divergence</span>, <span class="pl-s1">extra_event_dim</span>)</pre></div>
<p dir="auto">Here's independent_independent KL</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="@register_kl(Independent, Independent)
def _kl_independent_independent(p, q):
shared_ndims = min(p.reinterpreted_batch_ndims, q.reinterpreted_batch_ndims)
p_ndims = p.reinterpreted_batch_ndims - shared_ndims
q_ndims = q.reinterpreted_batch_ndims - shared_ndims
p = Independent(p.base_dist, p_ndims) if p_ndims else p.base_dist
q = Independent(q.base_dist, q_ndims) if q_ndims else q.base_dist
kl = kl_divergence(p, q)
if shared_ndims:
#this line gets called when base_dist is Gaussian
kl = sum_rightmost(kl, shared_ndims)
return kl"><pre class="notranslate"><span class="pl-en">@<span class="pl-en">register_kl</span>(<span class="pl-v">Independent</span>, <span class="pl-v">Independent</span>)</span>
<span class="pl-k">def</span> <span class="pl-en">_kl_independent_independent</span>(<span class="pl-s1">p</span>, <span class="pl-s1">q</span>):
<span class="pl-s1">shared_ndims</span> <span class="pl-c1">=</span> <span class="pl-en">min</span>(<span class="pl-s1">p</span>.<span class="pl-s1">reinterpreted_batch_ndims</span>, <span class="pl-s1">q</span>.<span class="pl-s1">reinterpreted_batch_ndims</span>)
<span class="pl-s1">p_ndims</span> <span class="pl-c1">=</span> <span class="pl-s1">p</span>.<span class="pl-s1">reinterpreted_batch_ndims</span> <span class="pl-c1">-</span> <span class="pl-s1">shared_ndims</span>
<span class="pl-s1">q_ndims</span> <span class="pl-c1">=</span> <span class="pl-s1">q</span>.<span class="pl-s1">reinterpreted_batch_ndims</span> <span class="pl-c1">-</span> <span class="pl-s1">shared_ndims</span>
<span class="pl-s1">p</span> <span class="pl-c1">=</span> <span class="pl-v">Independent</span>(<span class="pl-s1">p</span>.<span class="pl-s1">base_dist</span>, <span class="pl-s1">p_ndims</span>) <span class="pl-k">if</span> <span class="pl-s1">p_ndims</span> <span class="pl-k">else</span> <span class="pl-s1">p</span>.<span class="pl-s1">base_dist</span>
<span class="pl-s1">q</span> <span class="pl-c1">=</span> <span class="pl-v">Independent</span>(<span class="pl-s1">q</span>.<span class="pl-s1">base_dist</span>, <span class="pl-s1">q_ndims</span>) <span class="pl-k">if</span> <span class="pl-s1">q_ndims</span> <span class="pl-k">else</span> <span class="pl-s1">q</span>.<span class="pl-s1">base_dist</span>
<span class="pl-s1">kl</span> <span class="pl-c1">=</span> <span class="pl-en">kl_divergence</span>(<span class="pl-s1">p</span>, <span class="pl-s1">q</span>)
<span class="pl-k">if</span> <span class="pl-s1">shared_ndims</span>:
<span class="pl-c">#this line gets called when base_dist is Gaussian</span>
<span class="pl-s1">kl</span> <span class="pl-c1">=</span> <span class="pl-en">sum_rightmost</span>(<span class="pl-s1">kl</span>, <span class="pl-s1">shared_ndims</span>)
<span class="pl-k">return</span> <span class="pl-s1">kl</span></pre></div>
<h2 dir="auto">To Reproduce</h2>
<p dir="auto">I think something like this will do it:</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="p = Gaussian(torch.zeros(3, 10), torch.ones(3, 10)).independent(1)
q = Gaussian(torch.ones(3,10) *2, torch.ones(3,10).independent(1)
kl = kl_divergence(p, q)
print(kl.shape) #should be something like [3, 1] but will instead output []"><pre class="notranslate"><span class="pl-s1">p</span> <span class="pl-c1">=</span> <span class="pl-v">Gaussian</span>(<span class="pl-s1">torch</span>.<span class="pl-en">zeros</span>(<span class="pl-c1">3</span>, <span class="pl-c1">10</span>), <span class="pl-s1">torch</span>.<span class="pl-en">ones</span>(<span class="pl-c1">3</span>, <span class="pl-c1">10</span>)).<span class="pl-en">independent</span>(<span class="pl-c1">1</span>)
<span class="pl-s1">q</span> <span class="pl-c1">=</span> <span class="pl-v">Gaussian</span>(<span class="pl-s1">torch</span>.<span class="pl-en">ones</span>(<span class="pl-c1">3</span>,<span class="pl-c1">10</span>) <span class="pl-c1">*</span><span class="pl-c1">2</span>, <span class="pl-s1">torch</span>.<span class="pl-en">ones</span>(<span class="pl-c1">3</span>,<span class="pl-c1">10</span>).<span class="pl-en">independent</span>(<span class="pl-c1">1</span>)
<span class="pl-s1">kl</span> <span class="pl-c1">=</span> <span class="pl-en">kl_divergence</span>(<span class="pl-s1">p</span>, <span class="pl-s1">q</span>)
<span class="pl-s1">print</span>(<span class="pl-s1">kl</span>.<span class="pl-s1">shape</span>) <span class="pl-c">#should be something like [3, 1] but will instead output []</span></pre></div>
<h2 dir="auto">expected behavior</h2>
<p dir="auto">Maybe I misunderstand kl_divergence function, but I don't think it should be summing over batch.</p>
<ul dir="auto">
<li>PyTorch Version (e.g., 1.0): 1.2</li>
<li>OS (e.g., Linux): Ubuntu 16.04</li>
<li>How you installed PyTorch (<code class="notranslate">conda</code>, <code class="notranslate">pip</code>, source): conda/ pip</li>
<li>Build command you used (if compiling from source):</li>
<li>Python version: 3.6.9</li>
<li>CUDA/cuDNN version: 9.0</li>
<li>GPU models and configuration:</li>
<li>Any other relevant information:<br>
I originally found this using the MeanFieldTrace(...) class in the Pyro probabilistic programming language with normalizing flows. The Kl divergence function is used underneath the hood, and checks that the kl_divergence dimensions are the same as the q.batch_shape which in my case they were not.</li>
</ul> | 1 |
<p dir="auto">Please go to Stack Overflow for help and support:</p>
<p dir="auto"><a href="https://stackoverflow.com/questions/tagged/tensorflow" rel="nofollow">https://stackoverflow.com/questions/tagged/tensorflow</a></p>
<p dir="auto">If you open a GitHub issue, here is our policy:</p>
<ol dir="auto">
<li>It must be a bug, a feature request, or a significant problem with documentation (for small docs fixes please send a PR instead).</li>
<li>The form below must be filled out.</li>
<li>It shouldn't be a TensorBoard issue. Those go <a href="https://github.com/tensorflow/tensorboard/issues">here</a>.</li>
</ol>
<p dir="auto"><strong>Here's why we have that policy</strong>: TensorFlow developers respond to issues. We want to focus on work that benefits the whole community, e.g., fixing bugs and adding features. Support only helps individuals. GitHub also notifies thousands of people when issues are filed. We want them to see you communicating an interesting problem, rather than being redirected to Stack Overflow.</p>
<hr>
<h3 dir="auto">System information</h3>
<ul dir="auto">
<li><strong>Have I written custom code (as opposed to using a stock example script provided in TensorFlow)</strong>:</li>
<li><strong>OS Platform and Distribution (e.g., Linux Ubuntu 16.04)</strong>:</li>
<li><strong>Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device</strong>:</li>
<li><strong>TensorFlow installed from (source or binary)</strong>:</li>
<li><strong>TensorFlow version (use command below)</strong>:</li>
<li><strong>Python version</strong>:</li>
<li><strong>Bazel version (if compiling from source)</strong>:</li>
<li><strong>GCC/Compiler version (if compiling from source)</strong>:</li>
<li><strong>CUDA/cuDNN version</strong>:</li>
<li><strong>GPU model and memory</strong>:</li>
<li><strong>Exact command to reproduce</strong>:</li>
</ul>
<p dir="auto">You can collect some of this information using our environment capture script:</p>
<p dir="auto"><a href="https://github.com/tensorflow/tensorflow/tree/master/tools/tf_env_collect.sh">https://github.com/tensorflow/tensorflow/tree/master/tools/tf_env_collect.sh</a></p>
<p dir="auto">You can obtain the TensorFlow version with</p>
<p dir="auto">python -c "import tensorflow as tf; print(tf.GIT_VERSION, tf.VERSION)"</p>
<h3 dir="auto">Describe the problem</h3>
<p dir="auto">Describe the problem clearly here. Be sure to convey here why it's a bug in TensorFlow or a feature request.</p>
<h3 dir="auto">Source code / logs</h3>
<p dir="auto">Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached. Try to provide a reproducible test case that is the bare minimum necessary to generate the problem.</p> | <p dir="auto">I'm sure developers are working hard to catch up with Python 3.7.<br>
Is there any timeline?</p>
<p dir="auto">pip3 install tensorflow - apparently does not work, building from source:</p>
<p dir="auto">OS Platform and Distribution: Mac OS X 10.13.5<br>
Python: Python 3.7.0 (Homebrew)<br>
TensorFlow installed from: source (<a href="https://github.com/tensorflow/tensorflow.git">https://github.com/tensorflow/tensorflow.git</a>)<br>
TensorFlow version: TensorFlow 1.9.0-rc2<br>
Bazel version:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="Build label: 0.15.0-homebrew
Build target: bazel-out/darwin-opt/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer_deploy.jar
Build time: Tue Jun 26 12:42:27 2018 (1530016947)
Build timestamp: 1530016947
Build timestamp as int: 1530016947"><pre class="notranslate"><code class="notranslate">Build label: 0.15.0-homebrew
Build target: bazel-out/darwin-opt/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer_deploy.jar
Build time: Tue Jun 26 12:42:27 2018 (1530016947)
Build timestamp: 1530016947
Build timestamp as int: 1530016947
</code></pre></div>
<p dir="auto">CUDA/cuDNN version: None<br>
GPU model and memory: None<br>
Exact command to reproduce:<br>
<code class="notranslate">bazel build --config=opt //tensorflow/tools/pip_package:build_pip_package</code></p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="Starting local Bazel server and connecting to it...
...........................
WARNING: /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/BUILD:1992:1: in srcs attribute of cc_library rule @grpc//:grpc_nanopb: please do not import '@grpc//third_party/nanopb:pb_common.c' directly. You should either move the file to this package or depend on an appropriate rule there. Since this rule was created by the macro 'grpc_generate_one_off_targets', the error might have been caused by the macro implementation in /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/bazel/grpc_build_system.bzl:172:12
WARNING: /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/BUILD:1992:1: in srcs attribute of cc_library rule @grpc//:grpc_nanopb: please do not import '@grpc//third_party/nanopb:pb_decode.c' directly. You should either move the file to this package or depend on an appropriate rule there. Since this rule was created by the macro 'grpc_generate_one_off_targets', the error might have been caused by the macro implementation in /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/bazel/grpc_build_system.bzl:172:12
WARNING: /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/BUILD:1992:1: in srcs attribute of cc_library rule @grpc//:grpc_nanopb: please do not import '@grpc//third_party/nanopb:pb_encode.c' directly. You should either move the file to this package or depend on an appropriate rule there. Since this rule was created by the macro 'grpc_generate_one_off_targets', the error might have been caused by the macro implementation in /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/bazel/grpc_build_system.bzl:172:12
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/learn/BUILD:17:1: in py_library rule //tensorflow/contrib/learn:learn: target '//tensorflow/contrib/learn:learn' depends on deprecated target '//tensorflow/contrib/session_bundle:exporter': No longer supported. Switch to SavedModel immediately.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/learn/BUILD:17:1: in py_library rule //tensorflow/contrib/learn:learn: target '//tensorflow/contrib/learn:learn' depends on deprecated target '//tensorflow/contrib/session_bundle:gc': No longer supported. Switch to SavedModel immediately.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/timeseries/python/timeseries/BUILD:356:1: in py_library rule //tensorflow/contrib/timeseries/python/timeseries:ar_model: target '//tensorflow/contrib/timeseries/python/timeseries:ar_model' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/timeseries/python/timeseries/state_space_models/BUILD:73:1: in py_library rule //tensorflow/contrib/timeseries/python/timeseries/state_space_models:kalman_filter: target '//tensorflow/contrib/timeseries/python/timeseries/state_space_models:kalman_filter' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/timeseries/python/timeseries/state_space_models/BUILD:230:1: in py_library rule //tensorflow/contrib/timeseries/python/timeseries/state_space_models:filtering_postprocessor: target '//tensorflow/contrib/timeseries/python/timeseries/state_space_models:filtering_postprocessor' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/bayesflow/BUILD:17:1: in py_library rule //tensorflow/contrib/bayesflow:bayesflow_py: target '//tensorflow/contrib/bayesflow:bayesflow_py' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/seq2seq/BUILD:23:1: in py_library rule //tensorflow/contrib/seq2seq:seq2seq_py: target '//tensorflow/contrib/seq2seq:seq2seq_py' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/kfac/python/ops/BUILD:80:1: in py_library rule //tensorflow/contrib/kfac/python/ops:loss_functions: target '//tensorflow/contrib/kfac/python/ops:loss_functions' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/BUILD:14:1: in py_library rule //tensorflow/contrib:contrib_py: target '//tensorflow/contrib:contrib_py' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
INFO: Analysed target //tensorflow/tools/pip_package:build_pip_package (303 packages loaded).
INFO: Found 1 target...
INFO: From Linking external/grpc/libgrpc_base_c.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(endpoint_pair_uv.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(endpoint_pair_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(ev_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(fork_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(gethostname_fallback.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(gethostname_host_name_max.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(iocp_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(iomgr_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(pollset_set_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(pollset_uv.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(pollset_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(resolve_address_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(socket_utils_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(socket_utils_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(socket_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_client_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_server_utils_posix_noifaddrs.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_server_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_uv.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(timer_uv.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(unix_sockets_posix_noop.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(wakeup_fd_eventfd.o) has no symbols
INFO: From Linking external/grpc/libalts_util.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libalts_util.a(check_gcp_environment_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libalts_util.a(check_gcp_environment_windows.o) has no symbols
INFO: From Linking external/grpc/libtsi.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libtsi.a(ssl_session_openssl.o) has no symbols
INFO: From Linking external/grpc/libgrpc++_base.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc++_base.a(rpc_method.o) has no symbols
INFO: From Linking external/grpc/libgpr_base.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(cpu_iphone.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(cpu_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(cpu_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(env_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(env_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(log_android.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(log_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(log_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(string_util_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(string_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(sync_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(time_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(tls_pthread.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(tmpfile_msys.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(tmpfile_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(wrap_memcpy.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(thd_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(stap_timers.o) has no symbols
INFO: From Linking external/grpc/third_party/address_sorting/libaddress_sorting.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/third_party/address_sorting/libaddress_sorting.a(address_sorting_windows.o) has no symbols
ERROR: /Users/zardoz/Projects/tensorflow/tensorflow/python/BUILD:5315:1: Executing genrule //tensorflow/python:framework/fast_tensor_util.pyx_cython_translation failed (Exit 1)
Traceback (most recent call last):
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/execroot/org_tensorflow/bazel-out/host/bin/external/cython/cython_binary.runfiles/cython/cython.py", line 17, in <module>
main(command_line = 1)
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 720, in main
result = compile(sources, options)
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 695, in compile
return compile_multiple(source, options)
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 666, in compile_multiple
context = options.create_context()
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 590, in create_context
self.cplus, self.language_level, options=self)
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 75, in __init__
from . import Builtin, CythonScope
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/CythonScope.py", line 5, in <module>
from .UtilityCode import CythonUtilityCode
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/UtilityCode.py", line 3, in <module>
from .TreeFragment import parse_from_strings, StringParseContext
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/TreeFragment.py", line 17, in <module>
from .Visitor import VisitorTransform
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Visitor.py", line 15, in <module>
from . import ExprNodes
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/ExprNodes.py", line 2875
await = None
^
SyntaxError: invalid syntax
Target //tensorflow/tools/pip_package:build_pip_package failed to build
Use --verbose_failures to see the command lines of failed build steps.
INFO: Elapsed time: 179.318s, Critical Path: 6.38s
INFO: 413 processes: 413 local.
FAILED: Build did NOT complete successfully"><pre class="notranslate"><code class="notranslate">Starting local Bazel server and connecting to it...
...........................
WARNING: /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/BUILD:1992:1: in srcs attribute of cc_library rule @grpc//:grpc_nanopb: please do not import '@grpc//third_party/nanopb:pb_common.c' directly. You should either move the file to this package or depend on an appropriate rule there. Since this rule was created by the macro 'grpc_generate_one_off_targets', the error might have been caused by the macro implementation in /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/bazel/grpc_build_system.bzl:172:12
WARNING: /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/BUILD:1992:1: in srcs attribute of cc_library rule @grpc//:grpc_nanopb: please do not import '@grpc//third_party/nanopb:pb_decode.c' directly. You should either move the file to this package or depend on an appropriate rule there. Since this rule was created by the macro 'grpc_generate_one_off_targets', the error might have been caused by the macro implementation in /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/bazel/grpc_build_system.bzl:172:12
WARNING: /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/BUILD:1992:1: in srcs attribute of cc_library rule @grpc//:grpc_nanopb: please do not import '@grpc//third_party/nanopb:pb_encode.c' directly. You should either move the file to this package or depend on an appropriate rule there. Since this rule was created by the macro 'grpc_generate_one_off_targets', the error might have been caused by the macro implementation in /private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/grpc/bazel/grpc_build_system.bzl:172:12
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/learn/BUILD:17:1: in py_library rule //tensorflow/contrib/learn:learn: target '//tensorflow/contrib/learn:learn' depends on deprecated target '//tensorflow/contrib/session_bundle:exporter': No longer supported. Switch to SavedModel immediately.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/learn/BUILD:17:1: in py_library rule //tensorflow/contrib/learn:learn: target '//tensorflow/contrib/learn:learn' depends on deprecated target '//tensorflow/contrib/session_bundle:gc': No longer supported. Switch to SavedModel immediately.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/timeseries/python/timeseries/BUILD:356:1: in py_library rule //tensorflow/contrib/timeseries/python/timeseries:ar_model: target '//tensorflow/contrib/timeseries/python/timeseries:ar_model' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/timeseries/python/timeseries/state_space_models/BUILD:73:1: in py_library rule //tensorflow/contrib/timeseries/python/timeseries/state_space_models:kalman_filter: target '//tensorflow/contrib/timeseries/python/timeseries/state_space_models:kalman_filter' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/timeseries/python/timeseries/state_space_models/BUILD:230:1: in py_library rule //tensorflow/contrib/timeseries/python/timeseries/state_space_models:filtering_postprocessor: target '//tensorflow/contrib/timeseries/python/timeseries/state_space_models:filtering_postprocessor' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/bayesflow/BUILD:17:1: in py_library rule //tensorflow/contrib/bayesflow:bayesflow_py: target '//tensorflow/contrib/bayesflow:bayesflow_py' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/seq2seq/BUILD:23:1: in py_library rule //tensorflow/contrib/seq2seq:seq2seq_py: target '//tensorflow/contrib/seq2seq:seq2seq_py' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/kfac/python/ops/BUILD:80:1: in py_library rule //tensorflow/contrib/kfac/python/ops:loss_functions: target '//tensorflow/contrib/kfac/python/ops:loss_functions' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
WARNING: /Users/zardoz/Projects/tensorflow/tensorflow/contrib/BUILD:14:1: in py_library rule //tensorflow/contrib:contrib_py: target '//tensorflow/contrib:contrib_py' depends on deprecated target '//tensorflow/contrib/distributions:distributions_py': TensorFlow Distributions has migrated to TensorFlow Probability (https://github.com/tensorflow/probability). Deprecated copies remaining in tf.contrib.distributions are unmaintained, unsupported, and will be removed by late 2018. You should update all usage of `tf.contrib.distributions` to `tfp.distributions`.
INFO: Analysed target //tensorflow/tools/pip_package:build_pip_package (303 packages loaded).
INFO: Found 1 target...
INFO: From Linking external/grpc/libgrpc_base_c.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(endpoint_pair_uv.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(endpoint_pair_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(ev_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(fork_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(gethostname_fallback.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(gethostname_host_name_max.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(iocp_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(iomgr_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(pollset_set_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(pollset_uv.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(pollset_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(resolve_address_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(socket_utils_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(socket_utils_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(socket_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_client_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_server_utils_posix_noifaddrs.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_server_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_uv.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(tcp_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(timer_uv.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(unix_sockets_posix_noop.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc_base_c.a(wakeup_fd_eventfd.o) has no symbols
INFO: From Linking external/grpc/libalts_util.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libalts_util.a(check_gcp_environment_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libalts_util.a(check_gcp_environment_windows.o) has no symbols
INFO: From Linking external/grpc/libtsi.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libtsi.a(ssl_session_openssl.o) has no symbols
INFO: From Linking external/grpc/libgrpc++_base.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgrpc++_base.a(rpc_method.o) has no symbols
INFO: From Linking external/grpc/libgpr_base.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(cpu_iphone.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(cpu_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(cpu_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(env_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(env_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(log_android.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(log_linux.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(log_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(string_util_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(string_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(sync_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(time_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(tls_pthread.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(tmpfile_msys.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(tmpfile_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(wrap_memcpy.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(thd_windows.o) has no symbols
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/libgpr_base.a(stap_timers.o) has no symbols
INFO: From Linking external/grpc/third_party/address_sorting/libaddress_sorting.a [for host]:
/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/ranlib: file: bazel-out/host/bin/external/grpc/third_party/address_sorting/libaddress_sorting.a(address_sorting_windows.o) has no symbols
ERROR: /Users/zardoz/Projects/tensorflow/tensorflow/python/BUILD:5315:1: Executing genrule //tensorflow/python:framework/fast_tensor_util.pyx_cython_translation failed (Exit 1)
Traceback (most recent call last):
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/execroot/org_tensorflow/bazel-out/host/bin/external/cython/cython_binary.runfiles/cython/cython.py", line 17, in <module>
main(command_line = 1)
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 720, in main
result = compile(sources, options)
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 695, in compile
return compile_multiple(source, options)
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 666, in compile_multiple
context = options.create_context()
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 590, in create_context
self.cplus, self.language_level, options=self)
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Main.py", line 75, in __init__
from . import Builtin, CythonScope
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/CythonScope.py", line 5, in <module>
from .UtilityCode import CythonUtilityCode
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/UtilityCode.py", line 3, in <module>
from .TreeFragment import parse_from_strings, StringParseContext
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/TreeFragment.py", line 17, in <module>
from .Visitor import VisitorTransform
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/Visitor.py", line 15, in <module>
from . import ExprNodes
File "/private/var/tmp/_bazel_zardoz/5e080a8a46c0e2b2146c013eb1079337/external/cython/Cython/Compiler/ExprNodes.py", line 2875
await = None
^
SyntaxError: invalid syntax
Target //tensorflow/tools/pip_package:build_pip_package failed to build
Use --verbose_failures to see the command lines of failed build steps.
INFO: Elapsed time: 179.318s, Critical Path: 6.38s
INFO: 413 processes: 413 local.
FAILED: Build did NOT complete successfully
</code></pre></div> | 1 |
<p dir="auto"><strong>Migrated issue, originally created by Bastien Chatelard (<a href="https://github.com/bchatelard">@bchatelard</a>)</strong></p>
<p dir="auto">I am facing a regression on joinedload on polymorphic entity in 1.0.10.</p>
<p dir="auto">I have reproduced a minimal test case that works fine in 1.0.9 and does not in 1.0.10:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="import sqlalchemy as sa
import sqlalchemy.orm
from sqlalchemy.orm import joinedload
import sqlalchemy.ext.declarative
import sqlalchemy.ext.orderinglist
Base = sqlalchemy.ext.declarative.declarative_base()
class Person(Base):
__tablename__ = 'person'
id = sa.Column(sa.Integer, nullable=False, primary_key=True)
person_type = sa.Column(sa.String(20), nullable=False)
__mapper_args__ = {
'polymorphic_on': person_type,
'with_polymorphic': '*',
}
class User(Person):
__mapper_args__ = {'polymorphic_identity': 'user'}
class Admin(Person):
__mapper_args__ = {'polymorphic_identity': 'admin'}
class Action(Base):
__tablename__ = 'action'
id = sa.Column(sa.Integer, nullable=False, primary_key=True)
author_id = sa.Column(sa.Integer, sa.ForeignKey(User.id))
author = sa.orm.relationship('User')
post_type = sa.Column(sa.String(20), nullable=False)
__mapper_args__ = {
'polymorphic_on': post_type,
'polymorphic_identity': 'post',
'with_polymorphic': '*',
}
class Like(Action):
__tablename__ = 'like'
__mapper_args__ = {'polymorphic_identity': 'like'}
id = sa.Column(sa.Integer, sa.ForeignKey('action.id'), primary_key=True)
class Comment(Action):
__tablename__ = 'comment'
__mapper_args__ = {'polymorphic_identity': 'comment'}
id = sa.Column(sa.Integer, sa.ForeignKey('action.id'), primary_key=True)
article_id = sa.Column(sa.Integer, sa.ForeignKey('article.id'))
article = sa.orm.relationship('Article', primaryjoin='Article.id==Comment.article_id', backref='comments')
class Article(Base):
__tablename__ = 'article'
id = sa.Column(sa.Integer, nullable=False, primary_key=True)
engine = sa.create_engine('sqlite:///:memory:', echo=True)
Base.metadata.drop_all(engine)
Base.metadata.create_all(engine)
Session = sa.orm.sessionmaker(bind=engine)
session = Session()
user1 = User(id=1)
admin1 = Admin(id=2)
user2 = User(id=3)
article1 = Article(id=1)
comment1 = Comment(id=2, author=user1, article_id=1)
comment2 = Comment(id=3, author=user2, article_id=1)
session.add_all([user1, user2, admin1, article1, comment1, comment2])
session.commit()
session.expunge_all()
# Works with 1.0.9
# Not working with 1.0.10: raise AssertionError
try:
new_n1 = session.query(Article).options(
joinedload(Article.comments).joinedload(Action.author),
).all()
print new_n1
print new_n1[0].comments
print new_n1[0].comments[0].author
except AssertionError as e:
print e
# Works with 1.0.9
# Not working with 1.0.10: 2 queries
try:
new_n1 = session.query(Article).options(
joinedload('comments'),
joinedload('comments.author')
).all()
print new_n1
print new_n1[0].comments
print new_n1[0].comments[0].author
except AssertionError as e:
print e
# Works with 1.0.9
# Not working with 1.0.10: raise AssertionError
try:
new_n1 = session.query(Article).options(
joinedload('comments').joinedload('author')
).all()
print new_n1
print new_n1[0].comments
print new_n1[0].comments[0].author
except AssertionError as e:
print e
# Works with 1.0.9
# Not working with 1.0.10: raise AssertionError
try:
new_n1 = session.query(Article).options(
joinedload('comments').load_only('id', 'author_id').joinedload('author').load_only('id')
).all()
print new_n1
print new_n1[0].comments
print new_n1[0].comments[0].author
except AssertionError as e:
print e"><pre class="notranslate"><code class="notranslate">import sqlalchemy as sa
import sqlalchemy.orm
from sqlalchemy.orm import joinedload
import sqlalchemy.ext.declarative
import sqlalchemy.ext.orderinglist
Base = sqlalchemy.ext.declarative.declarative_base()
class Person(Base):
__tablename__ = 'person'
id = sa.Column(sa.Integer, nullable=False, primary_key=True)
person_type = sa.Column(sa.String(20), nullable=False)
__mapper_args__ = {
'polymorphic_on': person_type,
'with_polymorphic': '*',
}
class User(Person):
__mapper_args__ = {'polymorphic_identity': 'user'}
class Admin(Person):
__mapper_args__ = {'polymorphic_identity': 'admin'}
class Action(Base):
__tablename__ = 'action'
id = sa.Column(sa.Integer, nullable=False, primary_key=True)
author_id = sa.Column(sa.Integer, sa.ForeignKey(User.id))
author = sa.orm.relationship('User')
post_type = sa.Column(sa.String(20), nullable=False)
__mapper_args__ = {
'polymorphic_on': post_type,
'polymorphic_identity': 'post',
'with_polymorphic': '*',
}
class Like(Action):
__tablename__ = 'like'
__mapper_args__ = {'polymorphic_identity': 'like'}
id = sa.Column(sa.Integer, sa.ForeignKey('action.id'), primary_key=True)
class Comment(Action):
__tablename__ = 'comment'
__mapper_args__ = {'polymorphic_identity': 'comment'}
id = sa.Column(sa.Integer, sa.ForeignKey('action.id'), primary_key=True)
article_id = sa.Column(sa.Integer, sa.ForeignKey('article.id'))
article = sa.orm.relationship('Article', primaryjoin='Article.id==Comment.article_id', backref='comments')
class Article(Base):
__tablename__ = 'article'
id = sa.Column(sa.Integer, nullable=False, primary_key=True)
engine = sa.create_engine('sqlite:///:memory:', echo=True)
Base.metadata.drop_all(engine)
Base.metadata.create_all(engine)
Session = sa.orm.sessionmaker(bind=engine)
session = Session()
user1 = User(id=1)
admin1 = Admin(id=2)
user2 = User(id=3)
article1 = Article(id=1)
comment1 = Comment(id=2, author=user1, article_id=1)
comment2 = Comment(id=3, author=user2, article_id=1)
session.add_all([user1, user2, admin1, article1, comment1, comment2])
session.commit()
session.expunge_all()
# Works with 1.0.9
# Not working with 1.0.10: raise AssertionError
try:
new_n1 = session.query(Article).options(
joinedload(Article.comments).joinedload(Action.author),
).all()
print new_n1
print new_n1[0].comments
print new_n1[0].comments[0].author
except AssertionError as e:
print e
# Works with 1.0.9
# Not working with 1.0.10: 2 queries
try:
new_n1 = session.query(Article).options(
joinedload('comments'),
joinedload('comments.author')
).all()
print new_n1
print new_n1[0].comments
print new_n1[0].comments[0].author
except AssertionError as e:
print e
# Works with 1.0.9
# Not working with 1.0.10: raise AssertionError
try:
new_n1 = session.query(Article).options(
joinedload('comments').joinedload('author')
).all()
print new_n1
print new_n1[0].comments
print new_n1[0].comments[0].author
except AssertionError as e:
print e
# Works with 1.0.9
# Not working with 1.0.10: raise AssertionError
try:
new_n1 = session.query(Article).options(
joinedload('comments').load_only('id', 'author_id').joinedload('author').load_only('id')
).all()
print new_n1
print new_n1[0].comments
print new_n1[0].comments[0].author
except AssertionError as e:
print e
</code></pre></div>
<p dir="auto">It might be related to <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="384627684" data-permission-text="Title is private" data-url="https://github.com/sqlalchemy/sqlalchemy/issues/2714" data-hovercard-type="issue" data-hovercard-url="/sqlalchemy/sqlalchemy/issues/2714/hovercard" href="https://github.com/sqlalchemy/sqlalchemy/issues/2714">#2714</a> or <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="384635497" data-permission-text="Title is private" data-url="https://github.com/sqlalchemy/sqlalchemy/issues/3593" data-hovercard-type="issue" data-hovercard-url="/sqlalchemy/sqlalchemy/issues/3593/hovercard" href="https://github.com/sqlalchemy/sqlalchemy/issues/3593">#3593</a></p> | <p dir="auto"><strong>Migrated issue, originally created by Jorge Sierra</strong></p>
<p dir="auto">conn_str = string.Template('db2+ibm_db://$user:$password@localhost:50000/Sample').substitute(user = user , password = password)</p>
<p dir="auto">for t in metadata.sorted_tables:<br>
logger.info ("CreateTable %s %s" % (t.name, <strong>CreateTable</strong>(t).compile(db2)))</p>
<p dir="auto">When trying to run <strong>CreateTable</strong></p>
<p dir="auto"><strong>compiler.py</strong></p>
<p dir="auto">around line 2588<br>
under function</p>
<p dir="auto">def <strong>visit_foreign_key_constraint</strong>(self, constraint):</p>
<p dir="auto">this function is assuming <strong>f.parent</strong> is never 'None' and in my case it was, same as <strong>f.column</strong></p>
<p dir="auto">This is the fix below</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="list_text1 = []
for f in constraint.elements:
if f.parent is not None:
list_text1.append(preparer.quote(f.parent.name))
list_text2 = []
for f in constraint.elements:
if hasattr(f, 'column'):
if f.column is not None:
list_text2.append(preparer.quote(f.column.name))
text += "FOREIGN KEY(%s) REFERENCES %s (%s)" % (
', '.join(list_text1),
self.define_constraint_remote_table(
constraint, remote_table, preparer),
', '.join(list_text2))
text += self.define_constraint_match(constraint)
text += self.define_constraint_cascades(constraint)
text += self.define_constraint_deferrability(constraint)
return text
"><pre class="notranslate"><code class="notranslate">list_text1 = []
for f in constraint.elements:
if f.parent is not None:
list_text1.append(preparer.quote(f.parent.name))
list_text2 = []
for f in constraint.elements:
if hasattr(f, 'column'):
if f.column is not None:
list_text2.append(preparer.quote(f.column.name))
text += "FOREIGN KEY(%s) REFERENCES %s (%s)" % (
', '.join(list_text1),
self.define_constraint_remote_table(
constraint, remote_table, preparer),
', '.join(list_text2))
text += self.define_constraint_match(constraint)
text += self.define_constraint_cascades(constraint)
text += self.define_constraint_deferrability(constraint)
return text
</code></pre></div> | 0 |
<p dir="auto">We recently migrated our ES 1.1.1 cluster to ES 1.2.2, after that we found that one field 'displayPublishedDateTime', which is defined as dataOptionalTime type in the mapping, seems corrupted and not able to order by it in the query.<br>
In the query, when order by desc on it, its sort value is '\b', and order by asc, its sort value is shown as '' \u0001\u0000\u0000\u0000(aC3@X", see the following query and results. In other queries, which order by 'displayPublishedDateTime', the order actually didn't take effect. It looks like that Elasticsearch was unable to get the correct sort value for this field.</p>
<p dir="auto">Query</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="GET /doc-v2/_search
{
"query": {
"match": {
"_id": "AAUYRY"
}
},
"_source":[
"_id",
"_document.displayPublishedDateTime",
],
"sort": [
{
"_document.displayPublishedDateTime": {
"order": "desc"
}
}
]
}"><pre class="notranslate"><code class="notranslate">GET /doc-v2/_search
{
"query": {
"match": {
"_id": "AAUYRY"
}
},
"_source":[
"_id",
"_document.displayPublishedDateTime",
],
"sort": [
{
"_document.displayPublishedDateTime": {
"order": "desc"
}
}
]
}
</code></pre></div>
<p dir="auto">Results for 'desc'</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="{
"took": 46,
"timed_out": false,
"_shards": {
"total": 25,
"successful": 25,
"failed": 0
},
"hits": {
"total": 1,
"max_score": null,
"hits": [
{
"_index": "doc-v2",
"_type": "article",
"_id": "AAUYRY",
"_score": null,
"_source": {
"_document": {
"displayPublishedDateTime": "2014-05-20T06:58:47Z"
}
},
"sort": [
"\\\b"
]
}
]
}
}"><pre class="notranslate"><code class="notranslate">{
"took": 46,
"timed_out": false,
"_shards": {
"total": 25,
"successful": 25,
"failed": 0
},
"hits": {
"total": 1,
"max_score": null,
"hits": [
{
"_index": "doc-v2",
"_type": "article",
"_id": "AAUYRY",
"_score": null,
"_source": {
"_document": {
"displayPublishedDateTime": "2014-05-20T06:58:47Z"
}
},
"sort": [
"\\\b"
]
}
]
}
}
</code></pre></div>
<p dir="auto">Results for 'asc'</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="{
"took": 78,
"timed_out": false,
"_shards": {
"total": 25,
"successful": 25,
"failed": 0
},
"hits": {
"total": 1,
"max_score": null,
"hits": [
{
"_index": "doc-v2",
"_type": "article",
"_id": "AAUYRY",
"_score": null,
"_source": {
"_document": {
"displayPublishedDateTime": "2014-05-20T06:58:47Z"
}
},
"sort": [
" \u0001\u0000\u0000\u0000(aC3@X"
]
}
]
}
}"><pre class="notranslate"><code class="notranslate">{
"took": 78,
"timed_out": false,
"_shards": {
"total": 25,
"successful": 25,
"failed": 0
},
"hits": {
"total": 1,
"max_score": null,
"hits": [
{
"_index": "doc-v2",
"_type": "article",
"_id": "AAUYRY",
"_score": null,
"_source": {
"_document": {
"displayPublishedDateTime": "2014-05-20T06:58:47Z"
}
},
"sort": [
" \u0001\u0000\u0000\u0000(aC3@X"
]
}
]
}
}
</code></pre></div> | <p dir="auto">When querying with "preference": "_local", the _source is returned instead of the requested fields.</p>
<p dir="auto"><code class="notranslate">$ curl -XPUT 'http://ks398280.kimsufi.com:9200/test/test/1' -d '{"user": "foo", "amount": 35, "data": "some more data"}'</code></p>
<div class="highlight highlight-source-json notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="$ curl -XPOST 'http://ks398280.kimsufi.com:9200/test/test/_search?pretty' -d '{
"query": {
"match": {
"data": {
"query": "some"
}
}
},
"fields": ["user", "amount"]
}'
{
"took" : 1,
"timed_out" : false,
"_shards" : {
"total" : 5,
"successful" : 5,
"failed" : 0
},
"hits" : {
"total" : 1,
"max_score" : 0.15342641,
"hits" : [ {
"_index" : "test",
"_type" : "test",
"_id" : "1",
"_score" : 0.15342641,
"fields" : {
"amount" : 35,
"user" : "foo"
}
} ]
}
}"><pre class="notranslate"><span class="pl-ii">$ curl -XPOST 'http://ks398280.kimsufi.com:9200/test/test/_search?pretty' -d '{</span>
<span class="pl-ent">"query"</span>: {
<span class="pl-ent">"match"</span>: {
<span class="pl-ent">"data"</span>: {
<span class="pl-ent">"query"</span>: <span class="pl-s"><span class="pl-pds">"</span>some<span class="pl-pds">"</span></span>
}
}
},
<span class="pl-ent">"fields"</span>: [<span class="pl-s"><span class="pl-pds">"</span>user<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>amount<span class="pl-pds">"</span></span>]
<span class="pl-ii">}'</span>
{
<span class="pl-ent">"took"</span> : <span class="pl-c1">1</span>,
<span class="pl-ent">"timed_out"</span> : <span class="pl-c1">false</span>,
<span class="pl-ent">"_shards"</span> : {
<span class="pl-ent">"total"</span> : <span class="pl-c1">5</span>,
<span class="pl-ent">"successful"</span> : <span class="pl-c1">5</span>,
<span class="pl-ent">"failed"</span> : <span class="pl-c1">0</span>
},
<span class="pl-ent">"hits"</span> : {
<span class="pl-ent">"total"</span> : <span class="pl-c1">1</span>,
<span class="pl-ent">"max_score"</span> : <span class="pl-c1">0.15342641</span>,
<span class="pl-ent">"hits"</span> : [ {
<span class="pl-ent">"_index"</span> : <span class="pl-s"><span class="pl-pds">"</span>test<span class="pl-pds">"</span></span>,
<span class="pl-ent">"_type"</span> : <span class="pl-s"><span class="pl-pds">"</span>test<span class="pl-pds">"</span></span>,
<span class="pl-ent">"_id"</span> : <span class="pl-s"><span class="pl-pds">"</span>1<span class="pl-pds">"</span></span>,
<span class="pl-ent">"_score"</span> : <span class="pl-c1">0.15342641</span>,
<span class="pl-ent">"fields"</span> : {
<span class="pl-ent">"amount"</span> : <span class="pl-c1">35</span>,
<span class="pl-ent">"user"</span> : <span class="pl-s"><span class="pl-pds">"</span>foo<span class="pl-pds">"</span></span>
}
} ]
}
}</pre></div>
<div class="highlight highlight-source-json notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="$ curl -XPOST 'http://ks398280.kimsufi.com:9200/test/test/_search?pretty' -d '{
"query": {
"match": {
"data": {
"query": "some"
}
},
"preference": "_local"
},
"fields": ["user", "amount"],
}'
{
"took" : 1,
"timed_out" : false,
"_shards" : {
"total" : 5,
"successful" : 5,
"failed" : 0
},
"hits" : {
"total" : 1,
"max_score" : 0.15342641,
"hits" : [ {
"_index" : "test",
"_type" : "test",
"_id" : "1",
"_score" : 0.15342641, "_source" : {"user": "foo", "amount": 35, "data": "some more data"}
} ]
}
}"><pre class="notranslate"><span class="pl-ii">$ curl -XPOST 'http://ks398280.kimsufi.com:9200/test/test/_search?pretty' -d '{</span>
<span class="pl-ent">"query"</span>: {
<span class="pl-ent">"match"</span>: {
<span class="pl-ent">"data"</span>: {
<span class="pl-ent">"query"</span>: <span class="pl-s"><span class="pl-pds">"</span>some<span class="pl-pds">"</span></span>
}
},
<span class="pl-ent">"preference"</span>: <span class="pl-s"><span class="pl-pds">"</span>_local<span class="pl-pds">"</span></span>
},
<span class="pl-ent">"fields"</span>: [<span class="pl-s"><span class="pl-pds">"</span>user<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>amount<span class="pl-pds">"</span></span>],
<span class="pl-ii">}'</span>
{
<span class="pl-ent">"took"</span> : <span class="pl-c1">1</span>,
<span class="pl-ent">"timed_out"</span> : <span class="pl-c1">false</span>,
<span class="pl-ent">"_shards"</span> : {
<span class="pl-ent">"total"</span> : <span class="pl-c1">5</span>,
<span class="pl-ent">"successful"</span> : <span class="pl-c1">5</span>,
<span class="pl-ent">"failed"</span> : <span class="pl-c1">0</span>
},
<span class="pl-ent">"hits"</span> : {
<span class="pl-ent">"total"</span> : <span class="pl-c1">1</span>,
<span class="pl-ent">"max_score"</span> : <span class="pl-c1">0.15342641</span>,
<span class="pl-ent">"hits"</span> : [ {
<span class="pl-ent">"_index"</span> : <span class="pl-s"><span class="pl-pds">"</span>test<span class="pl-pds">"</span></span>,
<span class="pl-ent">"_type"</span> : <span class="pl-s"><span class="pl-pds">"</span>test<span class="pl-pds">"</span></span>,
<span class="pl-ent">"_id"</span> : <span class="pl-s"><span class="pl-pds">"</span>1<span class="pl-pds">"</span></span>,
<span class="pl-ent">"_score"</span> : <span class="pl-c1">0.15342641</span>, <span class="pl-ent">"_source"</span> : {<span class="pl-ent">"user"</span>: <span class="pl-s"><span class="pl-pds">"</span>foo<span class="pl-pds">"</span></span>, <span class="pl-ent">"amount"</span>: <span class="pl-c1">35</span>, <span class="pl-ent">"data"</span>: <span class="pl-s"><span class="pl-pds">"</span>some more data<span class="pl-pds">"</span></span>}
} ]
}
}</pre></div> | 0 |
<p dir="auto">In OSX Mavericks and Atom 0.121.0, there is a bar with nothing in the top of the screen, when Atom is in FullScreen mode.</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://cloud.githubusercontent.com/assets/5305627/3910845/b1a3897a-231b-11e4-86a7-7a71d4e80ad9.png"><img src="https://cloud.githubusercontent.com/assets/5305627/3910845/b1a3897a-231b-11e4-86a7-7a71d4e80ad9.png" alt="capture decran 2014-08-13 a 20 54 05" style="max-width: 100%;"></a></p>
<p dir="auto">Is this a bug?</p> | <ol dir="auto">
<li>Open an Atom window with a file or project. (For some reason, this bug didn't happen when I tried to reproduce from an empty Atom window.)</li>
<li>Enter fullscreen mode (Ctrl-Cmd-F)</li>
<li>Open another window of Atom. (e.g. with the command line, File > Open [and then select a folder], or with the Project Quick Open package)</li>
</ol>
<p dir="auto">The window that opens will have a vestigial title bar at the top that looks like this:<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://cloud.githubusercontent.com/assets/1977010/3835399/bbbdf11e-1dbe-11e4-8bf4-d9a0d2049336.png"><img src="https://cloud.githubusercontent.com/assets/1977010/3835399/bbbdf11e-1dbe-11e4-8bf4-d9a0d2049336.png" alt="screen shot 2014-08-06 at 4 08 32 pm" style="max-width: 100%;"></a></p>
<p dir="auto">It's not a big deal, but it does undermine the purpose of fullscreen mode a little bit. ;)</p>
<p dir="auto">This is on Atom 0.120.0 running on OS X 10.9.3. I was able to reproduce the problem in safe mode.</p> | 1 |
<p dir="auto">When building a Pipeline with custom transformers, what is the best way to serialize that for later use?</p>
<p dir="auto">If you use pickle, you need to define those functions in the new environment, so that doesn't seem like a solution to me. I ran into the same issue with dill and joblib.</p>
<p dir="auto">What is the best practice here?</p>
<p dir="auto">Thanks!</p> | <p dir="auto">pred=[1e-10, 0, 0]<br>
sol=[1, 0, 0]<br>
metrics.roc_auc_score(sol, pred) # 0.5, wrong, 1 is correct</p>
<p dir="auto">pred=[1, 0, 0]<br>
sol=[1, 0, 0]<br>
metrics.roc_auc_score(sol, pred) # 1 correct</p> | 0 |
<h3 dir="auto">Version</h3>
<p dir="auto">2.5.16</p>
<h3 dir="auto">Reproduction link</h3>
<p dir="auto"><a href="https://github.com/ondrej-janosik/vue-bug-props/blob/master/src/App.vue">https://github.com/ondrej-janosik/vue-bug-props/blob/master/src/App.vue</a></p>
<h3 dir="auto">Steps to reproduce</h3>
<ol dir="auto">
<li>Create a new project with VUE CLI, just add typescript<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/18070427/41981037-4a4b4282-7a28-11e8-812b-168bf93f5734.png"><img src="https://user-images.githubusercontent.com/18070427/41981037-4a4b4282-7a28-11e8-812b-168bf93f5734.png" alt="image" style="max-width: 100%;"></a></li>
<li>Open is VS code</li>
<li>Create a property in data in App.vue<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/18070427/41981060-5765ff02-7a28-11e8-88df-154083375ef4.png"><img src="https://user-images.githubusercontent.com/18070427/41981060-5765ff02-7a28-11e8-88df-154083375ef4.png" alt="image" style="max-width: 100%;"></a></li>
<li>Access this property in a computed method - works<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/18070427/41981083-62836e10-7a28-11e8-8d61-0a83ee132e93.png"><img src="https://user-images.githubusercontent.com/18070427/41981083-62836e10-7a28-11e8-8d61-0a83ee132e93.png" alt="image" style="max-width: 100%;"></a></li>
<li>Add prop with type array<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/18070427/41981104-6ac55822-7a28-11e8-898e-4681c0087410.png"><img src="https://user-images.githubusercontent.com/18070427/41981104-6ac55822-7a28-11e8-898e-4681c0087410.png" alt="image" style="max-width: 100%;"></a></li>
<li>Computed property context is broken - not work<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/18070427/41981194-9e4c6f78-7a28-11e8-8cf7-c114e3ab36d6.png"><img src="https://user-images.githubusercontent.com/18070427/41981194-9e4c6f78-7a28-11e8-8cf7-c114e3ab36d6.png" alt="image" style="max-width: 100%;"></a></li>
<li>Change prop type to for example to object - works<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/18070427/41981213-a92b4f22-7a28-11e8-89f2-d61da38bc687.png"><img src="https://user-images.githubusercontent.com/18070427/41981213-a92b4f22-7a28-11e8-89f2-d61da38bc687.png" alt="image" style="max-width: 100%;"></a></li>
</ol>
<h3 dir="auto">What is expected?</h3>
<p dir="auto">A computed property should have the same context even if prop with type array is presented</p>
<h3 dir="auto">What is actually happening?</h3>
<p dir="auto">Computed property context is missing</p> | <h3 dir="auto">Version</h3>
<p dir="auto">2.5.16</p>
<h3 dir="auto">Reproduction link</h3>
<p dir="auto"><a href="https://codesandbox.io/s/github/weihongyu12/vue-typescript-demo" rel="nofollow">https://codesandbox.io/s/github/weihongyu12/vue-typescript-demo</a></p>
<h3 dir="auto">Steps to reproduce</h3>
<p dir="auto">Uses @vue/cli to build the project. The configuration uses TypeScript. However, in the process of using, in the <code class="notranslate">methods</code>, there is no access to the <code class="notranslate">computed</code>. The construction does not exist on type 'Vue'.</p>
<div class="highlight highlight-text-html-vue notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="<template>
<ul
v-if="show"
class="picker">
<li
v-for="(item, index) of pickerList"
:key="index">{{ item }}</li>
</ul>
</template>
<script lang="ts">
import Vue from 'vue';
export default Vue.extend({
name : 'Picker',
props: {
columns: {
type: Array,
default: [],
},
show: {
type: Boolean ,
default: false,
},
},
computed: {
pickerList(): Array<string> {
const arr = [];
const columns: any = this.columns;
for (const item of columns) {
arr.push(item.title);
}
return arr;
},
pickerMap() : Map<string, number> {
const map = new Map();
const columns: any = this.columns;
for (const item of columns) {
map.set(item.title, item.id);
}
return map;
},
},
methods: {
onConfirm(value: string): void {
const resumeId = this.pickerMap.get(value); // Property 'pickerMap' does not exist on type 'Vue'.
this.$emit('confirm', resumeId);
},
},
});
</script>"><pre class="notranslate"><<span class="pl-ent">template</span>>
<<span class="pl-ent">ul</span>
<span class="pl-e">v-if</span>=<span class="pl-s1"><span class="pl-pds">"</span>show<span class="pl-pds">"</span></span>
<span class="pl-e">class</span>=<span class="pl-s"><span class="pl-pds">"</span>picker<span class="pl-pds">"</span></span>>
<<span class="pl-ent">li</span>
<span class="pl-e">v-for</span>=<span class="pl-s1"><span class="pl-pds">"</span>(item, index) <span class="pl-k">of</span> pickerList<span class="pl-pds">"</span></span>
:<span class="pl-e">key</span>=<span class="pl-s1"><span class="pl-pds">"</span>index<span class="pl-pds">"</span></span>>{{ item }}</<span class="pl-ent">li</span>>
</<span class="pl-ent">ul</span>>
</<span class="pl-ent">template</span>>
<span class="pl-s1"><<span class="pl-ent">script</span> <span class="pl-e">lang</span>=<span class="pl-s"><span class="pl-pds">"</span>ts<span class="pl-pds">"</span></span>></span>
<span class="pl-s1"><span class="pl-k">import</span> <span class="pl-smi">Vue</span> <span class="pl-k">from</span> <span class="pl-s"><span class="pl-pds">'</span>vue<span class="pl-pds">'</span></span>;</span>
<span class="pl-s1"></span>
<span class="pl-s1"><span class="pl-k">export</span> <span class="pl-k">default</span> <span class="pl-smi">Vue</span>.<span class="pl-en">extend</span>({</span>
<span class="pl-s1"> name : <span class="pl-s"><span class="pl-pds">'</span>Picker<span class="pl-pds">'</span></span>,</span>
<span class="pl-s1"> props: {</span>
<span class="pl-s1"> columns: {</span>
<span class="pl-s1"> type: <span class="pl-c1">Array</span>,</span>
<span class="pl-s1"> default: [],</span>
<span class="pl-s1"> },</span>
<span class="pl-s1"> show: {</span>
<span class="pl-s1"> type: <span class="pl-c1">Boolean</span> ,</span>
<span class="pl-s1"> default: <span class="pl-c1">false</span>,</span>
<span class="pl-s1"> },</span>
<span class="pl-s1"> },</span>
<span class="pl-s1"> computed: {</span>
<span class="pl-s1"> pickerList()<span class="pl-k">:</span> <span class="pl-en">Array</span><<span class="pl-c1">string</span>> {</span>
<span class="pl-s1"> <span class="pl-k"><span class="pl-k">const</span></span> arr <span class="pl-k">=</span> [];</span>
<span class="pl-s1"> <span class="pl-k"><span class="pl-k">const</span></span> columns<span class="pl-k">:</span> <span class="pl-c1">any</span> <span class="pl-k">=</span> <span class="pl-c1">this</span>.<span class="pl-smi">columns</span>;</span>
<span class="pl-s1"> <span class="pl-k">for</span> (<span class="pl-k"><span class="pl-k">const</span></span> item <span class="pl-k">of</span> <span class="pl-smi">columns</span>) {</span>
<span class="pl-s1"> <span class="pl-smi">arr</span>.<span class="pl-c1">push</span>(<span class="pl-smi">item</span>.<span class="pl-c1">title</span>);</span>
<span class="pl-s1"> }</span>
<span class="pl-s1"> <span class="pl-k">return</span> <span class="pl-smi">arr</span>;</span>
<span class="pl-s1"> },</span>
<span class="pl-s1"> pickerMap() <span class="pl-k">:</span> <span class="pl-en">Map</span><<span class="pl-c1">string</span>, <span class="pl-c1">number</span>> {</span>
<span class="pl-s1"> <span class="pl-k"><span class="pl-k">const</span></span> map <span class="pl-k">=</span> <span class="pl-k">new</span> <span class="pl-c1">Map</span>();</span>
<span class="pl-s1"> <span class="pl-k"><span class="pl-k">const</span></span> columns<span class="pl-k">:</span> <span class="pl-c1">any</span> <span class="pl-k">=</span> <span class="pl-c1">this</span>.<span class="pl-smi">columns</span>;</span>
<span class="pl-s1"> <span class="pl-k">for</span> (<span class="pl-k"><span class="pl-k">const</span></span> item <span class="pl-k">of</span> <span class="pl-smi">columns</span>) {</span>
<span class="pl-s1"> <span class="pl-smi">map</span>.<span class="pl-c1">set</span>(<span class="pl-smi">item</span>.<span class="pl-c1">title</span>, <span class="pl-smi">item</span>.<span class="pl-c1">id</span>);</span>
<span class="pl-s1"> }</span>
<span class="pl-s1"> <span class="pl-k">return</span> <span class="pl-smi">map</span>;</span>
<span class="pl-s1"> },</span>
<span class="pl-s1"> },</span>
<span class="pl-s1"> methods: {</span>
<span class="pl-s1"> onConfirm(<span class="pl-v">value</span><span class="pl-k">:</span> <span class="pl-c1">string</span>)<span class="pl-k">:</span> <span class="pl-c1">void</span> {</span>
<span class="pl-s1"> <span class="pl-k"><span class="pl-k">const</span></span> resumeId <span class="pl-k">=</span> <span class="pl-c1">this</span>.<span class="pl-smi">pickerMap</span>.<span class="pl-c1">get</span>(<span class="pl-smi">value</span>); <span class="pl-c"><span class="pl-c">//</span> Property 'pickerMap' does not exist on type 'Vue'.</span></span>
<span class="pl-s1"> <span class="pl-c1">this</span>.<span class="pl-en">$emit</span>(<span class="pl-s"><span class="pl-pds">'</span>confirm<span class="pl-pds">'</span></span>, <span class="pl-smi">resumeId</span>);</span>
<span class="pl-s1"> },</span>
<span class="pl-s1"> },</span>
<span class="pl-s1">});</span>
<span class="pl-s1"></<span class="pl-ent">script</span>></span></pre></div>
<h3 dir="auto">What is expected?</h3>
<p dir="auto">I don't know if it's because of my writing, or because of the configuration that caused me this error. I wrote the code based on the official document. I believe it can be build successfully.</p>
<h3 dir="auto">What is actually happening?</h3>
<p dir="auto">When I run <code class="notranslate">npm run build</code>, an error Property 'pickerMap' does not exist on type 'Vue'. The complete prompt is as follows:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="ERROR in E:/project/demo/src/components/picker.vue
44:29 Property 'pickerMap' does not exist on type 'Vue'.
42 | methods: {
43 | onConfirm(value: string): void {
> 44 | const resumeId = this.pickerMap.get(value);
| ^
45 | this.$emit('confirm', resumeId);
46 | },
47 | },"><pre class="notranslate"><code class="notranslate">ERROR in E:/project/demo/src/components/picker.vue
44:29 Property 'pickerMap' does not exist on type 'Vue'.
42 | methods: {
43 | onConfirm(value: string): void {
> 44 | const resumeId = this.pickerMap.get(value);
| ^
45 | this.$emit('confirm', resumeId);
46 | },
47 | },
</code></pre></div>
<hr>
<p dir="auto">I do not know of any reason for this error. It is imperative that I have a friend to help me solve this error, but I hope that the official can speed up the improvement of TypeScript's documentation, because I believe TypeScript will become popular. Thank you.</p> | 1 |
<p dir="auto">As I see gvfs starts with a PROPFIND (I'm using litmus_test_server.go).</p>
<p dir="auto">(By the way, at the moment go test fails, too).</p>
<p dir="auto">I'm not a DAV expert at all, but I'd really like to have this work with Windows (for camlistore.org :)).<br>
Where should I start? What's missing, or known lack of care?</p> | <pre class="notranslate">Hello,
Just like we have FormatFloat(...) string and AppendFloat(...) []byte, it would be nice
to avoid a costly string conversion when parsing numbers from a byteslice.
Simple code like:
func main() {
s := "123.45"
r := bytes.NewBufferString(s)
runtime.UpdateMemStats()
println(runtime.MemStats.Mallocs)
var v float64
fmt.Fscanf(r, "%f", &v)
runtime.UpdateMemStats()
println(runtime.MemStats.Mallocs)
}
says it's doing 7 allocations in the Scanf.</pre> | 0 |
<p dir="auto">Is this behavior of <code class="notranslate">np.double</code> desired? I'm wondering about the middle case which I would have expected to be <code class="notranslate">array([3.])</code> instead of <code class="notranslate">3.0</code>. I assume there is nothing special about <code class="notranslate">double</code> or <code class="notranslate">3</code> and that this is probably not a real issue, so it can be closed soon.</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content=">>> np.double(np.asarray(3))
3.0
>>> np.double(np.asarray([3]))
3.0
>>> np.double(np.asarray([3, 4]))
array([ 3., 4.])"><pre class="notranslate"><code class="notranslate">>>> np.double(np.asarray(3))
3.0
>>> np.double(np.asarray([3]))
3.0
>>> np.double(np.asarray([3, 4]))
array([ 3., 4.])
</code></pre></div> | <p dir="auto">Hi all, I'm reporting a problem when using <code class="notranslate">np.float64</code> as a constructor.</p>
<p dir="auto">For any <code class="notranslate">a = np.array([[scalar]])</code> assuming <code class="notranslate">type(scalar) in (int, float)</code>, this raises <code class="notranslate">AssertionError</code>:</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="assert np.float64(a).shape is a.shape"><pre class="notranslate"><span class="pl-k">assert</span> <span class="pl-s1">np</span>.<span class="pl-en">float64</span>(<span class="pl-s1">a</span>).<span class="pl-s1">shape</span> <span class="pl-c1">is</span> <span class="pl-s1">a</span>.<span class="pl-s1">shape</span></pre></div>
<p dir="auto">The shape after casting to <code class="notranslate">np.float64</code> is <code class="notranslate">()</code> instead of the expected <code class="notranslate">(1,1)</code>. We can see what's going on more clearly when only this last assertion fails:</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="scalar = 0.0 # example
a = np.array([[scalar]])
assert type(a) is np.ndarray
assert type(np.float64(a)) is np.ndarray #AssertionError"><pre class="notranslate"><span class="pl-s1">scalar</span> <span class="pl-c1">=</span> <span class="pl-c1">0.0</span> <span class="pl-c"># example</span>
<span class="pl-s1">a</span> <span class="pl-c1">=</span> <span class="pl-s1">np</span>.<span class="pl-en">array</span>([[<span class="pl-s1">scalar</span>]])
<span class="pl-k">assert</span> <span class="pl-en">type</span>(<span class="pl-s1">a</span>) <span class="pl-c1">is</span> <span class="pl-s1">np</span>.<span class="pl-s1">ndarray</span>
<span class="pl-k">assert</span> <span class="pl-en">type</span>(<span class="pl-s1">np</span>.<span class="pl-en">float64</span>(<span class="pl-s1">a</span>)) <span class="pl-c1">is</span> <span class="pl-s1">np</span>.<span class="pl-s1">ndarray</span> <span class="pl-c">#AssertionError</span></pre></div>
<p dir="auto">The problem is that <code class="notranslate">np.float64(a) is np.float64(scalar)</code>. Contrary to expectations, <code class="notranslate">np.float64(a) is not np.float64([[scalar]])</code></p>
<p dir="auto">This happens with <code class="notranslate">v1.14.2</code> as well as <code class="notranslate">master</code>. I've tested with python <code class="notranslate">3.6.4</code>/<code class="notranslate">2.7.11</code> on <code class="notranslate">ubuntu 16.04</code>/<code class="notranslate">Mac OS 10.12.3</code>.</p> | 1 |
<p dir="auto">As discussed <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="60182440" data-permission-text="Title is private" data-url="https://github.com/kubernetes/kubernetes/issues/5156" data-hovercard-type="issue" data-hovercard-url="/kubernetes/kubernetes/issues/5156/hovercard" href="https://github.com/kubernetes/kubernetes/issues/5156">#5156</a> and in other places, the load balancer creation should not be synchronous with resource creation (goes against the "dumb resources" model) but instead be applied by a controller both before and after creation.</p> | <p dir="auto">I'm trying to bring up a service and pod (<a href="https://github.com/youtube/vitess/examples/kubernetes/vtctld-up.sh">https://github.com/youtube/vitess/examples/kubernetes/vtctld-up.sh</a>). The service has createExternalLoadBalancer: true. The call nearly always returns:<br>
ERROR: F0306 15:59:47.443762 17632 create.go:81] Timeout: request did not complete within allowed duration.</p>
<p dir="auto">I then proceed to create a firewall rule (created successfully) and access the service, which occasionally works. Note the non-determinism:</p>
<ol dir="auto">
<li>I nearly always get the timeout error (> 90%).</li>
<li>Even if I get the timeout error, I can still create the firewall rule and it's a toss-up whether I can access the service or not.</li>
<li>If I automate these steps in a script (bring up cluster, bring up service, create firewall rule), I am virtually never able to access the service using the forwarded IP and port.</li>
</ol>
<p dir="auto">Our thoughts are that adding the external load balancer makes bringing up the service take longer, putting it at risk of timeout. I couldn't find a way to increase the timeout threshold in kubectl create.</p>
<p dir="auto">In addition there may be some hidden steps or waits between steps that are needed since the automation never works, but running the steps manually sometimes does.</p>
<p dir="auto">Also of note is that this seems to be new behavior. Prior to 3/5/2015, I hadn't seen any issues (if the timeout error occurred, it didn't have any impact), automation worked completely fine.</p> | 1 |
<pre class="notranslate">ok text/template/parse 0.007s
--- FAIL: TestOverflowRuntimeTimer (0.10 seconds)
sleep_test.go:402: runtime timer stuck: overflow in addtimer
FAIL
FAIL time 4.651s
Fails on several builds, FreeBSD, OpenBSD, Windows. I believe all these builders are
running inside vm's so this could be related
What is the expected output? What do you see instead?
Is this test failure an actual problem, or just noise by slow builders with a poor sense
of the passage of time ?
Please use labels and text to provide additional information.
<a href="http://build.golang.org/log/5467657de567ee1ff619b30da43ef06871b2698c" rel="nofollow">http://build.golang.org/log/5467657de567ee1ff619b30da43ef06871b2698c</a>
<a href="http://build.golang.org/log/977d7ed62d3cc11c571832016b98020056523485" rel="nofollow">http://build.golang.org/log/977d7ed62d3cc11c571832016b98020056523485</a></pre> | <pre class="notranslate">What steps will reproduce the problem?
panic messages should list the go version (from runtime.theVersion) as well as the
operating system and arch details as close to the top of the panic message as possible.
This would greatly aide in debugging problems reported on the mailing list and IRC as
those are the first details requested and the most often omitted.</pre> | 0 |
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I have searched the <a href="https://github.com/zeit/next.js/issues">issues</a> of this repository and believe that this is not a duplicate.</li>
</ul>
<h2 dir="auto">Expected Behavior</h2>
<p dir="auto">HTML pages generated by <code class="notranslate">next export</code> command should render initial view as quickly as possible.</p>
<h2 dir="auto">Current Behavior</h2>
<p dir="auto">JS is loaded before everything else:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="<!DOCTYPE html>
<html>
<head>
<meta charSet="utf-8" class="next-head"/>
<meta name="viewport" content="width=device-width, initial-scale=1" class="next-head"/>
<link rel="preload" href="/_next/[hash]/page/index.js" as="script"/>
<link rel="preload" href="/_next/[hash]/page/_error.js" as="script"/>
<link rel="preload" href="/_next/[hash]/app.js" as="script"/>
</head>"><pre class="notranslate"><code class="notranslate"><!DOCTYPE html>
<html>
<head>
<meta charSet="utf-8" class="next-head"/>
<meta name="viewport" content="width=device-width, initial-scale=1" class="next-head"/>
<link rel="preload" href="/_next/[hash]/page/index.js" as="script"/>
<link rel="preload" href="/_next/[hash]/page/_error.js" as="script"/>
<link rel="preload" href="/_next/[hash]/app.js" as="script"/>
</head>
</code></pre></div>
<p dir="auto">If those <code class="notranslate"><link rel="preload"/></code> tags are removed, the JS still loads a bit sooner than I would prefer.</p>
<p dir="auto">One solution is to replace these lines:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="<script async="" id="__NEXT_PAGE__/" type="text/javascript" src="/_next/[hash]/page/index.js"></script>
<script async="" id="__NEXT_PAGE__/_error" type="text/javascript" src="/_next/[hash]/page/_error.js"></script>
<script type="text/javascript" src="/_next/[hash]/app.js" async=""></script>
</body>
</html>"><pre class="notranslate"><code class="notranslate"><script async="" id="__NEXT_PAGE__/" type="text/javascript" src="/_next/[hash]/page/index.js"></script>
<script async="" id="__NEXT_PAGE__/_error" type="text/javascript" src="/_next/[hash]/page/_error.js"></script>
<script type="text/javascript" src="/_next/[hash]/app.js" async=""></script>
</body>
</html>
</code></pre></div>
<p dir="auto">With something like:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="<script>
setTimeout(function(){
var insertScript = function (src, id) {
var scriptEl = document.createElement('script');
id && (scriptEl.id = id);
scriptEl.src = src;
document.body.appendChild(scriptEl);
}
insertScript('/_next/[hash]/page/index.js', '__NEXT_PAGE__/');
insertScript('/_next/[hash]/page/_error.js', '__NEXT_PAGE__/_error');
insertScript('/_next/[hash]/app.js');
}, 0);
</script>
</body>
</html>"><pre class="notranslate"><code class="notranslate"><script>
setTimeout(function(){
var insertScript = function (src, id) {
var scriptEl = document.createElement('script');
id && (scriptEl.id = id);
scriptEl.src = src;
document.body.appendChild(scriptEl);
}
insertScript('/_next/[hash]/page/index.js', '__NEXT_PAGE__/');
insertScript('/_next/[hash]/page/_error.js', '__NEXT_PAGE__/_error');
insertScript('/_next/[hash]/app.js');
}, 0);
</script>
</body>
</html>
</code></pre></div>
<p dir="auto">I've cobbled together a crude / brittle approach to making both of these modifications:</p>
<ol dir="auto">
<li>
<p dir="auto"><code class="notranslate">npm install --save-dev replace-in-file</code></p>
</li>
<li>
<p dir="auto">Add the file in <a href="https://gist.github.com/mrienstra/218c0836208af352951a663652a6d7c7">this gist</a> to your project root.</p>
</li>
<li>
<p dir="auto">Add something like the following to <code class="notranslate">package.json</code>:</p>
</li>
</ol>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content=" "scripts": {
"export": "next export && node export_postprocess.js",
},"><pre class="notranslate"><code class="notranslate"> "scripts": {
"export": "next export && node export_postprocess.js",
},
</code></pre></div>
<h2 dir="auto">Your Environment</h2>
<table role="table">
<thead>
<tr>
<th>Tech</th>
<th>Version</th>
</tr>
</thead>
<tbody>
<tr>
<td>next</td>
<td>4.3.0-canary.1</td>
</tr>
<tr>
<td>node</td>
<td>8.9.4</td>
</tr>
<tr>
<td>OS</td>
<td>OS X 10.13.2</td>
</tr>
</tbody>
</table>
<h2 dir="auto">Notes</h2>
<p dir="auto">The HTML snippet shown above after "One solution is to replace these lines" is as seen in <code class="notranslate">next.js 4.3.0-canary.1</code>. In <code class="notranslate">4.2.1</code> these lines are slightly different (line 3 empty div, line 5 closing div):</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="<script async="" id="__NEXT_PAGE__/" type="text/javascript" src="/_next/[hash]/page/index.js"></script>
<script async="" id="__NEXT_PAGE__/_error" type="text/javascript" src="/_next/[hash]/page/_error.js"></script>
<div></div>
<script type="text/javascript" src="/_next/[hash]/app.js" async=""></script>
</div>
</body>
</html>"><pre class="notranslate"><code class="notranslate"><script async="" id="__NEXT_PAGE__/" type="text/javascript" src="/_next/[hash]/page/index.js"></script>
<script async="" id="__NEXT_PAGE__/_error" type="text/javascript" src="/_next/[hash]/page/_error.js"></script>
<div></div>
<script type="text/javascript" src="/_next/[hash]/app.js" async=""></script>
</div>
</body>
</html>
</code></pre></div>
<p dir="auto">The <code class="notranslate">export_postprocess.js</code> node script (see gist link above) takes this difference into account.</p>
<h2 dir="auto">Next Steps</h2>
<p dir="auto">Should these changes be applied (in a cleaner way) to the <code class="notranslate">export</code> tool? As options or as defaults?</p>
<p dir="auto">Would be interested in hearing how these changes affect the loading sequence for others.</p> | <p dir="auto">We are building an app with several dependencies being used in several pages, the thing is when building the application for production, some dependencies are being used in several of the chunks generated, here you can see how it looks:<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/996671/33784872-873b1dc0-dc30-11e7-94c9-b0e1ffeda64d.png"><img width="1280" alt="screen shot 2017-12-08 at 3 49 34 pm" src="https://user-images.githubusercontent.com/996671/33784872-873b1dc0-dc30-11e7-94c9-b0e1ffeda64d.png" style="max-width: 100%;"></a></p>
<p dir="auto">In this case draft-js, redux-form among others are repeated in many pages across the app.</p>
<p dir="auto">I wanted to know if it is possible to configure webpack to extract the repeated dependencies into a separated normal chunk to decrease the size of the pages.</p>
<ul dir="auto">
<li>[x ] I have searched the <a href="https://github.com/zeit/next.js/issues">issues</a> of this repository and believe that this is not a duplicate.</li>
</ul>
<h2 dir="auto">Your Environment</h2>
<table role="table">
<thead>
<tr>
<th>Tech</th>
<th>Version</th>
</tr>
</thead>
<tbody>
<tr>
<td>next</td>
<td>4.1.4</td>
</tr>
</tbody>
</table> | 0 |
<h4 dir="auto">Issue Description</h4>
<p dir="auto">It looks likes there's a bug in the Beta FreeCodeCamp's <a href="https://beta.freecodecamp.com/en/signup" rel="nofollow">signup</a></p>
<p dir="auto">When user signups for the first time, it directs to a blank page. I would've assumed it would return back to the homepage (or at least returns back to page where it was before clicking the sign-up page)</p>
<h4 dir="auto">Browser Information</h4>
<ul dir="auto">
<li>Browser Name, Version: Chrome/55.0.2883.95</li>
<li>Operating System: Macintosh; Intel Mac OS X 10_11_6</li>
<li>Mobile, Desktop, or Tablet: MacBookPro</li>
</ul>
<h4 dir="auto">Screenshot</h4>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://cloud.githubusercontent.com/assets/10217112/22176357/47cf7700-dfbd-11e6-9513-e55ae07a4758.png"><img width="1236" alt="screen shot 2017-01-21 at 9 24 42 am" src="https://cloud.githubusercontent.com/assets/10217112/22176357/47cf7700-dfbd-11e6-9513-e55ae07a4758.png" style="max-width: 100%;"></a></p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://cloud.githubusercontent.com/assets/10217112/22176361/4ffdcf3a-dfbd-11e6-8b8b-db01ca8243f0.png"><img width="623" alt="screen shot 2017-01-21 at 9 23 29 am" src="https://cloud.githubusercontent.com/assets/10217112/22176361/4ffdcf3a-dfbd-11e6-8b8b-db01ca8243f0.png" style="max-width: 100%;"></a></p> | <p dir="auto">After creating an user in beta(tried the same email I have FCC associated with). I was redirected to <a href="http://beta.freecodecamp.com/api/users" rel="nofollow">http://beta.freecodecamp.com/api/users</a></p>
<p dir="auto">And all I see, is <code class="notranslate">{}</code></p>
<p dir="auto">After refreshing this page, I am getting the error displayed: <code class="notranslate">The [email protected] email address is already associated with an account. Try signing in with it here instead.</code> With what I did log in afterwards</p>
<ul dir="auto">
<li>Browser Name, Version: Chrome Version 55.0.2883.87</li>
<li>Operating System: Windows 7 64bit</li>
</ul> | 1 |
<p dir="auto">Trying to run a test that extends <code class="notranslate">ESIntegTestCase</code> with Elasticsearch 2.0.0, getting this:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="java.lang.RuntimeException: found jar hell in test classpath
at org.elasticsearch.bootstrap.BootstrapForTesting.<clinit>(BootstrapForTesting.java:63)
at org.elasticsearch.test.ESTestCase.<clinit>(ESTestCase.java:106)
at java.lang.Class.forName0(Native Method)
at java.lang.Class.forName(Class.java:348)
at com.carrotsearch.randomizedtesting.RandomizedRunner$1.run(RandomizedRunner.java:573)
Caused by: java.lang.IllegalStateException: jar hell!
class: jdk.packager.services.UserJvmOptionsService
jar1: /Library/Java/JavaVirtualMachines/jdk1.8.0_66.jdk/Contents/Home/lib/ant-javafx.jar
jar2: /Library/Java/JavaVirtualMachines/jdk1.8.0_66.jdk/Contents/Home/lib/packager.jar
at org.elasticsearch.bootstrap.JarHell.checkClass(JarHell.java:267)
at org.elasticsearch.bootstrap.JarHell.checkJarHell(JarHell.java:185)
at org.elasticsearch.bootstrap.JarHell.checkJarHell(JarHell.java:86)
at org.elasticsearch.bootstrap.BootstrapForTesting.<clinit>(BootstrapForTesting.java:61)
... 4 more"><pre class="notranslate"><code class="notranslate">java.lang.RuntimeException: found jar hell in test classpath
at org.elasticsearch.bootstrap.BootstrapForTesting.<clinit>(BootstrapForTesting.java:63)
at org.elasticsearch.test.ESTestCase.<clinit>(ESTestCase.java:106)
at java.lang.Class.forName0(Native Method)
at java.lang.Class.forName(Class.java:348)
at com.carrotsearch.randomizedtesting.RandomizedRunner$1.run(RandomizedRunner.java:573)
Caused by: java.lang.IllegalStateException: jar hell!
class: jdk.packager.services.UserJvmOptionsService
jar1: /Library/Java/JavaVirtualMachines/jdk1.8.0_66.jdk/Contents/Home/lib/ant-javafx.jar
jar2: /Library/Java/JavaVirtualMachines/jdk1.8.0_66.jdk/Contents/Home/lib/packager.jar
at org.elasticsearch.bootstrap.JarHell.checkClass(JarHell.java:267)
at org.elasticsearch.bootstrap.JarHell.checkJarHell(JarHell.java:185)
at org.elasticsearch.bootstrap.JarHell.checkJarHell(JarHell.java:86)
at org.elasticsearch.bootstrap.BootstrapForTesting.<clinit>(BootstrapForTesting.java:61)
... 4 more
</code></pre></div>
<p dir="auto">Not sure I can do much about that.</p> | <p dir="auto">ES' hit highlighting abilities are really sophisticated, but I'd like to suggest the API be made a bit more flexible.</p>
<p dir="auto">One use case we have is the user performs their query against the <code class="notranslate">_all</code> field, and our UI needs to highlight matching terms in whatever specific fields the user happens to be viewing.</p>
<p dir="auto">The fact that the <code class="notranslate">{"highlight"}</code> API lets you wildcard field names is pretty great, until you simply specify <code class="notranslate">"*"</code>. What happens is that ES will complain that certain fields aren't necessarily compatible with the highlighting options specified. For example, you might see:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="[2014-04-09 16:48:07,009][DEBUG][action.search.type ] [Miles Warren] [6838] Failed to execute fetch phase
org.elasticsearch.ElasticsearchIllegalArgumentException: the field [edit_date] should be indexed with term vector with position offsets to be used with fast vector highlighter
at org.elasticsearch.search.highlight.FastVectorHighlighter.highlight(FastVectorHighlighter.java:68)
at org.elasticsearch.search.highlight.HighlightPhase.hitExecute(HighlightPhase.java:126)
at org.elasticsearch.search.fetch.FetchPhase.execute(FetchPhase.java:211)
at org.elasticsearch.search.SearchService.executeFetchPhase(SearchService.java:444)
at org.elasticsearch.search.action.SearchServiceTransportAction.sendExecuteFetch(SearchServiceTransportAction.java:406)
at org.elasticsearch.action.search.type.TransportSearchQueryThenFetchAction$AsyncAction.executeFetch(TransportSearchQueryThenFetchAction.java:150)
at org.elasticsearch.action.search.type.TransportSearchQueryThenFetchAction$AsyncAction$2.run(TransportSearchQueryThenFetchAction.java:134)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:744)"><pre class="notranslate"><code class="notranslate">[2014-04-09 16:48:07,009][DEBUG][action.search.type ] [Miles Warren] [6838] Failed to execute fetch phase
org.elasticsearch.ElasticsearchIllegalArgumentException: the field [edit_date] should be indexed with term vector with position offsets to be used with fast vector highlighter
at org.elasticsearch.search.highlight.FastVectorHighlighter.highlight(FastVectorHighlighter.java:68)
at org.elasticsearch.search.highlight.HighlightPhase.hitExecute(HighlightPhase.java:126)
at org.elasticsearch.search.fetch.FetchPhase.execute(FetchPhase.java:211)
at org.elasticsearch.search.SearchService.executeFetchPhase(SearchService.java:444)
at org.elasticsearch.search.action.SearchServiceTransportAction.sendExecuteFetch(SearchServiceTransportAction.java:406)
at org.elasticsearch.action.search.type.TransportSearchQueryThenFetchAction$AsyncAction.executeFetch(TransportSearchQueryThenFetchAction.java:150)
at org.elasticsearch.action.search.type.TransportSearchQueryThenFetchAction$AsyncAction$2.run(TransportSearchQueryThenFetchAction.java:134)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:744)
</code></pre></div>
<p dir="auto">("edit_date" is defined as a date field and of course has no "index_options" on it).</p>
<p dir="auto">And if you hack a bit at the ES sources to silently ignore these errors you'll see it also complains about ES' core fields such as <code class="notranslate">_id</code>, <code class="notranslate">_timestamp</code>, <code class="notranslate">_source</code>, etc.</p>
<p dir="auto">What I'd like to propose is an option in the <code class="notranslate">{ "highlight" }</code> API to be lenient about mismatches between field mapping definitions and highlight parameters. So if the highlight parameters aren't compatible with the mapping, ES will just move on to the next field, rather than bomb the entire request.</p>
<p dir="auto">Call it a best-efforts highlighting for "all" fields.</p>
<p dir="auto">Maybe it could fall back to decreasingly less sophisticated highlighters, until it just can't do anything. So if the request indicates the <code class="notranslate">fvh</code> be used, and that's not compatible with a particular field, it tries the <code class="notranslate">postings</code> highlighter, then the <code class="notranslate">plain</code> highlighter, then nothing.</p>
<p dir="auto">Additionally, I'd like to see a highlight parameter to exclude fields. Another use case we have is, "I want to highlight all matched fields (whatever they are), except this array of 'large text' fields that I'm not currently showing the user" -- this would really be a bit of a performance optimization, from the client. Our data model includes a couple of large text fields (10k+) per document that wouldn't need to be highlighted except at very specific times -- the other fields are small and generally need to be highlighted in one call to ES.</p>
<p dir="auto">Thoughts?</p>
<p dir="auto">If y'all think this is a good idea, I'm happy to fork, do the work, make a pull request, and work with the committer(s) until it's accepted. I basically need to do this anyways, so I'd rather have buy-in instead of maintaining my own patched version of ES.</p>
<p dir="auto">I thought about doing this as a plugin, but it looked like I'd have to duplicate quite a bit of core ES classes, which doesn't seem maintainable in the long run.</p> | 0 |
<pre class="notranslate">What steps will reproduce the problem?
If possible, include a link to a program on play.golang.org.
Compile & run the code in:
<a href="https://gist.github.com/ernestokarim/5033404">https://gist.github.com/ernestokarim/5033404</a>
It cannot be reproduced in the playground because it needs sqlite3.
What is the expected output?
Run without bugs.
What do you see instead?
Using any revision after this one:
<a href="https://code.google.com/p/go/source/detail?r=ddb9e6365e57" rel="nofollow">https://code.google.com/p/go/source/detail?r=ddb9e6365e57</a>
it starts to consume file descriptors without end, you can check it:
while [ 1 ] ; do ls /proc/6559/fd | wc -l ; sleep 1; done
even if I'm closing the prepared statements correctly. In the iteration 1021 it fails
with "unable to open database file".
Which compiler are you using (5g, 6g, 8g, gccgo)?
8g.
Which operating system are you using?
Fedora 17, 32 bits. SQLite3, the latest version from the website.
Which version are you using? (run 'go version')
go version devel +c0258d9a3589 Wed Feb 27 15:22:14 2013 -0800 linux/386
Please provide any additional information below.
<a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="11381442" data-permission-text="Title is private" data-url="https://github.com/mattn/go-sqlite3/issues/41" data-hovercard-type="issue" data-hovercard-url="/mattn/go-sqlite3/issues/41/hovercard" href="https://github.com/mattn/go-sqlite3/issues/41">mattn/go-sqlite3#41</a></pre> | <pre class="notranslate">What steps will reproduce the problem?
1. compile attached program and run it
What is the expected output?
the last value consumed by any of the multiple readers should be '9'
What do you see instead?
the last value generated before the channel was closed is 9, but more often than not the last value consumed by any of the chan's readers is 8.
Which compiler are you using (5g, 6g, 8g, gccgo)?
6g
Which operating system are you using?
OSX
Which revision are you using? (hg identify)
684ebd420f34+ tip
Please provide any additional information below.
The "9" does appear to be consumed occasionally:
$ while true; do ./6.out | egrep 'consumed: [89]'; done | egrep -n 'consumed: 9'
180:consumed: 9
356:consumed: 9
569:consumed: 9
984:consumed: 9
989:consumed: 9
2166:consumed: 9
2235:consumed: 9
2295:consumed: 9
2350:consumed: 9
2929:consumed: 9
3039:consumed: 9</pre>
<p dir="auto">Attachments:</p>
<ol dir="auto">
<li><a href="https://storage.googleapis.com/go-attachment/1183/0/t.go" rel="nofollow">t.go</a> (405 bytes)</li>
</ol> | 0 |
<p dir="auto">It seems to be something like this issue:<br>
<a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="580514624" data-permission-text="Title is private" data-url="https://github.com/mrdoob/three.js/issues/18893" data-hovercard-type="issue" data-hovercard-url="/mrdoob/three.js/issues/18893/hovercard" href="https://github.com/mrdoob/three.js/issues/18893">#18893</a></p>
<p dir="auto">In my case, the renderer <code class="notranslate">outputEncoding</code> is <code class="notranslate">GammaEncoding</code> and I'm using an <code class="notranslate">EffectComposer</code> with a <code class="notranslate">GammaCorrectionShader</code> pass at the end.</p>
<p dir="auto">If I use a reflector with the default <code class="notranslate">LinearEncoding</code> then it runs fine, but the reflection colors are screwed, they appear too dark. If I use <code class="notranslate">sRGBEncoding</code> as the reflector encoding, then the reflection colors are correct but it's running slow as hell, probably recompiling the shaders all the time.</p>
<p dir="auto">Is there a way I can get correct colors without recompiling the shaders all the time?</p> | <h5 dir="auto">Description of the problem</h5>
<p dir="auto">I'm experiencing horrible render performance when using the combination of the Water.js shader, a plane with a MeshStandardMaterial and the renderer.output using sRGBEncoding.</p>
<p dir="auto">Recreated <a href="https://threejs.org/examples/#webgl_water" rel="nofollow">water example</a> in jsfiddle and added <code class="notranslate">renderer.outputEncoding = THREE.sRGBEncoding</code> (line 119) in the fiddle: <a href="https://jsfiddle.net/4esfwtkh/" rel="nofollow">https://jsfiddle.net/4esfwtkh/</a></p>
<p dir="auto">The performance does not dip:</p>
<ul dir="auto">
<li>By not setting render.output to sRGBEncoding and using MeshStandardMaterial it will run ok (comment line 117)</li>
<li>By setting render.output to sRGBEncoding but using MeshBasicMaterial it will run ok (comment line 55 and uncomment line 56)</li>
<li>When the water is not visible in the camera frustum (i.e. by looking from below)</li>
</ul>
<h5 dir="auto">Three.js version</h5>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> Dev</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> r114</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> r113</li>
</ul> | 1 |
<h3 dir="auto">Apache Airflow version</h3>
<p dir="auto">2.4.2</p>
<h3 dir="auto">What happened</h3>
<p dir="auto">We upgraded airflow from 2.4.1 to 2.4.2 and immediately notice that every task log line is duplicated <em>into</em> CloudWatch. Comparing logs from tasks run before upgrade and after upgrade indicates that the issue is not in how the logs are displayed in Airflow, but rather that it now produces two log lines instead of one.</p>
<p dir="auto">When observing both the CloudWatch log streams and the Airflow UI, we can see duplicate log lines for <del><em>all</em></del> most log entries post upgrade, whilst seeing single log lines in tasks before upgrade.</p>
<p dir="auto">This happens <em>both</em> for tasks ran in a remote <code class="notranslate">EcsRunTaskOperator</code>'s as well as in regular <code class="notranslate">PythonOperator</code>'s.</p>
<h3 dir="auto">What you think should happen instead</h3>
<p dir="auto">A single non-duplicate log line should be produced into CloudWatch.</p>
<h3 dir="auto">How to reproduce</h3>
<p dir="auto">From my understanding now, any setup on 2.4.2 that uses CloudWatch remote logging will produce duplicate log lines. (But I have not been able to confirm other setups)</p>
<h3 dir="auto">Operating System</h3>
<p dir="auto">Docker: <code class="notranslate">apache/airflow:2.4.2-python3.9</code> - Running on AWS ECS Fargate</p>
<h3 dir="auto">Versions of Apache Airflow Providers</h3>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="apache-airflow[celery,postgres,apache.hive,jdbc,mysql,ssh,amazon,google,google_auth]==2.4.2
apache-airflow-providers-amazon==6.0.0"><pre class="notranslate"><code class="notranslate">apache-airflow[celery,postgres,apache.hive,jdbc,mysql,ssh,amazon,google,google_auth]==2.4.2
apache-airflow-providers-amazon==6.0.0
</code></pre></div>
<h3 dir="auto">Deployment</h3>
<p dir="auto">Other Docker-based deployment</p>
<h3 dir="auto">Deployment details</h3>
<p dir="auto">We are running a docker inside Fargate ECS on AWS.</p>
<p dir="auto">The following environment variables + config in CloudFormation control remote logging:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content=" - Name: AIRFLOW__LOGGING__REMOTE_LOGGING
Value: True
- Name: AIRFLOW__LOGGING__REMOTE_BASE_LOG_FOLDER
Value: !Sub "cloudwatch://${TasksLogGroup.Arn}""><pre class="notranslate"><code class="notranslate"> - Name: AIRFLOW__LOGGING__REMOTE_LOGGING
Value: True
- Name: AIRFLOW__LOGGING__REMOTE_BASE_LOG_FOLDER
Value: !Sub "cloudwatch://${TasksLogGroup.Arn}"
</code></pre></div>
<h3 dir="auto">Anything else</h3>
<p dir="auto">We did not change any other configuration during the upgrade, simply bumped the requirements for provider list + docker image from 2.4.1 to 2.4.2.</p>
<h3 dir="auto">Are you willing to submit PR?</h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> Yes I am willing to submit a PR!</li>
</ul>
<h3 dir="auto">Code of Conduct</h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I agree to follow this project's <a href="https://github.com/apache/airflow/blob/main/CODE_OF_CONDUCT.md">Code of Conduct</a></li>
</ul> | <p dir="auto"><strong>Apache Airflow version</strong>:1.10.12</p>
<p dir="auto"><strong>Apache Airflow Provider versions</strong> (please include all providers that are relevant to your bug):<br>
No other providers</p>
<p dir="auto"><strong>Kubernetes version (if you are using kubernetes)</strong> (use <code class="notranslate">kubectl version</code>):<br>
No using k8s</p>
<p dir="auto"><strong>Environment</strong>:<br>
AWS Linux</p>
<ul dir="auto">
<li><strong>Cloud provider or hardware configuration</strong>:AWS</li>
<li><strong>OS</strong> (e.g. from /etc/os-release):<br>
-NAME="Amazon Linux"<br>
VERSION="2"<br>
ID="amzn"<br>
ID_LIKE="centos rhel fedora"</li>
<li><strong>Kernel</strong> (e.g. <code class="notranslate">uname -a</code>): 4.14.186-146.268.amzn2.x86_64 <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="69689814" data-permission-text="Title is private" data-url="https://github.com/apache/airflow/issues/1" data-hovercard-type="pull_request" data-hovercard-url="/apache/airflow/pull/1/hovercard" href="https://github.com/apache/airflow/pull/1">#1</a> SMP Tue Jul 14 18:16:52 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux</li>
<li><strong>Install tools</strong>: yum, pip3</li>
<li><strong>Others</strong>:</li>
</ul>
<p dir="auto"><strong>What happened</strong>:<br>
I have been running airflow 1.10.12 for several months without any issues. Recently I updated airflow to use SSL cert. After that I reboot the airflow server and then restarted the airflow scheduler and Webserver as following:<br>
airflow Webserver &<br>
airflow scheduler &</p>
<p dir="auto">The jobs run normally at beginning. However, within one day, all the new jobs will immediately fail. From UI, I saw the task status as failed. However there is no log under the task when I checked the log from UI.</p>
<p dir="auto">I also checked the /logs/scheduler directory, except a warning of "psycopg2.errors.UniqueViolation: duplicate key value violates unique constraint "variable_key_key", which has existed for long time, I didn't see any other warning or error.</p>
<p dir="auto">Now this problem happened every day, i.e., if I restart airflow Webserver and scheduler, it works well. But within one day, all the new jobs will immediately fail with no log.</p>
<p dir="auto">Do you have any insight what can go wrong? Thanks.</p>
<p dir="auto"><strong>What you expected to happen</strong>:</p>
<p dir="auto"><strong>How to reproduce it</strong>:<br>
Not exactly sure how to reproduce. I just reboot airflow server.</p>
<p dir="auto"><strong>Anything else we need to know</strong>:</p>
<p dir="auto">How often does this problem occur? Once? Every time etc?<br>
I noticed the problem happens within 1 day.</p>
<p dir="auto">Any relevant logs to include? Put them here in side a detail tag:</p> | 0 |
<p dir="auto">Currently the carousel has only one transition effect which is a slide between pictures.<br>
It would be great if we can have a fade effect too.</p>
<p dir="auto">In a tiny project I made, you can see an exemple of a possible implementation :<br>
<a href="https://github.com/GuillaumeAmat/bootstrap-more/blob/master/bootstrap-more.css#L493">https://github.com/GuillaumeAmat/bootstrap-more/blob/master/bootstrap-more.css#L493</a></p> | <p dir="auto">With the addition of the fade transition on this js module, the most common options found on a website will be covered. For many types of images, instead of sliding the fade in would be a lot better. For all other cases, users will have to use a separate plugin.</p> | 1 |
<p dir="auto"><strong>Describe the bug</strong></p>
<p dir="auto">When calling <code class="notranslate">clock.getDelta()</code> <em>after</em> <code class="notranslate">clock.getElapsedTime()</code>, the returned value is not the delta between successive calls to <code class="notranslate">clock.getDelta()</code>.</p>
<p dir="auto"><em><strong>Live example</strong></em></p>
<p dir="auto"><a href="https://codesandbox.io/s/three-frame-delta-ev4d9" rel="nofollow">https://codesandbox.io/s/three-frame-delta-ev4d9</a></p>
<p dir="auto"><strong>Expected behavior</strong></p>
<p dir="auto">Sometimes I want to animate certain objects based on the elapsed time and others based on time deltas. I expect to be able to call <code class="notranslate">clock.getDelta()</code> once per frame and have the result represent the time between frames. However, given the implementation of <code class="notranslate">clock.getElapsedTime()</code>, this is not the case if you are also calling <code class="notranslate">clock.getElapsedTime</code> in the same loop.</p>
<p dir="auto">If you read closely, <a href="https://threejs.org/docs/#api/en/core/Clock.getDelta" rel="nofollow">the docs</a> do correctly describe this behavior ("Get the seconds passed since the time oldTime was set"), but IMO it is not the expected behavior.</p>
<p dir="auto">Or perhaps this is more of an enhancement request: could there be a separate function like <code class="notranslate">clock.getFrameDelta()</code> that is independent of <code class="notranslate">clock.getElapsedTime()</code>? Or is the suggested workaround to use multiple clocks here?</p>
<p dir="auto"><strong>Platform:</strong></p>
<ul dir="auto">
<li>Three.js version: r128</li>
</ul> | <p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://cloud.githubusercontent.com/assets/220033/12996799/0798051a-d0e8-11e5-8eb4-15c22988ed0c.png"><img src="https://cloud.githubusercontent.com/assets/220033/12996799/0798051a-d0e8-11e5-8eb4-15c22988ed0c.png" alt="screen shot 2016-02-11 at 5 46 00 pm" style="max-width: 100%;"></a></p>
<p dir="auto"><a href="https://dl.dropboxusercontent.com/u/1595444/shaka-player/stream-texture.html" rel="nofollow">https://dl.dropboxusercontent.com/u/1595444/shaka-player/stream-texture.html</a><br>
(on Chrome clicking play on the video element shows an upside down render of the video on the canvas element. In Safari/Firefox <code class="notranslate">texImage2D</code> is real upset)</p> | 0 |
<p dir="auto"><code class="notranslate">main.go</code>:</p>
<div class="highlight highlight-source-go notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="package main
const foo = `
//+build ignore
`"><pre class="notranslate"><span class="pl-k">package</span> main
<span class="pl-k">const</span> <span class="pl-s1">foo</span> <span class="pl-c1">=</span> <span class="pl-s">`</span>
<span class="pl-s">//+build ignore</span>
<span class="pl-s">`</span></pre></div>
<p dir="auto">Running <code class="notranslate">vet</code> produces:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="main.go:3: +build comment must appear before package clause and be followed by a blank line"><pre class="notranslate"><code class="notranslate">main.go:3: +build comment must appear before package clause and be followed by a blank line
</code></pre></div>
<p dir="auto">This also contributes to noise in the vet output for core, e.g.</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="./src/crypto/x509/root_darwin_arm_gen.go:181: +build comment must appear before package clause and be followed by a blank line
./src/crypto/x509/root_darwin_arm_gen.go:182: +build comment must appear before package clause and be followed by a blank line
./src/crypto/x509/root_darwin_arm_gen.go:183: +build comment must appear before package clause and be followed by a blank line"><pre class="notranslate"><code class="notranslate">./src/crypto/x509/root_darwin_arm_gen.go:181: +build comment must appear before package clause and be followed by a blank line
./src/crypto/x509/root_darwin_arm_gen.go:182: +build comment must appear before package clause and be followed by a blank line
./src/crypto/x509/root_darwin_arm_gen.go:183: +build comment must appear before package clause and be followed by a blank line
</code></pre></div> | <p dir="auto">I ran into a likely false positive from <code class="notranslate">go vet</code> when writing a program that auto-generates code.</p>
<p dir="auto"><a href="http://play.golang.org/p/kJgX_8nTCa" rel="nofollow">This program</a> generates a warning <code class="notranslate">+build comment must appear before package clause and be followed by a blank line</code></p>
<p dir="auto"><a href="http://play.golang.org/p/A-B8EklrHv" rel="nofollow">This equivalent program</a> without the <code class="notranslate">// autogenerated by...</code> comment in the first line doesn't</p>
<p dir="auto">In either case, i don't think <code class="notranslate">go vet</code> should be vetting contents of the string,</p> | 1 |
<h5 dir="auto">ISSUE TYPE</h5>
<ul dir="auto">
<li>Bug Report (same as <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="152893515" data-permission-text="Title is private" data-url="https://github.com/ansible/ansible/issues/15713" data-hovercard-type="issue" data-hovercard-url="/ansible/ansible/issues/15713/hovercard" href="https://github.com/ansible/ansible/issues/15713">#15713</a> and <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="128227674" data-permission-text="Title is private" data-url="https://github.com/ansible/ansible/issues/14082" data-hovercard-type="issue" data-hovercard-url="/ansible/ansible/issues/14082/hovercard" href="https://github.com/ansible/ansible/issues/14082">#14082</a>)</li>
</ul>
<h5 dir="auto">COMPONENT NAME</h5>
<ul dir="auto">
<li>Handlers and variable interpolation</li>
</ul>
<h5 dir="auto">ANSIBLE VERSION</h5>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="ansible 2.3.0 (devel 9962245b92) last updated 2016/10/06 09:55:33 (GMT -400)
lib/ansible/modules/core: (detached HEAD 0ee774ff15) last updated 2016/10/06 09:55:59 (GMT -400)
lib/ansible/modules/extras: (detached HEAD 5cc72c3f06) last updated 2016/10/06 09:55:59 (GMT -400)
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides"><pre class="notranslate"><code class="notranslate">ansible 2.3.0 (devel 9962245b92) last updated 2016/10/06 09:55:33 (GMT -400)
lib/ansible/modules/core: (detached HEAD 0ee774ff15) last updated 2016/10/06 09:55:59 (GMT -400)
lib/ansible/modules/extras: (detached HEAD 5cc72c3f06) last updated 2016/10/06 09:55:59 (GMT -400)
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
</code></pre></div>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="ansible-playbook 2.2.0.0 (stable-2.2 be6396d5e9) last updated 2016/10/05 17:40:02 (GMT -400)
lib/ansible/modules/core: (detached HEAD a6b8215e62) last updated 2016/10/05 17:45:10 (GMT -400)
lib/ansible/modules/extras: (detached HEAD d42975e068) last updated 2016/10/05 17:45:11 (GMT -400)
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
derek@derek-HP-EliteOne-800-G1-AiO:~$ ansible-playbook play.yaml
[WARNING]: provided hosts list is empty, only localhost is available"><pre class="notranslate"><code class="notranslate">ansible-playbook 2.2.0.0 (stable-2.2 be6396d5e9) last updated 2016/10/05 17:40:02 (GMT -400)
lib/ansible/modules/core: (detached HEAD a6b8215e62) last updated 2016/10/05 17:45:10 (GMT -400)
lib/ansible/modules/extras: (detached HEAD d42975e068) last updated 2016/10/05 17:45:11 (GMT -400)
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
derek@derek-HP-EliteOne-800-G1-AiO:~$ ansible-playbook play.yaml
[WARNING]: provided hosts list is empty, only localhost is available
</code></pre></div>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="ansible-playbook 2.1.2.0 (stable-2.1 a7d0cc6e61) last updated 2016/10/05 17:42:07 (GMT -400)
lib/ansible/modules/core: (detached HEAD 4602021670) last updated 2016/10/05 17:43:21 (GMT -400)
lib/ansible/modules/extras: (detached HEAD d312f34d9b) last updated 2016/10/05 17:43:22 (GMT -400)
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
derek@derek-HP-EliteOne-800-G1-AiO:~$ ansible-playbook play.yaml
[WARNING]: provided hosts list is empty, only localhost is available"><pre class="notranslate"><code class="notranslate">ansible-playbook 2.1.2.0 (stable-2.1 a7d0cc6e61) last updated 2016/10/05 17:42:07 (GMT -400)
lib/ansible/modules/core: (detached HEAD 4602021670) last updated 2016/10/05 17:43:21 (GMT -400)
lib/ansible/modules/extras: (detached HEAD d312f34d9b) last updated 2016/10/05 17:43:22 (GMT -400)
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
derek@derek-HP-EliteOne-800-G1-AiO:~$ ansible-playbook play.yaml
[WARNING]: provided hosts list is empty, only localhost is available
</code></pre></div>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="ansible 2.1.2.0 (detached HEAD 29f2f26278) last updated 2016/10/06 11:25:15 (GMT -400)
lib/ansible/modules/core: (detached HEAD 17ee1cfaf9) last updated 2016/10/06 11:30:32 (GMT -400)
lib/ansible/modules/extras: (detached HEAD d312f34d9b) last updated 2016/10/06 11:30:32 (GMT -400)
config file =
configured module search path = Default w/o overrides
derek@derek-HP-EliteOne-800-G1-AiO:~$ ansible-playbook --connection=local play.yaml
[WARNING]: Host file not found: /etc/ansible/hosts"><pre class="notranslate"><code class="notranslate">ansible 2.1.2.0 (detached HEAD 29f2f26278) last updated 2016/10/06 11:25:15 (GMT -400)
lib/ansible/modules/core: (detached HEAD 17ee1cfaf9) last updated 2016/10/06 11:30:32 (GMT -400)
lib/ansible/modules/extras: (detached HEAD d312f34d9b) last updated 2016/10/06 11:30:32 (GMT -400)
config file =
configured module search path = Default w/o overrides
derek@derek-HP-EliteOne-800-G1-AiO:~$ ansible-playbook --connection=local play.yaml
[WARNING]: Host file not found: /etc/ansible/hosts
</code></pre></div>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="ansible 2.1.1.0 (detached HEAD 780c363482) last updated 2016/10/06 11:32:54 (GMT -400)
lib/ansible/modules/core: (detached HEAD 242368e99b) last updated 2016/10/06 11:32:58 (GMT -400)
lib/ansible/modules/extras: (detached HEAD 14887a9ea8) last updated 2016/10/06 11:32:58 (GMT -400)
config file =
configured module search path = Default w/o overrides"><pre class="notranslate"><code class="notranslate">ansible 2.1.1.0 (detached HEAD 780c363482) last updated 2016/10/06 11:32:54 (GMT -400)
lib/ansible/modules/core: (detached HEAD 242368e99b) last updated 2016/10/06 11:32:58 (GMT -400)
lib/ansible/modules/extras: (detached HEAD 14887a9ea8) last updated 2016/10/06 11:32:58 (GMT -400)
config file =
configured module search path = Default w/o overrides
</code></pre></div>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="ansible 2.0.0.2 (stable-2.0.0.1 3b5d7400de) last updated 2016/10/06 11:11:52 (GMT -400)
lib/ansible/modules/core: (detached HEAD ce6619bf5d) last updated 2016/10/06 11:11:55 (GMT -400)
lib/ansible/modules/extras: (detached HEAD 29af26884e) last updated 2016/10/06 11:11:55 (GMT -400)
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides"><pre class="notranslate"><code class="notranslate">ansible 2.0.0.2 (stable-2.0.0.1 3b5d7400de) last updated 2016/10/06 11:11:52 (GMT -400)
lib/ansible/modules/core: (detached HEAD ce6619bf5d) last updated 2016/10/06 11:11:55 (GMT -400)
lib/ansible/modules/extras: (detached HEAD 29af26884e) last updated 2016/10/06 11:11:55 (GMT -400)
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
</code></pre></div>
<h5 dir="auto">CONFIGURATION</h5>
<p dir="auto">No change.</p>
<h5 dir="auto">OS / ENVIRONMENT</h5>
<ul dir="auto">
<li>Ansible manager: Lubuntu 16.04.1</li>
<li>Ansible worker: Ubuntu Server 16.04.1</li>
<li>Python 2.7.12</li>
</ul>
<h5 dir="auto">SUMMARY</h5>
<p dir="auto">Ansible 2.1.2.0, 2.1.1.0, and branches devel, stable-2.2, stable-2.1, and stable-2.0.0.1 do not resolve variable references in handler names.</p>
<h5 dir="auto">STEPS TO REPRODUCE</h5>
<ul dir="auto">
<li>See <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="152893515" data-permission-text="Title is private" data-url="https://github.com/ansible/ansible/issues/15713" data-hovercard-type="issue" data-hovercard-url="/ansible/ansible/issues/15713/hovercard" href="https://github.com/ansible/ansible/issues/15713">#15713</a>.</li>
</ul>
<h5 dir="auto">EXPECTED RESULTS</h5>
<ul dir="auto">
<li>See <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="152893515" data-permission-text="Title is private" data-url="https://github.com/ansible/ansible/issues/15713" data-hovercard-type="issue" data-hovercard-url="/ansible/ansible/issues/15713/hovercard" href="https://github.com/ansible/ansible/issues/15713">#15713</a>.</li>
</ul>
<p dir="auto">Ansible 2.1.0.0 produces the correct result:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="derek@derek-HP-EliteOne-800-G1-AiO:~$ ansible --version
ansible 2.1.0.0 (detached HEAD b599477242) last updated 2016/10/06 11:36:36 (GMT -400)
lib/ansible/modules/core: (detached HEAD 04a871d007) last updated 2016/10/06 11:36:38 (GMT -400)
lib/ansible/modules/extras: (detached HEAD b5fa2b3416) last updated 2016/10/06 11:36:38 (GMT -400)
config file =
configured module search path = Default w/o overrides
derek@derek-HP-EliteOne-800-G1-AiO:~$ ansible-playbook --connection=local play.yaml
[WARNING]: Host file not found: /etc/ansible/hosts
[WARNING]: provided hosts list is empty, only localhost is available
PLAY [localhost] ***************************************************************
TASK [setup] *******************************************************************
ok: [localhost]
TASK [bogus task with notify for someapp] **************************************
changed: [localhost]
RUNNING HANDLER [restart someapp] **********************************************
ok: [localhost] => {
"msg": "some bogus module call in handler someapp"
}
PLAY RECAP *********************************************************************
localhost : ok=3 changed=1 unreachable=0 failed=0
"><pre class="notranslate"><code class="notranslate">derek@derek-HP-EliteOne-800-G1-AiO:~$ ansible --version
ansible 2.1.0.0 (detached HEAD b599477242) last updated 2016/10/06 11:36:36 (GMT -400)
lib/ansible/modules/core: (detached HEAD 04a871d007) last updated 2016/10/06 11:36:38 (GMT -400)
lib/ansible/modules/extras: (detached HEAD b5fa2b3416) last updated 2016/10/06 11:36:38 (GMT -400)
config file =
configured module search path = Default w/o overrides
derek@derek-HP-EliteOne-800-G1-AiO:~$ ansible-playbook --connection=local play.yaml
[WARNING]: Host file not found: /etc/ansible/hosts
[WARNING]: provided hosts list is empty, only localhost is available
PLAY [localhost] ***************************************************************
TASK [setup] *******************************************************************
ok: [localhost]
TASK [bogus task with notify for someapp] **************************************
changed: [localhost]
RUNNING HANDLER [restart someapp] **********************************************
ok: [localhost] => {
"msg": "some bogus module call in handler someapp"
}
PLAY RECAP *********************************************************************
localhost : ok=3 changed=1 unreachable=0 failed=0
</code></pre></div>
<p dir="auto">Ansible 2.0.2.0 on branch stable-2.0 produces the correct result:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="derek@derek-HP-EliteOne-800-G1-AiO:~$ ansible --version
ansible 2.0.2.0 (stable-2.0 26078418e9) last updated 2016/10/06 11:13:05 (GMT -400)
lib/ansible/modules/core: (detached HEAD 1e68326ea6) last updated 2016/10/06 11:13:07 (GMT -400)
lib/ansible/modules/extras: (detached HEAD 4eb177e545) last updated 2016/10/06 11:13:07 (GMT -400)
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
derek@derek-HP-EliteOne-800-G1-AiO:~$ ansible-playbook --connection=local play.yaml
[WARNING]: provided hosts list is empty, only localhost is available
PLAY [localhost] ***************************************************************
TASK [setup] *******************************************************************
ok: [localhost]
TASK [bogus task with notify for someapp] **************************************
changed: [localhost]
RUNNING HANDLER [restart someapp] **********************************************
ok: [localhost] => {
"msg": "some bogus module call in handler someapp"
}
PLAY RECAP *********************************************************************
localhost : ok=3 changed=1 unreachable=0 failed=0
"><pre class="notranslate"><code class="notranslate">derek@derek-HP-EliteOne-800-G1-AiO:~$ ansible --version
ansible 2.0.2.0 (stable-2.0 26078418e9) last updated 2016/10/06 11:13:05 (GMT -400)
lib/ansible/modules/core: (detached HEAD 1e68326ea6) last updated 2016/10/06 11:13:07 (GMT -400)
lib/ansible/modules/extras: (detached HEAD 4eb177e545) last updated 2016/10/06 11:13:07 (GMT -400)
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
derek@derek-HP-EliteOne-800-G1-AiO:~$ ansible-playbook --connection=local play.yaml
[WARNING]: provided hosts list is empty, only localhost is available
PLAY [localhost] ***************************************************************
TASK [setup] *******************************************************************
ok: [localhost]
TASK [bogus task with notify for someapp] **************************************
changed: [localhost]
RUNNING HANDLER [restart someapp] **********************************************
ok: [localhost] => {
"msg": "some bogus module call in handler someapp"
}
PLAY RECAP *********************************************************************
localhost : ok=3 changed=1 unreachable=0 failed=0
</code></pre></div>
<h5 dir="auto">ACTUAL RESULTS</h5>
<p dir="auto">Results for Ansible 2.1.2.0, 2.1.1.0, and branches devel, stable-2.2, stable-2.1, and stable-2.0.0.1:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content=" [WARNING]: provided hosts list is empty, only localhost is available
PLAY [localhost] ***************************************************************
TASK [setup] *******************************************************************
ok: [localhost]
TASK [bogus task with notify for someapp] **************************************
changed: [localhost]
RUNNING HANDLER [restart {{ appname }}] ****************************************
ok: [localhost] => {
"msg": "some bogus module call in handler someapp"
}
PLAY RECAP *********************************************************************
localhost : ok=3 changed=1 unreachable=0 failed=0
"><pre class="notranslate"><code class="notranslate"> [WARNING]: provided hosts list is empty, only localhost is available
PLAY [localhost] ***************************************************************
TASK [setup] *******************************************************************
ok: [localhost]
TASK [bogus task with notify for someapp] **************************************
changed: [localhost]
RUNNING HANDLER [restart {{ appname }}] ****************************************
ok: [localhost] => {
"msg": "some bogus module call in handler someapp"
}
PLAY RECAP *********************************************************************
localhost : ok=3 changed=1 unreachable=0 failed=0
</code></pre></div> | <h5 dir="auto">ISSUE TYPE</h5>
<ul dir="auto">
<li>Bug Report</li>
</ul>
<h5 dir="auto">ANSIBLE VERSION</h5>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="ansible 2.0.0.2
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides"><pre class="notranslate"><code class="notranslate">ansible 2.0.0.2
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
</code></pre></div>
<h5 dir="auto">CONFIGURATION</h5>
<h5 dir="auto">OS / ENVIRONMENT</h5>
<p dir="auto">N/A</p>
<h5 dir="auto">COMPONENT NAME</h5>
<p dir="auto">playbook</p>
<h5 dir="auto">SUMMARY</h5>
<p dir="auto">Variables are not expanded in handler names.</p>
<p dir="auto">Starting from some point variables are being expanded in task names: <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="15060561" data-permission-text="Title is private" data-url="https://github.com/ansible/ansible/issues/3103" data-hovercard-type="issue" data-hovercard-url="/ansible/ansible/issues/3103/hovercard" href="https://github.com/ansible/ansible/issues/3103">#3103</a>, <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="95319669" data-permission-text="Title is private" data-url="https://github.com/ansible/ansible/issues/11605" data-hovercard-type="issue" data-hovercard-url="/ansible/ansible/issues/11605/hovercard" href="https://github.com/ansible/ansible/issues/11605">#11605</a>, I expect that it should work in handler names too.</p>
<p dir="auto">Tricky part may be here: handler name is handler identifier.</p>
<h5 dir="auto">STEPS TO REPRODUCE</h5>
<p dir="auto">Run following playbook:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="
---
- hosts: localhost
vars:
service_instance: "worker3"
tasks:
- name: configure {{ service_instance }}
debug: msg="Writing {{ service_instance }} configuration..."
notify:
- restart {{ service_instance }}
changed_when: True
handlers:
- name: restart {{ service_instance }}
debug: msg="Restarting {{ service_instance }}...""><pre class="notranslate"><code class="notranslate">
---
- hosts: localhost
vars:
service_instance: "worker3"
tasks:
- name: configure {{ service_instance }}
debug: msg="Writing {{ service_instance }} configuration..."
notify:
- restart {{ service_instance }}
changed_when: True
handlers:
- name: restart {{ service_instance }}
debug: msg="Restarting {{ service_instance }}..."
</code></pre></div>
<h5 dir="auto">EXPECTED RESULTS</h5>
<p dir="auto">Handler named with variable in it should have it expanded:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="RUNNING HANDLER [restart worker3] *******************************"><pre class="notranslate"><code class="notranslate">RUNNING HANDLER [restart worker3] *******************************
</code></pre></div>
<h5 dir="auto">ACTUAL RESULTS</h5>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="$ ansible-playbook playbook.yaml
[WARNING]: provided hosts list is empty, only localhost is available
PLAY ***************************************************************************
TASK [setup] *******************************************************************
ok: [localhost]
TASK [configure worker3] *******************************************************
ok: [localhost] => {
"msg": "Writing worker3 configuration..."
}
RUNNING HANDLER [restart {{ service_instance }}] *******************************
ok: [localhost] => {
"msg": "Restarting worker3..."
}
PLAY RECAP *********************************************************************
localhost : ok=3 changed=1 unreachable=0 failed=0 "><pre class="notranslate"><code class="notranslate">$ ansible-playbook playbook.yaml
[WARNING]: provided hosts list is empty, only localhost is available
PLAY ***************************************************************************
TASK [setup] *******************************************************************
ok: [localhost]
TASK [configure worker3] *******************************************************
ok: [localhost] => {
"msg": "Writing worker3 configuration..."
}
RUNNING HANDLER [restart {{ service_instance }}] *******************************
ok: [localhost] => {
"msg": "Restarting worker3..."
}
PLAY RECAP *********************************************************************
localhost : ok=3 changed=1 unreachable=0 failed=0
</code></pre></div> | 1 |
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I have searched the <a href="https://github.com/mui-org/material-ui/issues">issues</a> of this repository and believe that this is not a duplicate.</li>
</ul>
<h2 dir="auto">Expected Behaviour</h2>
<p dir="auto">Should not throw an error</p>
<h2 dir="auto">Current Behavior</h2>
<p dir="auto">Throws the following error:</p>
<p dir="auto">Menu.js:107 Uncaught TypeError: Cannot read property 'focus' of null<br>
at Menu._this.focus (Menu.js:107)<br>
at Object.Menu._this.handleEnter [as onEnter] (Menu.js:122)<br>
at Object.Popover._this.handleEnter (Popover.js:226)<br>
at Object.chainedFunction [as onEnter] (helpers.js:85)<br>
at Object.Grow._this.handleEnter [as onEnter] (Grow.js:86)<br>
at Object.CSSTransition._this.onEnter (CSSTransition.js:217)<br>
at Transition.performEnter (Transition.js:242)<br>
at Transition.updateStatus (Transition.js:215)<br>
at Transition.componentDidMount (Transition.js:157)<br>
at commitLifeCycles (react-dom.development.js:8770)</p>
<h2 dir="auto">Steps to Reproduce (for bugs)</h2>
<p dir="auto">Place a Select control within a Dialog and click on the control.</p>
<table role="table">
<thead>
<tr>
<th>Tech</th>
<th>Version</th>
</tr>
</thead>
<tbody>
<tr>
<td>Material-U I</td>
<td>v1.0.0-beta.24</td>
</tr>
<tr>
<td>React</td>
<td>v16.2.0</td>
</tr>
<tr>
<td>Cromium</td>
<td>Version 63.0.3239.84 (Official Build) Built on Ubuntu , running on Ubuntu 17.04 (64-bit)</td>
</tr>
</tbody>
</table> | <h2 dir="auto">Description</h2>
<p dir="auto">When you click on item «Phone ringtone» in list - you see broken layout of confirmation dialog.<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/25990573/33186688-2222177e-d094-11e7-88d5-f543a937334c.png"><img src="https://user-images.githubusercontent.com/25990573/33186688-2222177e-d094-11e7-88d5-f543a937334c.png" alt="selection_002" style="max-width: 100%;"></a></p>
<h2 dir="auto">Steps to Reproduce</h2>
<p dir="auto"><a href="https://material-ui-1dab0.firebaseapp.com/demos/dialogs/#confirmation-dialogs" rel="nofollow">https://material-ui-1dab0.firebaseapp.com/demos/dialogs/#confirmation-dialogs</a></p>
<h2 dir="auto">Environment</h2>
<table role="table">
<thead>
<tr>
<th>Tech</th>
<th>Version</th>
</tr>
</thead>
<tbody>
<tr>
<td>Material-UI</td>
<td>v1.0.0-beta.21</td>
</tr>
<tr>
<td>Browsers</td>
<td>Google Chrome 61.0.3163.100 (Official Build) (64-bit), Mozilla Firefox 57.0 (64-bit)</td>
</tr>
</tbody>
</table> | 0 |
<p dir="auto">This,</p>
<div class="highlight highlight-source-rust notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="/// A binary tree with nodes labelled with `T`
pub struct Tree<T> {
pub value: T,
pub left: Option<Box<Tree<T>>>,
pub right: Option<Box<Tree<T>>>,
}
impl<T> Tree<T> {
pub fn map_step<F, V>(self, f: &mut FnMut(T) -> V) -> Tree<V> {
Tree {
left: self.left.map(|x| { let a: Tree<V> = x.map_step(f);
Box::new(a) }),
right: self.right.map(|x| Box::new(x.map_step(f))),
value: f(self.value),
}
}
}
fn main() {}"><pre class="notranslate"><span class="pl-c">/// A binary tree with nodes labelled with `T`</span>
<span class="pl-k">pub</span> <span class="pl-k">struct</span> <span class="pl-smi">Tree</span><span class="pl-kos"><</span><span class="pl-smi">T</span><span class="pl-kos">></span> <span class="pl-kos">{</span>
<span class="pl-k">pub</span> <span class="pl-c1">value</span><span class="pl-kos">:</span> <span class="pl-smi">T</span><span class="pl-kos">,</span>
<span class="pl-k">pub</span> <span class="pl-c1">left</span><span class="pl-kos">:</span> <span class="pl-smi">Option</span><span class="pl-kos"><</span><span class="pl-smi">Box</span><span class="pl-kos"><</span><span class="pl-smi">Tree</span><span class="pl-kos"><</span><span class="pl-smi">T</span><span class="pl-kos">></span><span class="pl-kos">></span><span class="pl-kos">></span><span class="pl-kos">,</span>
<span class="pl-k">pub</span> <span class="pl-c1">right</span><span class="pl-kos">:</span> <span class="pl-smi">Option</span><span class="pl-kos"><</span><span class="pl-smi">Box</span><span class="pl-kos"><</span><span class="pl-smi">Tree</span><span class="pl-kos"><</span><span class="pl-smi">T</span><span class="pl-kos">></span><span class="pl-kos">></span><span class="pl-kos">></span><span class="pl-kos">,</span>
<span class="pl-kos">}</span>
<span class="pl-k">impl</span><span class="pl-kos"><</span><span class="pl-smi">T</span><span class="pl-kos">></span> <span class="pl-smi">Tree</span><span class="pl-kos"><</span><span class="pl-smi">T</span><span class="pl-kos">></span> <span class="pl-kos">{</span>
<span class="pl-k">pub</span> <span class="pl-k">fn</span> <span class="pl-en">map_step</span><span class="pl-kos"><</span><span class="pl-smi">F</span><span class="pl-kos">,</span> <span class="pl-smi">V</span><span class="pl-kos">></span><span class="pl-kos">(</span><span class="pl-smi">self</span><span class="pl-kos">,</span> <span class="pl-s1">f</span><span class="pl-kos">:</span> <span class="pl-c1">&</span><span class="pl-k">mut</span> <span class="pl-smi">FnMut</span><span class="pl-kos">(</span><span class="pl-smi">T</span><span class="pl-kos">)</span> -> <span class="pl-smi">V</span><span class="pl-kos">)</span> -> <span class="pl-smi">Tree</span><span class="pl-kos"><</span><span class="pl-smi">V</span><span class="pl-kos">></span> <span class="pl-kos">{</span>
<span class="pl-smi">Tree</span> <span class="pl-kos">{</span>
<span class="pl-c1">left</span><span class="pl-kos">:</span> <span class="pl-smi">self</span><span class="pl-kos">.</span><span class="pl-c1">left</span><span class="pl-kos">.</span><span class="pl-en">map</span><span class="pl-kos">(</span>|x| <span class="pl-kos">{</span> <span class="pl-k">let</span> a<span class="pl-kos">:</span> <span class="pl-smi">Tree</span><span class="pl-kos"><</span><span class="pl-smi">V</span><span class="pl-kos">></span> = x<span class="pl-kos">.</span><span class="pl-en">map_step</span><span class="pl-kos">(</span>f<span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-smi">Box</span><span class="pl-kos">::</span><span class="pl-en">new</span><span class="pl-kos">(</span>a<span class="pl-kos">)</span> <span class="pl-kos">}</span><span class="pl-kos">)</span><span class="pl-kos">,</span>
<span class="pl-c1">right</span><span class="pl-kos">:</span> <span class="pl-smi">self</span><span class="pl-kos">.</span><span class="pl-c1">right</span><span class="pl-kos">.</span><span class="pl-en">map</span><span class="pl-kos">(</span>|x| <span class="pl-smi">Box</span><span class="pl-kos">::</span><span class="pl-en">new</span><span class="pl-kos">(</span>x<span class="pl-kos">.</span><span class="pl-en">map_step</span><span class="pl-kos">(</span>f<span class="pl-kos">)</span><span class="pl-kos">)</span><span class="pl-kos">)</span><span class="pl-kos">,</span>
<span class="pl-c1">value</span><span class="pl-kos">:</span> <span class="pl-en">f</span><span class="pl-kos">(</span><span class="pl-smi">self</span><span class="pl-kos">.</span><span class="pl-c1">value</span><span class="pl-kos">)</span><span class="pl-kos">,</span>
<span class="pl-kos">}</span>
<span class="pl-kos">}</span>
<span class="pl-kos">}</span>
<span class="pl-k">fn</span> <span class="pl-en">main</span><span class="pl-kos">(</span><span class="pl-kos">)</span> <span class="pl-kos">{</span><span class="pl-kos">}</span></pre></div>
<p dir="auto">fails with,</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="hi.rs:12:58: 12:69 error: unable to infer enough type information about `_`; type annotations required [E0282]
hi.rs:12 left: self.left.map(|x| { let a: Tree<V> = x.map_step(f);
^~~~~~~~~~~"><pre class="notranslate"><code class="notranslate">hi.rs:12:58: 12:69 error: unable to infer enough type information about `_`; type annotations required [E0282]
hi.rs:12 left: self.left.map(|x| { let a: Tree<V> = x.map_step(f);
^~~~~~~~~~~
</code></pre></div>
<p dir="auto">All of my attempts at helping the compiler find the right types have failed, therefore I suspect it's a bug.</p> | <p dir="auto">This code producesses an unclear error</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="pub trait ValueGiver<T> {
fn give(_: Option<Self>) -> T;
}
struct m;
impl<f64> ValueGiver<f64> for m {
fn give(_:Option<m>) -> f64 {
0.5f64
}
}
fn main() {
let y : f64 = ValueGiver::give(None::<m>);
}"><pre class="notranslate"><code class="notranslate">pub trait ValueGiver<T> {
fn give(_: Option<Self>) -> T;
}
struct m;
impl<f64> ValueGiver<f64> for m {
fn give(_:Option<m>) -> f64 {
0.5f64
}
}
fn main() {
let y : f64 = ValueGiver::give(None::<m>);
}
</code></pre></div>
<p dir="auto">Here the error of course is the <code class="notranslate">impl<f64></code> that should be <code class="notranslate">impl</code>, but the error message is not clear :</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="c.rs:64:5: 68:6 error: unable to infer enough type information about `_`; type annotations required
c.rs:64 impl<f64> ValueGiver<f64> for m {
c.rs:65 fn give(_:Option<m>) -> f64 {
c.rs:66 0.5f64
c.rs:67 }
c.rs:68 }"><pre class="notranslate"><code class="notranslate">c.rs:64:5: 68:6 error: unable to infer enough type information about `_`; type annotations required
c.rs:64 impl<f64> ValueGiver<f64> for m {
c.rs:65 fn give(_:Option<m>) -> f64 {
c.rs:66 0.5f64
c.rs:67 }
c.rs:68 }
</code></pre></div>
<p dir="auto">Could this error message be more helpful ?</p> | 1 |
<p dir="auto">We have a lot of modules which a user might want to import, and it would be nice to show a list of these as a completion hint after <code class="notranslate">import {} from '</code><br>
Additionally, we have so many modules that walking the filesystem is not practical, and some of them are generated external module names, not files. So we would like to run a periodic indexing, and then somehow provide a list of these to the language services (eg. an RPC, or a file we stage on local disk, etc.)</p>
<p dir="auto">cc <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/mprobst/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/mprobst">@mprobst</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/rkirov/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/rkirov">@rkirov</a></p> | <p dir="auto">Hi,</p>
<p dir="auto">Visual Studio (2013 Ultimate) provides intellisense for the src attribute for script elements, by reading the file system and displaying available files or folders.</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://camo.githubusercontent.com/d806f3a8fc96af8601bdd892a447e3da24eb8c05ac2760edbfd31ee39bf1419e/687474703a2f2f7777772e696d61676575726c686f73742e636f6d2f696d616765732f69676934657964677130347037346c6c707537652e706e67"><img src="https://camo.githubusercontent.com/d806f3a8fc96af8601bdd892a447e3da24eb8c05ac2760edbfd31ee39bf1419e/687474703a2f2f7777772e696d61676575726c686f73742e636f6d2f696d616765732f69676934657964677130347037346c6c707537652e706e67" alt="Image" data-canonical-src="http://www.imageurlhost.com/images/igi4eydgq04p74llpu7e.png" style="max-width: 100%;"></a></p>
<p dir="auto">It will be quite useful if similar functionality can be provided for <code class="notranslate"><reference></code> and import statements:</p>
<div class="highlight highlight-source-ts notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="
<reference path="foo/ <--- here
import foo = require('foo/ <--- and here
"><pre class="notranslate"> <span class="pl-c1"><</span><span class="pl-s1">reference</span> <span class="pl-s1">path</span><span class="pl-c1">=</span>"foo/ <--- here
import foo = require('foo/ <--- and here</pre></div> | 1 |
<h2 dir="auto"><g-emoji class="g-emoji" alias="bug" fallback-src="https://github.githubassets.com/images/icons/emoji/unicode/1f41b.png">🐛</g-emoji> Bug</h2>
<p dir="auto">distributed/_pipeline/sync/skip/ tests have different arg parsers, introduced in <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="691591664" data-permission-text="Title is private" data-url="https://github.com/pytorch/pytorch/issues/44090" data-hovercard-type="pull_request" data-hovercard-url="/pytorch/pytorch/pull/44090/hovercard" href="https://github.com/pytorch/pytorch/pull/44090">#44090</a></p>
<h2 dir="auto">To Reproduce</h2>
<p dir="auto">Steps to reproduce the behavior:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="$ python test/run_test.py -i distributed/_pipeline/sync/skip/test_api -- -f
Running distributed/_pipeline/sync/skip/test_api ... [2020-10-23 11:29:49.748332]
Executing ['/usr/bin/python', '-m', 'pytest', 'distributed/_pipeline/sync/skip/test_api.py', '-f'] ... [2020-10-23 11:29:49.748382]
ERROR: usage: __main__.py [options] [file_or_dir] [file_or_dir] [...]
__main__.py: error: unrecognized arguments: -f
inifile: None
rootdir: /home/....../pytorch
Traceback (most recent call last):
File "test/run_test.py", line 862, in <module>
main()
File "test/run_test.py", line 845, in main
raise RuntimeError(err_message)
RuntimeError: distributed/_pipeline/sync/skip/test_api failed!"><pre class="notranslate"><code class="notranslate">$ python test/run_test.py -i distributed/_pipeline/sync/skip/test_api -- -f
Running distributed/_pipeline/sync/skip/test_api ... [2020-10-23 11:29:49.748332]
Executing ['/usr/bin/python', '-m', 'pytest', 'distributed/_pipeline/sync/skip/test_api.py', '-f'] ... [2020-10-23 11:29:49.748382]
ERROR: usage: __main__.py [options] [file_or_dir] [file_or_dir] [...]
__main__.py: error: unrecognized arguments: -f
inifile: None
rootdir: /home/....../pytorch
Traceback (most recent call last):
File "test/run_test.py", line 862, in <module>
main()
File "test/run_test.py", line 845, in main
raise RuntimeError(err_message)
RuntimeError: distributed/_pipeline/sync/skip/test_api failed!
</code></pre></div>
<p dir="auto"><code class="notranslate">-f</code> is a flag for fail fast which works for all other test modules but not tests introduced in this PR. There should be some consistency for all tests.</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="$ python test/run_test.py -i test_nn -- --help
Running test_nn ... [2020-10-23 11:27:00.250544]
Executing ['/usr/bin/python', 'test_nn.py', '--help'] ... [2020-10-23 11:27:00.250593]
usage: test_nn.py [-h] [-v] [-q] [--locals] [-f] [-c] [-b] [-k TESTNAMEPATTERNS] [tests [tests ...]]
positional arguments:
tests a list of any number of test modules, classes and test methods.
optional arguments:
-h, --help show this help message and exit
-v, --verbose Verbose output
-q, --quiet Quiet output
--locals Show local variables in tracebacks
-f, --failfast Stop on first fail or error
-c, --catch Catch Ctrl-C and display results so far
-b, --buffer Buffer stdout and stderr during tests
-k TESTNAMEPATTERNS Only run tests which match the given substring
Examples:
test_nn.py - run default set of tests
test_nn.py MyTestSuite - run suite 'MyTestSuite'
test_nn.py MyTestCase.testSomething - run MyTestCase.testSomething
test_nn.py MyTestCase - run all 'test*' test methods
in MyTestCase"><pre class="notranslate"><code class="notranslate">$ python test/run_test.py -i test_nn -- --help
Running test_nn ... [2020-10-23 11:27:00.250544]
Executing ['/usr/bin/python', 'test_nn.py', '--help'] ... [2020-10-23 11:27:00.250593]
usage: test_nn.py [-h] [-v] [-q] [--locals] [-f] [-c] [-b] [-k TESTNAMEPATTERNS] [tests [tests ...]]
positional arguments:
tests a list of any number of test modules, classes and test methods.
optional arguments:
-h, --help show this help message and exit
-v, --verbose Verbose output
-q, --quiet Quiet output
--locals Show local variables in tracebacks
-f, --failfast Stop on first fail or error
-c, --catch Catch Ctrl-C and display results so far
-b, --buffer Buffer stdout and stderr during tests
-k TESTNAMEPATTERNS Only run tests which match the given substring
Examples:
test_nn.py - run default set of tests
test_nn.py MyTestSuite - run suite 'MyTestSuite'
test_nn.py MyTestCase.testSomething - run MyTestCase.testSomething
test_nn.py MyTestCase - run all 'test*' test methods
in MyTestCase
</code></pre></div>
<p dir="auto">but</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="$ python test/run_test.py -i distributed/_pipeline/sync/skip/test_api -- --help
Running distributed/_pipeline/sync/skip/test_api ... [2020-10-23 11:25:01.118070]
Executing ['/usr/bin/python', '-m', 'pytest', 'distributed/_pipeline/sync/skip/test_api.py', '--help'] ... [2020-10-23 11:25:01.118119]
usage: __main__.py [options] [file_or_dir] [file_or_dir] [...]
positional arguments:
file_or_dir
general:
-k EXPRESSION only run tests which match the given substring expression. An expression is a python evaluatable expression where all names are substring-matched against test names and their parent classes.
Example: -k 'test_method or test_other' matches all test functions and classes whose name contains 'test_method' or 'test_other', while -k 'not test_method' matches those that don't contain
'test_method' in their names. -k 'not test_method and not test_other' will eliminate the matches. Additionally keywords are matched to classes and functions containing extra names in their
'extra_keyword_matches' set, as well as functions which have names assigned directly to them. The matching is case-insensitive.
-m MARKEXPR only run tests matching given mark expression.
For example: -m 'mark1 and not mark2'.
--markers show markers (builtin, plugin and per-project ones).
-x, --exitfirst exit instantly on first error or failed test.
--fixtures, --funcargs
show available fixtures, sorted by plugin appearance (fixtures with leading '_' are only shown with '-v')
--fixtures-per-test show fixtures per test
--pdb start the interactive Python debugger on errors or KeyboardInterrupt.
--pdbcls=modulename:classname
start a custom interactive Python debugger on errors. For example: --pdbcls=IPython.terminal.debugger:TerminalPdb
--trace Immediately break when running each test.
--capture=method per-test capturing method: one of fd|sys|no|tee-sys.
-s shortcut for --capture=no.
--runxfail report the results of xfail tests as if they were not marked
--lf, --last-failed rerun only the tests that failed at the last run (or all if none failed)
--ff, --failed-first run all tests, but run the last failures first.
This may re-order tests and thus lead to repeated fixture setup/teardown.
--nf, --new-first run tests from new files first, then the rest of the tests sorted by file mtime
--cache-show=[CACHESHOW]
show cache contents, don't perform collection or tests. Optional argument: glob (default: '*').
--cache-clear remove all cache contents at start of test run.
--lfnf={all,none}, --last-failed-no-failures={all,none}
which tests to run with no previously (known) failures.
--sw, --stepwise exit on test failure and continue from last failing test next time
--stepwise-skip ignore the first failing test but stop on the next failing test
reporting:
--durations=N show N slowest setup/test durations (N=0 for all).
--durations-min=N Minimal duration in seconds for inclusion in slowest list. Default 0.005
-v, --verbose increase verbosity.
--no-header disable header
--no-summary disable summary
-q, --quiet decrease verbosity.
--verbosity=VERBOSE set verbosity. Default is 0.
-r chars show extra test summary info as specified by chars: (f)ailed, (E)rror, (s)kipped, (x)failed, (X)passed, (p)assed, (P)assed with output, (a)ll except passed (p/P), or (A)ll. (w)arnings are enabled
by default (see --disable-warnings), 'N' can be used to reset the list. (default: 'fE').
--disable-warnings, --disable-pytest-warnings
disable warnings summary
-l, --showlocals show locals in tracebacks (disabled by default).
--tb=style traceback print mode (auto/long/short/line/native/no).
--show-capture={no,stdout,stderr,log,all}
Controls how captured stdout/stderr/log is shown on failed tests. Default is 'all'.
--full-trace don't cut any tracebacks (default is to cut).
--color=color color terminal output (yes/no/auto).
--code-highlight={yes,no}
Whether code should be highlighted (only if --color is also enabled)
--pastebin=mode send failed|all info to bpaste.net pastebin service.
--junit-xml=path create junit-xml style report file at given path.
--junit-prefix=str prepend prefix to classnames in junit-xml output
pytest-warnings:
-W PYTHONWARNINGS, --pythonwarnings=PYTHONWARNINGS
set which warnings to report, see -W option of python itself.
--maxfail=num exit after first num failures or errors.
--strict-config any warnings encountered while parsing the `pytest` section of the configuration file raise errors.
--strict-markers, --strict
markers not registered in the `markers` section of the configuration file raise errors.
-c file load configuration from `file` instead of trying to locate one of the implicit configuration files.
--continue-on-collection-errors
Force test execution even if collection errors occur.
--rootdir=ROOTDIR Define root directory for tests. Can be relative path: 'root_dir', './root_dir', 'root_dir/another_dir/'; absolute path: '/home/user/root_dir'; path with variables: '$HOME/root_dir'.
collection:
--collect-only, --co only collect tests, don't execute them.
--pyargs try to interpret all arguments as python packages.
--ignore=path ignore path during collection (multi-allowed).
--ignore-glob=path ignore path pattern during collection (multi-allowed).
--deselect=nodeid_prefix
deselect item (via node id prefix) during collection (multi-allowed).
--confcutdir=dir only load conftest.py's relative to specified dir.
--noconftest Don't load any conftest.py files.
--keep-duplicates Keep duplicate tests.
--collect-in-virtualenv
Don't ignore tests in a local virtualenv directory
--import-mode={prepend,append,importlib}
prepend/append to sys.path when importing test modules and conftest files, default is to prepend.
--doctest-modules run doctests in all .py modules
--doctest-report={none,cdiff,ndiff,udiff,only_first_failure}
choose another output format for diffs on doctest failure
--doctest-glob=pat doctests file matching pattern, default: test*.txt
--doctest-ignore-import-errors
ignore doctest ImportErrors
--doctest-continue-on-failure
for a given doctest, continue to run after the first failure
test session debugging and configuration:
--basetemp=dir base temporary directory for this test run.(warning: this directory is removed if it exists)
-V, --version display pytest version and information about plugins.When given twice, also display information about plugins.
-h, --help show help message and configuration info
-p name early-load given plugin module name or entry point (multi-allowed).
To avoid loading of plugins, use the `no:` prefix, e.g. `no:doctest`.
--trace-config trace considerations of conftest.py files.
--debug store internal tracing debug information in 'pytestdebug.log'.
-o OVERRIDE_INI, --override-ini=OVERRIDE_INI
override ini option with "option=value" style, e.g. `-o xfail_strict=True -o cache_dir=cache`.
--assert=MODE Control assertion debugging tools.
'plain' performs no assertion debugging.
'rewrite' (the default) rewrites assert statements in test modules on import to provide assert expression information.
--setup-only only setup fixtures, do not execute tests.
--setup-show show setup of fixtures while executing tests.
--setup-plan show what fixtures and tests would be executed but don't execute anything.
logging:
--log-level=LEVEL level of messages to catch/display.
Not set by default, so it depends on the root/parent log handler's effective level, where it is "WARNING" by default.
--log-format=LOG_FORMAT
log format as used by the logging module.
--log-date-format=LOG_DATE_FORMAT
log date format as used by the logging module.
--log-cli-level=LOG_CLI_LEVEL
cli logging level.
--log-cli-format=LOG_CLI_FORMAT
log format as used by the logging module.
--log-cli-date-format=LOG_CLI_DATE_FORMAT
log date format as used by the logging module.
--log-file=LOG_FILE path to a file when logging will be written to.
--log-file-level=LOG_FILE_LEVEL
log file logging level.
--log-file-format=LOG_FILE_FORMAT
log format as used by the logging module.
--log-file-date-format=LOG_FILE_DATE_FORMAT
log date format as used by the logging module.
--log-auto-indent=LOG_AUTO_INDENT
Auto-indent multiline messages passed to the logging module. Accepts true|on, false|off or an integer.
Hypothesis:
--hypothesis-profile=HYPOTHESIS_PROFILE
Load in a registered hypothesis.settings profile
--hypothesis-verbosity={quiet,normal,verbose,debug}
Override profile with verbosity setting specified
--hypothesis-show-statistics
Configure when statistics are printed
--hypothesis-seed=HYPOTHESIS_SEED
Set a seed to use for all Hypothesis tests
[pytest] ini-options in the first pytest.ini|tox.ini|setup.cfg file found:
markers (linelist): markers for test functions
empty_parameter_set_mark (string):
default marker for empty parametersets
norecursedirs (args): directory patterns to avoid for recursion
testpaths (args): directories to search for tests when no files or directories are given in the command line.
filterwarnings (linelist):
Each line specifies a pattern for warnings.filterwarnings. Processed after -W/--pythonwarnings.
usefixtures (args): list of default fixtures to be used with this project
python_files (args): glob-style file patterns for Python test module discovery
python_classes (args):
prefixes or glob names for Python test class discovery
python_functions (args):
prefixes or glob names for Python test function and method discovery
disable_test_id_escaping_and_forfeit_all_rights_to_community_support (bool):
disable string escape non-ascii characters, might cause unwanted side effects(use at your own risk)
console_output_style (string):
console output: "classic", or with additional progress information ("progress" (percentage) | "count").
xfail_strict (bool): default for the strict parameter of xfail markers when not given explicitly (default: False)
enable_assertion_pass_hook (bool):
Enables the pytest_assertion_pass hook.Make sure to delete any previously generated pyc cache files.
junit_suite_name (string):
Test suite name for JUnit report
junit_logging (string):
Write captured log messages to JUnit report: one of no|log|system-out|system-err|out-err|all
junit_log_passing_tests (bool):
Capture log information for passing tests to JUnit report:
junit_duration_report (string):
Duration time to report: one of total|call
junit_family (string):
Emit XML for schema: one of legacy|xunit1|xunit2
doctest_optionflags (args):
option flags for doctests
doctest_encoding (string):
encoding used for doctest files
cache_dir (string): cache directory path.
log_level (string): default value for --log-level
log_format (string): default value for --log-format
log_date_format (string):
default value for --log-date-format
log_cli (bool): enable log display during test run (also known as "live logging").
log_cli_level (string):
default value for --log-cli-level
log_cli_format (string):
default value for --log-cli-format
log_cli_date_format (string):
default value for --log-cli-date-format
log_file (string): default value for --log-file
log_file_level (string):
default value for --log-file-level
log_file_format (string):
default value for --log-file-format
log_file_date_format (string):
default value for --log-file-date-format
log_auto_indent (string):
default value for --log-auto-indent
faulthandler_timeout (string):
Dump the traceback of all threads if a test takes more than TIMEOUT seconds to finish.
addopts (args): extra command line options
minversion (string): minimally required pytest version
required_plugins (args):
plugins that must be present for pytest to run
environment variables:
PYTEST_ADDOPTS extra command line options
PYTEST_PLUGINS comma-separated plugins to load during startup
PYTEST_DISABLE_PLUGIN_AUTOLOAD set to disable plugin auto-loading
PYTEST_DEBUG set to enable debug tracing of pytest's internals
to see available markers type: pytest --markers
to see available fixtures type: pytest --fixtures
(shown according to specified file_or_dir or current dir if not specified; fixtures with leading '_' are only shown with the '-v' option"><pre class="notranslate"><code class="notranslate">$ python test/run_test.py -i distributed/_pipeline/sync/skip/test_api -- --help
Running distributed/_pipeline/sync/skip/test_api ... [2020-10-23 11:25:01.118070]
Executing ['/usr/bin/python', '-m', 'pytest', 'distributed/_pipeline/sync/skip/test_api.py', '--help'] ... [2020-10-23 11:25:01.118119]
usage: __main__.py [options] [file_or_dir] [file_or_dir] [...]
positional arguments:
file_or_dir
general:
-k EXPRESSION only run tests which match the given substring expression. An expression is a python evaluatable expression where all names are substring-matched against test names and their parent classes.
Example: -k 'test_method or test_other' matches all test functions and classes whose name contains 'test_method' or 'test_other', while -k 'not test_method' matches those that don't contain
'test_method' in their names. -k 'not test_method and not test_other' will eliminate the matches. Additionally keywords are matched to classes and functions containing extra names in their
'extra_keyword_matches' set, as well as functions which have names assigned directly to them. The matching is case-insensitive.
-m MARKEXPR only run tests matching given mark expression.
For example: -m 'mark1 and not mark2'.
--markers show markers (builtin, plugin and per-project ones).
-x, --exitfirst exit instantly on first error or failed test.
--fixtures, --funcargs
show available fixtures, sorted by plugin appearance (fixtures with leading '_' are only shown with '-v')
--fixtures-per-test show fixtures per test
--pdb start the interactive Python debugger on errors or KeyboardInterrupt.
--pdbcls=modulename:classname
start a custom interactive Python debugger on errors. For example: --pdbcls=IPython.terminal.debugger:TerminalPdb
--trace Immediately break when running each test.
--capture=method per-test capturing method: one of fd|sys|no|tee-sys.
-s shortcut for --capture=no.
--runxfail report the results of xfail tests as if they were not marked
--lf, --last-failed rerun only the tests that failed at the last run (or all if none failed)
--ff, --failed-first run all tests, but run the last failures first.
This may re-order tests and thus lead to repeated fixture setup/teardown.
--nf, --new-first run tests from new files first, then the rest of the tests sorted by file mtime
--cache-show=[CACHESHOW]
show cache contents, don't perform collection or tests. Optional argument: glob (default: '*').
--cache-clear remove all cache contents at start of test run.
--lfnf={all,none}, --last-failed-no-failures={all,none}
which tests to run with no previously (known) failures.
--sw, --stepwise exit on test failure and continue from last failing test next time
--stepwise-skip ignore the first failing test but stop on the next failing test
reporting:
--durations=N show N slowest setup/test durations (N=0 for all).
--durations-min=N Minimal duration in seconds for inclusion in slowest list. Default 0.005
-v, --verbose increase verbosity.
--no-header disable header
--no-summary disable summary
-q, --quiet decrease verbosity.
--verbosity=VERBOSE set verbosity. Default is 0.
-r chars show extra test summary info as specified by chars: (f)ailed, (E)rror, (s)kipped, (x)failed, (X)passed, (p)assed, (P)assed with output, (a)ll except passed (p/P), or (A)ll. (w)arnings are enabled
by default (see --disable-warnings), 'N' can be used to reset the list. (default: 'fE').
--disable-warnings, --disable-pytest-warnings
disable warnings summary
-l, --showlocals show locals in tracebacks (disabled by default).
--tb=style traceback print mode (auto/long/short/line/native/no).
--show-capture={no,stdout,stderr,log,all}
Controls how captured stdout/stderr/log is shown on failed tests. Default is 'all'.
--full-trace don't cut any tracebacks (default is to cut).
--color=color color terminal output (yes/no/auto).
--code-highlight={yes,no}
Whether code should be highlighted (only if --color is also enabled)
--pastebin=mode send failed|all info to bpaste.net pastebin service.
--junit-xml=path create junit-xml style report file at given path.
--junit-prefix=str prepend prefix to classnames in junit-xml output
pytest-warnings:
-W PYTHONWARNINGS, --pythonwarnings=PYTHONWARNINGS
set which warnings to report, see -W option of python itself.
--maxfail=num exit after first num failures or errors.
--strict-config any warnings encountered while parsing the `pytest` section of the configuration file raise errors.
--strict-markers, --strict
markers not registered in the `markers` section of the configuration file raise errors.
-c file load configuration from `file` instead of trying to locate one of the implicit configuration files.
--continue-on-collection-errors
Force test execution even if collection errors occur.
--rootdir=ROOTDIR Define root directory for tests. Can be relative path: 'root_dir', './root_dir', 'root_dir/another_dir/'; absolute path: '/home/user/root_dir'; path with variables: '$HOME/root_dir'.
collection:
--collect-only, --co only collect tests, don't execute them.
--pyargs try to interpret all arguments as python packages.
--ignore=path ignore path during collection (multi-allowed).
--ignore-glob=path ignore path pattern during collection (multi-allowed).
--deselect=nodeid_prefix
deselect item (via node id prefix) during collection (multi-allowed).
--confcutdir=dir only load conftest.py's relative to specified dir.
--noconftest Don't load any conftest.py files.
--keep-duplicates Keep duplicate tests.
--collect-in-virtualenv
Don't ignore tests in a local virtualenv directory
--import-mode={prepend,append,importlib}
prepend/append to sys.path when importing test modules and conftest files, default is to prepend.
--doctest-modules run doctests in all .py modules
--doctest-report={none,cdiff,ndiff,udiff,only_first_failure}
choose another output format for diffs on doctest failure
--doctest-glob=pat doctests file matching pattern, default: test*.txt
--doctest-ignore-import-errors
ignore doctest ImportErrors
--doctest-continue-on-failure
for a given doctest, continue to run after the first failure
test session debugging and configuration:
--basetemp=dir base temporary directory for this test run.(warning: this directory is removed if it exists)
-V, --version display pytest version and information about plugins.When given twice, also display information about plugins.
-h, --help show help message and configuration info
-p name early-load given plugin module name or entry point (multi-allowed).
To avoid loading of plugins, use the `no:` prefix, e.g. `no:doctest`.
--trace-config trace considerations of conftest.py files.
--debug store internal tracing debug information in 'pytestdebug.log'.
-o OVERRIDE_INI, --override-ini=OVERRIDE_INI
override ini option with "option=value" style, e.g. `-o xfail_strict=True -o cache_dir=cache`.
--assert=MODE Control assertion debugging tools.
'plain' performs no assertion debugging.
'rewrite' (the default) rewrites assert statements in test modules on import to provide assert expression information.
--setup-only only setup fixtures, do not execute tests.
--setup-show show setup of fixtures while executing tests.
--setup-plan show what fixtures and tests would be executed but don't execute anything.
logging:
--log-level=LEVEL level of messages to catch/display.
Not set by default, so it depends on the root/parent log handler's effective level, where it is "WARNING" by default.
--log-format=LOG_FORMAT
log format as used by the logging module.
--log-date-format=LOG_DATE_FORMAT
log date format as used by the logging module.
--log-cli-level=LOG_CLI_LEVEL
cli logging level.
--log-cli-format=LOG_CLI_FORMAT
log format as used by the logging module.
--log-cli-date-format=LOG_CLI_DATE_FORMAT
log date format as used by the logging module.
--log-file=LOG_FILE path to a file when logging will be written to.
--log-file-level=LOG_FILE_LEVEL
log file logging level.
--log-file-format=LOG_FILE_FORMAT
log format as used by the logging module.
--log-file-date-format=LOG_FILE_DATE_FORMAT
log date format as used by the logging module.
--log-auto-indent=LOG_AUTO_INDENT
Auto-indent multiline messages passed to the logging module. Accepts true|on, false|off or an integer.
Hypothesis:
--hypothesis-profile=HYPOTHESIS_PROFILE
Load in a registered hypothesis.settings profile
--hypothesis-verbosity={quiet,normal,verbose,debug}
Override profile with verbosity setting specified
--hypothesis-show-statistics
Configure when statistics are printed
--hypothesis-seed=HYPOTHESIS_SEED
Set a seed to use for all Hypothesis tests
[pytest] ini-options in the first pytest.ini|tox.ini|setup.cfg file found:
markers (linelist): markers for test functions
empty_parameter_set_mark (string):
default marker for empty parametersets
norecursedirs (args): directory patterns to avoid for recursion
testpaths (args): directories to search for tests when no files or directories are given in the command line.
filterwarnings (linelist):
Each line specifies a pattern for warnings.filterwarnings. Processed after -W/--pythonwarnings.
usefixtures (args): list of default fixtures to be used with this project
python_files (args): glob-style file patterns for Python test module discovery
python_classes (args):
prefixes or glob names for Python test class discovery
python_functions (args):
prefixes or glob names for Python test function and method discovery
disable_test_id_escaping_and_forfeit_all_rights_to_community_support (bool):
disable string escape non-ascii characters, might cause unwanted side effects(use at your own risk)
console_output_style (string):
console output: "classic", or with additional progress information ("progress" (percentage) | "count").
xfail_strict (bool): default for the strict parameter of xfail markers when not given explicitly (default: False)
enable_assertion_pass_hook (bool):
Enables the pytest_assertion_pass hook.Make sure to delete any previously generated pyc cache files.
junit_suite_name (string):
Test suite name for JUnit report
junit_logging (string):
Write captured log messages to JUnit report: one of no|log|system-out|system-err|out-err|all
junit_log_passing_tests (bool):
Capture log information for passing tests to JUnit report:
junit_duration_report (string):
Duration time to report: one of total|call
junit_family (string):
Emit XML for schema: one of legacy|xunit1|xunit2
doctest_optionflags (args):
option flags for doctests
doctest_encoding (string):
encoding used for doctest files
cache_dir (string): cache directory path.
log_level (string): default value for --log-level
log_format (string): default value for --log-format
log_date_format (string):
default value for --log-date-format
log_cli (bool): enable log display during test run (also known as "live logging").
log_cli_level (string):
default value for --log-cli-level
log_cli_format (string):
default value for --log-cli-format
log_cli_date_format (string):
default value for --log-cli-date-format
log_file (string): default value for --log-file
log_file_level (string):
default value for --log-file-level
log_file_format (string):
default value for --log-file-format
log_file_date_format (string):
default value for --log-file-date-format
log_auto_indent (string):
default value for --log-auto-indent
faulthandler_timeout (string):
Dump the traceback of all threads if a test takes more than TIMEOUT seconds to finish.
addopts (args): extra command line options
minversion (string): minimally required pytest version
required_plugins (args):
plugins that must be present for pytest to run
environment variables:
PYTEST_ADDOPTS extra command line options
PYTEST_PLUGINS comma-separated plugins to load during startup
PYTEST_DISABLE_PLUGIN_AUTOLOAD set to disable plugin auto-loading
PYTEST_DEBUG set to enable debug tracing of pytest's internals
to see available markers type: pytest --markers
to see available fixtures type: pytest --fixtures
(shown according to specified file_or_dir or current dir if not specified; fixtures with leading '_' are only shown with the '-v' option
</code></pre></div>
<h2 dir="auto">Expected behavior</h2>
<p dir="auto">No fail</p>
<h2 dir="auto">Environment</h2>
<p dir="auto">N/A</p>
<h2 dir="auto">Additional context</h2>
<p dir="auto">N/A</p>
<p dir="auto">cc <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/pietern/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/pietern">@pietern</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/mrshenli/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/mrshenli">@mrshenli</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/pritamdamania87/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/pritamdamania87">@pritamdamania87</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/zhaojuanmao/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/zhaojuanmao">@zhaojuanmao</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/satgera/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/satgera">@satgera</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/rohan-varma/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/rohan-varma">@rohan-varma</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/gqchen/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/gqchen">@gqchen</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/aazzolini/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/aazzolini">@aazzolini</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/xush6528/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/xush6528">@xush6528</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/osalpekar/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/osalpekar">@osalpekar</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/jiayisuse/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/jiayisuse">@jiayisuse</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/agolynski/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/agolynski">@agolynski</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/ptrblck/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/ptrblck">@ptrblck</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/mruberry/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/mruberry">@mruberry</a></p> | <ul dir="auto">
<li>System: Linux fit16 4.4.0-94-generic <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="182559240" data-permission-text="Title is private" data-url="https://github.com/pytorch/pytorch/issues/117" data-hovercard-type="pull_request" data-hovercard-url="/pytorch/pytorch/pull/117/hovercard" href="https://github.com/pytorch/pytorch/pull/117">#117</a>~14.04.1-Ubuntu SMP Wed Aug 30 06:50:25 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux</li>
<li>Instruction: <a href="https://caffe2.ai/docs/getting-started.html?platform=ubuntu&configuration=compile" rel="nofollow">https://caffe2.ai/docs/getting-started.html?platform=ubuntu&configuration=compile</a> failed at <code class="notranslate">make</code> step</li>
<li>no other flags. <code class="notranslate">cmake ..</code></li>
<li>CMake output:</li>
</ul>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="-- The CXX compiler identification is GNU 4.8.5
-- The C compiler identification is GNU 4.8.5
-- Check for working CXX compiler: /usr/bin/c++
-- Check for working CXX compiler: /usr/bin/c++ -- works
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Check for working C compiler: /usr/bin/cc
-- Check for working C compiler: /usr/bin/cc -- works
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Detecting C compile features
-- Detecting C compile features - done
-- Not forcing any particular BLAS to be found
-- Build type not set - defaulting to Release
-- Performing Test CAFFE2_LONG_IS_INT32_OR_64
-- Performing Test CAFFE2_LONG_IS_INT32_OR_64 - Success
-- Does not need to define long separately.
-- Performing Test CAFFE2_EXCEPTION_PTR_SUPPORTED
-- Performing Test CAFFE2_EXCEPTION_PTR_SUPPORTED - Success
-- std::exception_ptr is supported.
-- Performing Test CAFFE2_IS_NUMA_AVAILABLE
-- Performing Test CAFFE2_IS_NUMA_AVAILABLE - Failed
-- NUMA is not available
-- Performing Test CAFFE2_NEED_TO_TURN_OFF_DEPRECATION_WARNING
-- Performing Test CAFFE2_NEED_TO_TURN_OFF_DEPRECATION_WARNING - Failed
-- Turning off deprecation warning due to glog.
-- Performing Test CAFFE2_COMPILER_SUPPORTS_AVX2_EXTENSIONS
-- Performing Test CAFFE2_COMPILER_SUPPORTS_AVX2_EXTENSIONS - Success
-- Current compiler supports avx2 extention. Will build perfkernels.
-- Performing Test COMPILER_SUPPORTS_HIDDEN_VISIBILITY
-- Performing Test COMPILER_SUPPORTS_HIDDEN_VISIBILITY - Success
-- Performing Test COMPILER_SUPPORTS_HIDDEN_INLINE_VISIBILITY
-- Performing Test COMPILER_SUPPORTS_HIDDEN_INLINE_VISIBILITY - Success
-- Building using own protobuf under third_party per request.
-- Use custom protobuf build.
-- Looking for pthread.h
-- Looking for pthread.h - found
-- Looking for pthread_create
-- Looking for pthread_create - not found
-- Looking for pthread_create in pthreads
-- Looking for pthread_create in pthreads - not found
-- Looking for pthread_create in pthread
-- Looking for pthread_create in pthread - found
-- Found Threads: TRUE
-- Caffe2 protobuf include directory: $<BUILD_INTERFACE:/home/atlantix/pytorch/third_party/protobuf/src>$<INSTALL_INTERFACE:include>
-- Found Git: /usr/bin/git (found version "1.9.1")
-- The BLAS backend of choice:Eigen
-- Brace yourself, we are building NNPACK
-- The ASM compiler identification is GNU
-- Found assembler: /usr/bin/cc
-- Found PythonInterp: /usr/bin/python (found version "2.7.6")
-- Check if compiler accepts -pthread
-- Check if compiler accepts -pthread - yes
-- Caffe2: Cannot find gflags automatically. Using legacy find.
-- Found gflags: /usr/include
-- Caffe2: Found gflags (include: /usr/include, library: /usr/lib/x86_64-linux-gnu/libgflags.so)
-- Caffe2: Cannot find glog automatically. Using legacy find.
-- Found glog: /usr/include
-- Caffe2: Found glog (include: /usr/include, library: /usr/lib/x86_64-linux-gnu/libglog.so)
-- Failed to find LLVM FileCheck
-- git Version: v0.0.0
-- Version: 0.0.0
-- Performing Test HAVE_CXX_FLAG_STD_CXX11
-- Performing Test HAVE_CXX_FLAG_STD_CXX11 - Success
-- Performing Test HAVE_CXX_FLAG_WALL
-- Performing Test HAVE_CXX_FLAG_WALL - Success
-- Performing Test HAVE_CXX_FLAG_WEXTRA
-- Performing Test HAVE_CXX_FLAG_WEXTRA - Success
-- Performing Test HAVE_CXX_FLAG_WSHADOW
-- Performing Test HAVE_CXX_FLAG_WSHADOW - Success
-- Performing Test HAVE_CXX_FLAG_WERROR
-- Performing Test HAVE_CXX_FLAG_WERROR - Success
-- Performing Test HAVE_CXX_FLAG_PEDANTIC
-- Performing Test HAVE_CXX_FLAG_PEDANTIC - Success
-- Performing Test HAVE_CXX_FLAG_PEDANTIC_ERRORS
-- Performing Test HAVE_CXX_FLAG_PEDANTIC_ERRORS - Success
-- Performing Test HAVE_CXX_FLAG_WSHORTEN_64_TO_32
-- Performing Test HAVE_CXX_FLAG_WSHORTEN_64_TO_32 - Failed
-- Performing Test HAVE_CXX_FLAG_WFLOAT_EQUAL
-- Performing Test HAVE_CXX_FLAG_WFLOAT_EQUAL - Success
-- Performing Test HAVE_CXX_FLAG_FSTRICT_ALIASING
-- Performing Test HAVE_CXX_FLAG_FSTRICT_ALIASING - Success
-- Performing Test HAVE_CXX_FLAG_WSTRICT_ALIASING
-- Performing Test HAVE_CXX_FLAG_WSTRICT_ALIASING - Success
-- Performing Test HAVE_CXX_FLAG_WD654
-- Performing Test HAVE_CXX_FLAG_WD654 - Failed
-- Performing Test HAVE_CXX_FLAG_WTHREAD_SAFETY
-- Performing Test HAVE_CXX_FLAG_WTHREAD_SAFETY - Failed
-- Performing Test HAVE_CXX_FLAG_COVERAGE
-- Performing Test HAVE_CXX_FLAG_COVERAGE - Success
-- Performing Test HAVE_STD_REGEX
-- Performing Test HAVE_STD_REGEX
-- Performing Test HAVE_STD_REGEX -- compiled but failed to run
-- Performing Test HAVE_GNU_POSIX_REGEX
-- Performing Test HAVE_GNU_POSIX_REGEX
-- Performing Test HAVE_GNU_POSIX_REGEX -- failed to compile
-- Performing Test HAVE_POSIX_REGEX
-- Performing Test HAVE_POSIX_REGEX
-- Performing Test HAVE_POSIX_REGEX -- success
-- Performing Test HAVE_STEADY_CLOCK
-- Performing Test HAVE_STEADY_CLOCK
-- Performing Test HAVE_STEADY_CLOCK -- success
-- Found LMDB: /usr/include
-- Found lmdb (include: /usr/include, library: /usr/lib/x86_64-linux-gnu/liblmdb.so)
-- Found LevelDB: /usr/include
-- Found LevelDB (include: /usr/include, library: /usr/lib/x86_64-linux-gnu/libleveldb.so)
-- Found Snappy: /usr/include
-- Found Snappy (include: /usr/include, library: /usr/lib/libsnappy.so)
-- Could NOT find Numa (missing: Numa_INCLUDE_DIR Numa_LIBRARIES)
CMake Warning at cmake/Dependencies.cmake:243 (message):
Not compiling with NUMA. Suppress this warning with -DUSE_NUMA=OFF
Call Stack (most recent call first):
CMakeLists.txt:188 (include)
-- OpenCV found (/usr/local/share/OpenCV)
CMake Warning at cmake/Dependencies.cmake:309 (find_package):
By not providing "FindEigen3.cmake" in CMAKE_MODULE_PATH this project has
asked CMake to find a package configuration file provided by "Eigen3", but
CMake did not find one.
Could not find a package configuration file provided by "Eigen3" with any
of the following names:
Eigen3Config.cmake
eigen3-config.cmake
Add the installation prefix of "Eigen3" to CMAKE_PREFIX_PATH or set
"Eigen3_DIR" to a directory containing one of the above files. If "Eigen3"
provides a separate development package or SDK, be sure it has been
installed.
Call Stack (most recent call first):
CMakeLists.txt:188 (include)
-- Did not find system Eigen. Using third party subdirectory.
-- Found PythonInterp: /usr/bin/python (found suitable version "2.7.6", minimum required is "2.7")
-- Found PythonLibs: /usr/lib/x86_64-linux-gnu/libpython2.7.so (found suitable version "2.7.6", minimum required is "2.7")
-- Found NumPy: /usr/local/lib/python2.7/dist-packages/numpy/core/include (found version "1.13.3")
-- NumPy ver. 1.13.3 found (include: /usr/local/lib/python2.7/dist-packages/numpy/core/include)
-- Could NOT find pybind11 (missing: pybind11_INCLUDE_DIR)
-- Found MPI_C: /usr/lib/libmpi.so (found version "2.1")
-- Found MPI_CXX: /usr/lib/libmpi_cxx.so (found version "2.1")
-- Found MPI: TRUE (found version "2.1")
-- MPI support found
-- MPI compile flags: -pthread
-- MPI include path: /usr/lib/openmpi/include/openmpi/usr/lib/openmpi/include
-- MPI LINK flags path: -L/usr/lib/openmpi/lib -pthread
-- MPI libraries: /usr/lib/libmpi_cxx.so/usr/lib/libmpi.so/usr/lib/x86_64-linux-gnu/libdl.so/usr/lib/x86_64-linux-gnu/libhwloc.so
CMake Warning at cmake/Dependencies.cmake:364 (message):
OpenMPI found, but it is not built with CUDA support.
Call Stack (most recent call first):
CMakeLists.txt:188 (include)
-- Found CUDA: /usr/local/cuda-8.0 (found suitable version "8.0", minimum required is "7.0")
-- Found CUDNN: /usr/local/cuda-8.0/include
-- Caffe2: CUDA detected: 8.0
-- Found cuDNN: v6.0.21 (include: /usr/local/cuda-8.0/include, library: /usr/local/cuda-8.0/lib64/libcudnn.so)
-- Automatic GPU detection failed. Building for all known architectures.
-- Added CUDA NVCC flags for: sm_30 sm_35 sm_50 sm_52 sm_60 sm_61
-- Could NOT find NCCL (missing: NCCL_INCLUDE_DIRS NCCL_LIBRARIES)
-- Could NOT find CUB (missing: CUB_INCLUDE_DIR)
-- Could NOT find Gloo (missing: Gloo_INCLUDE_DIR Gloo_LIBRARY)
-- MPI include path: /usr/lib/openmpi/include/openmpi/usr/lib/openmpi/include
-- MPI libraries: /usr/lib/libmpi_cxx.so/usr/lib/libmpi.so/usr/lib/x86_64-linux-gnu/libdl.so/usr/lib/x86_64-linux-gnu/libhwloc.so
-- CUDA detected: 8.0
-- Found libcuda: /usr/lib/x86_64-linux-gnu/libcuda.so
-- Found libnvrtc: /usr/local/cuda-8.0/lib64/libnvrtc.so
-- Found nccl: /home/atlantix/pytorch/third_party/nccl/build/include
CMake Warning at cmake/Dependencies.cmake:551 (message):
mobile opengl is only used in android or ios builds.
Call Stack (most recent call first):
CMakeLists.txt:188 (include)
CMake Warning at cmake/Dependencies.cmake:627 (message):
Metal is only used in ios builds.
Call Stack (most recent call first):
CMakeLists.txt:188 (include)
-- GCC 4.8.5: Adding gcc and gcc_s libs to link line
-- Include NCCL operators
-- Excluding ideep operators as we are not using ideep
-- Including image processing operators
-- Excluding video processing operators due to no opencv
-- Excluding mkl operators as we are not using mkl
-- Include Observer library
-- Using lib/python2.7/dist-packages as python relative installation path
-- Automatically generating missing __init__.py files.
CMake Warning at CMakeLists.txt:318 (message):
Generated cmake files are only fully tested if one builds with system glog,
gflags, and protobuf. Other settings may generate files that are not well
tested.
--
-- ******** Summary ********
-- General:
-- CMake version : 3.11.2
-- CMake command : /usr/local/bin/cmake
-- Git version : v0.1.11-8572-g45cdb63
-- System : Linux
-- C++ compiler : /usr/bin/c++
-- C++ compiler version : 4.8.5
-- BLAS : Eigen
-- CXX flags : -Wno-deprecated -fvisibility-inlines-hidden -DONNX_NAMESPACE=onnx_c2 -O2 -fPIC -Wno-narrowing -Wno-invalid-partial-specialization
-- Build type : Release
-- Compile definitions :
--
-- BUILD_CAFFE2 : ON
-- BUILD_ATEN : OFF
-- BUILD_BINARY : ON
-- BUILD_CUSTOM_PROTOBUF : ON
-- Link local protobuf : ON
-- BUILD_DOCS : OFF
-- BUILD_PYTHON : ON
-- Python version : 2.7.6
-- Python includes : /usr/include/python2.7
-- BUILD_SHARED_LIBS : ON
-- BUILD_TEST : ON
-- USE_ASAN : OFF
-- USE_CUDA : ON
-- CUDA static link : OFF
-- USE_CUDNN : ON
-- CUDA version : 8.0
-- cuDNN version : 6.0.21
-- CUDA root directory : /usr/local/cuda-8.0
-- CUDA library : /usr/lib/x86_64-linux-gnu/libcuda.so
-- cudart library : /usr/local/cuda-8.0/lib64/libcudart_static.a;-pthread;dl;/usr/lib/x86_64-linux-gnu/librt.so
-- cublas library : /usr/local/cuda-8.0/lib64/libcublas.so;/usr/local/cuda-8.0/lib64/libcublas_device.a
-- cufft library : /usr/local/cuda-8.0/lib64/libcufft.so
-- curand library : /usr/local/cuda-8.0/lib64/libcurand.so
-- cuDNN library : /usr/local/cuda-8.0/lib64/libcudnn.so
-- nvrtc : /usr/local/cuda-8.0/lib64/libnvrtc.so
-- CUDA include path : /usr/local/cuda-8.0/include
-- NVCC executable : /usr/local/cuda-8.0/bin/nvcc
-- CUDA host compiler : /usr/bin/cc
-- USE_TENSORRT : OFF
-- USE_ROCM : OFF
-- USE_EIGEN_FOR_BLAS : ON
-- USE_FFMPEG : OFF
-- USE_GFLAGS : ON
-- USE_GLOG : ON
-- USE_GLOO : ON
-- USE_GLOO_IBVERBS : OFF
-- USE_LEVELDB : ON
-- LevelDB version : 1.15
-- Snappy version : 1.1.0
-- USE_LITE_PROTO : OFF
-- USE_LMDB : ON
-- LMDB version : 0.9.16
-- USE_METAL : OFF
-- USE_MKL :
-- USE_MOBILE_OPENGL : OFF
-- USE_MPI : ON
-- USE_NCCL : ON
-- USE_SYSTEM_NCCL : OFF
-- USE_NERVANA_GPU : OFF
-- USE_NNPACK : ON
-- USE_OBSERVERS : ON
-- USE_OPENCL : OFF
-- USE_OPENCV : ON
-- OpenCV version : 3.4.1
-- USE_OPENMP : OFF
-- USE_PROF : OFF
-- USE_REDIS : OFF
-- USE_ROCKSDB : OFF
-- USE_ZMQ : OFF
-- Configuring done
-- Generating done
-- Build files have been written to: /home/atlantix/pytorch/build"><pre class="notranslate"><code class="notranslate">-- The CXX compiler identification is GNU 4.8.5
-- The C compiler identification is GNU 4.8.5
-- Check for working CXX compiler: /usr/bin/c++
-- Check for working CXX compiler: /usr/bin/c++ -- works
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Check for working C compiler: /usr/bin/cc
-- Check for working C compiler: /usr/bin/cc -- works
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Detecting C compile features
-- Detecting C compile features - done
-- Not forcing any particular BLAS to be found
-- Build type not set - defaulting to Release
-- Performing Test CAFFE2_LONG_IS_INT32_OR_64
-- Performing Test CAFFE2_LONG_IS_INT32_OR_64 - Success
-- Does not need to define long separately.
-- Performing Test CAFFE2_EXCEPTION_PTR_SUPPORTED
-- Performing Test CAFFE2_EXCEPTION_PTR_SUPPORTED - Success
-- std::exception_ptr is supported.
-- Performing Test CAFFE2_IS_NUMA_AVAILABLE
-- Performing Test CAFFE2_IS_NUMA_AVAILABLE - Failed
-- NUMA is not available
-- Performing Test CAFFE2_NEED_TO_TURN_OFF_DEPRECATION_WARNING
-- Performing Test CAFFE2_NEED_TO_TURN_OFF_DEPRECATION_WARNING - Failed
-- Turning off deprecation warning due to glog.
-- Performing Test CAFFE2_COMPILER_SUPPORTS_AVX2_EXTENSIONS
-- Performing Test CAFFE2_COMPILER_SUPPORTS_AVX2_EXTENSIONS - Success
-- Current compiler supports avx2 extention. Will build perfkernels.
-- Performing Test COMPILER_SUPPORTS_HIDDEN_VISIBILITY
-- Performing Test COMPILER_SUPPORTS_HIDDEN_VISIBILITY - Success
-- Performing Test COMPILER_SUPPORTS_HIDDEN_INLINE_VISIBILITY
-- Performing Test COMPILER_SUPPORTS_HIDDEN_INLINE_VISIBILITY - Success
-- Building using own protobuf under third_party per request.
-- Use custom protobuf build.
-- Looking for pthread.h
-- Looking for pthread.h - found
-- Looking for pthread_create
-- Looking for pthread_create - not found
-- Looking for pthread_create in pthreads
-- Looking for pthread_create in pthreads - not found
-- Looking for pthread_create in pthread
-- Looking for pthread_create in pthread - found
-- Found Threads: TRUE
-- Caffe2 protobuf include directory: $<BUILD_INTERFACE:/home/atlantix/pytorch/third_party/protobuf/src>$<INSTALL_INTERFACE:include>
-- Found Git: /usr/bin/git (found version "1.9.1")
-- The BLAS backend of choice:Eigen
-- Brace yourself, we are building NNPACK
-- The ASM compiler identification is GNU
-- Found assembler: /usr/bin/cc
-- Found PythonInterp: /usr/bin/python (found version "2.7.6")
-- Check if compiler accepts -pthread
-- Check if compiler accepts -pthread - yes
-- Caffe2: Cannot find gflags automatically. Using legacy find.
-- Found gflags: /usr/include
-- Caffe2: Found gflags (include: /usr/include, library: /usr/lib/x86_64-linux-gnu/libgflags.so)
-- Caffe2: Cannot find glog automatically. Using legacy find.
-- Found glog: /usr/include
-- Caffe2: Found glog (include: /usr/include, library: /usr/lib/x86_64-linux-gnu/libglog.so)
-- Failed to find LLVM FileCheck
-- git Version: v0.0.0
-- Version: 0.0.0
-- Performing Test HAVE_CXX_FLAG_STD_CXX11
-- Performing Test HAVE_CXX_FLAG_STD_CXX11 - Success
-- Performing Test HAVE_CXX_FLAG_WALL
-- Performing Test HAVE_CXX_FLAG_WALL - Success
-- Performing Test HAVE_CXX_FLAG_WEXTRA
-- Performing Test HAVE_CXX_FLAG_WEXTRA - Success
-- Performing Test HAVE_CXX_FLAG_WSHADOW
-- Performing Test HAVE_CXX_FLAG_WSHADOW - Success
-- Performing Test HAVE_CXX_FLAG_WERROR
-- Performing Test HAVE_CXX_FLAG_WERROR - Success
-- Performing Test HAVE_CXX_FLAG_PEDANTIC
-- Performing Test HAVE_CXX_FLAG_PEDANTIC - Success
-- Performing Test HAVE_CXX_FLAG_PEDANTIC_ERRORS
-- Performing Test HAVE_CXX_FLAG_PEDANTIC_ERRORS - Success
-- Performing Test HAVE_CXX_FLAG_WSHORTEN_64_TO_32
-- Performing Test HAVE_CXX_FLAG_WSHORTEN_64_TO_32 - Failed
-- Performing Test HAVE_CXX_FLAG_WFLOAT_EQUAL
-- Performing Test HAVE_CXX_FLAG_WFLOAT_EQUAL - Success
-- Performing Test HAVE_CXX_FLAG_FSTRICT_ALIASING
-- Performing Test HAVE_CXX_FLAG_FSTRICT_ALIASING - Success
-- Performing Test HAVE_CXX_FLAG_WSTRICT_ALIASING
-- Performing Test HAVE_CXX_FLAG_WSTRICT_ALIASING - Success
-- Performing Test HAVE_CXX_FLAG_WD654
-- Performing Test HAVE_CXX_FLAG_WD654 - Failed
-- Performing Test HAVE_CXX_FLAG_WTHREAD_SAFETY
-- Performing Test HAVE_CXX_FLAG_WTHREAD_SAFETY - Failed
-- Performing Test HAVE_CXX_FLAG_COVERAGE
-- Performing Test HAVE_CXX_FLAG_COVERAGE - Success
-- Performing Test HAVE_STD_REGEX
-- Performing Test HAVE_STD_REGEX
-- Performing Test HAVE_STD_REGEX -- compiled but failed to run
-- Performing Test HAVE_GNU_POSIX_REGEX
-- Performing Test HAVE_GNU_POSIX_REGEX
-- Performing Test HAVE_GNU_POSIX_REGEX -- failed to compile
-- Performing Test HAVE_POSIX_REGEX
-- Performing Test HAVE_POSIX_REGEX
-- Performing Test HAVE_POSIX_REGEX -- success
-- Performing Test HAVE_STEADY_CLOCK
-- Performing Test HAVE_STEADY_CLOCK
-- Performing Test HAVE_STEADY_CLOCK -- success
-- Found LMDB: /usr/include
-- Found lmdb (include: /usr/include, library: /usr/lib/x86_64-linux-gnu/liblmdb.so)
-- Found LevelDB: /usr/include
-- Found LevelDB (include: /usr/include, library: /usr/lib/x86_64-linux-gnu/libleveldb.so)
-- Found Snappy: /usr/include
-- Found Snappy (include: /usr/include, library: /usr/lib/libsnappy.so)
-- Could NOT find Numa (missing: Numa_INCLUDE_DIR Numa_LIBRARIES)
CMake Warning at cmake/Dependencies.cmake:243 (message):
Not compiling with NUMA. Suppress this warning with -DUSE_NUMA=OFF
Call Stack (most recent call first):
CMakeLists.txt:188 (include)
-- OpenCV found (/usr/local/share/OpenCV)
CMake Warning at cmake/Dependencies.cmake:309 (find_package):
By not providing "FindEigen3.cmake" in CMAKE_MODULE_PATH this project has
asked CMake to find a package configuration file provided by "Eigen3", but
CMake did not find one.
Could not find a package configuration file provided by "Eigen3" with any
of the following names:
Eigen3Config.cmake
eigen3-config.cmake
Add the installation prefix of "Eigen3" to CMAKE_PREFIX_PATH or set
"Eigen3_DIR" to a directory containing one of the above files. If "Eigen3"
provides a separate development package or SDK, be sure it has been
installed.
Call Stack (most recent call first):
CMakeLists.txt:188 (include)
-- Did not find system Eigen. Using third party subdirectory.
-- Found PythonInterp: /usr/bin/python (found suitable version "2.7.6", minimum required is "2.7")
-- Found PythonLibs: /usr/lib/x86_64-linux-gnu/libpython2.7.so (found suitable version "2.7.6", minimum required is "2.7")
-- Found NumPy: /usr/local/lib/python2.7/dist-packages/numpy/core/include (found version "1.13.3")
-- NumPy ver. 1.13.3 found (include: /usr/local/lib/python2.7/dist-packages/numpy/core/include)
-- Could NOT find pybind11 (missing: pybind11_INCLUDE_DIR)
-- Found MPI_C: /usr/lib/libmpi.so (found version "2.1")
-- Found MPI_CXX: /usr/lib/libmpi_cxx.so (found version "2.1")
-- Found MPI: TRUE (found version "2.1")
-- MPI support found
-- MPI compile flags: -pthread
-- MPI include path: /usr/lib/openmpi/include/openmpi/usr/lib/openmpi/include
-- MPI LINK flags path: -L/usr/lib/openmpi/lib -pthread
-- MPI libraries: /usr/lib/libmpi_cxx.so/usr/lib/libmpi.so/usr/lib/x86_64-linux-gnu/libdl.so/usr/lib/x86_64-linux-gnu/libhwloc.so
CMake Warning at cmake/Dependencies.cmake:364 (message):
OpenMPI found, but it is not built with CUDA support.
Call Stack (most recent call first):
CMakeLists.txt:188 (include)
-- Found CUDA: /usr/local/cuda-8.0 (found suitable version "8.0", minimum required is "7.0")
-- Found CUDNN: /usr/local/cuda-8.0/include
-- Caffe2: CUDA detected: 8.0
-- Found cuDNN: v6.0.21 (include: /usr/local/cuda-8.0/include, library: /usr/local/cuda-8.0/lib64/libcudnn.so)
-- Automatic GPU detection failed. Building for all known architectures.
-- Added CUDA NVCC flags for: sm_30 sm_35 sm_50 sm_52 sm_60 sm_61
-- Could NOT find NCCL (missing: NCCL_INCLUDE_DIRS NCCL_LIBRARIES)
-- Could NOT find CUB (missing: CUB_INCLUDE_DIR)
-- Could NOT find Gloo (missing: Gloo_INCLUDE_DIR Gloo_LIBRARY)
-- MPI include path: /usr/lib/openmpi/include/openmpi/usr/lib/openmpi/include
-- MPI libraries: /usr/lib/libmpi_cxx.so/usr/lib/libmpi.so/usr/lib/x86_64-linux-gnu/libdl.so/usr/lib/x86_64-linux-gnu/libhwloc.so
-- CUDA detected: 8.0
-- Found libcuda: /usr/lib/x86_64-linux-gnu/libcuda.so
-- Found libnvrtc: /usr/local/cuda-8.0/lib64/libnvrtc.so
-- Found nccl: /home/atlantix/pytorch/third_party/nccl/build/include
CMake Warning at cmake/Dependencies.cmake:551 (message):
mobile opengl is only used in android or ios builds.
Call Stack (most recent call first):
CMakeLists.txt:188 (include)
CMake Warning at cmake/Dependencies.cmake:627 (message):
Metal is only used in ios builds.
Call Stack (most recent call first):
CMakeLists.txt:188 (include)
-- GCC 4.8.5: Adding gcc and gcc_s libs to link line
-- Include NCCL operators
-- Excluding ideep operators as we are not using ideep
-- Including image processing operators
-- Excluding video processing operators due to no opencv
-- Excluding mkl operators as we are not using mkl
-- Include Observer library
-- Using lib/python2.7/dist-packages as python relative installation path
-- Automatically generating missing __init__.py files.
CMake Warning at CMakeLists.txt:318 (message):
Generated cmake files are only fully tested if one builds with system glog,
gflags, and protobuf. Other settings may generate files that are not well
tested.
--
-- ******** Summary ********
-- General:
-- CMake version : 3.11.2
-- CMake command : /usr/local/bin/cmake
-- Git version : v0.1.11-8572-g45cdb63
-- System : Linux
-- C++ compiler : /usr/bin/c++
-- C++ compiler version : 4.8.5
-- BLAS : Eigen
-- CXX flags : -Wno-deprecated -fvisibility-inlines-hidden -DONNX_NAMESPACE=onnx_c2 -O2 -fPIC -Wno-narrowing -Wno-invalid-partial-specialization
-- Build type : Release
-- Compile definitions :
--
-- BUILD_CAFFE2 : ON
-- BUILD_ATEN : OFF
-- BUILD_BINARY : ON
-- BUILD_CUSTOM_PROTOBUF : ON
-- Link local protobuf : ON
-- BUILD_DOCS : OFF
-- BUILD_PYTHON : ON
-- Python version : 2.7.6
-- Python includes : /usr/include/python2.7
-- BUILD_SHARED_LIBS : ON
-- BUILD_TEST : ON
-- USE_ASAN : OFF
-- USE_CUDA : ON
-- CUDA static link : OFF
-- USE_CUDNN : ON
-- CUDA version : 8.0
-- cuDNN version : 6.0.21
-- CUDA root directory : /usr/local/cuda-8.0
-- CUDA library : /usr/lib/x86_64-linux-gnu/libcuda.so
-- cudart library : /usr/local/cuda-8.0/lib64/libcudart_static.a;-pthread;dl;/usr/lib/x86_64-linux-gnu/librt.so
-- cublas library : /usr/local/cuda-8.0/lib64/libcublas.so;/usr/local/cuda-8.0/lib64/libcublas_device.a
-- cufft library : /usr/local/cuda-8.0/lib64/libcufft.so
-- curand library : /usr/local/cuda-8.0/lib64/libcurand.so
-- cuDNN library : /usr/local/cuda-8.0/lib64/libcudnn.so
-- nvrtc : /usr/local/cuda-8.0/lib64/libnvrtc.so
-- CUDA include path : /usr/local/cuda-8.0/include
-- NVCC executable : /usr/local/cuda-8.0/bin/nvcc
-- CUDA host compiler : /usr/bin/cc
-- USE_TENSORRT : OFF
-- USE_ROCM : OFF
-- USE_EIGEN_FOR_BLAS : ON
-- USE_FFMPEG : OFF
-- USE_GFLAGS : ON
-- USE_GLOG : ON
-- USE_GLOO : ON
-- USE_GLOO_IBVERBS : OFF
-- USE_LEVELDB : ON
-- LevelDB version : 1.15
-- Snappy version : 1.1.0
-- USE_LITE_PROTO : OFF
-- USE_LMDB : ON
-- LMDB version : 0.9.16
-- USE_METAL : OFF
-- USE_MKL :
-- USE_MOBILE_OPENGL : OFF
-- USE_MPI : ON
-- USE_NCCL : ON
-- USE_SYSTEM_NCCL : OFF
-- USE_NERVANA_GPU : OFF
-- USE_NNPACK : ON
-- USE_OBSERVERS : ON
-- USE_OPENCL : OFF
-- USE_OPENCV : ON
-- OpenCV version : 3.4.1
-- USE_OPENMP : OFF
-- USE_PROF : OFF
-- USE_REDIS : OFF
-- USE_ROCKSDB : OFF
-- USE_ZMQ : OFF
-- Configuring done
-- Generating done
-- Build files have been written to: /home/atlantix/pytorch/build
</code></pre></div>
<ul dir="auto">
<li>Error message:</li>
</ul>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="[ 67%] Linking CXX shared library ../lib/libcaffe2.so
/usr/bin/ld: cannot find -lnvrtc
collect2: error: ld returned 1 exit status
make[2]: *** [lib/libcaffe2.so.1] Error 1
make[1]: *** [caffe2/CMakeFiles/caffe2.dir/all] Error 2
make: *** [all] Error 2"><pre class="notranslate"><code class="notranslate">[ 67%] Linking CXX shared library ../lib/libcaffe2.so
/usr/bin/ld: cannot find -lnvrtc
collect2: error: ld returned 1 exit status
make[2]: *** [lib/libcaffe2.so.1] Error 1
make[1]: *** [caffe2/CMakeFiles/caffe2.dir/all] Error 2
make: *** [all] Error 2
</code></pre></div>
<p dir="auto">My installation command is very simple: <code class="notranslate">cmake ..</code> and <code class="notranslate">make</code>, but it failed in the second step.<br>
I notice that <code class="notranslate">libnvrtc</code> can be found in <code class="notranslate">ldconfig</code>, and is found by cmake, but failed in make.<br>
Can anyone help?</p> | 0 |
<p dir="auto">Isn't the idea of 'inplace' that nothing needs to be returned? Additionally, when I have a lot of columns, the notebook's printout of the returned dataframe 'cramps my style'. ;)</p> | <p dir="auto">Is there any reason that whenever I'm doing inplace=True I always get self returned? Obviously not a huge deal, but this is kind of wart-y IMO. I wouldn't expect inplace to return anything.</p> | 1 |
<p dir="auto">I'm updating the Matrix4 docs and there's a method called Matrix4.makePerspective( fov, aspect, near, far )</p>
<p dir="auto">Internally this calculates the values of left, right, top and bottom and calls<br>
Matrix4.makeFrustum( left, right, bottom, top, near, far )<br>
and it looks like it is intended for use in PerspectiveCamera.updateProjectionMatrix()</p>
<p dir="auto">However the PerspectiveCamera calculates its own versions of left, right, top and bottom in a slightly more complex way, taking into account zoom, view (if set) and filmOffset, and then calls Matrix4.makeFrustum directly.</p>
<p dir="auto">So it seems that Matrix4.makePerspective() is superfluous - is there a use case that I am missing?</p> | <h5 dir="auto">Description of the problem</h5>
<p dir="auto">When using a NodeMaterial, setting <code class="notranslate">needsUpdate</code> to <code class="notranslate">true</code> because of a change made to the material will not yield the expected result as it's not recompiled. The change related to <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="525720159" data-permission-text="Title is private" data-url="https://github.com/mrdoob/three.js/issues/17968" data-hovercard-type="pull_request" data-hovercard-url="/mrdoob/three.js/pull/17968/hovercard" href="https://github.com/mrdoob/three.js/pull/17968">#17968</a> produces this result, and when reverting <a href="https://github.com/mrdoob/three.js/pull/17968/commits/b521945d95d261f200c11c1c6ecf5fe8840eda46#diff-e7ff19643959b771a35742c5e98f7c5fL120">this blob in specific</a>, the material updates as expected. This issue is present from r111 and up, possibly relating to the fact that the <code class="notranslate">needsUpdate</code> flag is never reset to <code class="notranslate">false</code>.<br>
In relation, <a href="https://github.com/mrdoob/three.js/issues/18347#issuecomment-572404500" data-hovercard-type="issue" data-hovercard-url="/mrdoob/three.js/issues/18347/hovercard">this proposed change by @sunag</a> (coming from <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="547265887" data-permission-text="Title is private" data-url="https://github.com/mrdoob/three.js/issues/18347" data-hovercard-type="issue" data-hovercard-url="/mrdoob/three.js/issues/18347/hovercard" href="https://github.com/mrdoob/three.js/issues/18347">#18347</a>) seems to work as well:</p>
<div class="highlight highlight-source-js notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="this.onBeforeCompile = function ( shader, renderer ) {
var materialProperties = renderer.properties.get( this );
if ( this.version !== materialProperties.__version ) {
this.build( { renderer: renderer } );
shader.uniforms = this.uniforms;
shader.vertexShader = this.vertexShader;
shader.fragmentShader = this.fragmentShader;
this.needsCompile = false;
}
};"><pre class="notranslate"><span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-en">onBeforeCompile</span> <span class="pl-c1">=</span> <span class="pl-k">function</span> <span class="pl-kos">(</span> <span class="pl-s1">shader</span><span class="pl-kos">,</span> <span class="pl-s1">renderer</span> <span class="pl-kos">)</span> <span class="pl-kos">{</span>
<span class="pl-k">var</span> <span class="pl-s1">materialProperties</span> <span class="pl-c1">=</span> <span class="pl-s1">renderer</span><span class="pl-kos">.</span><span class="pl-c1">properties</span><span class="pl-kos">.</span><span class="pl-en">get</span><span class="pl-kos">(</span> <span class="pl-smi">this</span> <span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-k">if</span> <span class="pl-kos">(</span> <span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-c1">version</span> <span class="pl-c1">!==</span> <span class="pl-s1">materialProperties</span><span class="pl-kos">.</span><span class="pl-c1">__version</span> <span class="pl-kos">)</span> <span class="pl-kos">{</span>
<span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-en">build</span><span class="pl-kos">(</span> <span class="pl-kos">{</span> <span class="pl-c1">renderer</span>: <span class="pl-s1">renderer</span> <span class="pl-kos">}</span> <span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-s1">shader</span><span class="pl-kos">.</span><span class="pl-c1">uniforms</span> <span class="pl-c1">=</span> <span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-c1">uniforms</span><span class="pl-kos">;</span>
<span class="pl-s1">shader</span><span class="pl-kos">.</span><span class="pl-c1">vertexShader</span> <span class="pl-c1">=</span> <span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-c1">vertexShader</span><span class="pl-kos">;</span>
<span class="pl-s1">shader</span><span class="pl-kos">.</span><span class="pl-c1">fragmentShader</span> <span class="pl-c1">=</span> <span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-c1">fragmentShader</span><span class="pl-kos">;</span>
<span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-c1">needsCompile</span> <span class="pl-c1">=</span> <span class="pl-c1">false</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span>
<span class="pl-kos">}</span><span class="pl-kos">;</span></pre></div>
<h5 dir="auto">Three.js version</h5>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> Dev</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> r113</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> r112</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> r111</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> r110</li>
</ul>
<h5 dir="auto">Browser</h5>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> All of them</li>
</ul>
<h5 dir="auto">OS</h5>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> All of them</li>
</ul>
<p dir="auto">Paging <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/sunag/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/sunag">@sunag</a>.</p> | 0 |
<p dir="auto">The using-nerv example is not working</p>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I have searched the <a href="https://github.com/zeit/next.js/issues">issues</a> of this repository and believe that this is not a duplicate.</li>
</ul>
<h2 dir="auto">Expected Behavior</h2>
<p dir="auto">can start the example server</p>
<h2 dir="auto">Current Behavior</h2>
<p dir="auto">Getting error:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="> [email protected] start /Users/preuss/next.js/examples/using-nerv
> NODE_ENV=production node server.js
module.js:557
throw err;
^
Error: Cannot find module 'nervjs/server'
at Function.Module._resolveFilename (module.js:555:15)
at Function.Module._resolveFilename (/Users/preuss/next.js/examples/using-nerv/node_modules/module-alias/index.js:43:29)
at Function.Module._load (module.js:482:25)
at Module.require (module.js:604:17)"><pre class="notranslate"><code class="notranslate">> [email protected] start /Users/preuss/next.js/examples/using-nerv
> NODE_ENV=production node server.js
module.js:557
throw err;
^
Error: Cannot find module 'nervjs/server'
at Function.Module._resolveFilename (module.js:555:15)
at Function.Module._resolveFilename (/Users/preuss/next.js/examples/using-nerv/node_modules/module-alias/index.js:43:29)
at Function.Module._load (module.js:482:25)
at Module.require (module.js:604:17)
</code></pre></div>
<h2 dir="auto">Steps to Reproduce (for bugs)</h2>
<ol dir="auto">
<li>cd examples/using-nerv</li>
<li>npm i</li>
<li>npm run build && npm run start</li>
</ol>
<h2 dir="auto">Your Environment</h2>
<table role="table">
<thead>
<tr>
<th>Tech</th>
<th>Version</th>
</tr>
</thead>
<tbody>
<tr>
<td>next</td>
<td>latest (5.0.0)</td>
</tr>
<tr>
<td>node</td>
<td>9.4.0</td>
</tr>
<tr>
<td>OS</td>
<td>osx</td>
</tr>
</tbody>
</table> | <ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I have searched the <a href="https://github.com/zeit/next.js/issues?q=is%3Aissue">issues</a> of this repository and believe that this is not a duplicate.</li>
</ul>
<h2 dir="auto">Expected Behavior</h2>
<p dir="auto">I am proxyfying requests handled by my Express app to my Next.js app, which is used programatically.</p>
<p dir="auto">I have express running on :3000 and next.js on :3001</p>
<p dir="auto">I get <code class="notranslate">TypeError: Cannot read property 'waitUntilReloaded' of undefined at HotReloader._callee7$ (/Users/vadorequest/dev/serverless-with-next/node_modules/next/dist/server/hot-reloader.js:658:44)</code> error message when loading any page.</p>
<div class="highlight highlight-source-js notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="const nextAppConfig = {
dev: !isHostedOnAWS() && process.env.NODE_ENV === 'development',
};
const nextApp = next(nextAppConfig);
const nextProxy = nextApp.getRequestHandler(); // Requests handled by nextProxy will be handled by the Next.js app
app.get('*', (req, res) => {
console.log('req from', req.protocol + '://' + req.get('host') + req.originalUrl);
nextProxy(req, res);
});"><pre class="notranslate"><span class="pl-k">const</span> <span class="pl-s1">nextAppConfig</span> <span class="pl-c1">=</span> <span class="pl-kos">{</span>
<span class="pl-c1">dev</span>: <span class="pl-c1">!</span><span class="pl-en">isHostedOnAWS</span><span class="pl-kos">(</span><span class="pl-kos">)</span> <span class="pl-c1">&&</span> <span class="pl-s1">process</span><span class="pl-kos">.</span><span class="pl-c1">env</span><span class="pl-kos">.</span><span class="pl-c1">NODE_ENV</span> <span class="pl-c1">===</span> <span class="pl-s">'development'</span><span class="pl-kos">,</span>
<span class="pl-kos">}</span><span class="pl-kos">;</span>
<span class="pl-k">const</span> <span class="pl-s1">nextApp</span> <span class="pl-c1">=</span> <span class="pl-en">next</span><span class="pl-kos">(</span><span class="pl-s1">nextAppConfig</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-k">const</span> <span class="pl-s1">nextProxy</span> <span class="pl-c1">=</span> <span class="pl-s1">nextApp</span><span class="pl-kos">.</span><span class="pl-en">getRequestHandler</span><span class="pl-kos">(</span><span class="pl-kos">)</span><span class="pl-kos">;</span> <span class="pl-c">// Requests handled by nextProxy will be handled by the Next.js app</span>
<span class="pl-s1">app</span><span class="pl-kos">.</span><span class="pl-en">get</span><span class="pl-kos">(</span><span class="pl-s">'*'</span><span class="pl-kos">,</span> <span class="pl-kos">(</span><span class="pl-s1">req</span><span class="pl-kos">,</span> <span class="pl-s1">res</span><span class="pl-kos">)</span> <span class="pl-c1">=></span> <span class="pl-kos">{</span>
<span class="pl-smi">console</span><span class="pl-kos">.</span><span class="pl-en">log</span><span class="pl-kos">(</span><span class="pl-s">'req from'</span><span class="pl-kos">,</span> <span class="pl-s1">req</span><span class="pl-kos">.</span><span class="pl-c1">protocol</span> <span class="pl-c1">+</span> <span class="pl-s">'://'</span> <span class="pl-c1">+</span> <span class="pl-s1">req</span><span class="pl-kos">.</span><span class="pl-en">get</span><span class="pl-kos">(</span><span class="pl-s">'host'</span><span class="pl-kos">)</span> <span class="pl-c1">+</span> <span class="pl-s1">req</span><span class="pl-kos">.</span><span class="pl-c1">originalUrl</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-s1">nextProxy</span><span class="pl-kos">(</span><span class="pl-s1">req</span><span class="pl-kos">,</span> <span class="pl-s1">res</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span><span class="pl-kos">)</span><span class="pl-kos">;</span></pre></div>
<p dir="auto">I don't reproduce this issue when running my app on AWS. See <a href="https://swn.dev.vadorequest.fr" rel="nofollow">https://swn.dev.vadorequest.fr</a></p>
<details>
<p dir="auto">sls TypeError: Cannot read property 'waitUntilReloaded' of undefined<br>
sls at HotReloader._callee7$ (/Users/vadorequest/dev/serverless-with-next/node_modules/next/dist/server/hot-reloader.js:658:44)<br>
sls at tryCatch (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:62:40)<br>
sls at GeneratorFunctionPrototype.invoke [as _invoke] (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:296:22)<br>
sls at GeneratorFunctionPrototype.prototype.(anonymous function) [as next] (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:114:21)<br>
sls at step (/Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:17:30)<br>
sls at /Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:35:14<br>
sls at Promise.F (/Users/vadorequest/dev/serverless-with-next/node_modules/core-js/library/modules/_export.js:35:28)<br>
sls at HotReloader. (/Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:14:12)<br>
sls at HotReloader.getCompilationErrors (/Users/vadorequest/dev/serverless-with-next/node_modules/next/dist/server/hot-reloader.js:837:23)<br>
sls at Server._callee25$ (/Users/vadorequest/dev/serverless-with-next/node_modules/next/dist/server/index.js:1322:41)<br>
sls at tryCatch (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:62:40)<br>
sls at GeneratorFunctionPrototype.invoke [as _invoke] (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:296:22)<br>
sls at GeneratorFunctionPrototype.prototype.(anonymous function) [as next] (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:114:21)<br>
sls at step (/Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:17:30)<br>
sls at /Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:35:14<br>
sls at Promise.F (/Users/vadorequest/dev/serverless-with-next/node_modules/core-js/library/modules/_export.js:35:28)<br>
sls at Server. (/Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:14:12)<br>
sls at Server.getCompilationError (/Users/vadorequest/dev/serverless-with-next/node_modules/next/dist/server/index.js:1346:23)<br>
sls at Server._callee20$ (/Users/vadorequest/dev/serverless-with-next/node_modules/next/dist/server/index.js:1040:29)<br>
sls at tryCatch (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:62:40)<br>
sls at GeneratorFunctionPrototype.invoke [as _invoke] (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:296:22)<br>
sls at GeneratorFunctionPrototype.prototype.(anonymous function) [as next] (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:114:21)<br>
sls at step (/Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:17:30)<br>
sls at /Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:35:14<br>
sls at Promise.F (/Users/vadorequest/dev/serverless-with-next/node_modules/core-js/library/modules/_export.js:35:28)<br>
sls at Server. (/Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:14:12)<br>
sls at Server.renderToHTML (/Users/vadorequest/dev/serverless-with-next/node_modules/next/dist/server/index.js:1088:23)<br>
sls at Server._callee19$ (/Users/vadorequest/dev/serverless-with-next/node_modules/next/dist/server/index.js:992:29)<br>
sls at tryCatch (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:62:40)<br>
sls at GeneratorFunctionPrototype.invoke [as _invoke] (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:296:22)<br>
sls at GeneratorFunctionPrototype.prototype.(anonymous function) [as next] (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:114:21)<br>
sls at step (/Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:17:30)<br>
sls at /Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:35:14<br>
sls at Promise.F (/Users/vadorequest/dev/serverless-with-next/node_modules/core-js/library/modules/_export.js:35:28)<br>
sls at Server. (/Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:14:12)<br>
sls at Server.render (/Users/vadorequest/dev/serverless-with-next/node_modules/next/dist/server/index.js:1020:23)<br>
sls at Server._callee16$ (/Users/vadorequest/dev/serverless-with-next/node_modules/next/dist/server/index.js:817:35)<br>
sls at tryCatch (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:62:40)<br>
sls at GeneratorFunctionPrototype.invoke [as _invoke] (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:296:22)<br>
sls at GeneratorFunctionPrototype.prototype.(anonymous function) [as next] (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:114:21)<br>
sls at step (/Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:17:30)<br>
sls at /Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:35:14<br>
sls at Promise.F (/Users/vadorequest/dev/serverless-with-next/node_modules/core-js/library/modules/_export.js:35:28)<br>
sls at Object. (/Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:14:12)<br>
sls at Object.fn (/Users/vadorequest/dev/serverless-with-next/node_modules/next/dist/server/index.js:828:27)<br>
sls at Router._callee$ (/Users/vadorequest/dev/serverless-with-next/node_modules/next/dist/server/router.js:81:60)<br>
sls at tryCatch (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:62:40)<br>
sls at GeneratorFunctionPrototype.invoke [as _invoke] (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:296:22)<br>
sls at GeneratorFunctionPrototype.prototype.(anonymous function) [as next] (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:114:21)<br>
sls at step (/Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:17:30)<br>
sls at /Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:35:14<br>
sls at Promise.F (/Users/vadorequest/dev/serverless-with-next/node_modules/core-js/library/modules/_export.js:35:28)<br>
sls at /Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:14:12<br>
sls at v (/Users/vadorequest/dev/serverless-with-next/node_modules/next/dist/server/router.js:92:29)<br>
sls at Server._callee18$ (/Users/vadorequest/dev/serverless-with-next/node_modules/next/dist/server/index.js:926:24)<br>
sls at tryCatch (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:62:40)<br>
sls at GeneratorFunctionPrototype.invoke [as _invoke] (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:296:22)<br>
sls at GeneratorFunctionPrototype.prototype.(anonymous function) [as next] (/Users/vadorequest/dev/serverless-with-next/node_modules/regenerator-runtime/runtime.js:114:21)<br>
sls at step (/Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:17:30)<br>
sls at /Users/vadorequest/dev/serverless-with-next/node_modules/babel-runtime/helpers/asyncToGenerator.js:28:13<br>
sls at process._tickDomainCallback (internal/process/next_tick.js:135:7)</p>
</details>
<h2 dir="auto">Steps to Reproduce (for bugs)</h2>
<p dir="auto">I've made a specific branch to reproduce this issue at:<br>
<a href="https://github.com/Vadorequest/serverless-with-next/tree/waitUntilReloaded">https://github.com/Vadorequest/serverless-with-next/tree/waitUntilReloaded</a></p>
<h2 dir="auto">Context</h2>
<p dir="auto">I first made it work against AWS and then ran into trouble with a local setup. The main difference between both environment is the <code class="notranslate">dev</code> which is false on AWS.</p>
<p dir="auto">On local env, I run Next.js as a standalone project to get HMR and automatic rebuilds to work.</p>
<h2 dir="auto">Your Environment</h2>
<table role="table">
<thead>
<tr>
<th>Tech</th>
<th>Version</th>
</tr>
</thead>
<tbody>
<tr>
<td>next</td>
<td>5.0.1-canary.9</td>
</tr>
<tr>
<td>node</td>
<td>6.10.3</td>
</tr>
</tbody>
</table> | 0 |
<p dir="auto">by <strong>peter.polakovic</strong>:</p>
<pre class="notranslate">When working with large images I found, that windows port doesn't free large chunks of
memory.
The following code
package main
import "fmt"
import "runtime"
func main() {
for i:=0; i<10; i++ {
a := make([]byte, 5000000)
if a == nil { }
a = nil
runtime.GC()
fmt.Println(i, runtime.MemStats.Alloc, runtime.MemStats.Sys, runtime.MemStats.HeapObjects)
}
fmt.Println()
for i:=0; i<10; i++ {
a := make([]byte, 20000000)
if a == nil { }
a = nil
runtime.GC()
fmt.Println(i, runtime.MemStats.Alloc, runtime.MemStats.Sys, runtime.MemStats.HeapObjects)
}
}
produces the following output
0 5176576 8887420 422
1 5213848 14564476 433
2 5213848 14564476 433
3 5213848 14564476 433
4 5213848 14564476 433
5 5213848 14564476 433
6 5213848 14564476 433
7 5213848 14564476 433
8 5213848 14564476 433
9 5213848 14564476 433
0 20250336 37125244 436
1 40288024 59686012 442
2 60325656 82246780 444
3 80363288 104807548 446
4 100400920 127368316 448
5 100437784 149929084 449
6 120438552 149929084 450
7 120475416 172489852 451
8 140476184 172489852 452
9 140513048 195050620 453
On OS X port first part of output seems similar to second one (allocated memory and
number of heap objects is the same in each iteration).
The same result when using runtime.Malloc() and runtime.Free().
I'm using binary distribution of release.r57.1 on Windows7 (32 bit).</pre> | <p dir="auto">by <strong>prencher</strong>:</p>
<pre class="notranslate">This was originally reported in issue #4829. I posted a comment on it, but as it has
been marked fixed, I don't know if that is monitored. If it is, please by all means
close this issue as a duplicate and accept my apology.
When trying to get packages that depend on cgo, I get the error "argument unused
during compilation: '-fno-eliminate-unused-debug-types'":
$ go get bazil.org/fuse
# bazil.org/fuse
clang: error: argument unused during compilation: '-fno-eliminate-unused-debug-types'
$ go version
go version devel +f4d1cb8d9a91 Thu Sep 19 22:34:33 2013 +1000 darwin/amd64
$ gcc --version
Configured with: --prefix=/Applications/Xcode.app/Contents/Developer/usr
--with-gxx-include-dir=/usr/include/c++/4.2.1
Apple LLVM version 5.0 (clang-500.2.76) (based on LLVM 3.3svn)
Target: x86_64-apple-darwin12.5.0
Thread model: posix
$ clang --version
Apple LLVM version 5.0 (clang-500.2.75) (based on LLVM 3.3svn)
Target: x86_64-apple-darwin12.5.0
Thread model: posix</pre> | 0 |
<p dir="auto"><strong>Migrated issue, originally created by Konsta Vesterinen (<a href="https://github.com/kvesteri">@kvesteri</a>)</strong></p>
<p dir="auto">I'm trying to do atomic SELECT or INSERT as explained here: <a href="http://stackoverflow.com/questions/6722344/select-or-insert-a-row-in-one-command/6722460#6722460" rel="nofollow">http://stackoverflow.com/questions/6722344/select-or-insert-a-row-in-one-command/6722460#6722460</a></p>
<p dir="auto">I'v been able to construct the query, however when I try to execute this query it gives me:</p>
<p dir="auto">AttributeError: 'CompoundSelect' object has no attribute '_returning'</p>
<p dir="auto">Here is a minimal test case for reproducing the issue:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="import sqlalchemy as sa
from sqlalchemy import create_engine
from sqlalchemy.dialects import postgresql
from sqlalchemy.ext.declarative import declarative_base
from sqlalchemy.sql.expression import CTE, exists
Model = declarative_base()
dns = 'postgres://postgres@localhost/my_test_db'
engine = create_engine(dns)
# engine.echo = True
connection = engine.connect()
class MyModel(Model):
__tablename__ = 'my_model'
id = sa.Column(sa.Integer, autoincrement=True, primary_key=True)
key = sa.Column(sa.Unicode(255), nullable=False)
sa.orm.configure_mappers()
Model.metadata.create_all(connection)
table = MyModel.__table__
value = 'some_key'
select = (
sa.select(['*'], from_obj=table)
.where(table.c.key == value)
)
insert = CTE(
table.insert()
.from_select(
[table.c.key],
sa.select([sa.text(value)])
.where(~ exists(select))
)
.returning(sa.text('*')),
name='new_row'
)
query = sa.select(['*'], from_obj=insert).union(select)
print connection.execute(query).fetchall()"><pre class="notranslate"><code class="notranslate">import sqlalchemy as sa
from sqlalchemy import create_engine
from sqlalchemy.dialects import postgresql
from sqlalchemy.ext.declarative import declarative_base
from sqlalchemy.sql.expression import CTE, exists
Model = declarative_base()
dns = 'postgres://postgres@localhost/my_test_db'
engine = create_engine(dns)
# engine.echo = True
connection = engine.connect()
class MyModel(Model):
__tablename__ = 'my_model'
id = sa.Column(sa.Integer, autoincrement=True, primary_key=True)
key = sa.Column(sa.Unicode(255), nullable=False)
sa.orm.configure_mappers()
Model.metadata.create_all(connection)
table = MyModel.__table__
value = 'some_key'
select = (
sa.select(['*'], from_obj=table)
.where(table.c.key == value)
)
insert = CTE(
table.insert()
.from_select(
[table.c.key],
sa.select([sa.text(value)])
.where(~ exists(select))
)
.returning(sa.text('*')),
name='new_row'
)
query = sa.select(['*'], from_obj=insert).union(select)
print connection.execute(query).fetchall()
</code></pre></div>
<p dir="auto">Edit: Another problem is that the CTE isn't adding parenthesis around the insert statement whereas it should be (otherwise PostgreSQL doesn't understand the query).</p> | <p dir="auto"><strong>Migrated issue, originally created by Michael Bayer (<a href="https://github.com/zzzeek">@zzzeek</a>)</strong></p>
<p dir="auto">apparently its all possible now. might as well make a new syntax bonanza out of it:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="WITH upsert AS (
UPDATE metric k SET k.count = k.count + 5
WHERE event = "foo" AND interval = "D" and date = "whatever"
RETURNING k.*
)
INSERT INTO metric (event, interval, date, count)
SELECT ("foo", "D", "whatever", 5)
WHERE NOT EXISTS (
SELECT 1 FROM upsert
);"><pre class="notranslate"><code class="notranslate">WITH upsert AS (
UPDATE metric k SET k.count = k.count + 5
WHERE event = "foo" AND interval = "D" and date = "whatever"
RETURNING k.*
)
INSERT INTO metric (event, interval, date, count)
SELECT ("foo", "D", "whatever", 5)
WHERE NOT EXISTS (
SELECT 1 FROM upsert
);
</code></pre></div>
<p dir="auto"><a href="http://www.postgresql.org/docs/9.1/static/sql-insert.html" rel="nofollow">http://www.postgresql.org/docs/9.1/static/sql-insert.html</a></p>
<hr>
<p dir="auto">Attachments: <a href="../wiki/imported_issue_attachments/2551/ticket_2551.patch">ticket_2551.patch</a> | <a href="../wiki/imported_issue_attachments/2551/CTE.sqlalchemy">CTE.sqlalchemy</a></p> | 1 |
<p dir="auto">To reproduce: run the material gallery demo; select the "Drop Down Button" from the drawer; click on the drop down button labelled "Free"--get red screen and the stack trace below.</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="stork:~/flutter/examples/material_gallery> flutter start && flutter logs --clear
Installing SkyShell.apk on device.
android: --------- beginning of main
android: --------- beginning of system
android: I/flutter : ------------------------------------------------------------------------
android: I/flutter : Exception caught while building _DropDownMenu
android: I/flutter : 'packages/flutter/src/widgets/basic.dart': Failed assertion: line 749: '() {
android: I/flutter : "Positioned must placed inside a Stack";
android: I/flutter : return ancestor is Stack;
android: I/flutter : }' is not true.
android: I/flutter : Stack trace:
android: I/flutter : #0 _AssertionError._throwNew (dart:core-patch/errors_patch.dart:27)
android: I/flutter : #1 Positioned.debugValidateAncestor (packages/flutter/src/widgets/basic.dart:749)
android: I/flutter : #2 ParentDataElement.mount.<anonymous closure> (packages/flutter/src/widgets/framework.dart:1262)
android: I/flutter : #3 ParentDataElement.mount (packages/flutter/src/widgets/framework.dart:1252)
android: I/flutter : #4 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #5 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #6 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #7 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #8 ComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1083)
android: I/flutter : #9 StatefulComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1177)
android: I/flutter : #10 ComponentElement.mount (packages/flutter/src/widgets/framework.dart:1078)
android: I/flutter : #11 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #12 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #13 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #14 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #15 ComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1083)
android: I/flutter : #16 ComponentElement.mount (packages/flutter/src/widgets/framework.dart:1078)
android: I/flutter : #17 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #18 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #19 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #20 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #21 ComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1083)
android: I/flutter : #22 ComponentElement.mount (packages/flutter/src/widgets/framework.dart:1078)
android: I/flutter : #23 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #24 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #25 OneChildRenderObjectElement.mount (packages/flutter/src/widgets/framework.dart:1608)
android: I/flutter : #26 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #27 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #28 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #29 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #30 ComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1083)
android: I/flutter : #31 ComponentElement.mount (packages/flutter/src/widgets/framework.dart:1078)
android: I/flutter : #32 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #33 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #34 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #35 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #36 ComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1083)
android: I/flutter : #37 StatefulComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1177)
android: I/flutter : #38 ComponentElement.mount (packages/flutter/src/widgets/framework.dart:1078)
android: I/flutter : #39 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #40 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #41 OneChildRenderObjectElement.mount (packages/flutter/src/widgets/framework.dart:1608)
android: I/flutter : #42 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #43 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #44 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #45 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #46 ComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1083)
android: I/flutter : #47 StatefulComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1177)
android: I/flutter : #48 ComponentElement.mount (packages/flutter/src/widgets/framework.dart:1078)
android: I/flutter : #49 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #50 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #51 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #52 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #53 ComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1083)
android: I/flutter : #54 ComponentElement.mount (packages/flutter/src/widgets/framework.dart:1078)
android: I/flutter : #55 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #56 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #57 RenderObjectElement.updateChildren (packages/flutter/src/widgets/framework.dart:1487)
android: I/flutter : #58 MultiChildRenderObjectElement.update (packages/flutter/src/widgets/framework.dart:1698)
android: I/flutter : #59 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #60 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #61 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #62 StatefulComponentElement.update (packages/flutter/src/widgets/framework.dart:1195)
android: I/flutter : #63 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #64 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #65 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #66 WidgetFlutterBinding.buildDirtyElements.<anonymous closure> (packages/flutter/src/widgets/binding.dart:58)
android: I/flutter : #67 BuildableElement.lockState (packages/flutter/src/widgets/framework.dart:970)
android: I/flutter : #68 WidgetFlutterBinding.buildDirtyElements (packages/flutter/src/widgets/binding.dart:53)
android: I/flutter : #69 WidgetFlutterBinding.beginFrame (packages/flutter/src/widgets/binding.dart:28)
android: I/flutter : #70 FlutterBinding._handlePersistentFrameCallback (packages/flutter/src/rendering/binding.dart:186)
android: I/flutter : #71 Scheduler.invokeCallback (packages/flutter/src/scheduler/scheduler.dart:221)
android: I/flutter : #72 Scheduler.beginFrame (packages/flutter/src/scheduler/scheduler.dart:203)
android: I/flutter : #73 _beginFrame (file:///src/flutter/engine/src/out/android_Release/gen/sky/bindings/Customhooks.dart:45)
android: I/flutter : ------------------------------------------------------------------------
android: I/flutter : ------------------------------------------------------------------------
android: I/flutter : Exception caught while building _DropDownMenu
android: I/flutter : 'packages/flutter/src/widgets/basic.dart': Failed assertion: line 749: '() {
android: I/flutter : "Positioned must placed inside a Stack";
android: I/flutter : return ancestor is Stack;
android: I/flutter : }' is not true.
android: I/flutter : Stack trace:
android: I/flutter : #0 _AssertionError._throwNew (dart:core-patch/errors_patch.dart:27)
android: I/flutter : #1 Positioned.debugValidateAncestor (packages/flutter/src/widgets/basic.dart:749)
android: I/flutter : #2 ParentDataElement.mount.<anonymous closure> (packages/flutter/src/widgets/framework.dart:1262)
android: I/flutter : #3 ParentDataElement.mount (packages/flutter/src/widgets/framework.dart:1252)
android: I/flutter : #4 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #5 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #6 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #7 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #8 StatefulComponentElement.update (packages/flutter/src/widgets/framework.dart:1195)
android: I/flutter : #9 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #10 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #11 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #12 ProxyElement.update (packages/flutter/src/widgets/framework.dart:1242)
android: I/flutter : #13 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #14 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #15 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #16 StatelessComponentElement.update (packages/flutter/src/widgets/framework.dart:1141)
android: I/flutter : #17 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #18 OneChildRenderObjectElement.update (packages/flutter/src/widgets/framework.dart:1614)
android: I/flutter : #19 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #20 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #21 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #22 ProxyElement.update (packages/flutter/src/widgets/framework.dart:1242)
android: I/flutter : #23 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #24 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #25 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #26 StatefulComponentElement.update (packages/flutter/src/widgets/framework.dart:1195)
android: I/flutter : #27 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #28 OneChildRenderObjectElement.update (packages/flutter/src/widgets/framework.dart:1614)
android: I/flutter : #29 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #30 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #31 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #32 WidgetFlutterBinding.buildDirtyElements.<anonymous closure> (packages/flutter/src/widgets/binding.dart:58)
android: I/flutter : #33 BuildableElement.lockState (packages/flutter/src/widgets/framework.dart:970)
android: I/flutter : #34 WidgetFlutterBinding.buildDirtyElements (packages/flutter/src/widgets/binding.dart:53)
android: I/flutter : #35 WidgetFlutterBinding.beginFrame (packages/flutter/src/widgets/binding.dart:28)
android: I/flutter : #36 FlutterBinding._handlePersistentFrameCallback (packages/flutter/src/rendering/binding.dart:186)
android: I/flutter : #37 Scheduler.invokeCallback (packages/flutter/src/scheduler/scheduler.dart:221)
android: I/flutter : #38 Scheduler.beginFrame (packages/flutter/src/scheduler/scheduler.dart:203)
android: I/flutter : #39 _beginFrame (file:///src/flutter/engine/src/out/android_Release/gen/sky/bindings/Customhooks.dart:45)
android: I/flutter : ------------------------------------------------------------------------"><pre class="notranslate"><code class="notranslate">stork:~/flutter/examples/material_gallery> flutter start && flutter logs --clear
Installing SkyShell.apk on device.
android: --------- beginning of main
android: --------- beginning of system
android: I/flutter : ------------------------------------------------------------------------
android: I/flutter : Exception caught while building _DropDownMenu
android: I/flutter : 'packages/flutter/src/widgets/basic.dart': Failed assertion: line 749: '() {
android: I/flutter : "Positioned must placed inside a Stack";
android: I/flutter : return ancestor is Stack;
android: I/flutter : }' is not true.
android: I/flutter : Stack trace:
android: I/flutter : #0 _AssertionError._throwNew (dart:core-patch/errors_patch.dart:27)
android: I/flutter : #1 Positioned.debugValidateAncestor (packages/flutter/src/widgets/basic.dart:749)
android: I/flutter : #2 ParentDataElement.mount.<anonymous closure> (packages/flutter/src/widgets/framework.dart:1262)
android: I/flutter : #3 ParentDataElement.mount (packages/flutter/src/widgets/framework.dart:1252)
android: I/flutter : #4 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #5 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #6 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #7 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #8 ComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1083)
android: I/flutter : #9 StatefulComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1177)
android: I/flutter : #10 ComponentElement.mount (packages/flutter/src/widgets/framework.dart:1078)
android: I/flutter : #11 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #12 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #13 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #14 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #15 ComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1083)
android: I/flutter : #16 ComponentElement.mount (packages/flutter/src/widgets/framework.dart:1078)
android: I/flutter : #17 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #18 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #19 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #20 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #21 ComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1083)
android: I/flutter : #22 ComponentElement.mount (packages/flutter/src/widgets/framework.dart:1078)
android: I/flutter : #23 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #24 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #25 OneChildRenderObjectElement.mount (packages/flutter/src/widgets/framework.dart:1608)
android: I/flutter : #26 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #27 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #28 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #29 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #30 ComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1083)
android: I/flutter : #31 ComponentElement.mount (packages/flutter/src/widgets/framework.dart:1078)
android: I/flutter : #32 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #33 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #34 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #35 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #36 ComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1083)
android: I/flutter : #37 StatefulComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1177)
android: I/flutter : #38 ComponentElement.mount (packages/flutter/src/widgets/framework.dart:1078)
android: I/flutter : #39 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #40 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #41 OneChildRenderObjectElement.mount (packages/flutter/src/widgets/framework.dart:1608)
android: I/flutter : #42 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #43 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #44 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #45 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #46 ComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1083)
android: I/flutter : #47 StatefulComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1177)
android: I/flutter : #48 ComponentElement.mount (packages/flutter/src/widgets/framework.dart:1078)
android: I/flutter : #49 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #50 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #51 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #52 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #53 ComponentElement._firstBuild (packages/flutter/src/widgets/framework.dart:1083)
android: I/flutter : #54 ComponentElement.mount (packages/flutter/src/widgets/framework.dart:1078)
android: I/flutter : #55 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #56 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #57 RenderObjectElement.updateChildren (packages/flutter/src/widgets/framework.dart:1487)
android: I/flutter : #58 MultiChildRenderObjectElement.update (packages/flutter/src/widgets/framework.dart:1698)
android: I/flutter : #59 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #60 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #61 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #62 StatefulComponentElement.update (packages/flutter/src/widgets/framework.dart:1195)
android: I/flutter : #63 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #64 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #65 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #66 WidgetFlutterBinding.buildDirtyElements.<anonymous closure> (packages/flutter/src/widgets/binding.dart:58)
android: I/flutter : #67 BuildableElement.lockState (packages/flutter/src/widgets/framework.dart:970)
android: I/flutter : #68 WidgetFlutterBinding.buildDirtyElements (packages/flutter/src/widgets/binding.dart:53)
android: I/flutter : #69 WidgetFlutterBinding.beginFrame (packages/flutter/src/widgets/binding.dart:28)
android: I/flutter : #70 FlutterBinding._handlePersistentFrameCallback (packages/flutter/src/rendering/binding.dart:186)
android: I/flutter : #71 Scheduler.invokeCallback (packages/flutter/src/scheduler/scheduler.dart:221)
android: I/flutter : #72 Scheduler.beginFrame (packages/flutter/src/scheduler/scheduler.dart:203)
android: I/flutter : #73 _beginFrame (file:///src/flutter/engine/src/out/android_Release/gen/sky/bindings/Customhooks.dart:45)
android: I/flutter : ------------------------------------------------------------------------
android: I/flutter : ------------------------------------------------------------------------
android: I/flutter : Exception caught while building _DropDownMenu
android: I/flutter : 'packages/flutter/src/widgets/basic.dart': Failed assertion: line 749: '() {
android: I/flutter : "Positioned must placed inside a Stack";
android: I/flutter : return ancestor is Stack;
android: I/flutter : }' is not true.
android: I/flutter : Stack trace:
android: I/flutter : #0 _AssertionError._throwNew (dart:core-patch/errors_patch.dart:27)
android: I/flutter : #1 Positioned.debugValidateAncestor (packages/flutter/src/widgets/basic.dart:749)
android: I/flutter : #2 ParentDataElement.mount.<anonymous closure> (packages/flutter/src/widgets/framework.dart:1262)
android: I/flutter : #3 ParentDataElement.mount (packages/flutter/src/widgets/framework.dart:1252)
android: I/flutter : #4 Element._inflateWidget (packages/flutter/src/widgets/framework.dart:785)
android: I/flutter : #5 Element.updateChild (packages/flutter/src/widgets/framework.dart:664)
android: I/flutter : #6 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #7 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #8 StatefulComponentElement.update (packages/flutter/src/widgets/framework.dart:1195)
android: I/flutter : #9 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #10 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #11 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #12 ProxyElement.update (packages/flutter/src/widgets/framework.dart:1242)
android: I/flutter : #13 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #14 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #15 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #16 StatelessComponentElement.update (packages/flutter/src/widgets/framework.dart:1141)
android: I/flutter : #17 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #18 OneChildRenderObjectElement.update (packages/flutter/src/widgets/framework.dart:1614)
android: I/flutter : #19 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #20 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #21 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #22 ProxyElement.update (packages/flutter/src/widgets/framework.dart:1242)
android: I/flutter : #23 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #24 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #25 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #26 StatefulComponentElement.update (packages/flutter/src/widgets/framework.dart:1195)
android: I/flutter : #27 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #28 OneChildRenderObjectElement.update (packages/flutter/src/widgets/framework.dart:1614)
android: I/flutter : #29 Element.updateChild (packages/flutter/src/widgets/framework.dart:657)
android: I/flutter : #30 ComponentElement.performRebuild (packages/flutter/src/widgets/framework.dart:1108)
android: I/flutter : #31 BuildableElement.rebuild (packages/flutter/src/widgets/framework.dart:1037)
android: I/flutter : #32 WidgetFlutterBinding.buildDirtyElements.<anonymous closure> (packages/flutter/src/widgets/binding.dart:58)
android: I/flutter : #33 BuildableElement.lockState (packages/flutter/src/widgets/framework.dart:970)
android: I/flutter : #34 WidgetFlutterBinding.buildDirtyElements (packages/flutter/src/widgets/binding.dart:53)
android: I/flutter : #35 WidgetFlutterBinding.beginFrame (packages/flutter/src/widgets/binding.dart:28)
android: I/flutter : #36 FlutterBinding._handlePersistentFrameCallback (packages/flutter/src/rendering/binding.dart:186)
android: I/flutter : #37 Scheduler.invokeCallback (packages/flutter/src/scheduler/scheduler.dart:221)
android: I/flutter : #38 Scheduler.beginFrame (packages/flutter/src/scheduler/scheduler.dart:203)
android: I/flutter : #39 _beginFrame (file:///src/flutter/engine/src/out/android_Release/gen/sky/bindings/Customhooks.dart:45)
android: I/flutter : ------------------------------------------------------------------------
</code></pre></div> | 1 |
|
<p dir="auto">This is scrapy's default <code class="notranslate">Dupefilter</code> class method <code class="notranslate">request_seen</code></p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="class RFPDupeFilter(BaseDupeFilter):
def request_seen(self, request):
fp = self.request_fingerprint(request)
if fp in self.fingerprints:
return True
self.fingerprints.add(fp)
if self.file:
self.file.write(fp + os.linesep)"><pre class="notranslate"><code class="notranslate">class RFPDupeFilter(BaseDupeFilter):
def request_seen(self, request):
fp = self.request_fingerprint(request)
if fp in self.fingerprints:
return True
self.fingerprints.add(fp)
if self.file:
self.file.write(fp + os.linesep)
</code></pre></div>
<p dir="auto">While implementing a custom dupefilter. i cannot retrieve the <code class="notranslate">spider</code> object from this class unlike other scrapy middleware</p>
<p dir="auto">Is there any way i can know which <code class="notranslate">spider</code> object this is? so i can customize it via a spider on spider basis?</p>
<p dir="auto">Also i cannot just implement a middleware which reads urls and puts it into a list & checks duplicates instead of a custom dupefilter. This is because i need to pause/resume crawls and need scrapy to store the request fingerprint by default using the <code class="notranslate">JOBDIR</code> setting</p> | <h2 dir="auto">Summary</h2>
<p dir="auto">Let us use DI framework in the pipelines and spiders. My personal favourite: <a href="https://python-dependency-injector.ets-labs.org/introduction/di_in_python.html" rel="nofollow">https://python-dependency-injector.ets-labs.org/introduction/di_in_python.html</a></p>
<h2 dir="auto">Motivation</h2>
<p dir="auto">We have a lot of common code we want to reuse in our scraping framework. Using DI to keep them tidy is nice to have.</p>
<h2 dir="auto">Describe alternatives you've considered</h2>
<p dir="auto">Duplicate all the code and constructions. But only idiots likes duplicated codes or make global variables.</p> | 0 |
<p dir="auto">From <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/dogfish182/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/dogfish182">@dogfish182</a></p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://cloud.githubusercontent.com/assets/5390419/12177701/d68119de-b56d-11e5-9a9d-d92e1bdca938.png"><img src="https://cloud.githubusercontent.com/assets/5390419/12177701/d68119de-b56d-11e5-9a9d-d92e1bdca938.png" alt="2016-01-07_18-35-46" style="max-width: 100%;"></a></p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="unable to get local issuer certificate"><pre class="notranslate"><code class="notranslate">unable to get local issuer certificate
</code></pre></div> | <ul dir="auto">
<li>VSCode Version: 1.5</li>
<li>OS Version: 10.11.6</li>
</ul>
<p dir="auto">Steps to Reproduce:</p>
<ol dir="auto">
<li>Install a transparent man-in-the-middle firewall with a self signed cert and be sure that all traffic is being intercepted.</li>
<li>Install VSCode 1.4.</li>
<li>Notice that VSCode can't update extensions due to MITM.</li>
<li>Set http.proxyStrictSSL=false</li>
<li>Notice that VSCode can now update extensions or search extensions, etc.</li>
<li>Upgrade to 1.5</li>
<li>Notice that VSCode can no longer speak to the internet when trying to search for an extension</li>
</ol> | 1 |
<p dir="auto"><strong><a href="https://jira.spring.io/secure/ViewProfile.jspa?name=magnus" rel="nofollow">Magnus Heino</a></strong> opened <strong><a href="https://jira.spring.io/browse/SPR-7478?redirect=false" rel="nofollow">SPR-7478</a></strong> and commented</p>
<p dir="auto">Hi. I just upgraded from 3.0.2 to 3.0.4 and my log is flooded with WARN messages. The app still seems to work, but jms communication seems to take some extra time...</p>
<p dir="auto">Looking at the code in the debugger, and the error message, a session seems to be null where expected not to. Is there a workaround, or something that I have missed in my config that is exposed now in 3.0.4, or is this a bug?</p>
<p dir="auto"><bean id="transactionalMessageListenerContainer" class="org.springframework.jms.listener.DefaultMessageListenerContainer" abstract="true"><br>
<property name="cacheLevelName" value="CACHE_NONE"/><br>
<property name="concurrentConsumers" value="1"/><br>
<property name="connectionFactory" ref="connectionFactory"/><br>
<property name="sessionTransacted" value="true"/><br>
<property name="transactionManager" ref="transactionManager"/><br>
<property name="autoStartup" value="false" /><br>
</bean><br>
<bean id="addressDefectTrackerContainer" parent="transactionalMessageListenerContainer"><br>
<property name="concurrentConsumers" value="#{settings.addressDefectTracker.messageListener.concurrentConsumers}" /><br>
<property name="maxConcurrentConsumers" value="#{settings.addressDefectTracker.messageListener.maxConcurrentConsumers}" /><br>
<property name="idleTaskExecutionLimit" value="#{settings.addressDefectTracker.messageListener.idleTaskExecutionLimit}" /><br>
<property name="maxMessagesPerTask" value="#{settings.addressDefectTracker.messageListener.maxMessagesPerTask}" /><br>
<property name="destination" ref="jmsAddressDefectTrackerChannel" /><br>
</bean></p>
<p dir="auto"><si-jms:message-driven-channel-adapter<br>
channel="defectTrackerChannel" container="addressDefectTrackerContainer" acknowledge="transacted" connection-factory="connectionFactory"/></p>
<p dir="auto">2010-08-20 12:00:44,312 [addressDefectTrackerContainer-2] DEBUG Initiating transaction rollback on listener exception at org.springframework.jms.listener.DefaultMessageListenerContainer.rollbackOnException(AbstractPollingMessageListenerContainer.java:388)<br>
java.lang.NullPointerException<br>
at org.springframework.jms.listener.AbstractMessageListenerContainer.commitIfNecessary(AbstractMessageListenerContainer.java:571)<br>
at org.springframework.jms.listener.AbstractPollingMessageListenerContainer.doReceiveAndExecute(AbstractPollingMessageListenerContainer.java:354)<br>
at org.springframework.jms.listener.AbstractPollingMessageListenerContainer.receiveAndExecute(AbstractPollingMessageListenerContainer.java:241)<br>
at org.springframework.jms.listener.DefaultMessageListenerContainer$AsyncMessageListenerInvoker.invokeListener(DefaultMessageListenerContainer.java:1056)<br>
at org.springframework.jms.listener.DefaultMessageListenerContainer$AsyncMessageListenerInvoker.executeOngoingLoop(DefaultMessageListenerContainer.java:1048)<br>
at org.springframework.jms.listener.DefaultMessageListenerContainer$AsyncMessageListenerInvoker.run(DefaultMessageListenerContainer.java:947)<br>
at java.lang.Thread.run(Thread.java:619)<br>
2010-08-20 12:00:49,349 [addressDefectTrackerContainer-2] WARN Setup of JMS message listener invoker failed for destination 'Queue[LM.UTV_maghei.ORIGO.GEODATA.ADDRESS.DEFECTTRACKER]' - trying to recover. Cause: null at org.springframework.jms.listener.DefaultMessageListenerContainer.handleListenerSetupFailure(DefaultMessageListenerContainer.java:818)<br>
java.lang.NullPointerException<br>
at org.springframework.jms.listener.AbstractMessageListenerContainer.commitIfNecessary(AbstractMessageListenerContainer.java:571)<br>
at org.springframework.jms.listener.AbstractPollingMessageListenerContainer.doReceiveAndExecute(AbstractPollingMessageListenerContainer.java:354)<br>
at org.springframework.jms.listener.AbstractPollingMessageListenerContainer.receiveAndExecute(AbstractPollingMessageListenerContainer.java:241)<br>
at org.springframework.jms.listener.DefaultMessageListenerContainer$AsyncMessageListenerInvoker.invokeListener(DefaultMessageListenerContainer.java:1056)<br>
at org.springframework.jms.listener.DefaultMessageListenerContainer$AsyncMessageListenerInvoker.executeOngoingLoop(DefaultMessageListenerContainer.java:1048)<br>
at org.springframework.jms.listener.DefaultMessageListenerContainer$AsyncMessageListenerInvoker.run(DefaultMessageListenerContainer.java:947)<br>
at java.lang.Thread.run(Thread.java:619)<br>
2010-08-20 12:00:49,350 [addressDefectTrackerContainer-2] INFO Successfully refreshed JMS Connection at org.springframework.jms.listener.DefaultMessageListenerContainer.refreshConnectionUntilSuccessful(DefaultMessageListenerContainer.java:862)</p>
<hr>
<p dir="auto"><strong>Affects:</strong> 3.0.4</p>
<p dir="auto"><strong>Issue Links:</strong></p>
<ul dir="auto">
<li><a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="398108115" data-permission-text="Title is private" data-url="https://github.com/spring-projects/spring-framework/issues/12306" data-hovercard-type="issue" data-hovercard-url="/spring-projects/spring-framework/issues/12306/hovercard" href="https://github.com/spring-projects/spring-framework/issues/12306">#12306</a> NPE in DefaultMessageListenerContainer while committing and no message was received (<em><strong>"is duplicated by"</strong></em>)</li>
<li><a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="398107789" data-permission-text="Title is private" data-url="https://github.com/spring-projects/spring-framework/issues/12259" data-hovercard-type="issue" data-hovercard-url="/spring-projects/spring-framework/issues/12259/hovercard" href="https://github.com/spring-projects/spring-framework/issues/12259">#12259</a> NPE in AbstractPollingMessageListenerContainer when committing transaction after message receive timeout (<em><strong>"is duplicated by"</strong></em>)</li>
<li><a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="398106175" data-permission-text="Title is private" data-url="https://github.com/spring-projects/spring-framework/issues/12024" data-hovercard-type="issue" data-hovercard-url="/spring-projects/spring-framework/issues/12024/hovercard" href="https://github.com/spring-projects/spring-framework/issues/12024">#12024</a> DefaultMessageListenerContainer transacted session race condition if no messages for a while when using shared/cached sessions</li>
</ul>
<p dir="auto"><strong>Referenced from:</strong> commits <a class="commit-link" data-hovercard-type="commit" data-hovercard-url="https://github.com/spring-projects/spring-framework/commit/3e5aca86d58f37ffd225c9e43e08296bdd12a638/hovercard" href="https://github.com/spring-projects/spring-framework/commit/3e5aca86d58f37ffd225c9e43e08296bdd12a638"><tt>3e5aca8</tt></a></p>
<p dir="auto">4 votes, 9 watchers</p> | <p dir="auto"><strong><a href="https://jira.spring.io/secure/ViewProfile.jspa?name=cbeams" rel="nofollow">Chris Beams</a></strong> opened <strong><a href="https://jira.spring.io/browse/SPR-8203?redirect=false" rel="nofollow">SPR-8203</a></strong> and commented</p>
<p dir="auto">Early in Spring 3.1 M1 a reserved 'default' profile name was introduced. We backed it out under the rationale that it's less magical to express defaults explicitly through <code class="notranslate">spring.profiles.default</code> or <code class="notranslate">ConfigurableEnvironment#setDefaultProfiles</code>.</p>
<p dir="auto">This can be (re-)introduced; just wanted to keep it simple to start and see if it's truly wanted/needed.</p>
<hr>
<p dir="auto"><strong>Affects:</strong> 3.1 M1</p>
<p dir="auto"><strong>Issue Links:</strong></p>
<ul dir="auto">
<li><a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="398118358" data-permission-text="Title is private" data-url="https://github.com/spring-projects/spring-framework/issues/13919" data-hovercard-type="issue" data-hovercard-url="/spring-projects/spring-framework/issues/13919/hovercard" href="https://github.com/spring-projects/spring-framework/issues/13919">#13919</a> Add ability to specify whether a bean profile is active by default (<em><strong>"is duplicated by"</strong></em>)</li>
<li><a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="398108948" data-permission-text="Title is private" data-url="https://github.com/spring-projects/spring-framework/issues/12434" data-hovercard-type="issue" data-hovercard-url="/spring-projects/spring-framework/issues/12434/hovercard" href="https://github.com/spring-projects/spring-framework/issues/12434">#12434</a> Support 'default' profile (<em><strong>"supersedes"</strong></em>)</li>
</ul>
<p dir="auto"><strong>Referenced from:</strong> commits <a class="commit-link" data-hovercard-type="commit" data-hovercard-url="https://github.com/spring-projects/spring-framework/commit/c4a13507f0bcb10bd870543e8a1f1f130505c321/hovercard" href="https://github.com/spring-projects/spring-framework/commit/c4a13507f0bcb10bd870543e8a1f1f130505c321"><tt>c4a1350</tt></a></p>
<p dir="auto">4 votes, 6 watchers</p> | 0 |
<p dir="auto">if all historicals (and peon) respond within the query timeout specified and it is broker spending time doing the merge (usually in the case of merging lots of sketches) then timeout wouldn't be honored irrespective of however long broker takes to merge things.</p>
<p dir="auto">on the client side, clients typically use the same timeout to close the connection... so broker doing all the work is eventually wasted.</p>
<p dir="auto">one possibly simple (maybe crude) fix is that broker puts query processing <code class="notranslate">startTime</code> in query context and all things taking time to process the query continuously watch the time spent and fail if it exceeds. I'm imagining probably in QueryToolChest.mergeResults(..) impl and may be in the QueryResource where it is streaming the results to response.</p>
<p dir="auto">another alternative to <code class="notranslate">startTime</code> would be to do query processing in a thread pool and cancel the future on query timeout .... druid code that takes time in processing the query would continuously check if the thread is interrupted.</p> | <p dir="auto">Currently If I specify a <code class="notranslate">timeout</code> query context when querying a broker, that timeout is not honored. It is entirely possible to set a timeout of 90s and have the broker hog resources for the query for 15m while it tries to complete the request.</p>
<p dir="auto">The expected behavior is that the broker terminates the returning data stream once <code class="notranslate">timeout</code> has expired, and also that it cancels any outstanding query efforts on the historicals related to the query.</p>
<p dir="auto">The current state of timeouts seems to be enforced on a per-segment basis, and not a per-query basis. Meaning that if historical nodes take an abnormally long time on queries, the overall query can take a very long time compared to the timeout.</p>
<p dir="auto">This stems from 2 things:</p>
<ol dir="auto">
<li>Lack of checking in QueryResource for commands taking too long. (see <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="84279140" data-permission-text="Title is private" data-url="https://github.com/apache/druid/issues/1416" data-hovercard-type="pull_request" data-hovercard-url="/apache/druid/pull/1416/hovercard" href="https://github.com/apache/druid/pull/1416">#1416</a> )</li>
<li>Lack of ability to interrupt streaming results properly.</li>
</ol>
<p dir="auto">Part of this comes from the assumption stated in <code class="notranslate">io.druid.query.AsyncQueryRunner::run</code></p>
<div class="highlight highlight-source-java notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content=" //Note: this is assumed that baseRunner does most of the work eagerly on call to the
//run() method and resulting sequence accumulate/yield is fast."><pre class="notranslate"> <span class="pl-c">//Note: this is assumed that baseRunner does most of the work eagerly on call to the</span>
<span class="pl-c">//run() method and resulting sequence accumulate/yield is fast.</span></pre></div>
<p dir="auto">because the a lot of the query workload is actually lazily done in yield.</p> | 1 |
<blockquote>
<p dir="auto">Issue originally made by <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/ondras/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/ondras">@ondras</a></p>
</blockquote>
<h3 dir="auto">Bug information</h3>
<ul dir="auto">
<li><strong>Babel version:</strong> 6.4.0</li>
<li><strong>Node version:</strong> 0.10.25</li>
<li><strong>npm version:</strong> 3.5.2</li>
</ul>
<h3 dir="auto">Input code</h3>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="export function a() {
alert("a");
}
function b() {
a();
}
b();
"><pre class="notranslate"><code class="notranslate">export function a() {
alert("a");
}
function b() {
a();
}
b();
</code></pre></div>
<h3 dir="auto">Description</h3>
<p dir="auto">Exported and non-exported functions end up in different scopes, which makes exported functions uncallable from within non-exported functions.</p>
<p dir="auto">Output code (using SystemJS module output):</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="System.register("...", [], function (_export, _context) {
function b() {
a();
}
return {
setters: [],
execute: function () {
function a() {
alert("a");
}
_export("a", a);
b();
}
};
});
"><pre class="notranslate"><code class="notranslate">System.register("...", [], function (_export, _context) {
function b() {
a();
}
return {
setters: [],
execute: function () {
function a() {
alert("a");
}
_export("a", a);
b();
}
};
});
</code></pre></div> | <p dir="auto">So, we have various files that are being compiled into a single library, many of which depend on core-js. The syntax <code class="notranslate">import core from 'core-js'</code> is invalid because core-js does not export default. However, compiling <code class="notranslate">import * as core from 'core-js'</code> in multiple files throws an error: Duplicate declaration "core".</p> | 0 |
<p dir="auto">When you select an item from the affix menu (see the bootstrap website) The item you selected scrolls under the navbar and can not be seen.</p> | <p dir="auto">I have a navbar-fixed-top navbar. Links that include "scroll to id" (as in href="#myid") don't adjust scroll position to account for the 50px of the navbar.</p>
<p dir="auto">Can you add some JS to detect this situation on the URL and adjust the scroll position back by the height of the navbar?</p> | 1 |
<h5 dir="auto">ISSUE TYPE</h5>
<ul dir="auto">
<li>Bug Report</li>
</ul>
<h5 dir="auto">COMPONENT NAME</h5>
<p dir="auto">ec2_group</p>
<h5 dir="auto">ANSIBLE VERSION</h5>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="ansible 2.2.1.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides"><pre class="notranslate"><code class="notranslate">ansible 2.2.1.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
</code></pre></div>
<p dir="auto">(but code review suggests that later versions likely also susceptible)</p>
<h5 dir="auto">CONFIGURATION</h5>
<p dir="auto">n/a</p>
<h5 dir="auto">OS / ENVIRONMENT</h5>
<p dir="auto">n/a</p>
<h5 dir="auto">SUMMARY</h5>
<p dir="auto">A playbook using an <code class="notranslate">ec2_group</code> module with a rule whose <code class="notranslate">cidr_ip</code> value has any host bits set will execute successfully at most once, but will fail to execute thereafter.</p>
<h5 dir="auto">STEPS TO REPRODUCE</h5>
<div class="highlight highlight-source-yaml notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="- hosts: localhost
tasks:
- name: Test case
ec2_group:
description: Test SG
name: Test SG
region: us-east-1
rules: [{proto: all, cidr_ip: '10.0.0.1/8'}]
rules_egress: [{proto: all, cidr_ip: '0.0.0.0/0'}]
vpc_id: FIXME
connection: local
become: false
gather_facts: false"><pre class="notranslate">- <span class="pl-ent">hosts</span>: <span class="pl-s">localhost</span>
<span class="pl-ent">tasks</span>:
- <span class="pl-ent">name</span>: <span class="pl-s">Test case</span>
<span class="pl-ent">ec2_group</span>:
<span class="pl-ent">description</span>: <span class="pl-s">Test SG</span>
<span class="pl-ent">name</span>: <span class="pl-s">Test SG</span>
<span class="pl-ent">region</span>: <span class="pl-s">us-east-1</span>
<span class="pl-ent">rules</span>: <span class="pl-s">[{proto: all, cidr_ip: '10.0.0.1/8'}]</span>
<span class="pl-ent">rules_egress</span>: <span class="pl-s">[{proto: all, cidr_ip: '0.0.0.0/0'}]</span>
<span class="pl-ent">vpc_id</span>: <span class="pl-s">FIXME</span>
<span class="pl-ent">connection</span>: <span class="pl-s">local</span>
<span class="pl-ent">become</span>: <span class="pl-c1">false</span>
<span class="pl-ent">gather_facts</span>: <span class="pl-c1">false</span></pre></div>
<p dir="auto">(replace <code class="notranslate">vpc_id</code>, and maybe <code class="notranslate">region</code>, as applicable)</p>
<h5 dir="auto">EXPECTED RESULTS</h5>
<p dir="auto"><em>Preferably</em> the rule's validation should fail entirely every time.</p>
<p dir="auto">Alternatively, the CIDR could be sanitized by Ansible at run-time <em>before</em> it compares the configured rules with the ones currently active in AWS.</p>
<h5 dir="auto">ACTUAL RESULTS</h5>
<p dir="auto">Initial run executes fine; the rule is created on the security group, however the CIDR is masked so that only the network bits are set (ie 10.0.0.0/8). Thus what shows up in AWS is slightly different than what the user requested, but it works (at least for the first time).</p>
<p dir="auto">Subsequent runs, however, will all fail with this error:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="Traceback (most recent call last):
File "/tmp/ansible_y6BGXK/ansible_module_ec2_group.py", line 478, in <module>
main()
File "/tmp/ansible_y6BGXK/ansible_module_ec2_group.py", line 373, in main
group.authorize(rule['proto'], rule['from_port'], rule['to_port'], thisip, grantGroup)
File "/usr/lib/python2.7/dist-packages/boto/ec2/securitygroup.py", line 203, in authorize
dry_run=dry_run)
File "/usr/lib/python2.7/dist-packages/boto/ec2/connection.py", line 3192, in authorize_security_group
params, verb='POST')
File "/usr/lib/python2.7/dist-packages/boto/connection.py", line 1223, in get_status
raise self.ResponseError(response.status, response.reason, body)
boto.exception.EC2ResponseError: EC2ResponseError: 400 Bad Request
<?xml version="1.0" encoding="UTF-8"?>
<Response><Errors><Error><Code>InvalidPermission.Duplicate</Code><Message>the specified rule "peer: 10.0.0.0/8, ALL, ALLOW" already exists</Message></Error></Errors><RequestID>963c0e1b-bcdb-4ab3-815a-89c7ce1b2204</RequestID></Response>
fatal: [localhost]: FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_name": "ec2_group"
},
"module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_y6BGXK/ansible_module_ec2_group.py\", line 478, in <module>\n main()\n File \"/tmp/ansible_y6BGXK/ansible_module_ec2_group.py\", line 373, in main\n group.authorize(rule['proto'], rule['from_port'], rule['to_port'], thisip, grantGroup)\n File \"/usr/lib/python2.7/dist-packages/boto/ec2/securitygroup.py\", line 203, in authorize\n dry_run=dry_run)\n File \"/usr/lib/python2.7/dist-packages/boto/ec2/connection.py\", line 3192, in authorize_security_group\n params, verb='POST')\n File \"/usr/lib/python2.7/dist-packages/boto/connection.py\", line 1223, in get_status\n raise self.ResponseError(response.status, response.reason, body)\nboto.exception.EC2ResponseError: EC2ResponseError: 400 Bad Request\n<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n<Response><Errors><Error><Code>InvalidPermission.Duplicate</Code><Message>the specified rule \"peer: 10.0.0.0/8, ALL, ALLOW\" already exists</Message></Error></Errors><RequestID>963c0e1b-bcdb-4ab3-815a-89c7ce1b2204</RequestID></Response>\n",
"module_stdout": "",
"msg": "MODULE FAILURE"
}"><pre lang="An" class="notranslate"><code class="notranslate">Traceback (most recent call last):
File "/tmp/ansible_y6BGXK/ansible_module_ec2_group.py", line 478, in <module>
main()
File "/tmp/ansible_y6BGXK/ansible_module_ec2_group.py", line 373, in main
group.authorize(rule['proto'], rule['from_port'], rule['to_port'], thisip, grantGroup)
File "/usr/lib/python2.7/dist-packages/boto/ec2/securitygroup.py", line 203, in authorize
dry_run=dry_run)
File "/usr/lib/python2.7/dist-packages/boto/ec2/connection.py", line 3192, in authorize_security_group
params, verb='POST')
File "/usr/lib/python2.7/dist-packages/boto/connection.py", line 1223, in get_status
raise self.ResponseError(response.status, response.reason, body)
boto.exception.EC2ResponseError: EC2ResponseError: 400 Bad Request
<?xml version="1.0" encoding="UTF-8"?>
<Response><Errors><Error><Code>InvalidPermission.Duplicate</Code><Message>the specified rule "peer: 10.0.0.0/8, ALL, ALLOW" already exists</Message></Error></Errors><RequestID>963c0e1b-bcdb-4ab3-815a-89c7ce1b2204</RequestID></Response>
fatal: [localhost]: FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_name": "ec2_group"
},
"module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_y6BGXK/ansible_module_ec2_group.py\", line 478, in <module>\n main()\n File \"/tmp/ansible_y6BGXK/ansible_module_ec2_group.py\", line 373, in main\n group.authorize(rule['proto'], rule['from_port'], rule['to_port'], thisip, grantGroup)\n File \"/usr/lib/python2.7/dist-packages/boto/ec2/securitygroup.py\", line 203, in authorize\n dry_run=dry_run)\n File \"/usr/lib/python2.7/dist-packages/boto/ec2/connection.py\", line 3192, in authorize_security_group\n params, verb='POST')\n File \"/usr/lib/python2.7/dist-packages/boto/connection.py\", line 1223, in get_status\n raise self.ResponseError(response.status, response.reason, body)\nboto.exception.EC2ResponseError: EC2ResponseError: 400 Bad Request\n<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n<Response><Errors><Error><Code>InvalidPermission.Duplicate</Code><Message>the specified rule \"peer: 10.0.0.0/8, ALL, ALLOW\" already exists</Message></Error></Errors><RequestID>963c0e1b-bcdb-4ab3-815a-89c7ce1b2204</RequestID></Response>\n",
"module_stdout": "",
"msg": "MODULE FAILURE"
}
</code></pre></div> | <h5 dir="auto">Issue Type: Bug Report</h5>
<h5 dir="auto">Component: core</h5>
<h5 dir="auto">Ansible Version:</h5>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="ansible 2.0.0.2
config file =
configured module search path = Default w/o overrides"><pre class="notranslate"><code class="notranslate">ansible 2.0.0.2
config file =
configured module search path = Default w/o overrides
</code></pre></div>
<h5 dir="auto">Ansible Configuration:</h5>
<p dir="auto">n/a</p>
<h5 dir="auto">Environment:</h5>
<p dir="auto">python==2.7.10<br>
osx==10.10.5</p>
<h5 dir="auto">Summary:</h5>
<p dir="auto">Role dependencies are run more than once.</p>
<h5 dir="auto">Steps To Reproduce:</h5>
<p dir="auto">here are my files:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="# inventory
localhost"><pre class="notranslate"><code class="notranslate"># inventory
localhost
</code></pre></div>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="# play.yml
---
- name: dependency test with role tags
hosts: localhost
connection: local
roles:
- role: base
tags:
- always
- role: feature
tags:
- always"><pre class="notranslate"><code class="notranslate"># play.yml
---
- name: dependency test with role tags
hosts: localhost
connection: local
roles:
- role: base
tags:
- always
- role: feature
tags:
- always
</code></pre></div>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="# roles/base/meta/main.yml
---
allow_duplicates: no"><pre class="notranslate"><code class="notranslate"># roles/base/meta/main.yml
---
allow_duplicates: no
</code></pre></div>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="# roles/base/tasks/main.yml
---
- name: base role
debug:
msg: base role"><pre class="notranslate"><code class="notranslate"># roles/base/tasks/main.yml
---
- name: base role
debug:
msg: base role
</code></pre></div>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="# roles/feature/meta/main.yml
---
allow_duplicates: no
dependencies:
- role: base
tags:
- always"><pre class="notranslate"><code class="notranslate"># roles/feature/meta/main.yml
---
allow_duplicates: no
dependencies:
- role: base
tags:
- always
</code></pre></div>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="# roles/feature/tasks/main.yml
---
- name: feature role
debug:
msg: feature role"><pre class="notranslate"><code class="notranslate"># roles/feature/tasks/main.yml
---
- name: feature role
debug:
msg: feature role
</code></pre></div>
<p dir="auto">I call it with:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="ansible-playbook -i inventory play.yml --list-tasks"><pre class="notranslate"><code class="notranslate">ansible-playbook -i inventory play.yml --list-tasks
</code></pre></div>
<h5 dir="auto">Expected Results:</h5>
<p dir="auto">The base role is executed once.</p>
<p dir="auto">Using ansible==1.9.4 the play works as expected:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="playbook: play.yml
play #1 (dependency test with role tags): TAGS: []
base role TAGS: [always]
feature role TAGS: [always]"><pre class="notranslate"><code class="notranslate">playbook: play.yml
play #1 (dependency test with role tags): TAGS: []
base role TAGS: [always]
feature role TAGS: [always]
</code></pre></div>
<h5 dir="auto">Actual Results:</h5>
<p dir="auto">The base role is executed twice.</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="playbook: play.yml
play #1 (localhost): dependency test with role tags TAGS: []
tasks:
base : base role TAGS: [always]
base : base role TAGS: [always]
feature : feature role TAGS: [always]"><pre class="notranslate"><code class="notranslate">playbook: play.yml
play #1 (localhost): dependency test with role tags TAGS: []
tasks:
base : base role TAGS: [always]
base : base role TAGS: [always]
feature : feature role TAGS: [always]
</code></pre></div> | 0 |
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://cloud.githubusercontent.com/assets/18270980/18379550/7454aae2-7673-11e6-9e98-7e8f340cd7ad.png"><img src="https://cloud.githubusercontent.com/assets/18270980/18379550/7454aae2-7673-11e6-9e98-7e8f340cd7ad.png" alt="bug" style="max-width: 100%;"></a><br>
Challenge <a href="https://www.freecodecamp.com/en/challenges/html5-and-css/change-the-color-of-text" rel="nofollow">change-the-color-of-text</a> has an issue.<br>
User Agent is: <code class="notranslate">Mozilla/5.0 (Windows NT 6.3; WOW64; rv:48.0) Gecko/20100101 Firefox/48.0</code>.</p>
<p dir="auto">I've been trying to solve this, but I think everything is alright. I have no idea how to fix it.</p>
<p dir="auto">`</p><h2 dir="auto">CatPhotoApp</h2><p dir="auto"></p>
<p dir="auto">Kitty ipsum dolor sit amet, shed everywhere shed everywhere stretching attack your ankles chase the red dot, hairball run catnip eat the grass sniff.</p>` | <p dir="auto">Challenge <a href="https://www.freecodecamp.com/en/challenges/html5-and-css/change-the-color-of-text" rel="nofollow">change-the-color-of-text</a> has an issue.<br>
User Agent is: <code class="notranslate">Mozilla/5.0 (Windows NT 10.0; WOW64; rv:48.0) Gecko/20100101 Firefox/48.0</code>.<br>
Please describe how to reproduce this issue, and include links to screenshots if possible.</p>
<div class="highlight highlight-text-html-basic notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="
<h2>CatPhotoApp</h2>
<p>Kitty ipsum dolor sit amet, shed everywhere shed everywhere stretching attack your ankles chase the red dot, hairball run catnip eat the grass sniff.</p>
<h2>CatPhotoApp</h2>
<p>Kitty ipsum dolor sit amet, shed everywhere shed everywhere stretching attack your ankles chase the red dot, hairball run catnip eat the grass sniff.</p>
"><pre class="notranslate"><span class="pl-kos"><</span><span class="pl-ent">h2</span><span class="pl-kos">></span>CatPhotoApp<span class="pl-kos"></</span><span class="pl-ent">h2</span><span class="pl-kos">></span>
<span class="pl-kos"><</span><span class="pl-ent">p</span><span class="pl-kos">></span>Kitty ipsum dolor sit amet, shed everywhere shed everywhere stretching attack your ankles chase the red dot, hairball run catnip eat the grass sniff.<span class="pl-kos"></</span><span class="pl-ent">p</span><span class="pl-kos">></span>
<span class="pl-kos"><</span><span class="pl-ent">h2</span><span class="pl-kos">></span>CatPhotoApp<span class="pl-kos"></</span><span class="pl-ent">h2</span><span class="pl-kos">></span>
<span class="pl-kos"><</span><span class="pl-ent">p</span><span class="pl-kos">></span>Kitty ipsum dolor sit amet, shed everywhere shed everywhere stretching attack your ankles chase the red dot, hairball run catnip eat the grass sniff.<span class="pl-kos"></</span><span class="pl-ent">p</span><span class="pl-kos">></span></pre></div>
<div class="highlight highlight-text-html-basic notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="
<h2 style="colr:red;">CatPhotoApp</h2>
<p>Kitty ipsum dolor sit amet, shed everywhere shed everywhere stretching attack your ankles chase the red dot, hairball run catnip eat the grass sniff.</p>
<h2 style="colr:red;">CatPhotoApp</h2>
<p>Kitty ipsum dolor sit amet, shed everywhere shed everywhere stretching attack your ankles chase the red dot, hairball run catnip eat the grass sniff.</p>
"><pre class="notranslate"><span class="pl-kos"><</span><span class="pl-ent">h2</span> <span class="pl-c1">style</span>="<span class="pl-s">colr:red;</span>"<span class="pl-kos">></span>CatPhotoApp<span class="pl-kos"></</span><span class="pl-ent">h2</span><span class="pl-kos">></span>
<span class="pl-kos"><</span><span class="pl-ent">p</span><span class="pl-kos">></span>Kitty ipsum dolor sit amet, shed everywhere shed everywhere stretching attack your ankles chase the red dot, hairball run catnip eat the grass sniff.<span class="pl-kos"></</span><span class="pl-ent">p</span><span class="pl-kos">></span>
<span class="pl-kos"><</span><span class="pl-ent">h2</span> <span class="pl-c1">style</span>="<span class="pl-s">colr:red;</span>"<span class="pl-kos">></span>CatPhotoApp<span class="pl-kos"></</span><span class="pl-ent">h2</span><span class="pl-kos">></span>
<span class="pl-kos"><</span><span class="pl-ent">p</span><span class="pl-kos">></span>Kitty ipsum dolor sit amet, shed everywhere shed everywhere stretching attack your ankles chase the red dot, hairball run catnip eat the grass sniff.<span class="pl-kos"></</span><span class="pl-ent">p</span><span class="pl-kos">></span>
</pre></div>
<p dir="auto">This had previously passed. Resubmitted now its failing. Cleared my code and tried to reenter now it doesnt accept the code at all. It resents and clears my text inputed when I click run code.</p> | 1 |
<p dir="auto">I'm unable to make a simple IPC communication between node and spawned electron process working on linux ubuntu. It works fine on Windows.</p>
<p dir="auto">node.js code</p>
<div class="highlight highlight-source-js notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="var electron = require('electron-prebuilt')
var spawn = require('child_process').spawn
var child = spawn(electron, ['electron-script.js'], {
stdio: ['ipc']
})
child.on('message', function (m) {
console.log('Yes it works!')
})"><pre class="notranslate"><span class="pl-k">var</span> <span class="pl-s1">electron</span> <span class="pl-c1">=</span> <span class="pl-en">require</span><span class="pl-kos">(</span><span class="pl-s">'electron-prebuilt'</span><span class="pl-kos">)</span>
<span class="pl-k">var</span> <span class="pl-s1">spawn</span> <span class="pl-c1">=</span> <span class="pl-en">require</span><span class="pl-kos">(</span><span class="pl-s">'child_process'</span><span class="pl-kos">)</span><span class="pl-kos">.</span><span class="pl-c1">spawn</span>
<span class="pl-k">var</span> <span class="pl-s1">child</span> <span class="pl-c1">=</span> <span class="pl-s1">spawn</span><span class="pl-kos">(</span><span class="pl-s1">electron</span><span class="pl-kos">,</span> <span class="pl-kos">[</span><span class="pl-s">'electron-script.js'</span><span class="pl-kos">]</span><span class="pl-kos">,</span> <span class="pl-kos">{</span>
<span class="pl-c1">stdio</span>: <span class="pl-kos">[</span><span class="pl-s">'ipc'</span><span class="pl-kos">]</span>
<span class="pl-kos">}</span><span class="pl-kos">)</span>
<span class="pl-s1">child</span><span class="pl-kos">.</span><span class="pl-en">on</span><span class="pl-kos">(</span><span class="pl-s">'message'</span><span class="pl-kos">,</span> <span class="pl-k">function</span> <span class="pl-kos">(</span><span class="pl-s1">m</span><span class="pl-kos">)</span> <span class="pl-kos">{</span>
<span class="pl-smi">console</span><span class="pl-kos">.</span><span class="pl-en">log</span><span class="pl-kos">(</span><span class="pl-s">'Yes it works!'</span><span class="pl-kos">)</span>
<span class="pl-kos">}</span><span class="pl-kos">)</span></pre></div>
<p dir="auto">electron script</p>
<div class="highlight highlight-source-js notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="process.send('I am ok')
process.exit(0);"><pre class="notranslate"><span class="pl-s1">process</span><span class="pl-kos">.</span><span class="pl-en">send</span><span class="pl-kos">(</span><span class="pl-s">'I am ok'</span><span class="pl-kos">)</span>
<span class="pl-s1">process</span><span class="pl-kos">.</span><span class="pl-en">exit</span><span class="pl-kos">(</span><span class="pl-c1">0</span><span class="pl-kos">)</span><span class="pl-kos">;</span></pre></div>
<p dir="auto">On windows I get back <code class="notranslate">Yes it works!</code> to console. There seems to be no communication happening and nothing is printed on ubuntu. Thank you for the help.</p>
<ul dir="auto">
<li>Electron version: ~ 1.3.1</li>
<li>Operating system: Ubuntu</li>
</ul> | <ul dir="auto">
<li>Electron version:<br>
1.8.4</li>
<li>Operating system:<br>
Arch Linux</li>
<li>Freetype version:<br>
2.9.0</li>
</ul>
<h3 dir="auto">Expected behavior</h3>
<p dir="auto">Electron should render fonts using subpixel rendering</p>
<h3 dir="auto">Actual behavior</h3>
<p dir="auto">Fonts are rendered in greyscale mode.</p>
<h3 dir="auto">How to reproduce</h3>
<p dir="auto">I guess the issue is related to freetype. It started after the new freetype release. It happens to every electron-based application (vscode, atom, ... ).</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/9977153/38354561-4eb45290-38cf-11e8-9856-51591d344963.png"><img src="https://user-images.githubusercontent.com/9977153/38354561-4eb45290-38cf-11e8-9856-51591d344963.png" alt="image" style="max-width: 100%;"></a><br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/9977153/38354586-5e71319e-38cf-11e8-8b7d-d80b4cded9a1.png"><img src="https://user-images.githubusercontent.com/9977153/38354586-5e71319e-38cf-11e8-8b7d-d80b4cded9a1.png" alt="image" style="max-width: 100%;"></a></p> | 0 |
<p dir="auto">Hi,</p>
<p dir="auto">I'm trying to run this model on a system with GeForce GTX 960M:<br>
python /usr/local/lib/python2.7/dist-packages/tensorflow/models/image/cifar10/cifar10_train.py<br>
and get the following output:</p>
<p dir="auto">I tensorflow/stream_executor/dso_loader.cc:101] successfully opened CUDA library libcublas.so.7.0 locally<br>
I tensorflow/stream_executor/dso_loader.cc:101] successfully opened CUDA library libcudnn.so.6.5 locally<br>
I tensorflow/stream_executor/dso_loader.cc:101] successfully opened CUDA library libcufft.so.7.0 locally<br>
I tensorflow/stream_executor/dso_loader.cc:101] successfully opened CUDA library libcuda.so locally<br>
I tensorflow/stream_executor/dso_loader.cc:101] successfully opened CUDA library libcurand.so.7.0 locally<br>
I tensorflow/core/common_runtime/local_device.cc:40] Local device intra op parallelism threads: 4<br>
I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:909] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero<br>
I tensorflow/core/common_runtime/gpu/gpu_init.cc:103] Found device 0 with properties:<br>
name: GeForce GTX 960M<br>
major: 5 minor: 0 memoryClockRate (GHz) 1.176<br>
pciBusID 0000:01:00.0<br>
Total memory: 2.00GiB<br>
Free memory: 1.69GiB<br>
I tensorflow/core/common_runtime/gpu/gpu_init.cc:127] DMA: 0<br>
I tensorflow/core/common_runtime/gpu/gpu_init.cc:137] 0: Y<br>
I tensorflow/core/common_runtime/gpu/gpu_device.cc:702] Creating TensorFlow device (/gpu:0) -> (device: 0, name: GeForce GTX 960M, pci bus id: 0000:01:00.0)<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:42] Allocating 1.50GiB bytes.<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:52] GPU 0 memory begins at 0x501a40000 extends to 0x5616a0000<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 1.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 2.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 4.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 8.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 16.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 32.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 64.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 128.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 256.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 512.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 1.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 2.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 4.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 8.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 16.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 32.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 64.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 128.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 256.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 512.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 1.00GiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 2.00GiB<br>
I tensorflow/core/common_runtime/direct_session.cc:58] Direct session inter op parallelism threads: 4<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: Identity/_79 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/cpu:0", send_device="/job:localhost/replica:0/task:0/gpu:0", send_device_incarnation=1, tensor_name="edge_344_Identity", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: Variable/initial_value/_81 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/cpu:0", send_device="/job:localhost/replica:0/task:0/gpu:0", send_device_incarnation=1, tensor_name="edge_1185_Variable/initial_value", tensor_type=DT_INT32, _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: conv1/biases/Assign/_67 = _Send<a href="conv1/biases/Assign/_66">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_304_conv1/biases/Assign", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: Identity_2/_63 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/cpu:0", send_device="/job:localhost/replica:0/task:0/gpu:0", send_device_incarnation=1, tensor_name="edge_293_Identity_2", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: Identity_3/_55 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/cpu:0", send_device="/job:localhost/replica:0/task:0/gpu:0", send_device_incarnation=1, tensor_name="edge_275_Identity_3", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: conv2/biases/Assign/_51 = _Send<a href="conv2/biases/Assign/_50">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_273_conv2/biases/Assign", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: Identity_4/_47 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/cpu:0", send_device="/job:localhost/replica:0/task:0/gpu:0", send_device_incarnation=1, tensor_name="edge_262_Identity_4", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: Identity_5/_39 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/cpu:0", send_device="/job:localhost/replica:0/task:0/gpu:0", send_device_incarnation=1, tensor_name="edge_233_Identity_5", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: Identity_6/_31 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/cpu:0", send_device="/job:localhost/replica:0/task:0/gpu:0", send_device_incarnation=1, tensor_name="edge_220_Identity_6", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: local3/biases/Assign/_35 = _Send<a href="local3/biases/Assign/_34">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_231_local3/biases/Assign", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: Identity_7/_23 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/cpu:0", send_device="/job:localhost/replica:0/task:0/gpu:0", send_device_incarnation=1, tensor_name="edge_191_Identity_7", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: local4/biases/Assign/_19 = _Send<a href="local4/biases/Assign/_18">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_189_local4/biases/Assign", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: Identity_8/_15 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/cpu:0", send_device="/job:localhost/replica:0/task:0/gpu:0", send_device_incarnation=1, tensor_name="edge_178_Identity_8", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: Identity_9/_7 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/cpu:0", send_device="/job:localhost/replica:0/task:0/gpu:0", send_device_incarnation=1, tensor_name="edge_160_Identity_9", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: softmax_linear/biases/Assign/_3 = _Send<a href="softmax_linear/biases/Assign/_2">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_158_softmax_linear/biases/Assign", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: Identity_1/_71 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/cpu:0", send_device="/job:localhost/replica:0/task:0/gpu:0", send_device_incarnation=1, tensor_name="edge_306_Identity_1", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: conv1/weights/Assign/_75 = _Send<a href="conv1/weights/Assign/_74">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_342_conv1/weights/Assign", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: softmax_linear/weights/Assign/_11 = _Send<a href="softmax_linear/weights/Assign/_10">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_176_softmax_linear/weights/Assign", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: conv1/biases/_64 = _Send<a href="conv1/biases">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_303_conv1/biases", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: conv2/biases/_48 = _Send<a href="conv2/biases">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_272_conv2/biases", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: local3/biases/_32 = _Send<a href="local3/biases">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_230_local3/biases", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: local4/biases/_16 = _Send<a href="local4/biases">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_188_local4/biases", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: softmax_linear/biases/_0 = _Send<a href="softmax_linear/biases">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_157_softmax_linear/biases", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x32983e0 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: init/NoOp_1/_88 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_1192_init/NoOp_1", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/gpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x32983e0 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: Variable/Assign/_84 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_1186_Variable/Assign", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/gpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x32983e0 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: conv1/weights/Assign/_76 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_342_conv1/weights/Assign", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/gpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x32983e0 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: conv1/biases/Assign/_68 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_304_conv1/biases/Assign", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/gpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x32983e0 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: conv2/weights/Assign/_60 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_291_conv2/weights/Assign", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/gpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x32983e0 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: conv2/biases/Assign/_52 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_273_conv2/biases/Assign", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/gpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x32983e0 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: local3/weights/Assign/_44 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_260_local3/weights/Assign", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/gpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x32983e0 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: local3/biases/Assign/_36 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_231_local3/biases/Assign", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/gpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x32983e0 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: local4/weights/Assign/_28 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_218_local4/weights/Assign", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/gpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x32983e0 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: local4/biases/Assign/_20 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_189_local4/biases/Assign", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/gpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x32983e0 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: softmax_linear/weights/Assign/_12 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_176_softmax_linear/weights/Assign", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/gpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x32983e0 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: softmax_linear/biases/Assign/_4 = _Recv<a href="">client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_158_softmax_linear/biases/Assign", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/gpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: conv2/weights/Assign/_59 = _Send<a href="conv2/weights/Assign/_58">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_291_conv2/weights/Assign", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: local4/weights/Assign/_27 = _Send<a href="local4/weights/Assign/_26">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_218_local4/weights/Assign", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x322d680 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: local3/weights/Assign/_43 = _Send<a href="local3/weights/Assign/_42">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_260_local3/weights/Assign", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
W tensorflow/core/common_runtime/executor.cc:1076] 0x32983e0 Compute status: Failed precondition: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]<br>
[[Node: Variable/initial_value/_80 = _HostSend<a href="Variable/initial_value">T=DT_INT32, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/cpu:0", send_device="/job:localhost/replica:0/task:0/gpu:0", send_device_incarnation=1, tensor_name="edge_1185_Variable/initial_value", _device="/job:localhost/replica:0/task:0/gpu:0"</a>]]<br>
Filling queue with 20000 CIFAR images before starting to train. This will take a few minutes.<br>
Traceback (most recent call last):<br>
File "/usr/local/lib/python2.7/dist-packages/tensorflow/models/image/cifar10/cifar10_train.py", line 138, in <br>
tf.app.run()<br>
File "/usr/local/lib/python2.7/dist-packages/tensorflow/python/platform/default/_app.py", line 30, in run<br>
sys.exit(main(sys.argv))<br>
File "/usr/local/lib/python2.7/dist-packages/tensorflow/models/image/cifar10/cifar10_train.py", line 134, in main<br>
train()<br>
File "/usr/local/lib/python2.7/dist-packages/tensorflow/models/image/cifar10/cifar10_train.py", line 94, in train<br>
sess.run(init)<br>
File "/usr/local/lib/python2.7/dist-packages/tensorflow/python/client/session.py", line 368, in run<br>
results = self._do_run(target_list, unique_fetch_targets, feed_dict_string)<br>
File "/usr/local/lib/python2.7/dist-packages/tensorflow/python/client/session.py", line 444, in _do_run<br>
e.code)<br>
tensorflow.python.framework.errors.FailedPreconditionError: Attempting to use uninitialized value conv1/weights<br>
[[Node: conv1/weights/_72 = _Send<a href="conv1/weights">T=DT_FLOAT, client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_341_conv1/weights", _device="/job:localhost/replica:0/task:0/cpu:0"</a>]]</p>
<p dir="auto">Can somebody tell me what's wrong?</p>
<p dir="auto">The mnist model: tensorflow/models/image/mnist/convolutional.py<br>
seems to work well:</p>
<p dir="auto">I tensorflow/stream_executor/dso_loader.cc:101] successfully opened CUDA library libcublas.so.7.0 locally<br>
I tensorflow/stream_executor/dso_loader.cc:101] successfully opened CUDA library libcudnn.so.6.5 locally<br>
I tensorflow/stream_executor/dso_loader.cc:101] successfully opened CUDA library libcufft.so.7.0 locally<br>
I tensorflow/stream_executor/dso_loader.cc:101] successfully opened CUDA library libcuda.so locally<br>
I tensorflow/stream_executor/dso_loader.cc:101] successfully opened CUDA library libcurand.so.7.0 locally<br>
I tensorflow/core/common_runtime/local_device.cc:40] Local device intra op parallelism threads: 4<br>
I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:909] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero<br>
I tensorflow/core/common_runtime/gpu/gpu_init.cc:103] Found device 0 with properties:<br>
name: GeForce GTX 960M<br>
major: 5 minor: 0 memoryClockRate (GHz) 1.176<br>
pciBusID 0000:01:00.0<br>
Total memory: 2.00GiB<br>
Free memory: 1.71GiB<br>
I tensorflow/core/common_runtime/gpu/gpu_init.cc:127] DMA: 0<br>
I tensorflow/core/common_runtime/gpu/gpu_init.cc:137] 0: Y<br>
I tensorflow/core/common_runtime/gpu/gpu_device.cc:702] Creating TensorFlow device (/gpu:0) -> (device: 0, name: GeForce GTX 960M, pci bus id: 0000:01:00.0)<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:42] Allocating 1.51GiB bytes.<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:52] GPU 0 memory begins at 0x501a40000 extends to 0x562600000<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 1.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 2.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 4.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 8.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 16.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 32.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 64.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 128.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 256.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 512.0KiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 1.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 2.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 4.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 8.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 16.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 32.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 64.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 128.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 256.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 512.00MiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 1.00GiB<br>
I tensorflow/core/common_runtime/gpu/gpu_bfc_allocator.cc:66] Creating bin of max chunk size 2.00GiB<br>
I tensorflow/core/common_runtime/direct_session.cc:58] Direct session inter op parallelism threads: 4<br>
Extracting data/train-images-idx3-ubyte.gz<br>
Extracting data/train-labels-idx1-ubyte.gz<br>
Extracting data/t10k-images-idx3-ubyte.gz<br>
Extracting data/t10k-labels-idx1-ubyte.gz<br>
Initialized!<br>
Epoch 0.00<br>
Minibatch loss: 12.054, learning rate: 0.010000<br>
Minibatch error: 90.6%<br>
Validation error: 84.6%<br>
Epoch 0.12<br>
Minibatch loss: 3.297, learning rate: 0.010000<br>
Minibatch error: 6.2%<br>
Validation error: 7.5%</p> | <h3 dir="auto">System information</h3>
<ul dir="auto">
<li><strong>Have I written custom code (as opposed to using a stock example script provided in TensorFlow)</strong>:</li>
<li><strong>OS Platform and Distribution (e.g., Linux Ubuntu 16.04)</strong>: Ubuntu 16.04.4 x64</li>
<li><strong>TensorFlow installed from (source or binary)</strong>: source master and source r1.9</li>
<li><strong>TensorFlow version (use command below)</strong>: 1.9.0rc0 and 1.9.0rc2</li>
<li><strong>Python version</strong>: 3.6.6</li>
<li><strong>Bazel version (if compiling from source)</strong>: 0.15.0</li>
<li><strong>GCC/Compiler version (if compiling from source)</strong>: 5.4.0</li>
<li><strong>CUDA/cuDNN version</strong>: 9.2/7.1</li>
<li><strong>GPU model and memory</strong>: GTX1080Ti 11GB X 2</li>
<li><strong>Exact command to reproduce</strong>:</li>
</ul>
<p dir="auto">checkout to r1.9 or master and ./configure</p>
<p dir="auto">after configure cuda and cudnn</p>
<p dir="auto">and configure NCCL ( default 1.3 or 2.2, whatever I choose, same )</p>
<p dir="auto">the process hangs forever</p> | 0 |
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I have searched the <a href="https://github.com/apache/dubbo/issues">issues</a> of this repository and believe that this is not a duplicate.</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I have checked the <a href="https://github.com/apache/dubbo/blob/master/FAQ.md">FAQ</a> of this repository and believe that this is not a duplicate.</li>
</ul>
<h3 dir="auto">Environment</h3>
<ul dir="auto">
<li>Dubbo version: 2.7.7</li>
<li>Operating System version: windows 10</li>
<li>Java version: 1.8</li>
</ul>
<p dir="auto">现象:<br>
在测试cluster: zone-aware时,ZoneAwareClusterInvoker类的doInvoke方法抛出了java.lang.ClassCastException,异常信息为:<br>
java.lang.ClassCastException: org.apache.dubbo.registry.integration.RegistryDirectory$InvokerDelegate cannot be cast to org.apache.dubbo.rpc.cluster.support.wrapper.MockClusterInvoker<br>
at org.apache.dubbo.rpc.cluster.support.registry.ZoneAwareClusterInvoker.doInvoke(ZoneAwareClusterInvoker.java:63)</p>
<p dir="auto">配置:<br>
<strong>Provider#1</strong><br>
server:<br>
port: 8090<br>
dubbo:<br>
scan:<br>
base-packages: cn.service.test.api<br>
application:<br>
name: dubbo-service<br>
config:<br>
multiple: true<br>
protocol:<br>
name: dubbo<br>
port: 2080<br>
threads: 200<br>
provider:<br>
version: 1.0.0<br>
registry:<br>
check: false<br>
timeout: 20000<br>
use-as-config-center: false<br>
address: zookeeper://localhost:2181<br>
zone: zone-1</p>
<p dir="auto"><strong>Provider#2</strong><br>
server:<br>
port: 8091<br>
dubbo:<br>
scan:<br>
base-packages: cn.service.test.api<br>
application:<br>
name: dubbo-service<br>
config:<br>
multiple: true<br>
protocol:<br>
name: dubbo<br>
port: 2081<br>
threads: 200<br>
provider:<br>
version: 1.0.0<br>
registry:<br>
check: false<br>
timeout: 20000<br>
use-as-config-center: false<br>
address: nacos://nacos.ym:80<br>
zone: zone-2</p>
<p dir="auto"><strong>Consumer</strong><br>
server:<br>
port: 8092<br>
dubbo:<br>
scan:<br>
base-packages: cn.service.test.api<br>
application:<br>
name: dubbo-client<br>
protocol:<br>
name: dubbo<br>
port: 2082<br>
threads: 200<br>
config:<br>
multiple: true<br>
provider:<br>
version: 1.0.0<br>
registries:<br>
zone1:<br>
address: zookeeper://localhost:2181<br>
timeout: 20000<br>
use-as-config-center: false<br>
cluster: zone-aware<br>
zone2:<br>
address: nacos://nacos.ym:80<br>
timeout: 20000<br>
preferred: true<br>
use-as-config-center: false<br>
cluster: zone-aware</p>
<p dir="auto">问题<br>
目前dubbo的文档中未发现zone-aware的使用方式,请问是配置错误导致吗?</p> | <ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> I have searched the <a href="https://github.com/apache/incubator-dubbo/issues">issues</a> of this repository and believe that this is not a duplicate.</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> I have checked the <a href="https://github.com/apache/incubator-dubbo/blob/master/FAQ.md">FAQ</a> of this repository and believe that this is not a duplicate.</li>
</ul>
<h3 dir="auto">Environment</h3>
<ul dir="auto">
<li>Dubbo version: all</li>
<li>Operating System version: all</li>
<li>Java version: all</li>
</ul>
<h3 dir="auto">Steps to reproduce this issue</h3>
<ol dir="auto">
<li>create two ReferenceConfig for same service with different configuration.<br>
<dubbo:reference id="demoService" interface="com.alibaba.dubbo.DemoService"><br>
</dubbo:reference><br>
<dubbo:reference id="demoServiceAsync" interface="com.alibaba.dubbo.DemoService" async="true"><br>
</dubbo:reference></li>
<li>startup the consumer application.</li>
<li>get the ConsumerModel from ApplicationModel.</li>
</ol>
<p dir="auto">Pls. provide [GitHub address] to reproduce this issue.</p>
<h3 dir="auto">Expected Result</h3>
<p dir="auto">ApplicationModel.allConsumerModels() should returns two ConsumerModel.</p>
<h3 dir="auto">Actual Result</h3>
<p dir="auto">ApplicationModel.allConsumerModels() only returns the first ConsumerModel.</p>
<p dir="auto">If there is an exception, please attach the exception trace:</p>
<p dir="auto">With future investigation, ReferenceConfig cannot put the second ConsumerModel as they are in same serviceName.</p>
<p dir="auto">Also, this is the same issue for provider with below configuration.<br>
<br>
<dubbo:service interface="com.alibaba.dubbo.DemoService" ref="demoService"><br>
</dubbo:service><br>
<br>
<dubbo:service interface="com.alibaba.dubbo.DemoService" ref="demoService2"><br>
</dubbo:service></p>
<p dir="auto">Here is my PR: <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="365094989" data-permission-text="Title is private" data-url="https://github.com/apache/dubbo/issues/2582" data-hovercard-type="pull_request" data-hovercard-url="/apache/dubbo/pull/2582/hovercard" href="https://github.com/apache/dubbo/pull/2582">#2582</a></p> | 0 |
<ul dir="auto">
<li>VSCode Version: v1.1</li>
<li>OS Version: windows 10</li>
<li>nodejs Version: v6</li>
</ul>
<p dir="auto">Steps to Reproduce:<br>
( a little bit complicate)</p>
<ol dir="auto">
<li>mkdir try-bugs && cd try-bugs<br>
2.npm install -g vue-cli<br>
2.vue init webpack<br>
3.npm install<br>
4.set up a task in task.json, whose command is <code class="notranslate">npm</code>, args is <code class="notranslate">["run", "build"]</code></li>
<li>press <code class="notranslate">contrl + shift + b</code> to build the project<br>
6.then babel of webpack does not work in the right way<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://cloud.githubusercontent.com/assets/3315274/15247779/e2c30e2a-1948-11e6-85ef-065a9c4bb524.png"><img src="https://cloud.githubusercontent.com/assets/3315274/15247779/e2c30e2a-1948-11e6-85ef-065a9c4bb524.png" alt="image" style="max-width: 100%;"></a></li>
</ol>
<p dir="auto">Further description and guess:<br>
I have install a extension of vscode called <a href="https://marketplace.visualstudio.com/items?itemName=odonno.new-cmd" rel="nofollow">new-cmd </a>, which can enable you to open a new terminal from current work space directory by typing command <code class="notranslate">cmd</code> after pressing <code class="notranslate">f1</code> in vscode.<br>
If I open a terminal using this extension and then run <code class="notranslate">npm run build</code> in this terminal, it will show me the same bug as run the build task from vscode.<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://cloud.githubusercontent.com/assets/3315274/15247757/cc4b29d4-1948-11e6-98da-07de74a8b956.png"><img src="https://cloud.githubusercontent.com/assets/3315274/15247757/cc4b29d4-1948-11e6-98da-07de74a8b956.png" alt="image" style="max-width: 100%;"></a></p>
<p dir="auto">However, if I open a terminal manually by right click and run the same command, everything goes well.<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://cloud.githubusercontent.com/assets/3315274/15247842/53734e96-1949-11e6-9f70-fee948c954c0.png"><img src="https://cloud.githubusercontent.com/assets/3315274/15247842/53734e96-1949-11e6-9f70-fee948c954c0.png" alt="image" style="max-width: 100%;"></a></p>
<p dir="auto">I notice that there is one difference between the two terminal is that the disk name. The disk name in former terminal is <code class="notranslate">d:\</code> while the later one is <code class="notranslate">D:\</code>.<br>
In the project, from one hand, there are some code, the config of webpack specifically, related to <code class="notranslate">path.join(__dirname, ...)</code>, which will generate an absolute path to files begin with <code class="notranslate">D:\</code>. From the other hand, there are some other code use relative path to locate file, which will result in path to file begin with <code class="notranslate">d:\</code>.<br>
I wonder whether this caused my problem or not.<br>
By the way, if I repeat the process with node5, the bug disappear.</p> | <p dir="auto">OS Version: <strong>Windows</strong></p>
<ul dir="auto">
<li>generate a source map where the 'sourceRoot' uses an absolute path starting with a drive letter.</li>
<li>set a breakpoint in the source and try to hit it</li>
<li>edit the source map and change the case of the drive letter</li>
<li>restart the debug session and try to hit it again</li>
</ul>
<p dir="auto">Observe: the breakpoint will be only hit once</p>
<p dir="auto">The problem is a path comparison in node-debug that doesn't ignore the case of the drive letter.</p> | 1 |
<p dir="auto">BUG REPORT</p>
<p dir="auto"><strong>Kubernetes version</strong> : v1.3.2</p>
<p dir="auto"><strong>Environment</strong>:</p>
<ul dir="auto">
<li>setup kubernetes cluster on 4 VMs of ubuntu 14.04, 1 master and 3 nodes</li>
<li>bring up by kube-up.sh with PROVIDER=ubuntu</li>
<li><strong>Kernel</strong> : 3.13.0-92-generic</li>
</ul>
<p dir="auto"><strong>What happened</strong>:<br>
when do kubectl get nodes returned:</p>
<p dir="auto">NAME STATUS AGE<br>
*.205 Ready 1h<br>
*.205 NotReady 14d<br>
*.206 Ready 14d<br>
*.207 Ready 14d</p>
<p dir="auto">so there are two nodes with the same name *.205 and one is Notready</p>
<p dir="auto"><strong>What you expected to happen</strong>:<br>
this probably happened after I restart kube-controller and api-server</p>
<p dir="auto">restart kubelet and tail kubelet.log on *.205 shows:<br>
I0819 09:54:08.146564 17698 factory.go:54] Registering systemd factory<br>
I0819 09:54:08.148318 17698 factory.go:86] Registering Raw factory<br>
I0819 09:54:08.150438 17698 manager.go:1072] Started watching for new ooms in manager<br>
I0819 09:54:08.150552 17698 oomparser.go:200] OOM parser using kernel log file: "/var/log/kern.log"<br>
I0819 09:54:08.151820 17698 manager.go:281] Starting recovery of all containers<br>
I0819 09:54:08.162525 17698 manager.go:286] Recovery completed<br>
I0819 09:54:08.256878 17698 kubelet.go:1187] *<em>Node <em>.205 was previously registered</em></em></p>
<p dir="auto">if add --register-node=false to /etc/default/kubelet and tail log shows:<br>
I0819 09:56:52.330062 18421 factory.go:54] Registering systemd factory<br>
I0819 09:56:52.331063 18421 factory.go:86] Registering Raw factory<br>
I0819 09:56:52.332105 18421 manager.go:1072] Started watching for new ooms in manager<br>
I0819 09:56:52.332225 18421 oomparser.go:200] OOM parser using kernel log file: "/var/log/kern.log"<br>
I0819 09:56:52.333547 18421 manager.go:281] Starting recovery of all containers<br>
I0819 09:56:52.348541 18421 manager.go:286] <strong>Recovery completed</strong></p>
<p dir="auto">the node *205 is working now but not stable and when I do setup of weave-scope which launch pod on each node would have two pods on two duplicate *205 node and the one who is <strong>Not Ready</strong> has the pod pending</p>
<p dir="auto"><strong>Is there any way to remove that dupliacte useless node?</strong></p> | <p dir="auto"><strong>What:</strong></p>
<p dir="auto">Propose refactor pkg/proxy/config/config.go</p>
<p dir="auto"><a href="https://github.com/kubernetes/kubernetes/blob/master/pkg/proxy/config/config.go">https://github.com/kubernetes/kubernetes/blob/master/pkg/proxy/config/config.go</a></p>
<p dir="auto"><strong>Why:</strong></p>
<p dir="auto">In pkg/proxy/config/config.go, there are many identical codes between Service and Endpoints.</p>
<ul dir="auto">
<li>ServiceUpdate VS. EndpointsUpdate</li>
</ul>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="type ServiceUpdate struct {
Services []api.Service
Op Operation
}"><pre class="notranslate"><code class="notranslate">type ServiceUpdate struct {
Services []api.Service
Op Operation
}
</code></pre></div>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="type EndpointsUpdate struct {
Endpoints []api.Endpoints
Op Operation
}"><pre class="notranslate"><code class="notranslate">type EndpointsUpdate struct {
Endpoints []api.Endpoints
Op Operation
}
</code></pre></div>
<ul dir="auto">
<li>ServiceConfigHandler VS. EndpointsConfigHandler</li>
</ul>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="type ServiceConfigHandler interface {
OnServiceUpdate(services []api.Service)
}"><pre class="notranslate"><code class="notranslate">type ServiceConfigHandler interface {
OnServiceUpdate(services []api.Service)
}
</code></pre></div>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="type EndpointsConfigHandler interface {
OnEndpointsUpdate(endpoints []api.Endpoints)
}"><pre class="notranslate"><code class="notranslate">type EndpointsConfigHandler interface {
OnEndpointsUpdate(endpoints []api.Endpoints)
}
</code></pre></div>
<ul dir="auto">
<li>EndpointsConfig VS. ServiceConfig</li>
</ul>
<p dir="auto">and almost the same <code class="notranslate">NewServiceConfig()</code> VS. <code class="notranslate">NewEndpointsConfig()</code>, <code class="notranslate">endpointsStore.Merge()</code> VS. <code class="notranslate">serviceStore.Merge()</code> and so on.</p>
<p dir="auto">So I think we can introduce an abstract layer which can generify both Service and Endpoints here. In this way, reuse some codes.</p>
<p dir="auto"><strong>How:</strong></p>
<p dir="auto">Move Service and Endpoints to interface{}, like:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="type ObjectUpdate struct {
Objects []interface{}
Op Operation
}"><pre class="notranslate"><code class="notranslate">type ObjectUpdate struct {
Objects []interface{}
Op Operation
}
</code></pre></div>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="type ObjectConfigHandler interface {
OnObjectUpdate(objects []interface{})
}"><pre class="notranslate"><code class="notranslate">type ObjectConfigHandler interface {
OnObjectUpdate(objects []interface{})
}
</code></pre></div>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="type ObjectConfig struct {
mux *config.Mux
bcaster *config.Broadcaster
store *objectStore
}"><pre class="notranslate"><code class="notranslate">type ObjectConfig struct {
mux *config.Mux
bcaster *config.Broadcaster
store *objectStore
}
</code></pre></div>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="func (s *objectStore) Merge(source string, change interface{}) error {
...
}"><pre class="notranslate"><code class="notranslate">func (s *objectStore) Merge(source string, change interface{}) error {
...
}
</code></pre></div>
<p dir="auto">...</p>
<p dir="auto"><a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/thockin/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/thockin">@thockin</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/smarterclayton/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/smarterclayton">@smarterclayton</a></p>
<p dir="auto">What do you think about it?</p>
<p dir="auto">Thanks!</p> | 0 |
<p dir="auto"><strong>Context:</strong></p>
<ul dir="auto">
<li>Playwright Version: 1.3.0</li>
<li>Operating System: Windows 7</li>
<li>Node.js version: 12.18</li>
<li>Browser:Chromium</li>
</ul>
<p dir="auto"><strong>Code Snippet</strong></p>
<div class="highlight highlight-source-js notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="const { chromium } = require('playwright');
(async () => {
const browser = await chromium.launch();
const context = await browser.newContext();
const page = await context.newPage();
await browser.close();
})();"><pre class="notranslate"><span class="pl-k">const</span> <span class="pl-kos">{</span> chromium <span class="pl-kos">}</span> <span class="pl-c1">=</span> <span class="pl-en">require</span><span class="pl-kos">(</span><span class="pl-s">'playwright'</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-kos">(</span><span class="pl-k">async</span> <span class="pl-kos">(</span><span class="pl-kos">)</span> <span class="pl-c1">=></span> <span class="pl-kos">{</span>
<span class="pl-k">const</span> <span class="pl-s1">browser</span> <span class="pl-c1">=</span> <span class="pl-k">await</span> <span class="pl-s1">chromium</span><span class="pl-kos">.</span><span class="pl-en">launch</span><span class="pl-kos">(</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-k">const</span> <span class="pl-s1">context</span> <span class="pl-c1">=</span> <span class="pl-k">await</span> <span class="pl-s1">browser</span><span class="pl-kos">.</span><span class="pl-en">newContext</span><span class="pl-kos">(</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-k">const</span> <span class="pl-s1">page</span> <span class="pl-c1">=</span> <span class="pl-k">await</span> <span class="pl-s1">context</span><span class="pl-kos">.</span><span class="pl-en">newPage</span><span class="pl-kos">(</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-k">await</span> <span class="pl-s1">browser</span><span class="pl-kos">.</span><span class="pl-en">close</span><span class="pl-kos">(</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span><span class="pl-kos">)</span><span class="pl-kos">(</span><span class="pl-kos">)</span><span class="pl-kos">;</span></pre></div>
<p dir="auto"><strong>Describe the bug</strong><br>
Chromium doesn't work, but other browsers seems to work fine</p>
<p dir="auto"><strong>Console Output:</strong></p>
<div class="highlight highlight-source-shell notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="G:\GitHub\quran-api>set "DEBUG=pw:api" && node dircheck.js
(node:16052) UnhandledPromiseRejectionWarning: Error: browserType.launch: Host s
ystem is missing dependencies!
Full list of missing libraries:
api-ms-win-core-winrt-error-l1-1-0.dll
Note: use DEBUG=pw:api environment variable and rerun to capture Playwright logs
.
at validateDependenciesWindows (G:\GitHub\quran-api\node_modules\playwright\
lib\server\validateDependencies.js:78:11)
at async validateDependencies (G:\GitHub\quran-api\node_modules\playwright\l
ib\server\validateDependencies.js:47:16)
at async Object.validateHostRequirements (G:\GitHub\quran-api\node_modules\p
laywright\lib\server\validateDependencies.js:34:5)
at async Chromium._launchServer (G:\GitHub\quran-api\node_modules\playwright
\lib\server\browserType.js:154:13)
at async Chromium._innerLaunch (G:\GitHub\quran-api\node_modules\playwright\
lib\server\browserType.js:76:61)
at async ProgressController.run (G:\GitHub\quran-api\node_modules\playwright
\lib\progress.js:75:28)
at async Chromium.launch (G:\GitHub\quran-api\node_modules\playwright\lib\se
rver\browserType.js:62:25)
at async G:\GitHub\quran-api\dircheck.js:4:19
(node:16052) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This
error originated either by throwing inside of an async function without a catch
block, or by rejecting a promise which was not handled with .catch(). To termin
ate the node process on unhandled promise rejection, use the CLI flag `--unhandl
ed-rejections=strict` (see https://nodejs.org/api/cli.html#cli_unhandled_rejecti
ons_mode). (rejection id: 1)
(node:16052) [DEP0018] DeprecationWarning: Unhandled promise rejections are depr
ecated. In the future, promise rejections that are not handled will terminate th
e Node.js process with a non-zero exit code."><pre class="notranslate">G:<span class="pl-cce">\G</span>itHub<span class="pl-cce">\q</span>uran-api<span class="pl-k">></span>set <span class="pl-s"><span class="pl-pds">"</span>DEBUG=pw:api<span class="pl-pds">"</span></span> <span class="pl-k">&&</span> node dircheck.js
(node:16052) UnhandledPromiseRejectionWarning: Error: browserType.launch: Host s
ystem is missing dependencies<span class="pl-k">!</span>
Full list of missing libraries:
api-ms-win-core-winrt-error-l1-1-0.dll
Note: use DEBUG=pw:api environment variable and rerun to capture Playwright logs
<span class="pl-c1">.</span>
at validateDependenciesWindows (G:<span class="pl-cce">\G</span>itHub<span class="pl-cce">\q</span>uran-api<span class="pl-cce">\n</span>ode_modules<span class="pl-cce">\p</span>laywright\
lib<span class="pl-cce">\s</span>erver<span class="pl-cce">\v</span>alidateDependencies.js:78:11)
at async validateDependencies (G:<span class="pl-cce">\G</span>itHub<span class="pl-cce">\q</span>uran-api<span class="pl-cce">\n</span>ode_modules<span class="pl-cce">\p</span>laywright<span class="pl-cce">\l</span>
ib<span class="pl-cce">\s</span>erver<span class="pl-cce">\v</span>alidateDependencies.js:47:16)
at async Object.validateHostRequirements (G:<span class="pl-cce">\G</span>itHub<span class="pl-cce">\q</span>uran-api<span class="pl-cce">\n</span>ode_modules<span class="pl-cce">\p</span>
laywright<span class="pl-cce">\l</span>ib<span class="pl-cce">\s</span>erver<span class="pl-cce">\v</span>alidateDependencies.js:34:5)
at async Chromium._launchServer (G:<span class="pl-cce">\G</span>itHub<span class="pl-cce">\q</span>uran-api<span class="pl-cce">\n</span>ode_modules<span class="pl-cce">\p</span>laywright
<span class="pl-cce">\l</span>ib<span class="pl-cce">\s</span>erver<span class="pl-cce">\b</span>rowserType.js:154:13)
at async Chromium._innerLaunch (G:<span class="pl-cce">\G</span>itHub<span class="pl-cce">\q</span>uran-api<span class="pl-cce">\n</span>ode_modules<span class="pl-cce">\p</span>laywright\
lib<span class="pl-cce">\s</span>erver<span class="pl-cce">\b</span>rowserType.js:76:61)
at async ProgressController.run (G:<span class="pl-cce">\G</span>itHub<span class="pl-cce">\q</span>uran-api<span class="pl-cce">\n</span>ode_modules<span class="pl-cce">\p</span>laywright
<span class="pl-cce">\l</span>ib<span class="pl-cce">\p</span>rogress.js:75:28)
at async Chromium.launch (G:<span class="pl-cce">\G</span>itHub<span class="pl-cce">\q</span>uran-api<span class="pl-cce">\n</span>ode_modules<span class="pl-cce">\p</span>laywright<span class="pl-cce">\l</span>ib<span class="pl-cce">\s</span>e
rver<span class="pl-cce">\b</span>rowserType.js:62:25)
at async G:<span class="pl-cce">\G</span>itHub<span class="pl-cce">\q</span>uran-api<span class="pl-cce">\d</span>ircheck.js:4:19
(node:16052) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This
error originated either by throwing inside of an async <span class="pl-k">function</span> <span class="pl-en">without</span> a catch
block, or by rejecting a promise which was not handled with <span class="pl-en">.catch</span>(). To termin
ate the node process on unhandled promise rejection, use the CLI flag <span class="pl-s"><span class="pl-pds">`</span>--unhandl</span>
<span class="pl-s">ed-rejections=strict<span class="pl-pds">`</span></span> (see https://nodejs.org/api/cli.html#cli_unhandled_rejecti
ons_mode). (rejection id: 1)
(node:16052) [DEP0018] DeprecationWarning: Unhandled promise rejections are depr
ecated. In the future, promise rejections that are not handled will terminate th
e Node.js process with a non-zero <span class="pl-c1">exit</span> code.</pre></div> | <p dir="auto"><strong>Context:</strong></p>
<ul dir="auto">
<li>Playwright Version: 1.3.0</li>
<li>Operating System: Windows</li>
<li>Node.js version: 12.18.1</li>
<li>Browser: All</li>
</ul>
<p dir="auto"><strong>Code Snippet</strong></p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="const playwright = require('playwright');
const browserType = 'chromium';
async function botfunc() {
let browser = playwright[browserType].launch();
const context = await browser.newContext();
const page = await context.newPage();
await page.goto("http://google.com");
}
botfunc();"><pre class="notranslate"><code class="notranslate">const playwright = require('playwright');
const browserType = 'chromium';
async function botfunc() {
let browser = playwright[browserType].launch();
const context = await browser.newContext();
const page = await context.newPage();
await page.goto("http://google.com");
}
botfunc();
</code></pre></div>
<p dir="auto"><strong>Describe the bug</strong><br>
Error log -</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="UnhandledPromiseRejectionWarning: Error: browserType.launch: Host system is missing dependencies!
Full list of missing libraries:
api-ms-win-core-winrt-error-l1-1-0.dll
"><pre class="notranslate"><code class="notranslate">UnhandledPromiseRejectionWarning: Error: browserType.launch: Host system is missing dependencies!
Full list of missing libraries:
api-ms-win-core-winrt-error-l1-1-0.dll
</code></pre></div>
<p dir="auto">It was solved when I used 1.0.2 version.</p> | 1 |
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I have searched the <a href="https://github.com/callemall/material-ui/issues">issues</a> of this repository and believe that this is not a duplicate.</li>
</ul>
<h2 dir="auto">Expected Behavior</h2>
<p dir="auto">The label stays floated, if the user enters non numeric values to a TextField with type="number"</p>
<h2 dir="auto">Current Behavior</h2>
<p dir="auto">The label stays not floated.</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/8149519/30313488-d2256ae8-979d-11e7-8269-718ddfa1051f.PNG"><img src="https://user-images.githubusercontent.com/8149519/30313488-d2256ae8-979d-11e7-8269-718ddfa1051f.PNG" alt="label_stays_not_floated" style="max-width: 100%;"></a></p>
<p dir="auto">With only numeric values the label stays floated</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/8149519/30313565-0aa601e8-979e-11e7-8cb3-ebefbd623717.PNG"><img src="https://user-images.githubusercontent.com/8149519/30313565-0aa601e8-979e-11e7-8cb3-ebefbd623717.PNG" alt="label_stays_floated" style="max-width: 100%;"></a></p>
<p dir="auto">Also there are some wierd characters if the TextField is empty:</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/8149519/30313528-f5f358d6-979d-11e7-8869-97204fed6edc.PNG"><img src="https://user-images.githubusercontent.com/8149519/30313528-f5f358d6-979d-11e7-8869-97204fed6edc.PNG" alt="wierd_chars" style="max-width: 100%;"></a></p>
<h2 dir="auto">Steps to Reproduce (for bugs)</h2>
<ol dir="auto">
<li>create a TextField component and use it with type="number"</li>
</ol>
<h2 dir="auto">Context</h2>
<h2 dir="auto">Your Environment</h2>
<table role="table">
<thead>
<tr>
<th>Tech</th>
<th>Version</th>
</tr>
</thead>
<tbody>
<tr>
<td>Material-UI</td>
<td>v1.0.0-beta.8</td>
</tr>
<tr>
<td>React</td>
<td>15.6.1</td>
</tr>
<tr>
<td>browser</td>
<td>Firefox 54.0</td>
</tr>
</tbody>
</table> | <p dir="auto">console logs <code class="notranslate">null, undefined, undefined</code></p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="<TextField
ref={(a, b, c) => {
console.log(a, b, c);
}}
..."><pre class="notranslate"><code class="notranslate"><TextField
ref={(a, b, c) => {
console.log(a, b, c);
}}
...
</code></pre></div>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I have searched the <a href="https://github.com/callemall/material-ui/issues">issues</a> of this repository and believe that this is not a duplicate.</li>
</ul>
<table role="table">
<thead>
<tr>
<th>Tech</th>
<th>Version</th>
</tr>
</thead>
<tbody>
<tr>
<td>Material-UI</td>
<td>^1.0.0-beta.17</td>
</tr>
<tr>
<td>React</td>
<td>^15.0.2</td>
</tr>
<tr>
<td>browser</td>
<td>chrome os x</td>
</tr>
<tr>
<td>etc</td>
<td></td>
</tr>
</tbody>
</table> | 0 |
<p dir="auto">Currently (at least on 0.3.11, have checked master docs but not implementation), abs does not behave in a mathematically correct manner: abs(typemin(Int)) == typemin(Int) on my Linux x86-64 machine.</p>
<p dir="auto">This problem plagues the C standard library abs(), llabs(), etc as well due to subtle asymmetry in 2's complement arithmetic. The C standard leaves behavior on INT_MIN, etc undefined, and this is fortunately documented, though easily missed by many programs.</p>
<p dir="auto">I thus see 2 possible approaches:</p>
<ol dir="auto">
<li>Follow the C standard behavior, give no guarantee for abs(typemin), and document it.</li>
<li>Do something special with typemin, and document it.</li>
</ol> | <p dir="auto">I saw the discussion in <a href="https://groups.google.com/forum/#!topic/julia-users/46i-NNUxpa8" rel="nofollow">https://groups.google.com/forum/#!topic/julia-users/46i-NNUxpa8</a>, and did not find an issue on github discussing this.</p>
<p dir="auto">Most users will probably expect <code class="notranslate">abs(x)>=0</code> and failing this assumption will probably create bugs that will be really hard to find (especially for the typical Julia user).</p>
<p dir="auto"><a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/StefanKarpinski/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/StefanKarpinski">@StefanKarpinski</a> suggested to throw an error for these numbers, and that seems like a good option. The problem is that performance will suffer so much that some users might want access to the simple inlinable <code class="notranslate">abs</code> function so that we have to add a <code class="notranslate">abs_raw</code>.</p> | 1 |
<ul dir="auto">
<li>Electron version: any</li>
<li>Operating system: Linux</li>
</ul>
<h3 dir="auto">Expected behavior</h3>
<p dir="auto">Latest Ubuntu comes with Wayland a default display server.<br>
On embedded devices, wayland provides usually the best rendering performance, at least compared to X11 where the performance is usually pretty poor.</p>
<h3 dir="auto">Actual behavior</h3>
<p dir="auto">Current linux / arm / arm64 binary releases are great but run pretty bad from my experience on embedded devices. They all link to libx11 / gtk which makes them unusable for wayland.</p>
<p dir="auto">libGtk can be build with --wayland-backend and then electron can be built against it</p>
<p dir="auto">I would really like electron team to consider providing prebuilt binaries usable with wayland, that would open a new range of possibilities, especially on embedded devices.</p> | <ul dir="auto">
<li>Electron version: 1.6.7</li>
<li>Operating system: Win 7</li>
</ul>
<p dir="auto">Hi!<br>
Several users reported that they can't use screensharing, because there is no available sources. For this cases <code class="notranslate">getSources</code> method is called, but callback is not triggered.<br>
I don't know how desktopCapturer works inside, so maybe you have ideas what's going wrong. May it be related to system preferences or drivers? May it be issue of 56 chromium?<br>
Unfortunately I can't check this issue on 1.7.* for now.</p>
<p dir="auto">Thanks!</p> | 0 |
<p dir="auto">I am trying to ingest messages from Kafka which were written by Divolte.</p>
<p dir="auto">Divolte uses Avro encoding for the messages it writes to Kafka. I believe it builds the Avro messages in a non-standard way. This from <a href="http://divolte-releases.s3-website-eu-west-1.amazonaws.com/divolte-collector/0.9.0/userdoc/html/getting_started.html#writing-to-kafka" rel="nofollow">Divolte's docs</a>:</p>
<blockquote>
<h2 dir="auto">Data in Kafka</h2>
<p dir="auto">Avro files on HDFS are written with the schema in the header. Unfortunately Kafka doesn’t really have a clear way of passing along the schema. For the messages on Kafka queues we expect the consumer to know the schema in advance, meaning that the messages that are passed onto the queue only contain the raw bytes of the serialized Avro record without any metadata. The key of each message is the party ID that for the event. Divolte Collector provides a small helper library to easily create Kafka consumers in Java using Avro’s code generation support. There is an example Kafka consumer with step by step instruction on getting it up and running in our usage examples repository here: <a href="https://github.com/divolte/divolte-examples/tree/master/tcp-kafka-consumer">https://github.com/divolte/divolte-examples/tree/master/tcp-kafka-consumer</a>.</p>
</blockquote>
<p dir="auto">And the messages Divolte is putting on the Kafka queue look like this:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="key: 0:jro4m5d8:xilY4lXNlasIzvVbEYDFECnbepfZOb_J
value: \x00\x00\x00\xB2\xED\xF0\xA2\x97Z\xDA\xCD\xF0\xA2\x97Z\x14172.25.0.1\x00\x02\x90\x02file:///Users/.../frontend/test/index.html\x02\x80\x1E\x02\xEE\n\x02\x80\x1E\x02\xB2\x12\x02V0:jro4m5d8:xilY4lXNlasIzvVbEYDFECnbepfZOb_J\x02V0:jro4m5d8:SPubeXNlImFqTIBWRTI5Kpt5NfWZ7ic1\x02D0:UdmHBajNonE_ONddXwP2679ks0EYxySf\x1AmyCustomEvent\x02\xF0\x01Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/71.0.3578.98 Safari/537.36\x02\fChrome\x02\fChrome\x02\x16Google Inc.\x02\x0EBrowser\x02\x1871.0.3578.98\x02\"Personal computer\x02\bOS X\x02\x0E10.14.0\x02(Apple Computer, Inc."><pre class="notranslate"><code class="notranslate">key: 0:jro4m5d8:xilY4lXNlasIzvVbEYDFECnbepfZOb_J
value: \x00\x00\x00\xB2\xED\xF0\xA2\x97Z\xDA\xCD\xF0\xA2\x97Z\x14172.25.0.1\x00\x02\x90\x02file:///Users/.../frontend/test/index.html\x02\x80\x1E\x02\xEE\n\x02\x80\x1E\x02\xB2\x12\x02V0:jro4m5d8:xilY4lXNlasIzvVbEYDFECnbepfZOb_J\x02V0:jro4m5d8:SPubeXNlImFqTIBWRTI5Kpt5NfWZ7ic1\x02D0:UdmHBajNonE_ONddXwP2679ks0EYxySf\x1AmyCustomEvent\x02\xF0\x01Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/71.0.3578.98 Safari/537.36\x02\fChrome\x02\fChrome\x02\x16Google Inc.\x02\x0EBrowser\x02\x1871.0.3578.98\x02\"Personal computer\x02\bOS X\x02\x0E10.14.0\x02(Apple Computer, Inc.
</code></pre></div>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="key: 0:jro4m5d8:xilY4lXNlasIzvVbEYDFECnbepfZOb_J
value: \x00\x00\x00\x9A\x91\xF1\xA2\x97Z\xE6\xF3\xF0\xA2\x97Z\x14172.25.0.1\x00\x02\x90\x02file:///Users/.../frontend/test/index.html\x02\x80\x1E\x02\xEE\n\x02\x80\x1E\x02\xB2\x12\x02V0:jro4m5d8:xilY4lXNlasIzvVbEYDFECnbepfZOb_J\x02V0:jro4m5d8:SPubeXNlImFqTIBWRTI5Kpt5NfWZ7ic1\x02D0:UdmHBajNonE_ONddXwP2679ks0EYxySf\x1AmyCustomEvent\x02\xF0\x01Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/71.0.3578.98 Safari/537.36\x02\fChrome\x02\fChrome\x02\x16Google Inc.\x02\x0EBrowser\x02\x1871.0.3578.98\x02\"Personal computer\x02\bOS X\x02\x0E10.14.0\x02(Apple Computer, Inc."><pre class="notranslate"><code class="notranslate">key: 0:jro4m5d8:xilY4lXNlasIzvVbEYDFECnbepfZOb_J
value: \x00\x00\x00\x9A\x91\xF1\xA2\x97Z\xE6\xF3\xF0\xA2\x97Z\x14172.25.0.1\x00\x02\x90\x02file:///Users/.../frontend/test/index.html\x02\x80\x1E\x02\xEE\n\x02\x80\x1E\x02\xB2\x12\x02V0:jro4m5d8:xilY4lXNlasIzvVbEYDFECnbepfZOb_J\x02V0:jro4m5d8:SPubeXNlImFqTIBWRTI5Kpt5NfWZ7ic1\x02D0:UdmHBajNonE_ONddXwP2679ks0EYxySf\x1AmyCustomEvent\x02\xF0\x01Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/71.0.3578.98 Safari/537.36\x02\fChrome\x02\fChrome\x02\x16Google Inc.\x02\x0EBrowser\x02\x1871.0.3578.98\x02\"Personal computer\x02\bOS X\x02\x0E10.14.0\x02(Apple Computer, Inc.
</code></pre></div>
<p dir="auto">Following the great tutorial by <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/Fokko/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/Fokko">@Fokko</a> <a href="https://blog.godatadriven.com/divolte-kafka-druid-superset" rel="nofollow">here</a>, I opted to use Druid's <code class="notranslate">schema_inline</code> Avro decoder in my supervisor spec.</p>
<p dir="auto">However, I am still getting parse errors in my ingestion reports.</p>
<p dir="auto">I'm wondering if something changed in the way the Avro extension works since <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/Fokko/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/Fokko">@Fokko</a> wrote that tutorial?</p>
<p dir="auto">Am I using the wrong Avro decoder?</p>
<p dir="auto">Is there simply no longer a decoder in the Druid Avro extension which can handle messages encoded this way?</p>
<h2 dir="auto">Supervisor</h2>
<div class="highlight highlight-source-json notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="{
"type" : "kafka",
"ioConfig": {
"topic": "livestream",
"consumerProperties": {
"bootstrap.servers": "10.0.1.32:9092"
},
"taskCount": 1,
"replicas": 1,
"taskDuration": "PT5M"
},
"dataSchema": {
"dataSource": "livestream",
"parser": {
"type": "avro_stream",
"avroBytesDecoder": {
"type": "schema_inline",
"schema": {
"namespace": "io.divolte.record",
"type": "record",
"name": "DefaultEventRecord",
"fields": [
{ "name": "detectedDuplicate", "type": "boolean" },
{ "name": "detectedCorruption", "type": "boolean" },
{ "name": "firstInSession", "type": "boolean" },
{ "name": "timestamp", "type": "long" },
{ "name": "remoteHost", "type": "string" },
{ "name": "referer", "type": ["null", "string"], "default": null },
{ "name": "location", "type": ["null", "string"], "default": null },
{ "name": "viewportPixelWidth", "type": ["null", "int"], "default": null },
{ "name": "viewportPixelHeight", "type": ["null", "int"], "default": null },
{ "name": "screenPixelWidth", "type": ["null", "int"], "default": null },
{ "name": "screenPixelHeight", "type": ["null", "int"], "default": null },
{ "name": "partyId", "type": ["null", "string"], "default": null },
{ "name": "sessionId", "type": ["null", "string"], "default": null },
{ "name": "pageViewId", "type": ["null", "string"], "default": null },
{ "name": "eventType", "type": "string", "default": "unknown" },
{ "name": "userAgentString", "type": ["null", "string"], "default": null },
{ "name": "userAgentName", "type": ["null", "string"], "default": null },
{ "name": "userAgentFamily", "type": ["null", "string"], "default": null },
{ "name": "userAgentVendor", "type": ["null", "string"], "default": null },
{ "name": "userAgentType", "type": ["null", "string"], "default": null },
{ "name": "userAgentVersion", "type": ["null", "string"], "default": null },
{ "name": "userAgentDeviceCategory", "type": ["null", "string"], "default": null },
{ "name": "userAgentOsFamily", "type": ["null", "string"], "default": null },
{ "name": "userAgentOsVersion", "type": ["null", "string"], "default": null },
{ "name": "userAgentOsVendor", "type": ["null", "string"], "default": null },
{ "name": "technology", "type": ["null", "string"], "default": null }
]
}
},
"parseSpec": {
"format": "timeAndDims",
"timestampSpec": {
"column": "timestamp",
"format": "iso"
},
"dimensionsSpec": {
"dimensions": [
"detectedDuplicate",
"detectedCorruption",
"firstInSession",
"remoteHost",
"referer",
"location",
"viewportPixelWidth",
"viewportPixelHeight",
"screenPixelWidth",
"screenPixelHeight",
"partyId",
"sessionId",
"pageViewId",
"eventType",
"userAgentString",
"userAgentName",
"userAgentFamily",
"userAgentVendor",
"userAgentType",
"userAgentVersion",
"userAgentDeviceCategory",
"userAgentOsFamily",
"userAgentOsVersion",
"userAgentOsVendor",
"technology"
]
}
}
},
"metricsSpec": [{
"name": "count",
"type": "count"
}],
"granularitySpec": {
"type": "uniform",
"segmentGranularity": "HOUR",
"queryGranularity": "NONE"
}
},
"tuningConfig": {
"type": "kafka",
"maxRowsPerSegment": 5000000,
"logParseExceptions": true,
"maxSavedParseExceptions": 100
}
}"><pre class="notranslate">{
<span class="pl-ent">"type"</span> : <span class="pl-s"><span class="pl-pds">"</span>kafka<span class="pl-pds">"</span></span>,
<span class="pl-ent">"ioConfig"</span>: {
<span class="pl-ent">"topic"</span>: <span class="pl-s"><span class="pl-pds">"</span>livestream<span class="pl-pds">"</span></span>,
<span class="pl-ent">"consumerProperties"</span>: {
<span class="pl-ent">"bootstrap.servers"</span>: <span class="pl-s"><span class="pl-pds">"</span>10.0.1.32:9092<span class="pl-pds">"</span></span>
},
<span class="pl-ent">"taskCount"</span>: <span class="pl-c1">1</span>,
<span class="pl-ent">"replicas"</span>: <span class="pl-c1">1</span>,
<span class="pl-ent">"taskDuration"</span>: <span class="pl-s"><span class="pl-pds">"</span>PT5M<span class="pl-pds">"</span></span>
},
<span class="pl-ent">"dataSchema"</span>: {
<span class="pl-ent">"dataSource"</span>: <span class="pl-s"><span class="pl-pds">"</span>livestream<span class="pl-pds">"</span></span>,
<span class="pl-ent">"parser"</span>: {
<span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>avro_stream<span class="pl-pds">"</span></span>,
<span class="pl-ent">"avroBytesDecoder"</span>: {
<span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>schema_inline<span class="pl-pds">"</span></span>,
<span class="pl-ent">"schema"</span>: {
<span class="pl-ent">"namespace"</span>: <span class="pl-s"><span class="pl-pds">"</span>io.divolte.record<span class="pl-pds">"</span></span>,
<span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>record<span class="pl-pds">"</span></span>,
<span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>DefaultEventRecord<span class="pl-pds">"</span></span>,
<span class="pl-ent">"fields"</span>: [
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>detectedDuplicate<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>boolean<span class="pl-pds">"</span></span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>detectedCorruption<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>boolean<span class="pl-pds">"</span></span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>firstInSession<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>boolean<span class="pl-pds">"</span></span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>timestamp<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>long<span class="pl-pds">"</span></span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>remoteHost<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>referer<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>location<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>viewportPixelWidth<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>int<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>viewportPixelHeight<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>int<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>screenPixelWidth<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>int<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>screenPixelHeight<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>int<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>partyId<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>sessionId<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>pageViewId<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>eventType<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>, <span class="pl-ent">"default"</span>: <span class="pl-s"><span class="pl-pds">"</span>unknown<span class="pl-pds">"</span></span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentString<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentName<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentFamily<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentVendor<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentType<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentVersion<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentDeviceCategory<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentOsFamily<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentOsVersion<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentOsVendor<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>technology<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> }
]
}
},
<span class="pl-ent">"parseSpec"</span>: {
<span class="pl-ent">"format"</span>: <span class="pl-s"><span class="pl-pds">"</span>timeAndDims<span class="pl-pds">"</span></span>,
<span class="pl-ent">"timestampSpec"</span>: {
<span class="pl-ent">"column"</span>: <span class="pl-s"><span class="pl-pds">"</span>timestamp<span class="pl-pds">"</span></span>,
<span class="pl-ent">"format"</span>: <span class="pl-s"><span class="pl-pds">"</span>iso<span class="pl-pds">"</span></span>
},
<span class="pl-ent">"dimensionsSpec"</span>: {
<span class="pl-ent">"dimensions"</span>: [
<span class="pl-s"><span class="pl-pds">"</span>detectedDuplicate<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>detectedCorruption<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>firstInSession<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>remoteHost<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>referer<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>location<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>viewportPixelWidth<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>viewportPixelHeight<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>screenPixelWidth<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>screenPixelHeight<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>partyId<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>sessionId<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>pageViewId<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>eventType<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>userAgentString<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>userAgentName<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>userAgentFamily<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>userAgentVendor<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>userAgentType<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>userAgentVersion<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>userAgentDeviceCategory<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>userAgentOsFamily<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>userAgentOsVersion<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>userAgentOsVendor<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>technology<span class="pl-pds">"</span></span>
]
}
}
},
<span class="pl-ent">"metricsSpec"</span>: [{
<span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>count<span class="pl-pds">"</span></span>,
<span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>count<span class="pl-pds">"</span></span>
}],
<span class="pl-ent">"granularitySpec"</span>: {
<span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>uniform<span class="pl-pds">"</span></span>,
<span class="pl-ent">"segmentGranularity"</span>: <span class="pl-s"><span class="pl-pds">"</span>HOUR<span class="pl-pds">"</span></span>,
<span class="pl-ent">"queryGranularity"</span>: <span class="pl-s"><span class="pl-pds">"</span>NONE<span class="pl-pds">"</span></span>
}
},
<span class="pl-ent">"tuningConfig"</span>: {
<span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>kafka<span class="pl-pds">"</span></span>,
<span class="pl-ent">"maxRowsPerSegment"</span>: <span class="pl-c1">5000000</span>,
<span class="pl-ent">"logParseExceptions"</span>: <span class="pl-c1">true</span>,
<span class="pl-ent">"maxSavedParseExceptions"</span>: <span class="pl-c1">100</span>
}
}</pre></div>
<h2 dir="auto">Divolte Conf</h2>
<h3 dir="auto">DefaultEventRecord.avsc</h3>
<div class="highlight highlight-source-json notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="{
"namespace": "io.divolte.record",
"name": "DefaultEventRecord",
"type": "record",
"fields": [
{ "name": "detectedDuplicate", "type": "boolean" },
{ "name": "detectedCorruption", "type": "boolean" },
{ "name": "firstInSession", "type": "boolean" },
{ "name": "timestamp", "type": "long" },
{ "name": "remoteHost", "type": "string" },
{ "name": "referer", "type": ["null", "string"], "default": null },
{ "name": "location", "type": ["null", "string"], "default": null },
{ "name": "viewportPixelWidth", "type": ["null", "int"], "default": null },
{ "name": "viewportPixelHeight", "type": ["null", "int"], "default": null },
{ "name": "screenPixelWidth", "type": ["null", "int"], "default": null },
{ "name": "screenPixelHeight", "type": ["null", "int"], "default": null },
{ "name": "partyId", "type": ["null", "string"], "default": null },
{ "name": "sessionId", "type": ["null", "string"], "default": null },
{ "name": "pageViewId", "type": ["null", "string"], "default": null },
{ "name": "eventType", "type": "string", "default": "unknown" },
{ "name": "userAgentString", "type": ["null", "string"], "default": null },
{ "name": "userAgentName", "type": ["null", "string"], "default": null },
{ "name": "userAgentFamily", "type": ["null", "string"], "default": null },
{ "name": "userAgentVendor", "type": ["null", "string"], "default": null },
{ "name": "userAgentType", "type": ["null", "string"], "default": null },
{ "name": "userAgentVersion", "type": ["null", "string"], "default": null },
{ "name": "userAgentDeviceCategory", "type": ["null", "string"], "default": null },
{ "name": "userAgentOsFamily", "type": ["null", "string"], "default": null },
{ "name": "userAgentOsVersion", "type": ["null", "string"], "default": null },
{ "name": "userAgentOsVendor", "type": ["null", "string"], "default": null },
{ "name": "technology", "type": ["null", "string"], "default": null }
]
}"><pre class="notranslate">{
<span class="pl-ent">"namespace"</span>: <span class="pl-s"><span class="pl-pds">"</span>io.divolte.record<span class="pl-pds">"</span></span>,
<span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>DefaultEventRecord<span class="pl-pds">"</span></span>,
<span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>record<span class="pl-pds">"</span></span>,
<span class="pl-ent">"fields"</span>: [
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>detectedDuplicate<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>boolean<span class="pl-pds">"</span></span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>detectedCorruption<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>boolean<span class="pl-pds">"</span></span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>firstInSession<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>boolean<span class="pl-pds">"</span></span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>timestamp<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>long<span class="pl-pds">"</span></span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>remoteHost<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>referer<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>location<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>viewportPixelWidth<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>int<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>viewportPixelHeight<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>int<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>screenPixelWidth<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>int<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>screenPixelHeight<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>int<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>partyId<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>sessionId<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>pageViewId<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>eventType<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>, <span class="pl-ent">"default"</span>: <span class="pl-s"><span class="pl-pds">"</span>unknown<span class="pl-pds">"</span></span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentString<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentName<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentFamily<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentVendor<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentType<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentVersion<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentDeviceCategory<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentOsFamily<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentOsVersion<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>userAgentOsVendor<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> },
{ <span class="pl-ent">"name"</span>: <span class="pl-s"><span class="pl-pds">"</span>technology<span class="pl-pds">"</span></span>, <span class="pl-ent">"type"</span>: [<span class="pl-s"><span class="pl-pds">"</span>null<span class="pl-pds">"</span></span>, <span class="pl-s"><span class="pl-pds">"</span>string<span class="pl-pds">"</span></span>], <span class="pl-ent">"default"</span>: <span class="pl-c1">null</span> }
]
}</pre></div>
<h3 dir="auto">mapping.groovy</h3>
<div class="highlight highlight-source-groovy notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="mapping {
map duplicate() onto 'detectedDuplicate'
map corrupt() onto 'detectedCorruption'
map firstInSession() onto 'firstInSession'
map timestamp() onto 'timestamp'
map remoteHost() onto 'remoteHost'
map referer() onto 'referer'
map location() onto 'location'
map viewportPixelWidth() onto 'viewportPixelWidth'
map viewportPixelHeight() onto 'viewportPixelHeight'
map screenPixelWidth() onto 'screenPixelWidth'
map screenPixelHeight() onto 'screenPixelHeight'
map partyId() onto 'partyId'
map sessionId() onto 'sessionId'
map pageViewId() onto 'pageViewId'
map eventType() onto 'eventType'
map userAgentString() onto 'userAgentString'
def ua = userAgent()
map ua.name() onto 'userAgentName'
map ua.family() onto 'userAgentFamily'
map ua.vendor() onto 'userAgentVendor'
map ua.type() onto 'userAgentType'
map ua.version() onto 'userAgentVersion'
map ua.deviceCategory() onto 'userAgentDeviceCategory'
map ua.osFamily() onto 'userAgentOsFamily'
map ua.osVersion() onto 'userAgentOsVersion'
map ua.osVendor() onto 'userAgentOsVendor'
map eventParameter('technology') onto 'technology'
}"><pre class="notranslate">mapping {
map duplicate() onto <span class="pl-s"><span class="pl-pds">'</span>detectedDuplicate<span class="pl-pds">'</span></span>
map corrupt() onto <span class="pl-s"><span class="pl-pds">'</span>detectedCorruption<span class="pl-pds">'</span></span>
map firstInSession() onto <span class="pl-s"><span class="pl-pds">'</span>firstInSession<span class="pl-pds">'</span></span>
map timestamp() onto <span class="pl-s"><span class="pl-pds">'</span>timestamp<span class="pl-pds">'</span></span>
map remoteHost() onto <span class="pl-s"><span class="pl-pds">'</span>remoteHost<span class="pl-pds">'</span></span>
map referer() onto <span class="pl-s"><span class="pl-pds">'</span>referer<span class="pl-pds">'</span></span>
map location() onto <span class="pl-s"><span class="pl-pds">'</span>location<span class="pl-pds">'</span></span>
map viewportPixelWidth() onto <span class="pl-s"><span class="pl-pds">'</span>viewportPixelWidth<span class="pl-pds">'</span></span>
map viewportPixelHeight() onto <span class="pl-s"><span class="pl-pds">'</span>viewportPixelHeight<span class="pl-pds">'</span></span>
map screenPixelWidth() onto <span class="pl-s"><span class="pl-pds">'</span>screenPixelWidth<span class="pl-pds">'</span></span>
map screenPixelHeight() onto <span class="pl-s"><span class="pl-pds">'</span>screenPixelHeight<span class="pl-pds">'</span></span>
map partyId() onto <span class="pl-s"><span class="pl-pds">'</span>partyId<span class="pl-pds">'</span></span>
map sessionId() onto <span class="pl-s"><span class="pl-pds">'</span>sessionId<span class="pl-pds">'</span></span>
map pageViewId() onto <span class="pl-s"><span class="pl-pds">'</span>pageViewId<span class="pl-pds">'</span></span>
map eventType() onto <span class="pl-s"><span class="pl-pds">'</span>eventType<span class="pl-pds">'</span></span>
map userAgentString() onto <span class="pl-s"><span class="pl-pds">'</span>userAgentString<span class="pl-pds">'</span></span>
<span class="pl-k">def</span> ua <span class="pl-k">=</span> userAgent()
map ua<span class="pl-k">.</span>name() onto <span class="pl-s"><span class="pl-pds">'</span>userAgentName<span class="pl-pds">'</span></span>
map ua<span class="pl-k">.</span>family() onto <span class="pl-s"><span class="pl-pds">'</span>userAgentFamily<span class="pl-pds">'</span></span>
map ua<span class="pl-k">.</span>vendor() onto <span class="pl-s"><span class="pl-pds">'</span>userAgentVendor<span class="pl-pds">'</span></span>
map ua<span class="pl-k">.</span>type() onto <span class="pl-s"><span class="pl-pds">'</span>userAgentType<span class="pl-pds">'</span></span>
map ua<span class="pl-k">.</span>version() onto <span class="pl-s"><span class="pl-pds">'</span>userAgentVersion<span class="pl-pds">'</span></span>
map ua<span class="pl-k">.</span>deviceCategory() onto <span class="pl-s"><span class="pl-pds">'</span>userAgentDeviceCategory<span class="pl-pds">'</span></span>
map ua<span class="pl-k">.</span>osFamily() onto <span class="pl-s"><span class="pl-pds">'</span>userAgentOsFamily<span class="pl-pds">'</span></span>
map ua<span class="pl-k">.</span>osVersion() onto <span class="pl-s"><span class="pl-pds">'</span>userAgentOsVersion<span class="pl-pds">'</span></span>
map ua<span class="pl-k">.</span>osVendor() onto <span class="pl-s"><span class="pl-pds">'</span>userAgentOsVendor<span class="pl-pds">'</span></span>
map eventParameter(<span class="pl-s"><span class="pl-pds">'</span>technology<span class="pl-pds">'</span></span>) onto <span class="pl-s"><span class="pl-pds">'</span>technology<span class="pl-pds">'</span></span>
}</pre></div>
<h2 dir="auto">Ingestion Report</h2>
<div class="highlight highlight-source-json notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="{
"ingestionStatsAndErrors": {
"taskId": "index_kafka_livestream_0db516aaf980d12_bjfgbipd",
"payload": {
"ingestionState": "COMPLETED",
"unparseableEvents": {
"buildSegments": [
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!",
"Fail to decode avro message!"
]
},
"rowStats": {
"buildSegments": {
"processed": 0,
"processedWithError": 0,
"thrownAway": 0,
"unparseable": 24
}
},
"errorMsg": null
},
"type": "ingestionStatsAndErrors"
}
}"><pre class="notranslate">{
<span class="pl-ent">"ingestionStatsAndErrors"</span>: {
<span class="pl-ent">"taskId"</span>: <span class="pl-s"><span class="pl-pds">"</span>index_kafka_livestream_0db516aaf980d12_bjfgbipd<span class="pl-pds">"</span></span>,
<span class="pl-ent">"payload"</span>: {
<span class="pl-ent">"ingestionState"</span>: <span class="pl-s"><span class="pl-pds">"</span>COMPLETED<span class="pl-pds">"</span></span>,
<span class="pl-ent">"unparseableEvents"</span>: {
<span class="pl-ent">"buildSegments"</span>: [
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>,
<span class="pl-s"><span class="pl-pds">"</span>Fail to decode avro message!<span class="pl-pds">"</span></span>
]
},
<span class="pl-ent">"rowStats"</span>: {
<span class="pl-ent">"buildSegments"</span>: {
<span class="pl-ent">"processed"</span>: <span class="pl-c1">0</span>,
<span class="pl-ent">"processedWithError"</span>: <span class="pl-c1">0</span>,
<span class="pl-ent">"thrownAway"</span>: <span class="pl-c1">0</span>,
<span class="pl-ent">"unparseable"</span>: <span class="pl-c1">24</span>
}
},
<span class="pl-ent">"errorMsg"</span>: <span class="pl-c1">null</span>
},
<span class="pl-ent">"type"</span>: <span class="pl-s"><span class="pl-pds">"</span>ingestionStatsAndErrors<span class="pl-pds">"</span></span>
}
}</pre></div> | <p dir="auto">Apache Druid 24.0.1 is a bug fix release that fixes some issues in the 24.0 release.<br>
<a href="https://github.com/apache/druid/milestone/47?closed=1">See the complete set of changes for additional details</a>.</p>
<h2 dir="auto"><a name="user-content-24-1-bug-fixes" href="#24-1-bug-fixes">#</a> Notable Bug fixes</h2>
<p dir="auto"><a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="1405368637" data-permission-text="Title is private" data-url="https://github.com/apache/druid/issues/13214" data-hovercard-type="pull_request" data-hovercard-url="/apache/druid/pull/13214/hovercard" href="https://github.com/apache/druid/pull/13214">#13214</a> to fix SQL planning when using the JSON_VALUE function.<br>
<a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="1432783586" data-permission-text="Title is private" data-url="https://github.com/apache/druid/issues/13297" data-hovercard-type="pull_request" data-hovercard-url="/apache/druid/pull/13297/hovercard" href="https://github.com/apache/druid/pull/13297">#13297</a> to fix values that match a range filter on nested columns.<br>
<a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="1370404448" data-permission-text="Title is private" data-url="https://github.com/apache/druid/issues/13077" data-hovercard-type="pull_request" data-hovercard-url="/apache/druid/pull/13077/hovercard" href="https://github.com/apache/druid/pull/13077">#13077</a> to fix detection of nested objects while generating an MSQ SQL in the web-console.<br>
<a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="1394603647" data-permission-text="Title is private" data-url="https://github.com/apache/druid/issues/13172" data-hovercard-type="pull_request" data-hovercard-url="/apache/druid/pull/13172/hovercard" href="https://github.com/apache/druid/pull/13172">#13172</a> to correctly handle overlord leader election even when tasks cannot be reacquired.<br>
<a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="1426857914" data-permission-text="Title is private" data-url="https://github.com/apache/druid/issues/13273" data-hovercard-type="pull_request" data-hovercard-url="/apache/druid/pull/13273/hovercard" href="https://github.com/apache/druid/pull/13273">#13273</a> to fix overlord API failures by de-duplicating task entries in memory.<br>
<a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="1365283581" data-permission-text="Title is private" data-url="https://github.com/apache/druid/issues/13049" data-hovercard-type="pull_request" data-hovercard-url="/apache/druid/pull/13049/hovercard" href="https://github.com/apache/druid/pull/13049">#13049</a> to fix a race condition while processing query context.<br>
<a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="1389053793" data-permission-text="Title is private" data-url="https://github.com/apache/druid/issues/13151" data-hovercard-type="pull_request" data-hovercard-url="/apache/druid/pull/13151/hovercard" href="https://github.com/apache/druid/pull/13151">#13151</a> to fix assertion error in SQL planning.</p>
<h2 dir="auto"><a name="user-content-24-1-credits" href="#24-1-credits">#</a> Credits</h2>
<p dir="auto">Thanks to everyone who contributed to this release!</p>
<p dir="auto"><a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/abhishekagarwal87/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/abhishekagarwal87">@abhishekagarwal87</a><br>
<a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/AmatyaAvadhanula/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/AmatyaAvadhanula">@AmatyaAvadhanula</a><br>
<a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/clintropolis/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/clintropolis">@clintropolis</a><br>
<a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/gianm/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/gianm">@gianm</a><br>
<a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/kfaraz/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/kfaraz">@kfaraz</a><br>
<a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/LakshSingla/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/LakshSingla">@LakshSingla</a><br>
<a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/paul-rogers/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/paul-rogers">@paul-rogers</a><br>
<a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/vogievetsky/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/vogievetsky">@vogievetsky</a></p> | 0 |
<p dir="auto">As far as I can tell, the static <code class="notranslate">getInitialProps</code> doesn't provide a uniform way to access url params on the client and server. On the server, you can access <code class="notranslate">req</code> and parse the url yourself. On the client, we don't seem to have access to the route information within the <code class="notranslate">getInitialProps</code> function. The uses for this are fairly obvious, i think, such as loading a particular user profile based on an <code class="notranslate">:id</code> param, and so on. I'm also wondering if auth could somehow be implemented here too -- thought I'm less clear if that's the best API.</p> | <ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I have searched the <a href="https://github.com/zeit/next.js/issues">issues</a> of this repository and believe that this is not a duplicate.</li>
</ul>
<p dir="auto">This <code class="notranslate">pages/_document.js</code> does change color only after 2 triggered HMRs:</p>
<div class="highlight highlight-source-js notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="import Document, {Head, Main, NextScript} from 'next/document';
export default class MyDocument extends Document {
static getInitialProps(ctx) {
return super.getInitialProps(ctx);
}
render() {
return (
<html>
<Head>
<style jsx global>{`
body {
color: green;
}
`}</style>
</Head>
<body>
<Main />
<NextScript />
</body>
</html>
);
}
};"><pre class="notranslate"><span class="pl-k">import</span> <span class="pl-v">Document</span><span class="pl-kos">,</span> <span class="pl-kos">{</span><span class="pl-v">Head</span><span class="pl-kos">,</span> <span class="pl-v">Main</span><span class="pl-kos">,</span> <span class="pl-v">NextScript</span><span class="pl-kos">}</span> <span class="pl-k">from</span> <span class="pl-s">'next/document'</span><span class="pl-kos">;</span>
<span class="pl-k">export</span> <span class="pl-k">default</span> <span class="pl-k">class</span> <span class="pl-v">MyDocument</span> <span class="pl-k">extends</span> <span class="pl-v">Document</span> <span class="pl-kos">{</span>
<span class="pl-k">static</span> <span class="pl-en">getInitialProps</span><span class="pl-kos">(</span><span class="pl-s1">ctx</span><span class="pl-kos">)</span> <span class="pl-kos">{</span>
<span class="pl-k">return</span> <span class="pl-smi">super</span><span class="pl-kos">.</span><span class="pl-en">getInitialProps</span><span class="pl-kos">(</span><span class="pl-s1">ctx</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span>
<span class="pl-en">render</span><span class="pl-kos">(</span><span class="pl-kos">)</span> <span class="pl-kos">{</span>
<span class="pl-k">return</span> <span class="pl-kos">(</span>
<span class="pl-c1"><</span><span class="pl-ent">html</span><span class="pl-c1">></span>
<span class="pl-c1"><</span><span class="pl-ent">Head</span><span class="pl-c1">></span>
<span class="pl-c1"><</span><span class="pl-ent">style</span> <span class="pl-c1">jsx</span> <span class="pl-c1">global</span><span class="pl-c1">></span><span class="pl-kos">{</span><span class="pl-s">`</span>
<span class="pl-s"> body {</span>
<span class="pl-s"> color: green;</span>
<span class="pl-s"> }</span>
<span class="pl-s"> `</span><span class="pl-kos">}</span><span class="pl-c1"><</span><span class="pl-c1">/</span><span class="pl-ent">style</span><span class="pl-c1">></span>
<span class="pl-c1"><</span><span class="pl-c1">/</span><span class="pl-ent">Head</span><span class="pl-c1">></span>
<span class="pl-c1"><</span><span class="pl-ent">body</span><span class="pl-c1">></span>
<span class="pl-c1"><</span><span class="pl-ent">Main</span> <span class="pl-c1">/</span><span class="pl-c1">></span>
<span class="pl-c1"><</span><span class="pl-ent">NextScript</span> <span class="pl-c1">/</span><span class="pl-c1">></span>
<span class="pl-c1"><</span><span class="pl-c1">/</span><span class="pl-ent">body</span><span class="pl-c1">></span>
<span class="pl-c1"><</span><span class="pl-c1">/</span><span class="pl-ent">html</span><span class="pl-c1">></span>
<span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span>
<span class="pl-kos">}</span><span class="pl-kos">;</span></pre></div>
<p dir="auto">This is in a fresh project (<code class="notranslate">yarn add next@^3.0.1-beta react react-dom</code>, hello world <code class="notranslate">pages/index.js</code> and the above in <code class="notranslate">pages/_document.js</code>), running server with <code class="notranslate">yarn next</code>.</p>
<h2 dir="auto">Expected Behavior</h2>
<p dir="auto">When I edit <code class="notranslate">color: green</code> to another color (e.g. red) and save, page updates to show the new color.</p>
<h2 dir="auto">Current Behavior</h2>
<p dir="auto">When I edit <code class="notranslate">color: green</code> to another color (e.g. red) and save, <em>page stays green</em>. Then I edit the color again (e.g. black) and save, <em>page becomes red</em>.</p>
<h2 dir="auto">Context</h2>
<p dir="auto">Trying to get a new project with some global styling going.</p>
<h2 dir="auto">Your Environment</h2>
<table role="table">
<thead>
<tr>
<th>Tech</th>
<th>Version</th>
</tr>
</thead>
<tbody>
<tr>
<td>next</td>
<td>2.4.7 and 3.0.1-beta.13 (using styled-jsx 1.0.8)</td>
</tr>
<tr>
<td>node</td>
<td>8.1.3</td>
</tr>
<tr>
<td>OS</td>
<td>OS X 10.9.5</td>
</tr>
<tr>
<td>browser</td>
<td>Chrome 59</td>
</tr>
<tr>
<td>editor</td>
<td>Sublime Text 3 (build 3141)</td>
</tr>
</tbody>
</table> | 0 |
<h3 dir="auto">Bug report</h3>
<p dir="auto"><strong>Bug summary</strong></p>
<p dir="auto">Calling show() after generating a simple line plot causes my machine to log my user out in what appears to be a forced shutdown. This problem only occurs when running python through terminal and not through a Jupyter Notebook (though in a notebook I have to run the cell twice to see any results).</p>
<p dir="auto"><strong>Code for reproduction</strong></p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="import numpy as np
import matplotlib.pyplot as plt
cond1 = np.array([25]*10)
plt.plot(cond1)
plt.show()"><pre class="notranslate"><span class="pl-k">import</span> <span class="pl-s1">numpy</span> <span class="pl-k">as</span> <span class="pl-s1">np</span>
<span class="pl-k">import</span> <span class="pl-s1">matplotlib</span>.<span class="pl-s1">pyplot</span> <span class="pl-k">as</span> <span class="pl-s1">plt</span>
<span class="pl-s1">cond1</span> <span class="pl-c1">=</span> <span class="pl-s1">np</span>.<span class="pl-en">array</span>([<span class="pl-c1">25</span>]<span class="pl-c1">*</span><span class="pl-c1">10</span>)
<span class="pl-s1">plt</span>.<span class="pl-en">plot</span>(<span class="pl-s1">cond1</span>)
<span class="pl-s1">plt</span>.<span class="pl-en">show</span>()</pre></div>
<p dir="auto"><strong>Actual outcome</strong></p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="(base) Johns-MacBook-Pro:Work_Code johnwilson$ python condition_numbers_new.py
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.644 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.644 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.644 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.649 python[11487:59261] HIToolbox: received notification of WindowServer event port death.
2019-08-27 16:28:51.649 python[11487:59261] port matched the WindowServer port created in BindCGSToRunLoop"><pre class="notranslate"><code class="notranslate">(base) Johns-MacBook-Pro:Work_Code johnwilson$ python condition_numbers_new.py
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.643 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.644 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.644 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.644 python[11487:59261] CGSTrackingRegionSetIsEnabled returned CG error 268435459
2019-08-27 16:28:51.649 python[11487:59261] HIToolbox: received notification of WindowServer event port death.
2019-08-27 16:28:51.649 python[11487:59261] port matched the WindowServer port created in BindCGSToRunLoop
</code></pre></div>
<p dir="auto"><strong>Expected outcome</strong></p>
<p dir="auto">This code should produce the following plot (generated by running the above code in a jupyter notebook) instead of shutting down.<br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/29575095/63810335-2464d080-c8ea-11e9-97ea-42167f8dd08b.png"><img width="388" alt="Screen Shot 2019-08-27 at 4 45 45 PM" src="https://user-images.githubusercontent.com/29575095/63810335-2464d080-c8ea-11e9-97ea-42167f8dd08b.png" style="max-width: 100%;"></a></p>
<p dir="auto"><strong>Matplotlib version</strong></p>
<ul dir="auto">
<li>Operating system: macOS Mojave, Version 10.14.6</li>
<li>Matplotlib version: 3.1.0</li>
<li>Matplotlib backend: module://ipykernel.pylab.backend_inline</li>
<li>Python version: Python 3.7.3</li>
<li>Jupyter version:
<ul dir="auto">
<li>jupyter==1.0.0</li>
<li>jupyter-client==5.3.1</li>
<li>jupyter-console==6.0.0</li>
<li>jupyter-core==4.5.0</li>
</ul>
</li>
<li>Other libraries:
<ul dir="auto">
<li>numpy==1.16.4</li>
</ul>
</li>
</ul>
<p dir="auto">I installed everything through the Anaconda distribution using default settings just last week.</p> | <p dir="auto">The weekly build with nightly wheels from numpy and pandas<br>
has failed. Check the logs for any updates that need to be<br>
made in matplotlib.<br>
<a href="https://github.com/matplotlib/matplotlib/actions/runs/4390878412">https://github.com/matplotlib/matplotlib/actions/runs/4390878412</a></p> | 0 |
<p dir="auto">I'm receiving a weird behavior which is: when using tab key to switch between <code class="notranslate">TextField</code> for the <strong>first</strong> time, the position of the <code class="notranslate">TextField</code> changes a little. The following screen shot might better explain the situation. I think it's because of 'FloatingLabelText` but what I don't understand is why only tab key's event causes this problem not click event, and why does this behavior only happens the first time.</p>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I have searched the <a href="https://github.com/mui-org/material-ui/issues">issues</a> of this repository and believe that this is not a duplicate.</li>
</ul>
<h2 dir="auto">Expected Behavior</h2>
<p dir="auto">When focus on a textfield using tab key, the position should stay the same.</p>
<h2 dir="auto">Current Behavior</h2>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/19363363/34689918-0b936ee4-f46c-11e7-87c2-110472a28ff6.gif"><img src="https://user-images.githubusercontent.com/19363363/34689918-0b936ee4-f46c-11e7-87c2-110472a28ff6.gif" alt="giphy" data-animated-image="" style="max-width: 100%;"></a><br>
As shown above, only the position of TextField named 'treatmentType' is not changed because it's focused by click event rather than tab key. And if the same <code class="notranslate">TextField</code> was focused by tab before, its position won't be changed.</p>
<h2 dir="auto">Steps to Reproduce (for bugs)</h2>
<ol dir="auto">
<li>Render component: Table > TableBody >TableRow >TableRowColumn > TextField</li>
<li>Click on the first <code class="notranslate">TextField</code> in the first <code class="notranslate">TableRowColumn</code></li>
<li>Switch focus on the other<code class="notranslate">TextField</code> by pressing Tab key.</li>
</ol>
<div class="highlight highlight-source-js notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="class TreatmentContent extends React.Component {
constructor(props) {
super(props);
this.state = {
type: '',
notes: '',
status: '',
location: '',
food: '',
firstName: '',
lastName: '',
email: '',
day: ''
};
this.notesToAdd = {
type: 'text',
notes: 'text',
status: 'text',
location: 'text',
food: 'text',
firstName: 'text',
lastName: 'text',
email: 'text',
day: 'number',
};
this.reorgTableRow = this.reorgTableRow.bind(this);
}
reorgTableRow() {
const notesKeys = _.keys(this.notesToAdd);
const chunkedKeys = _.chunk(notesKeys, 3);
return _.map(chunkedKeys, (keyChunk, i) => (
<TableRow selectable={false} key={i} style={{border: 'None'}}>
{_.map(keyChunk, (keyName, j) => (
<TableRowColumn key={j}>
<TextField
hintText={keyName}
floatingLabelText={keyName}
type={this.notesToAdd[keyName]}
value={this.state[keyName]}
/>
</TableRowColumn>
))}
</TableRow>
));
}
render() {
return (
<div>
<Table>
<TableBody displayRowCheckbox={false}>
{this.reorgTableRow()}
</TableBody>
</Table>
</div>
);
}
}"><pre class="notranslate"><span class="pl-k">class</span> <span class="pl-v">TreatmentContent</span> <span class="pl-k">extends</span> <span class="pl-v">React</span><span class="pl-kos">.</span><span class="pl-c1">Component</span> <span class="pl-kos">{</span>
<span class="pl-en">constructor</span><span class="pl-kos">(</span><span class="pl-s1">props</span><span class="pl-kos">)</span> <span class="pl-kos">{</span>
<span class="pl-smi">super</span><span class="pl-kos">(</span><span class="pl-s1">props</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-c1">state</span> <span class="pl-c1">=</span> <span class="pl-kos">{</span>
<span class="pl-c1">type</span>: <span class="pl-s">''</span><span class="pl-kos">,</span>
<span class="pl-c1">notes</span>: <span class="pl-s">''</span><span class="pl-kos">,</span>
<span class="pl-c1">status</span>: <span class="pl-s">''</span><span class="pl-kos">,</span>
<span class="pl-c1">location</span>: <span class="pl-s">''</span><span class="pl-kos">,</span>
<span class="pl-c1">food</span>: <span class="pl-s">''</span><span class="pl-kos">,</span>
<span class="pl-c1">firstName</span>: <span class="pl-s">''</span><span class="pl-kos">,</span>
<span class="pl-c1">lastName</span>: <span class="pl-s">''</span><span class="pl-kos">,</span>
<span class="pl-c1">email</span>: <span class="pl-s">''</span><span class="pl-kos">,</span>
<span class="pl-c1">day</span>: <span class="pl-s">''</span>
<span class="pl-kos">}</span><span class="pl-kos">;</span>
<span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-c1">notesToAdd</span> <span class="pl-c1">=</span> <span class="pl-kos">{</span>
<span class="pl-c1">type</span>: <span class="pl-s">'text'</span><span class="pl-kos">,</span>
<span class="pl-c1">notes</span>: <span class="pl-s">'text'</span><span class="pl-kos">,</span>
<span class="pl-c1">status</span>: <span class="pl-s">'text'</span><span class="pl-kos">,</span>
<span class="pl-c1">location</span>: <span class="pl-s">'text'</span><span class="pl-kos">,</span>
<span class="pl-c1">food</span>: <span class="pl-s">'text'</span><span class="pl-kos">,</span>
<span class="pl-c1">firstName</span>: <span class="pl-s">'text'</span><span class="pl-kos">,</span>
<span class="pl-c1">lastName</span>: <span class="pl-s">'text'</span><span class="pl-kos">,</span>
<span class="pl-c1">email</span>: <span class="pl-s">'text'</span><span class="pl-kos">,</span>
<span class="pl-c1">day</span>: <span class="pl-s">'number'</span><span class="pl-kos">,</span>
<span class="pl-kos">}</span><span class="pl-kos">;</span>
<span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-c1">reorgTableRow</span> <span class="pl-c1">=</span> <span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-c1">reorgTableRow</span><span class="pl-kos">.</span><span class="pl-en">bind</span><span class="pl-kos">(</span><span class="pl-smi">this</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span>
<span class="pl-en">reorgTableRow</span><span class="pl-kos">(</span><span class="pl-kos">)</span> <span class="pl-kos">{</span>
<span class="pl-k">const</span> <span class="pl-s1">notesKeys</span> <span class="pl-c1">=</span> <span class="pl-s1">_</span><span class="pl-kos">.</span><span class="pl-en">keys</span><span class="pl-kos">(</span><span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-c1">notesToAdd</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-k">const</span> <span class="pl-s1">chunkedKeys</span> <span class="pl-c1">=</span> <span class="pl-s1">_</span><span class="pl-kos">.</span><span class="pl-en">chunk</span><span class="pl-kos">(</span><span class="pl-s1">notesKeys</span><span class="pl-kos">,</span> <span class="pl-c1">3</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-k">return</span> <span class="pl-s1">_</span><span class="pl-kos">.</span><span class="pl-en">map</span><span class="pl-kos">(</span><span class="pl-s1">chunkedKeys</span><span class="pl-kos">,</span> <span class="pl-kos">(</span><span class="pl-s1">keyChunk</span><span class="pl-kos">,</span> <span class="pl-s1">i</span><span class="pl-kos">)</span> <span class="pl-c1">=></span> <span class="pl-kos">(</span>
<span class="pl-c1"><</span><span class="pl-ent">TableRow</span> <span class="pl-c1">selectable</span><span class="pl-c1">=</span><span class="pl-kos">{</span><span class="pl-c1">false</span><span class="pl-kos">}</span> <span class="pl-c1">key</span><span class="pl-c1">=</span><span class="pl-kos">{</span><span class="pl-s1">i</span><span class="pl-kos">}</span> <span class="pl-c1">style</span><span class="pl-c1">=</span><span class="pl-kos">{</span><span class="pl-kos">{</span><span class="pl-c1">border</span>: <span class="pl-s">'None'</span><span class="pl-kos">}</span><span class="pl-kos">}</span><span class="pl-c1">></span>
<span class="pl-kos">{</span><span class="pl-s1">_</span><span class="pl-kos">.</span><span class="pl-en">map</span><span class="pl-kos">(</span><span class="pl-s1">keyChunk</span><span class="pl-kos">,</span> <span class="pl-kos">(</span><span class="pl-s1">keyName</span><span class="pl-kos">,</span> <span class="pl-s1">j</span><span class="pl-kos">)</span> <span class="pl-c1">=></span> <span class="pl-kos">(</span>
<span class="pl-c1"><</span><span class="pl-ent">TableRowColumn</span> <span class="pl-c1">key</span><span class="pl-c1">=</span><span class="pl-kos">{</span><span class="pl-s1">j</span><span class="pl-kos">}</span><span class="pl-c1">></span>
<span class="pl-c1"><</span><span class="pl-ent">TextField</span>
<span class="pl-c1">hintText</span><span class="pl-c1">=</span><span class="pl-kos">{</span><span class="pl-s1">keyName</span><span class="pl-kos">}</span>
<span class="pl-c1">floatingLabelText</span><span class="pl-c1">=</span><span class="pl-kos">{</span><span class="pl-s1">keyName</span><span class="pl-kos">}</span>
<span class="pl-c1">type</span><span class="pl-c1">=</span><span class="pl-kos">{</span><span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-c1">notesToAdd</span><span class="pl-kos">[</span><span class="pl-s1">keyName</span><span class="pl-kos">]</span><span class="pl-kos">}</span>
<span class="pl-c1">value</span><span class="pl-c1">=</span><span class="pl-kos">{</span><span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-c1">state</span><span class="pl-kos">[</span><span class="pl-s1">keyName</span><span class="pl-kos">]</span><span class="pl-kos">}</span>
<span class="pl-c1">/</span><span class="pl-c1">></span>
<span class="pl-c1"><</span><span class="pl-c1">/</span><span class="pl-ent">TableRowColumn</span><span class="pl-c1">></span>
<span class="pl-kos">)</span><span class="pl-kos">)</span><span class="pl-kos">}</span>
<span class="pl-c1"><</span><span class="pl-c1">/</span><span class="pl-ent">TableRow</span><span class="pl-c1">></span>
<span class="pl-kos">)</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span>
<span class="pl-en">render</span><span class="pl-kos">(</span><span class="pl-kos">)</span> <span class="pl-kos">{</span>
<span class="pl-k">return</span> <span class="pl-kos">(</span>
<span class="pl-c1"><</span><span class="pl-ent">div</span><span class="pl-c1">></span>
<span class="pl-c1"><</span><span class="pl-ent">Table</span><span class="pl-c1">></span>
<span class="pl-c1"><</span><span class="pl-ent">TableBody</span> <span class="pl-c1">displayRowCheckbox</span><span class="pl-c1">=</span><span class="pl-kos">{</span><span class="pl-c1">false</span><span class="pl-kos">}</span><span class="pl-c1">></span>
<span class="pl-kos">{</span><span class="pl-smi">this</span><span class="pl-kos">.</span><span class="pl-en">reorgTableRow</span><span class="pl-kos">(</span><span class="pl-kos">)</span><span class="pl-kos">}</span>
<span class="pl-c1"><</span><span class="pl-c1">/</span><span class="pl-ent">TableBody</span><span class="pl-c1">></span>
<span class="pl-c1"><</span><span class="pl-c1">/</span><span class="pl-ent">Table</span><span class="pl-c1">></span>
<span class="pl-c1"><</span><span class="pl-c1">/</span><span class="pl-ent">div</span><span class="pl-c1">></span>
<span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span>
<span class="pl-kos">}</span></pre></div>
<h2 dir="auto">Context</h2>
<h2 dir="auto">Your Environment</h2>
<p dir="auto">"react": "^15.6.1",<br>
"material-ui": "latest",<br>
browser: Chrome v63</p>
<table role="table">
<thead>
<tr>
<th>Tech</th>
<th>Version</th>
</tr>
</thead>
<tbody>
<tr>
<td>Material-UI</td>
<td></td>
</tr>
<tr>
<td>React</td>
<td></td>
</tr>
<tr>
<td>browser</td>
<td></td>
</tr>
<tr>
<td>etc</td>
<td></td>
</tr>
</tbody>
</table> | <ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I have searched the <a href="https://github.com/callemall/material-ui/issues">issues</a> of this repository and believe that this is not a duplicate.</li>
</ul>
<h2 dir="auto">Expected Behavior</h2>
<p dir="auto">The tab indicator must always show the selected tab</p>
<h2 dir="auto">Current Behavior</h2>
<p dir="auto">When the number of tabs in a tab menu changes the indicator have a wrong position until next interaction such as clicking a new tab.</p>
<h2 dir="auto">Steps to Reproduce (for bugs)</h2>
<ol dir="auto">
<li>Go to <a href="https://codesandbox.io/s/2pm5jzmk10" rel="nofollow">https://codesandbox.io/s/2pm5jzmk10</a></li>
<li>Select a random tab</li>
<li>Click switch</li>
<li>Chehck indicator position</li>
</ol>
<h2 dir="auto">Context</h2>
<p dir="auto">I Noticed this issue when trying to remove some tabs based on the users permission role. I quickly realised that the indicator was not in sync with the amount of tabs inside the menu.</p>
<h2 dir="auto">Reproduction Environment</h2>
<table role="table">
<thead>
<tr>
<th>Tech</th>
<th>Version</th>
</tr>
</thead>
<tbody>
<tr>
<td>Material-UI</td>
<td>1.0.0-beta.11</td>
</tr>
<tr>
<td>React</td>
<td>15.5.3</td>
</tr>
<tr>
<td>browser</td>
<td>Google Chrome v61</td>
</tr>
</tbody>
</table>
<h2 dir="auto">Your Environment</h2>
<table role="table">
<thead>
<tr>
<th>Tech</th>
<th>Version</th>
</tr>
</thead>
<tbody>
<tr>
<td>Material-UI</td>
<td>1.0.0-beta.6</td>
</tr>
<tr>
<td>React</td>
<td>15.6.1</td>
</tr>
<tr>
<td>browser</td>
<td>Google Chrome v61</td>
</tr>
</tbody>
</table> | 0 |
<p dir="auto">Describe what you were doing when the bug occurred:</p>
<ol dir="auto">
<li>I was watching the Flamegraph</li>
<li>I clicked on a node</li>
<li>I clicked on the Ranked tab and the bug occurred</li>
</ol>
<hr>
<h2 dir="auto">Please do not remove the text below this line</h2>
<p dir="auto">DevTools version: 4.2.1-3816ae7c3</p>
<p dir="auto">Call stack: at chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:40:157108<br>
at Map.forEach ()<br>
at commitIndex (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:40:157054)<br>
at e.getRankedChartData (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:40:157577)<br>
at vl (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:40:314907)<br>
at gi (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:32:59907)<br>
at jl (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:32:107381)<br>
at Lc (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:32:92715)<br>
at Pc (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:32:92640)<br>
at wc (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:32:89544)</p>
<p dir="auto">Component stack: in vl<br>
in div<br>
in div<br>
in div<br>
in wo<br>
in Unknown<br>
in n<br>
in Unknown<br>
in div<br>
in div<br>
in Li<br>
in $e<br>
in dn<br>
in Ca<br>
in Pc</p> | <p dir="auto">PLEASE INCLUDE REPRO INSTRUCTIONS AND EXAMPLE CODE</p>
<p dir="auto">I got this error when I click 'Ranked'.</p>
<hr>
<h2 dir="auto">Please do not remove the text below this line</h2>
<p dir="auto">DevTools version: 4.0.4-3c6a219</p>
<p dir="auto">Call stack: at chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:11:11441<br>
at Map.forEach ()<br>
at commitIndex (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:11:11387)<br>
at e.getRankedChartData (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:11:11920)<br>
at _i (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:56:277123)<br>
at Ha (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:43:55890)<br>
at Xl (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:43:98280)<br>
at Hl (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:43:84255)<br>
at Fl (chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:43:81285)<br>
at chrome-extension://fmkadmapgofadopljbjfkapdkoienihi/build/main.js:43:25363</p>
<p dir="auto">Component stack: in _i<br>
in div<br>
in div<br>
in div<br>
in Or<br>
in Unknown<br>
in n<br>
in Unknown<br>
in div<br>
in div<br>
in Ha<br>
in le<br>
in ve<br>
in ko<br>
in Ul</p> | 1 |
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> I have searched the <a href="https://github.com/apache/dubbo/issues">issues</a> of this repository and believe that this is not a duplicate.</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> I have checked the <a href="https://github.com/apache/dubbo/blob/master/FAQ.md">FAQ</a> of this repository and believe that this is not a duplicate.</li>
</ul>
<h3 dir="auto">Environment</h3>
<ul dir="auto">
<li>Dubbo version: xxx</li>
<li>Operating System version: xxx</li>
<li>Java version: xxx</li>
</ul>
<h3 dir="auto">Steps to reproduce this issue</h3>
<ol dir="auto">
<li>xxx</li>
<li>xxx</li>
<li>xxx</li>
</ol>
<p dir="auto">Pls. provide [GitHub address] to reproduce this issue.</p>
<h3 dir="auto">Expected Result</h3>
<p dir="auto">What do you expected from the above steps?</p>
<h3 dir="auto">Actual Result</h3>
<p dir="auto">What actually happens?</p>
<p dir="auto">If there is an exception, please attach the exception trace:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="Just put your stack trace here!"><pre class="notranslate"><code class="notranslate">Just put your stack trace here!
</code></pre></div> | <ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I have searched the <a href="https://github.com/apache/dubbo/issues">issues</a> of this repository and believe that this is not a duplicate.</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> I have checked the <a href="https://github.com/apache/dubbo/blob/master/FAQ.md">FAQ</a> of this repository and believe that this is not a duplicate.</li>
</ul>
<h3 dir="auto">Environment</h3>
<ul dir="auto">
<li>Dubbo version: 2.7.7</li>
<li>Operating System version: MacOS 10.14.6</li>
<li>Java version: 1.8.0_231</li>
</ul>
<h3 dir="auto">Steps to reproduce this issue</h3>
<ol dir="auto">
<li>参数验证信息被吞掉了</li>
</ol>
<p dir="auto"><a href="https://github.com/apache/dubbo-samples/blob/master/java/dubbo-samples-validation/src/main/java/org/apache/dubbo/samples/validation/ValidationConsumer.java">https://github.com/apache/dubbo-samples/blob/master/java/dubbo-samples-validation/src/main/java/org/apache/dubbo/samples/validation/ValidationConsumer.java</a></p>
<h3 dir="auto">Expected Result</h3>
<p dir="auto">What do you expected from the above steps?</p>
<h3 dir="auto">Actual Result</h3>
<p dir="auto">What actually happens?</p>
<p dir="auto">If there is an exception, please attach the exception trace:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="Just put your stack trace here!"><pre class="notranslate"><code class="notranslate">Just put your stack trace here!
</code></pre></div>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/16585330/82537560-d03b0400-9b7c-11ea-8f34-d8bdc080519a.png"><img src="https://user-images.githubusercontent.com/16585330/82537560-d03b0400-9b7c-11ea-8f34-d8bdc080519a.png" alt="image" style="max-width: 100%;"></a><br>
<a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/16585330/82537804-37f14f00-9b7d-11ea-92fa-02054d0f99df.png"><img src="https://user-images.githubusercontent.com/16585330/82537804-37f14f00-9b7d-11ea-92fa-02054d0f99df.png" alt="image" style="max-width: 100%;"></a></p>
<p dir="auto"><a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="533089340" data-permission-text="Title is private" data-url="https://github.com/apache/dubbo/issues/5432" data-hovercard-type="issue" data-hovercard-url="/apache/dubbo/issues/5432/hovercard" href="https://github.com/apache/dubbo/issues/5432">#5432</a></p> | 0 |
<p dir="auto">I have been using Julia-1.0 at the Terminal and under Jupyter for some time under the previous MacOS. I recently upgraded to Catalina 10.15.2. I can still use Julia 1.0 under Terminal, but when I try to run the new Julia-1.3, I get the alert:</p>
<blockquote>
<p dir="auto">“Julia-1.3” can’t be opened because Apple cannot check it for malicious software.</p>
</blockquote>
<p dir="auto">Update: I am able to get around this by approving within "System Preferences / Security & Privacy".</p>
<p dir="auto">Not sure, but I am guessing this is related to the new required app notarization under Catalina. Anyone want to fight with this? Maybe someone can update the instructions in the meantime.</p> | <p dir="auto">macOS 10.15+ is going to start requiring us to "notarize" our apps by sending them to Apple to be scanned. If we do not, we will get scary errors like the following:</p>
<blockquote>
<p dir="auto">"Julia-1.0" can't be opened because Apple cannot check it for malicious software.<br>
<sub>This software needs to be updated. Contact the developer for more information.</sub></p>
</blockquote>
<p dir="auto">To temporarily work around this, right click on the application and select "open."</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/130920/65284720-44ee0800-daef-11e9-96a9-f359830baa32.png"><img src="https://user-images.githubusercontent.com/130920/65284720-44ee0800-daef-11e9-96a9-f359830baa32.png" alt="image" style="max-width: 100%;"></a></p>
<p dir="auto">Note that this is distinct from the "cannot be verified" error when an app is not signed; notarization protects against malware by forcing developers to upload their app bundles to Apple, where Apple can scan them with some algorithm, determine there is nothing harmful in there, then pass back a cryptographic notarization. This must be done for every single build of Julia, if we want the <code class="notranslate">.app</code> to be openable. (We could conceivably not do this for nightly builds, and expect users to bypass the scary dialog, if this turns out to be a flaky process)</p>
<p dir="auto">The workflow we'll have to adopt is:</p>
<ul dir="auto">
<li>
<p dir="auto">Build and package Julia, just like normal.</p>
</li>
<li>
<p dir="auto">Sign julia, but opt in to the "<a href="https://developer.apple.com/documentation/security/hardened_runtime_entitlements" rel="nofollow">hardened runtime</a>". Note that this, by default, disables things like being able to jump into writable pages, so we'll have to figure out the right set of "entitlements" to request, in order to peel back the appropriate layers of protection to get our JIT to still work. This is done by passing <code class="notranslate">--options runtime</code> to <code class="notranslate">codesign</code>.</p>
</li>
<li>
<p dir="auto">Submit the packaged <code class="notranslate">.dmg</code> to Apple Notary service. This should be doable by <code class="notranslate">/usr/bin/xcrun altool --notarize-app --primary-bundle-id "$BUNDLE_ID" --username "$APPLE_ID_EMAIL" --password "$APPLE_ID_PASSWORD" --file "$JULIA_DMG" --output-format "xml"</code></p>
</li>
<li>
<p dir="auto">Poll Apple until the notarization is complete. We do that with something like <code class="notranslate">/usr/bin/xcrun altool --notarization-info --username "$APPLE_ID_EMAIL" --password "$APPLE_ID_PASSWORD" --output-format "xml"</code>, then parse the XML. Note that this notarization can take multiple hours to complete, it is not a fast process, so I think we're going to need to run this as an asynchronous job, perhaps in parallel to testing.</p>
</li>
<li>
<p dir="auto">Staple notarization ticket to our app. <code class="notranslate">xcrun stapler staple Julia.app</code>. Note that this is performed on an <code class="notranslate">.app</code>, not a <code class="notranslate">.dmg</code>. We need to download the DMG, mount it as read/write (probably by using shadow mounts, like we do when needing to <a href="https://gist.github.com/staticfloat/62369d705f265c0b35ff8888b2e1d76c">resign something</a>), staple the ticket to the .app, then generate a <em>new</em> <code class="notranslate">.dmg</code>.</p>
</li>
</ul>
<p dir="auto">References:</p>
<ul dir="auto">
<li>
<p dir="auto"><a href="https://stackoverflow.com/questions/56890749/macos-notarize-in-script" rel="nofollow">https://stackoverflow.com/questions/56890749/macos-notarize-in-script</a></p>
</li>
<li>
<p dir="auto"><a href="https://developer.apple.com/documentation/security/hardened_runtime_entitlements" rel="nofollow">https://developer.apple.com/documentation/security/hardened_runtime_entitlements</a></p>
</li>
<li>
<p dir="auto"><a href="https://gist.github.com/staticfloat/62369d705f265c0b35ff8888b2e1d76c">https://gist.github.com/staticfloat/62369d705f265c0b35ff8888b2e1d76c</a></p>
</li>
</ul> | 1 |
<h2 dir="auto">Bug Report</h2>
<p dir="auto"><strong>Input Code</strong></p>
<div class="highlight highlight-source-js notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="const path = require('path');
const {transformSync} = require('@babel/core');
transformSync('', {
filename: path.resolve('filename.js'),
plugins: [
() => ({
visitor: {
Program(path) {
throw path.buildCodeFrameError('example error');
}
}
})
]
});"><pre class="notranslate"><span class="pl-k">const</span> <span class="pl-s1">path</span> <span class="pl-c1">=</span> <span class="pl-en">require</span><span class="pl-kos">(</span><span class="pl-s">'path'</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-k">const</span> <span class="pl-kos">{</span>transformSync<span class="pl-kos">}</span> <span class="pl-c1">=</span> <span class="pl-en">require</span><span class="pl-kos">(</span><span class="pl-s">'@babel/core'</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-en">transformSync</span><span class="pl-kos">(</span><span class="pl-s">''</span><span class="pl-kos">,</span> <span class="pl-kos">{</span>
<span class="pl-c1">filename</span>: <span class="pl-s1">path</span><span class="pl-kos">.</span><span class="pl-en">resolve</span><span class="pl-kos">(</span><span class="pl-s">'filename.js'</span><span class="pl-kos">)</span><span class="pl-kos">,</span>
<span class="pl-c1">plugins</span>: <span class="pl-kos">[</span>
<span class="pl-kos">(</span><span class="pl-kos">)</span> <span class="pl-c1">=></span> <span class="pl-kos">(</span><span class="pl-kos">{</span>
<span class="pl-c1">visitor</span>: <span class="pl-kos">{</span>
<span class="pl-en">Program</span><span class="pl-kos">(</span><span class="pl-s1">path</span><span class="pl-kos">)</span> <span class="pl-kos">{</span>
<span class="pl-k">throw</span> <span class="pl-s1">path</span><span class="pl-kos">.</span><span class="pl-en">buildCodeFrameError</span><span class="pl-kos">(</span><span class="pl-s">'example error'</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span>
<span class="pl-kos">}</span>
<span class="pl-kos">}</span><span class="pl-kos">)</span>
<span class="pl-kos">]</span>
<span class="pl-kos">}</span><span class="pl-kos">)</span><span class="pl-kos">;</span></pre></div>
<p dir="auto"><strong>Expected behavior/code</strong><br>
The first line of error.message should be <code class="notranslate">/path/to/filename.js: example error</code></p>
<p dir="auto"><strong>Current Behavior</strong><br>
The first line of error.message is <code class="notranslate">/path/to/filename.js: /path/to/filename.js: example error</code></p>
<p dir="auto"><strong>Babel Configuration (.babelrc, package.json, cli command)</strong></p>
<p dir="auto">None / see code.</p>
<p dir="auto"><strong>Environment</strong></p>
<ul dir="auto">
<li>Babel version(s): 7.6.3</li>
<li>Node/npm version: node 8+</li>
<li>OS: all</li>
<li>Monorepo: no</li>
<li>How you are using Babel: API</li>
</ul>
<p dir="auto"><strong>Possible Solution</strong><br>
Since packages/babel-core/src/transformation/index.js now prepends the error with the message maybe <code class="notranslate">path.buildCodeFrameError</code> should not do the same.</p>
<p dir="auto"><strong>Additional context/Screenshots</strong><br>
Introduced by <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="500519854" data-permission-text="Title is private" data-url="https://github.com/babel/babel/issues/10511" data-hovercard-type="pull_request" data-hovercard-url="/babel/babel/pull/10511/hovercard" href="https://github.com/babel/babel/pull/10511">#10511</a>. CC <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/JLHwung/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/JLHwung">@JLHwung</a></p> | <h2 dir="auto">Bug Report</h2>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox"> I would like to work on a fix!</li>
</ul>
<p dir="auto"><strong>Current behavior</strong></p>
<p dir="auto"><a href="https://unpkg.com/@babel/[email protected]/helpers/createSuper/_index.mjs" rel="nofollow">https://unpkg.com/@babel/[email protected]/helpers/createSuper/_index.mjs</a></p>
<div class="highlight highlight-source-js notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="import getPrototypeOf from "@babel/runtime/helpers/getPrototypeOf";
import isNativeReflectConstruct from "@babel/runtime/helpers/isNativeReflectConstruct";
import possibleConstructorReturn from "@babel/runtime/helpers/possibleConstructorReturn";
export default function _createSuper(Derived) {
var hasNativeReflectConstruct = isNativeReflectConstruct();
return function _createSuperInternal() {
var Super = getPrototypeOf(Derived),
result;
if (hasNativeReflectConstruct) {
var NewTarget = getPrototypeOf(this).constructor;
result = Reflect.construct(Super, arguments, NewTarget);
} else {
result = Super.apply(this, arguments);
}
return possibleConstructorReturn(this, result);
};
}"><pre class="notranslate"><span class="pl-k">import</span> <span class="pl-s1">getPrototypeOf</span> <span class="pl-k">from</span> <span class="pl-s">"@babel/runtime/helpers/getPrototypeOf"</span><span class="pl-kos">;</span>
<span class="pl-k">import</span> <span class="pl-s1">isNativeReflectConstruct</span> <span class="pl-k">from</span> <span class="pl-s">"@babel/runtime/helpers/isNativeReflectConstruct"</span><span class="pl-kos">;</span>
<span class="pl-k">import</span> <span class="pl-s1">possibleConstructorReturn</span> <span class="pl-k">from</span> <span class="pl-s">"@babel/runtime/helpers/possibleConstructorReturn"</span><span class="pl-kos">;</span>
<span class="pl-k">export</span> <span class="pl-k">default</span> <span class="pl-k">function</span> <span class="pl-en">_createSuper</span><span class="pl-kos">(</span><span class="pl-v">Derived</span><span class="pl-kos">)</span> <span class="pl-kos">{</span>
<span class="pl-k">var</span> <span class="pl-s1">hasNativeReflectConstruct</span> <span class="pl-c1">=</span> <span class="pl-en">isNativeReflectConstruct</span><span class="pl-kos">(</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-k">return</span> <span class="pl-k">function</span> <span class="pl-en">_createSuperInternal</span><span class="pl-kos">(</span><span class="pl-kos">)</span> <span class="pl-kos">{</span>
<span class="pl-k">var</span> <span class="pl-v">Super</span> <span class="pl-c1">=</span> <span class="pl-en">getPrototypeOf</span><span class="pl-kos">(</span><span class="pl-v">Derived</span><span class="pl-kos">)</span><span class="pl-kos">,</span>
<span class="pl-s1">result</span><span class="pl-kos">;</span>
<span class="pl-k">if</span> <span class="pl-kos">(</span><span class="pl-s1">hasNativeReflectConstruct</span><span class="pl-kos">)</span> <span class="pl-kos">{</span>
<span class="pl-k">var</span> <span class="pl-v">NewTarget</span> <span class="pl-c1">=</span> <span class="pl-en">getPrototypeOf</span><span class="pl-kos">(</span><span class="pl-smi">this</span><span class="pl-kos">)</span><span class="pl-kos">.</span><span class="pl-c1">constructor</span><span class="pl-kos">;</span>
<span class="pl-s1">result</span> <span class="pl-c1">=</span> <span class="pl-v">Reflect</span><span class="pl-kos">.</span><span class="pl-en">construct</span><span class="pl-kos">(</span><span class="pl-v">Super</span><span class="pl-kos">,</span> <span class="pl-smi">arguments</span><span class="pl-kos">,</span> <span class="pl-v">NewTarget</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span> <span class="pl-k">else</span> <span class="pl-kos">{</span>
<span class="pl-s1">result</span> <span class="pl-c1">=</span> <span class="pl-v">Super</span><span class="pl-kos">.</span><span class="pl-en">apply</span><span class="pl-kos">(</span><span class="pl-smi">this</span><span class="pl-kos">,</span> <span class="pl-smi">arguments</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span>
<span class="pl-k">return</span> <span class="pl-en">possibleConstructorReturn</span><span class="pl-kos">(</span><span class="pl-smi">this</span><span class="pl-kos">,</span> <span class="pl-s1">result</span><span class="pl-kos">)</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span><span class="pl-kos">;</span>
<span class="pl-kos">}</span></pre></div>
<p dir="auto"><strong>Expected behavior</strong></p>
<p dir="auto">Should it be?</p>
<div class="highlight highlight-source-diff notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="```js
- import getPrototypeOf from "@babel/runtime/helpers/getPrototypeOf";
- import isNativeReflectConstruct from "@babel/runtime/helpers/isNativeReflectConstruct";
- import possibleConstructorReturn from "@babel/runtime/helpers/possibleConstructorReturn";
+ import getPrototypeOf from "@babel/runtime/helpers/esm/getPrototypeOf";
+ import isNativeReflectConstruct from "@babel/runtime/helpers/esm/isNativeReflectConstruct";
+ import possibleConstructorReturn from "@babel/runtime/helpers/esm/possibleConstructorReturn";
export default function _createSuper(Derived) {
var hasNativeReflectConstruct = isNativeReflectConstruct();
return function _createSuperInternal() {
var Super = getPrototypeOf(Derived),
result;
if (hasNativeReflectConstruct) {
var NewTarget = getPrototypeOf(this).constructor;
result = Reflect.construct(Super, arguments, NewTarget);
} else {
result = Super.apply(this, arguments);
}
return possibleConstructorReturn(this, result);
};
}"><pre class="notranslate">```js
<span class="pl-md"><span class="pl-md">-</span> import getPrototypeOf from "@babel/runtime/helpers/getPrototypeOf";</span>
<span class="pl-md"><span class="pl-md">-</span> import isNativeReflectConstruct from "@babel/runtime/helpers/isNativeReflectConstruct";</span>
<span class="pl-md"><span class="pl-md">-</span> import possibleConstructorReturn from "@babel/runtime/helpers/possibleConstructorReturn";</span>
<span class="pl-mi1"><span class="pl-mi1">+</span> import getPrototypeOf from "@babel/runtime/helpers/esm/getPrototypeOf";</span>
<span class="pl-mi1"><span class="pl-mi1">+</span> import isNativeReflectConstruct from "@babel/runtime/helpers/esm/isNativeReflectConstruct";</span>
<span class="pl-mi1"><span class="pl-mi1">+</span> import possibleConstructorReturn from "@babel/runtime/helpers/esm/possibleConstructorReturn";</span>
export default function _createSuper(Derived) {
var hasNativeReflectConstruct = isNativeReflectConstruct();
return function _createSuperInternal() {
var Super = getPrototypeOf(Derived),
result;
if (hasNativeReflectConstruct) {
var NewTarget = getPrototypeOf(this).constructor;
result = Reflect.construct(Super, arguments, NewTarget);
} else {
result = Super.apply(this, arguments);
}
return possibleConstructorReturn(this, result);
};
}</pre></div>
<p dir="auto">like: <a href="https://unpkg.com/browse/@babel/[email protected]/helpers/esm/createSuper.js" rel="nofollow">https://unpkg.com/browse/@babel/[email protected]/helpers/esm/createSuper.js</a></p>
<p dir="auto"><strong>Environment</strong></p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content=" System:
OS: macOS 10.15.7
Binaries:
Node: 15.3.0 - ~/.nvm/versions/node/v15.3.0/bin/node
Yarn: 1.22.10 - ~/.nvm/versions/node/v15.3.0/bin/yarn
npm: 7.0.14 - ~/.nvm/versions/node/v15.3.0/bin/npm
npmPackages:
@babel/runtime: ^7.12.5 => 7.13.2
babel-plugin-add-react-displayname: ^0.0.5 => 0.0.5
eslint: ^7.9.0 => 7.20.0
eslint-plugin-babel: ^5.3.0 => 5.3.1
jest: ^26.0.0 => 26.6.3 "><pre class="notranslate"><code class="notranslate"> System:
OS: macOS 10.15.7
Binaries:
Node: 15.3.0 - ~/.nvm/versions/node/v15.3.0/bin/node
Yarn: 1.22.10 - ~/.nvm/versions/node/v15.3.0/bin/yarn
npm: 7.0.14 - ~/.nvm/versions/node/v15.3.0/bin/npm
npmPackages:
@babel/runtime: ^7.12.5 => 7.13.2
babel-plugin-add-react-displayname: ^0.0.5 => 0.0.5
eslint: ^7.9.0 => 7.20.0
eslint-plugin-babel: ^5.3.0 => 5.3.1
jest: ^26.0.0 => 26.6.3
</code></pre></div>
<ul dir="auto">
<li>Babel version(s): 7.13.3</li>
<li>Node/npm version: 15.3.0</li>
<li>OS: [e.g. macOS 10.15.4, Windows 10]</li>
<li>Monorepo: [e.g. yes/no/Lerna]</li>
<li>How you are using Babel: [e.g. <code class="notranslate">webpack</code>, <code class="notranslate">rollup</code>, <code class="notranslate">parcel</code>, <code class="notranslate">babel-register</code>]</li>
</ul>
<p dir="auto"><strong>Possible Solution</strong></p>
<p dir="auto"><strong>Additional context</strong></p> | 0 |
<h2 dir="auto">Description</h2>
<p dir="auto">I find that it would be practical to be able to set up <strong>application group</strong> with a windows position that can be configured in the same way as the <strong>application groups</strong> set up on the <em>Surface Duo</em> :</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://camo.githubusercontent.com/29bb4816ae8761f93789c032e18f2f1c3e10181ad70a0fc8e07121e3d53fe819/68747470733a2f2f6d73706f776572757365722e636f6d2f77702d636f6e74656e742f75706c6f6164732f323032302f30382f656467652d616e642d6f6e656e6e6f74652d6170702d67726f75702e6a7067"><img src="https://camo.githubusercontent.com/29bb4816ae8761f93789c032e18f2f1c3e10181ad70a0fc8e07121e3d53fe819/68747470733a2f2f6d73706f776572757365722e636f6d2f77702d636f6e74656e742f75706c6f6164732f323032302f30382f656467652d616e642d6f6e656e6e6f74652d6170702d67726f75702e6a7067" alt="App group" data-canonical-src="https://mspoweruser.com/wp-content/uploads/2020/08/edge-and-onennote-app-group.jpg" style="max-width: 100%;"></a></p>
<p dir="auto">All this can be configured with a positioning based on <strong>FancyZones</strong> as well as an <strong>application group</strong> icon in the taskbar :</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/40165444/94113682-8d00e280-fe47-11ea-8257-a7fe65aa7079.png"><img src="https://user-images.githubusercontent.com/40165444/94113682-8d00e280-fe47-11ea-8257-a7fe65aa7079.png" alt="Taskbar app group" style="max-width: 100%;"></a></p>
<p dir="auto">Fancyzone amelioration of <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="677687348" data-permission-text="Title is private" data-url="https://github.com/microsoft/PowerToys/issues/5901" data-hovercard-type="issue" data-hovercard-url="/microsoft/PowerToys/issues/5901/hovercard" href="https://github.com/microsoft/PowerToys/issues/5901">#5901</a></p>
<hr>
<p dir="auto">If you'd like to see this feature implemented, add a 👍 reaction to this post.</p> | <h1 dir="auto">Environment</h1>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="Windows build number: Microsoft Windows [Version 10.0.18363.752]
PowerToys version: v0.16.0
PowerToy module for which you are reporting the bug (if applicable): FancyZones"><pre class="notranslate"><code class="notranslate">Windows build number: Microsoft Windows [Version 10.0.18363.752]
PowerToys version: v0.16.0
PowerToy module for which you are reporting the bug (if applicable): FancyZones
</code></pre></div>
<h1 dir="auto">Steps to reproduce</h1>
<ol dir="auto">
<li>Press Win + '</li>
<li>Attempt to scale zones to fit full screen</li>
</ol>
<h1 dir="auto">Expected behavior</h1>
<p dir="auto">Zones to fit complete screen</p>
<h1 dir="auto">Actual behavior</h1>
<p dir="auto">There is a gap along the right side of the screen and along the bottom of the screen</p>
<h1 dir="auto">Screenshots</h1>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/48316637/78113936-66ae3d00-73f8-11ea-8f2e-4ccde9e1090c.png"><img src="https://user-images.githubusercontent.com/48316637/78113936-66ae3d00-73f8-11ea-8f2e-4ccde9e1090c.png" alt="image" style="max-width: 100%;"></a></p> | 0 |
<p dir="auto"><strong>Apache Airflow version</strong>: 2.0.0</p>
<p dir="auto"><strong>Kubernetes version (if you are using kubernetes)</strong> (use <code class="notranslate">kubectl version</code>): 1.19.4</p>
<p dir="auto"><strong>Environment</strong>:</p>
<ul dir="auto">
<li><strong>Cloud provider or hardware configuration</strong>: Google Cloud Platform/GKE</li>
</ul>
<p dir="auto"><strong>What happened</strong>:</p>
<p dir="auto">I successfully cleared the state of a failed task using the graph view UI, but when I attempted to re-run the cleared task instance in graph view manually by selecting the task instance and clicking "Run", I received the following error:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="Something bad has happened.
Please consider letting us know by creating a bug report using GitHub.
Python version: 3.8.7
Airflow version: 2.0.0
Node: mr_node
-------------------------------------------------------------------------------
Traceback (most recent call last):
File "/home/airflow/.local/lib/python3.8/site-packages/flask/app.py", line 2447, in wsgi_app
response = self.full_dispatch_request()
File "/home/airflow/.local/lib/python3.8/site-packages/flask/app.py", line 1952, in full_dispatch_request
rv = self.handle_user_exception(e)
File "/home/airflow/.local/lib/python3.8/site-packages/flask/app.py", line 1821, in handle_user_exception
reraise(exc_type, exc_value, tb)
File "/home/airflow/.local/lib/python3.8/site-packages/flask/_compat.py", line 39, in reraise
raise value
File "/home/airflow/.local/lib/python3.8/site-packages/flask/app.py", line 1950, in full_dispatch_request
rv = self.dispatch_request()
File "/home/airflow/.local/lib/python3.8/site-packages/flask/app.py", line 1936, in dispatch_request
return self.view_functions[rule.endpoint](**req.view_args)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/www/auth.py", line 34, in decorated
return func(*args, **kwargs)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/www/decorators.py", line 60, in wrapper
return f(*args, **kwargs)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/www/views.py", line 1366, in run
executor.start()
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/executors/kubernetes_executor.py", line 493, in start
raise AirflowException("Could not get scheduler_job_id")
airflow.exceptions.AirflowException: Could not get scheduler_job_id"><pre class="notranslate"><code class="notranslate">Something bad has happened.
Please consider letting us know by creating a bug report using GitHub.
Python version: 3.8.7
Airflow version: 2.0.0
Node: mr_node
-------------------------------------------------------------------------------
Traceback (most recent call last):
File "/home/airflow/.local/lib/python3.8/site-packages/flask/app.py", line 2447, in wsgi_app
response = self.full_dispatch_request()
File "/home/airflow/.local/lib/python3.8/site-packages/flask/app.py", line 1952, in full_dispatch_request
rv = self.handle_user_exception(e)
File "/home/airflow/.local/lib/python3.8/site-packages/flask/app.py", line 1821, in handle_user_exception
reraise(exc_type, exc_value, tb)
File "/home/airflow/.local/lib/python3.8/site-packages/flask/_compat.py", line 39, in reraise
raise value
File "/home/airflow/.local/lib/python3.8/site-packages/flask/app.py", line 1950, in full_dispatch_request
rv = self.dispatch_request()
File "/home/airflow/.local/lib/python3.8/site-packages/flask/app.py", line 1936, in dispatch_request
return self.view_functions[rule.endpoint](**req.view_args)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/www/auth.py", line 34, in decorated
return func(*args, **kwargs)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/www/decorators.py", line 60, in wrapper
return f(*args, **kwargs)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/www/views.py", line 1366, in run
executor.start()
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/executors/kubernetes_executor.py", line 493, in start
raise AirflowException("Could not get scheduler_job_id")
airflow.exceptions.AirflowException: Could not get scheduler_job_id
</code></pre></div>
<p dir="auto"><strong>What you expected to happen</strong>:</p>
<p dir="auto">I expected the task instance to be scheduled and begin running again.</p>
<p dir="auto"><strong>How to reproduce it</strong>:</p>
<p dir="auto">Configure Airflow 2.0.0 to run on GCP, clear the state of a finished task instance using the UI (I was able to reproduce the error on a task instance maked "Success" as well), and again use the Web UI to "Run" the task.</p>
<p dir="auto"><strong>Anything else we need to know</strong>:</p>
<p dir="auto">One important item to note is that when I <em>only</em> clear task instance and do not attempt to run it manually using the UI, the task does queue and is placing in a <code class="notranslate">running</code> state, but quickly fails with the following error:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="[2021-01-08 21:08:16,140] {taskinstance.py:1396} ERROR - (0)
Reason: Handshake status 500 Internal Server Error
Traceback (most recent call last):
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/stream/ws_client.py", line 296, in websocket_call
client = WSClient(configuration, get_websocket_url(url), headers, capture_all)
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/stream/ws_client.py", line 94, in __init__
self.sock.connect(url, header=header)
File "/home/airflow/.local/lib/python3.8/site-packages/websocket/_core.py", line 226, in connect
self.handshake_response = handshake(self.sock, *addrs, **options)
File "/home/airflow/.local/lib/python3.8/site-packages/websocket/_handshake.py", line 80, in handshake
status, resp = _get_resp_headers(sock)
File "/home/airflow/.local/lib/python3.8/site-packages/websocket/_handshake.py", line 165, in _get_resp_headers
raise WebSocketBadStatusException("Handshake status %d %s", status, status_message, resp_headers)
websocket._exceptions.WebSocketBadStatusException: Handshake status 500 Internal Server Error
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/models/taskinstance.py", line 1086, in _run_raw_task
self._prepare_and_execute_task_with_callbacks(context, task)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/models/taskinstance.py", line 1260, in _prepare_and_execute_task_with_callbacks
result = self._execute_task(context, task_copy)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/models/taskinstance.py", line 1300, in _execute_task
result = task_copy.execute(context=context)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/providers/cncf/kubernetes/operators/kubernetes_pod.py", line 335, in execute
final_state, result = self.handle_pod_overlap(
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/providers/cncf/kubernetes/operators/kubernetes_pod.py", line 375, in handle_pod_overlap
final_state, result = self.monitor_launched_pod(launcher, pod)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/providers/cncf/kubernetes/operators/kubernetes_pod.py", line 513, in monitor_launched_pod
(final_state, result) = launcher.monitor_pod(pod, get_logs=self.get_logs)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/kubernetes/pod_launcher.py", line 151, in monitor_pod
result = self._extract_xcom(pod)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/kubernetes/pod_launcher.py", line 246, in _extract_xcom
resp = kubernetes_stream(
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/stream/stream.py", line 35, in stream
return func(*args, **kwargs)
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/client/api/core_v1_api.py", line 841, in connect_get_namespaced_pod_exec
(data) = self.connect_get_namespaced_pod_exec_with_http_info(name, namespace, **kwargs) # noqa: E501
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/client/api/core_v1_api.py", line 927, in connect_get_namespaced_pod_exec_with_http_info
return self.api_client.call_api(
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/client/api_client.py", line 340, in call_api
return self.__call_api(resource_path, method,
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/client/api_client.py", line 172, in __call_api
response_data = self.request(
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/stream/stream.py", line 30, in _intercept_request_call
return ws_client.websocket_call(config, *args, **kwargs)
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/stream/ws_client.py", line 302, in websocket_call
raise ApiException(status=0, reason=str(e))
kubernetes.client.rest.ApiException: (0)
Reason: Handshake status 500 Internal Server Error"><pre class="notranslate"><code class="notranslate">[2021-01-08 21:08:16,140] {taskinstance.py:1396} ERROR - (0)
Reason: Handshake status 500 Internal Server Error
Traceback (most recent call last):
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/stream/ws_client.py", line 296, in websocket_call
client = WSClient(configuration, get_websocket_url(url), headers, capture_all)
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/stream/ws_client.py", line 94, in __init__
self.sock.connect(url, header=header)
File "/home/airflow/.local/lib/python3.8/site-packages/websocket/_core.py", line 226, in connect
self.handshake_response = handshake(self.sock, *addrs, **options)
File "/home/airflow/.local/lib/python3.8/site-packages/websocket/_handshake.py", line 80, in handshake
status, resp = _get_resp_headers(sock)
File "/home/airflow/.local/lib/python3.8/site-packages/websocket/_handshake.py", line 165, in _get_resp_headers
raise WebSocketBadStatusException("Handshake status %d %s", status, status_message, resp_headers)
websocket._exceptions.WebSocketBadStatusException: Handshake status 500 Internal Server Error
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/models/taskinstance.py", line 1086, in _run_raw_task
self._prepare_and_execute_task_with_callbacks(context, task)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/models/taskinstance.py", line 1260, in _prepare_and_execute_task_with_callbacks
result = self._execute_task(context, task_copy)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/models/taskinstance.py", line 1300, in _execute_task
result = task_copy.execute(context=context)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/providers/cncf/kubernetes/operators/kubernetes_pod.py", line 335, in execute
final_state, result = self.handle_pod_overlap(
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/providers/cncf/kubernetes/operators/kubernetes_pod.py", line 375, in handle_pod_overlap
final_state, result = self.monitor_launched_pod(launcher, pod)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/providers/cncf/kubernetes/operators/kubernetes_pod.py", line 513, in monitor_launched_pod
(final_state, result) = launcher.monitor_pod(pod, get_logs=self.get_logs)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/kubernetes/pod_launcher.py", line 151, in monitor_pod
result = self._extract_xcom(pod)
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/kubernetes/pod_launcher.py", line 246, in _extract_xcom
resp = kubernetes_stream(
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/stream/stream.py", line 35, in stream
return func(*args, **kwargs)
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/client/api/core_v1_api.py", line 841, in connect_get_namespaced_pod_exec
(data) = self.connect_get_namespaced_pod_exec_with_http_info(name, namespace, **kwargs) # noqa: E501
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/client/api/core_v1_api.py", line 927, in connect_get_namespaced_pod_exec_with_http_info
return self.api_client.call_api(
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/client/api_client.py", line 340, in call_api
return self.__call_api(resource_path, method,
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/client/api_client.py", line 172, in __call_api
response_data = self.request(
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/stream/stream.py", line 30, in _intercept_request_call
return ws_client.websocket_call(config, *args, **kwargs)
File "/home/airflow/.local/lib/python3.8/site-packages/kubernetes/stream/ws_client.py", line 302, in websocket_call
raise ApiException(status=0, reason=str(e))
kubernetes.client.rest.ApiException: (0)
Reason: Handshake status 500 Internal Server Error
</code></pre></div>
<p dir="auto">Thanks!</p> | <p dir="auto"><strong>Apache Airflow version:</strong></p>
<p dir="auto">2.0.0</p>
<p dir="auto"><strong>Kubernetes version (if you are using kubernetes) (use kubectl version):</strong></p>
<p dir="auto">1.18.3</p>
<p dir="auto"><strong>Environment:</strong></p>
<p dir="auto">Cloud provider or hardware configuration: AWS</p>
<p dir="auto"><strong>What happened:</strong></p>
<p dir="auto">When trying to run a DAG, it gets scheduled, but task is never run. When attempting to run task manually, it shows an error:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="Something bad has happened.
Please consider letting us know by creating a bug report using GitHub.
Python version: 3.8.7
Airflow version: 2.0.0
Node: airflow-web-ffdd89d6-h98vj
-------------------------------------------------------------------------------
Traceback (most recent call last):
File "/usr/local/lib/python3.8/site-packages/flask/app.py", line 2447, in wsgi_app
response = self.full_dispatch_request()
File "/usr/local/lib/python3.8/site-packages/flask/app.py", line 1952, in full_dispatch_request
rv = self.handle_user_exception(e)
File "/usr/local/lib/python3.8/site-packages/flask/app.py", line 1821, in handle_user_exception
reraise(exc_type, exc_value, tb)
File "/usr/local/lib/python3.8/site-packages/flask/_compat.py", line 39, in reraise
raise value
File "/usr/local/lib/python3.8/site-packages/flask/app.py", line 1950, in full_dispatch_request
rv = self.dispatch_request()
File "/usr/local/lib/python3.8/site-packages/flask/app.py", line 1936, in dispatch_request
return self.view_functions[rule.endpoint](**req.view_args)
File "/usr/local/lib/python3.8/site-packages/airflow/www/auth.py", line 34, in decorated
return func(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/airflow/www/decorators.py", line 60, in wrapper
return f(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/airflow/www/views.py", line 1366, in run
executor.start()
File "/usr/local/lib/python3.8/site-packages/airflow/executors/kubernetes_executor.py", line 493, in start
raise AirflowException("Could not get scheduler_job_id")
airflow.exceptions.AirflowException: Could not get scheduler_job_id"><pre class="notranslate"><code class="notranslate">Something bad has happened.
Please consider letting us know by creating a bug report using GitHub.
Python version: 3.8.7
Airflow version: 2.0.0
Node: airflow-web-ffdd89d6-h98vj
-------------------------------------------------------------------------------
Traceback (most recent call last):
File "/usr/local/lib/python3.8/site-packages/flask/app.py", line 2447, in wsgi_app
response = self.full_dispatch_request()
File "/usr/local/lib/python3.8/site-packages/flask/app.py", line 1952, in full_dispatch_request
rv = self.handle_user_exception(e)
File "/usr/local/lib/python3.8/site-packages/flask/app.py", line 1821, in handle_user_exception
reraise(exc_type, exc_value, tb)
File "/usr/local/lib/python3.8/site-packages/flask/_compat.py", line 39, in reraise
raise value
File "/usr/local/lib/python3.8/site-packages/flask/app.py", line 1950, in full_dispatch_request
rv = self.dispatch_request()
File "/usr/local/lib/python3.8/site-packages/flask/app.py", line 1936, in dispatch_request
return self.view_functions[rule.endpoint](**req.view_args)
File "/usr/local/lib/python3.8/site-packages/airflow/www/auth.py", line 34, in decorated
return func(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/airflow/www/decorators.py", line 60, in wrapper
return f(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/airflow/www/views.py", line 1366, in run
executor.start()
File "/usr/local/lib/python3.8/site-packages/airflow/executors/kubernetes_executor.py", line 493, in start
raise AirflowException("Could not get scheduler_job_id")
airflow.exceptions.AirflowException: Could not get scheduler_job_id
</code></pre></div>
<p dir="auto"><strong>What you expected to happen:</strong></p>
<p dir="auto">The task to be run successfully without</p>
<p dir="auto"><strong>How to reproduce it:</strong></p>
<p dir="auto">Haven't pinpointed what causes the issue, besides an attempted upgrade from Airflow 1.10.14 to Airflow 2.0.0</p>
<p dir="auto"><strong>Anything else we need to know:</strong></p>
<p dir="auto">This error is encountered in an upgrade of Airflow from 1.10.14 to Airflow 2.0.0</p>
<p dir="auto">EDIT: Formatted to fit the issue template</p> | 1 |
<p dir="auto">Following is what I recently made using ThreeJS by importing OBJ files:<br>
<a href="http://devplace.in/~harman/geometry_viewer/mug.html" rel="nofollow">http://devplace.in/~harman/geometry_viewer/mug.html</a><br>
<a href="http://devplace.in/~harman/geometry_viewer/cylinder.html" rel="nofollow">http://devplace.in/~harman/geometry_viewer/cylinder.html</a></p>
<p dir="auto">Users report following problem:</p>
<ul dir="auto">
<li>The model appears (briefly flashing) while user rotates. It appeared just fine before I applied material. User just get a blue background and the performance monitor in the bottom left corner. If he rotate the invisible model really quickly, it'll display the model briefly for a frame or two, but then goes back to blue.</li>
</ul>
<p dir="auto">Similar issue has been noticed at threeJS Editor: <a href="http://threejs.org/editor/" rel="nofollow">http://threejs.org/editor/</a> Model disappears on changing renderer. However, I'm not sure whether both of these problems are same.</p>
<ul dir="auto">
<li>Some other set of users encountered edges / lines on the model surface even when wireframe flag was not set. See:</li>
</ul>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://camo.githubusercontent.com/bba60fe38d257def0c4444181a547f0c12e1bc46f7679077b80615cf7569c135/68747470733a2f2f662e636c6f75642e6769746875622e636f6d2f6173736574732f313039343435372f3837323732312f37656633326464342d663836382d313165322d383736372d3536306466303561663837302e706e67"><img src="https://camo.githubusercontent.com/bba60fe38d257def0c4444181a547f0c12e1bc46f7679077b80615cf7569c135/68747470733a2f2f662e636c6f75642e6769746875622e636f6d2f6173736574732f313039343435372f3837323732312f37656633326464342d663836382d313165322d383736372d3536306466303561663837302e706e67" alt="screenshot from 2013-07-27 00_31_39" data-canonical-src="https://f.cloud.github.com/assets/1094457/872721/7ef32dd4-f868-11e2-8767-560df05af870.png" style="max-width: 100%;"></a></p>
<p dir="auto">and</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://camo.githubusercontent.com/ede6d730bd0592f482f9f56adea9f0721d07d984a834028a7247130b1b48a25c/68747470733a2f2f662e636c6f75642e6769746875622e636f6d2f6173736574732f313039343435372f3837323636362f38323964306633632d663836372d313165322d393431312d3662383437623066613264302e6a7067"><img src="https://camo.githubusercontent.com/ede6d730bd0592f482f9f56adea9f0721d07d984a834028a7247130b1b48a25c/68747470733a2f2f662e636c6f75642e6769746875622e636f6d2f6173736574732f313039343435372f3837323636362f38323964306633632d663836372d313165322d393431312d3662383437623066613264302e6a7067" alt="jrqs84" data-canonical-src="https://f.cloud.github.com/assets/1094457/872666/829d0f3c-f867-11e2-9411-6b847b0fa2d0.jpg" style="max-width: 100%;"></a></p>
<p dir="auto">It looks like a renderer problem, but still not sure.</p>
<ul dir="auto">
<li>From above image of mug, one can see the inner yellow bottom from the outside of the mug which is unexpected.</li>
</ul>
<p dir="auto">So is it a bug(s) in ThreeJS or it should be something fixable by me?</p> | <p dir="auto">Following is what I recently made using ThreeJS by importing OBJ files:<br>
<a href="http://devplace.in/~harman/geometry_viewer/mug.html" rel="nofollow">http://devplace.in/~harman/geometry_viewer/mug.html</a><br>
<a href="http://devplace.in/~harman/geometry_viewer/cylinder.html" rel="nofollow">http://devplace.in/~harman/geometry_viewer/cylinder.html</a></p>
<p dir="auto">Users report following problems:</p>
<ul dir="auto">
<li>User just get a blue background and the performance monitor in the bottom left corner. If he rotate the invisible model really quickly, it'll display the model briefly for a frame or two, but then goes back to blue. It appeared just fine before I assigned material to models.</li>
</ul>
<p dir="auto">Similar problem has been noticed in ThreeJS Editor: <a href="http://threejs.org/editor/" rel="nofollow">http://threejs.org/editor/</a>. When one changes the renderer from the drop down, model disappears. However, I'm not sure whether both these problems are same.</p>
<ul dir="auto">
<li>Other user encountered edges / lines on model surface even when the wireframe flag is not true. See:</li>
</ul>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://camo.githubusercontent.com/eca42c721182a705f904b09fa5d4e7aef63b6e1262ad3350c8896702499fbb57/68747470733a2f2f662e636c6f75642e6769746875622e636f6d2f6173736574732f313039343435372f3837333037372f32336566333265302d663837302d313165322d393935392d3933373463663335616638302e706e67"><img src="https://camo.githubusercontent.com/eca42c721182a705f904b09fa5d4e7aef63b6e1262ad3350c8896702499fbb57/68747470733a2f2f662e636c6f75642e6769746875622e636f6d2f6173736574732f313039343435372f3837333037372f32336566333265302d663837302d313165322d393935392d3933373463663335616638302e706e67" alt="screenshot from 2013-07-27 00_31_39" data-canonical-src="https://f.cloud.github.com/assets/1094457/873077/23ef32e0-f870-11e2-9959-9374cf35af80.png" style="max-width: 100%;"></a></p>
<p dir="auto">and</p>
<p dir="auto"><a target="_blank" rel="noopener noreferrer nofollow" href="https://camo.githubusercontent.com/827b497e2c9c80410b73fefa02e734f343e2567a0ab22afb55fb38a62be2e5cc/68747470733a2f2f662e636c6f75642e6769746875622e636f6d2f6173736574732f313039343435372f3837333037382f32613334323664382d663837302d313165322d393966362d3939343737323137393639632e6a7067"><img src="https://camo.githubusercontent.com/827b497e2c9c80410b73fefa02e734f343e2567a0ab22afb55fb38a62be2e5cc/68747470733a2f2f662e636c6f75642e6769746875622e636f6d2f6173736574732f313039343435372f3837333037382f32613334323664382d663837302d313165322d393966362d3939343737323137393639632e6a7067" alt="jrqs84" data-canonical-src="https://f.cloud.github.com/assets/1094457/873078/2a3426d8-f870-11e2-99f6-99477217969c.jpg" style="max-width: 100%;"></a></p>
<p dir="auto">This may be the renderer problem, but still not sure.</p>
<ul dir="auto">
<li>Also in above image of mug, one can see the inner yellow bottom from outside of the mug which is unexpected.</li>
</ul>
<p dir="auto">So, is it a bug(s) in ThreeJS or it should be something fixable by me?</p> | 1 |
<p dir="auto">E2E tests <code class="notranslate">Deployment deployment should label adopted RSs and pods</code> fails with:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/deployment.go:71 Expected error: <*errors.errorString | 0xc2084c0e60>: { s: "failed to wait for pods running: [gave up waiting for pod 'nginx-controller-cceta' to be 'running' after 5m0s gave up waiting for pod 'nginx-controller-h8inh' to be 'running' after 5m0s]", } failed to wait for pods running: [gave up waiting for pod 'nginx-controller-cceta' to be 'running' after 5m0s gave up waiting for pod 'nginx-controller-h8inh' to be 'running' after 5m0s] not to have occurred"><pre class="notranslate"><code class="notranslate">/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/deployment.go:71 Expected error: <*errors.errorString | 0xc2084c0e60>: { s: "failed to wait for pods running: [gave up waiting for pod 'nginx-controller-cceta' to be 'running' after 5m0s gave up waiting for pod 'nginx-controller-h8inh' to be 'running' after 5m0s]", } failed to wait for pods running: [gave up waiting for pod 'nginx-controller-cceta' to be 'running' after 5m0s gave up waiting for pod 'nginx-controller-h8inh' to be 'running' after 5m0s] not to have occurred
</code></pre></div>
<p dir="auto">Log from <a href="http://kubekins.dls.corp.google.com/view/Critical%20Builds/job/kubernetes-e2e-gke/1976/" rel="nofollow">http://kubekins.dls.corp.google.com/view/Critical%20Builds/job/kubernetes-e2e-gke/1976/</a>:</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="14:55:55 ------------------------------
14:55:55 [BeforeEach] Deployment
14:55:55 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework.go:83
14:55:55 STEP: Creating a kubernetes client
14:55:55 Feb 29 14:45:40.139: INFO: >>> testContext.KubeConfig: /var/lib/jenkins/jobs/kubernetes-e2e-gke/workspace/.kube/config
14:55:55
14:55:55 STEP: Building a namespace api object
14:55:55 Feb 29 14:45:40.176: INFO: Waiting up to 2m0s for service account default to be provisioned in ns e2e-tests-deployment-ezytf
14:55:55 Feb 29 14:45:40.198: INFO: Service account default in ns e2e-tests-deployment-ezytf had 0 secrets, ignoring for 2s: <nil>
14:55:55 Feb 29 14:45:42.201: INFO: Service account default in ns e2e-tests-deployment-ezytf with secrets found. (2.025365056s)
14:55:55 STEP: Waiting for a default service account to be provisioned in namespace
14:55:55 Feb 29 14:45:42.201: INFO: Waiting up to 2m0s for service account default to be provisioned in ns e2e-tests-deployment-ezytf
14:55:55 Feb 29 14:45:42.206: INFO: Service account default in ns e2e-tests-deployment-ezytf with secrets found. (4.670131ms)
14:55:55 [It] deployment should label adopted RSs and pods
14:55:55 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/deployment.go:71
14:55:55 Feb 29 14:45:42.226: INFO: Pod name nginx: Found 0 pods out of 3
14:55:55 Feb 29 14:45:47.230: INFO: Pod name nginx: Found 3 pods out of 3
14:55:55 STEP: ensuring each pod is running
14:55:55 Feb 29 14:45:47.230: INFO: Waiting up to 5m0s for pod nginx-controller-cceta status to be running
14:55:55 Feb 29 14:45:47.233: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2.621607ms elapsed)
14:55:55 Feb 29 14:45:49.236: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2.006058051s elapsed)
14:55:55 Feb 29 14:45:51.240: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4.009845765s elapsed)
14:55:55 Feb 29 14:45:53.252: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (6.022251049s elapsed)
14:55:55 Feb 29 14:45:55.258: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (8.027621505s elapsed)
14:55:55 Feb 29 14:45:57.261: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (10.031267285s elapsed)
14:55:55 Feb 29 14:45:59.268: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (12.037638231s elapsed)
14:55:55 Feb 29 14:46:01.280: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (14.049614949s elapsed)
14:55:55 Feb 29 14:46:03.283: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (16.053400403s elapsed)
14:55:55 Feb 29 14:46:05.287: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (18.057109564s elapsed)
14:55:55 Feb 29 14:46:07.291: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (20.060796787s elapsed)
14:55:55 Feb 29 14:46:09.294: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (22.06440015s elapsed)
14:55:55 Feb 29 14:46:11.298: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (24.06814485s elapsed)
14:55:55 Feb 29 14:46:13.305: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (26.074887193s elapsed)
14:55:55 Feb 29 14:46:15.310: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (28.080003644s elapsed)
14:55:55 Feb 29 14:46:17.317: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (30.087297487s elapsed)
14:55:55 Feb 29 14:46:19.325: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (32.09515456s elapsed)
14:55:55 Feb 29 14:46:21.334: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (34.104373139s elapsed)
14:55:55 Feb 29 14:46:23.338: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (36.108334597s elapsed)
14:55:55 Feb 29 14:46:25.346: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (38.116396703s elapsed)
14:55:55 Feb 29 14:46:27.351: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (40.121301689s elapsed)
14:55:55 Feb 29 14:46:29.355: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (42.124813176s elapsed)
14:55:55 Feb 29 14:46:31.359: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (44.12878585s elapsed)
14:55:55 Feb 29 14:46:33.365: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (46.134847862s elapsed)
14:55:55 Feb 29 14:46:35.368: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (48.138450891s elapsed)
14:55:55 Feb 29 14:46:37.372: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (50.142281963s elapsed)
14:55:55 Feb 29 14:46:39.377: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (52.146617091s elapsed)
14:55:55 Feb 29 14:46:41.381: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (54.15122489s elapsed)
14:55:55 Feb 29 14:46:43.388: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (56.158110274s elapsed)
14:55:55 Feb 29 14:46:45.394: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (58.164077418s elapsed)
14:55:55 Feb 29 14:46:47.398: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m0.167681085s elapsed)
14:55:55 Feb 29 14:46:49.401: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m2.171265049s elapsed)
14:55:55 Feb 29 14:46:51.406: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m4.175785784s elapsed)
14:55:55 Feb 29 14:46:53.409: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m6.179102726s elapsed)
14:55:55 Feb 29 14:46:55.422: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m8.191951719s elapsed)
14:55:55 Feb 29 14:46:57.430: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m10.20031798s elapsed)
14:55:55 Feb 29 14:46:59.434: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m12.204123329s elapsed)
14:55:55 Feb 29 14:47:01.439: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m14.208718392s elapsed)
14:55:55 Feb 29 14:47:03.442: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m16.212384665s elapsed)
14:55:55 Feb 29 14:47:05.446: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m18.216284818s elapsed)
14:55:55 Feb 29 14:47:07.450: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m20.220123499s elapsed)
14:55:55 Feb 29 14:47:09.454: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m22.224276995s elapsed)
14:55:55 Feb 29 14:47:11.458: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m24.228176691s elapsed)
14:55:55 Feb 29 14:47:13.462: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m26.23182926s elapsed)
14:55:55 Feb 29 14:47:15.466: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m28.235893296s elapsed)
14:55:55 Feb 29 14:47:17.470: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m30.239743747s elapsed)
14:55:55 Feb 29 14:47:19.482: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m32.252012755s elapsed)
14:55:55 Feb 29 14:47:21.490: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m34.260158566s elapsed)
14:55:55 Feb 29 14:47:23.494: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m36.263887374s elapsed)
14:55:55 Feb 29 14:47:25.498: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m38.267798471s elapsed)
14:55:55 Feb 29 14:47:27.502: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m40.271853684s elapsed)
14:55:55 Feb 29 14:47:29.513: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m42.283292405s elapsed)
14:55:55 Feb 29 14:47:31.517: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m44.287226687s elapsed)
14:55:55 Feb 29 14:47:33.528: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m46.298316069s elapsed)
14:55:55 Feb 29 14:47:35.533: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m48.303456523s elapsed)
14:55:55 Feb 29 14:47:37.538: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m50.308487927s elapsed)
14:55:55 Feb 29 14:47:39.543: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m52.312857219s elapsed)
14:55:55 Feb 29 14:47:41.550: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m54.32038014s elapsed)
14:55:55 Feb 29 14:47:43.558: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m56.327828258s elapsed)
14:55:55 Feb 29 14:47:45.562: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m58.332220565s elapsed)
14:55:55 Feb 29 14:47:47.569: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m0.339085895s elapsed)
14:55:55 Feb 29 14:47:49.573: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m2.34325204s elapsed)
14:55:55 Feb 29 14:47:51.588: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m4.35849625s elapsed)
14:55:55 Feb 29 14:47:53.593: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m6.363014495s elapsed)
14:55:55 Feb 29 14:47:55.597: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m8.366630623s elapsed)
14:55:55 Feb 29 14:47:57.600: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m10.370008163s elapsed)
14:55:55 Feb 29 14:47:59.605: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m12.374953605s elapsed)
14:55:55 Feb 29 14:48:01.609: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m14.378645389s elapsed)
14:55:55 Feb 29 14:48:03.613: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m16.382921602s elapsed)
14:55:55 Feb 29 14:48:05.625: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m18.395545053s elapsed)
14:55:55 Feb 29 14:48:07.628: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m20.398245389s elapsed)
14:55:55 Feb 29 14:48:09.631: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m22.401138921s elapsed)
14:55:55 Feb 29 14:48:11.634: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m24.404201567s elapsed)
14:55:55 Feb 29 14:48:13.637: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m26.407054385s elapsed)
14:55:55 Feb 29 14:48:15.641: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m28.411133696s elapsed)
14:55:55 Feb 29 14:48:17.661: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m30.430611995s elapsed)
14:55:55 Feb 29 14:48:19.663: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m32.433593617s elapsed)
14:55:55 Feb 29 14:48:21.673: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m34.443030852s elapsed)
14:55:55 Feb 29 14:48:23.677: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m36.446616458s elapsed)
14:55:55 Feb 29 14:48:25.685: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m38.455065752s elapsed)
14:55:55 Feb 29 14:48:27.693: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m40.462934495s elapsed)
14:55:55 Feb 29 14:48:29.696: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m42.466417547s elapsed)
14:55:55 Feb 29 14:48:31.700: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m44.47012654s elapsed)
14:55:55 Feb 29 14:48:33.704: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m46.473811221s elapsed)
14:55:55 Feb 29 14:48:35.707: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m48.477187473s elapsed)
14:55:55 Feb 29 14:48:37.710: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m50.480501037s elapsed)
14:55:55 Feb 29 14:48:39.714: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m52.484141754s elapsed)
14:55:55 Feb 29 14:48:41.718: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m54.487909755s elapsed)
14:55:55 Feb 29 14:48:43.721: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m56.491389362s elapsed)
14:55:55 Feb 29 14:48:45.724: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m58.494594496s elapsed)
14:55:55 Feb 29 14:48:47.730: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m0.499728488s elapsed)
14:55:55 Feb 29 14:48:49.736: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m2.50568688s elapsed)
14:55:55 Feb 29 14:48:51.745: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m4.515130367s elapsed)
14:55:55 Feb 29 14:48:53.750: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m6.520567985s elapsed)
14:55:55 Feb 29 14:48:55.754: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m8.524143912s elapsed)
14:55:55 Feb 29 14:48:57.758: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m10.527734184s elapsed)
14:55:55 Feb 29 14:48:59.761: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m12.531404579s elapsed)
14:55:55 Feb 29 14:49:01.765: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m14.534863411s elapsed)
14:55:55 Feb 29 14:49:03.768: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m16.538322454s elapsed)
14:55:55 Feb 29 14:49:05.772: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m18.542429464s elapsed)
14:55:55 Feb 29 14:49:07.776: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m20.546443927s elapsed)
14:55:55 Feb 29 14:49:09.780: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m22.550147462s elapsed)
14:55:55 Feb 29 14:49:11.784: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m24.554155997s elapsed)
14:55:55 Feb 29 14:49:13.788: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m26.557779943s elapsed)
14:55:55 Feb 29 14:49:15.793: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m28.562762235s elapsed)
14:55:55 Feb 29 14:49:17.797: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m30.566609002s elapsed)
14:55:55 Feb 29 14:49:19.800: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m32.570050505s elapsed)
14:55:55 Feb 29 14:49:21.804: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m34.573663773s elapsed)
14:55:55 Feb 29 14:49:23.817: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m36.587287962s elapsed)
14:55:55 Feb 29 14:49:25.821: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m38.59092768s elapsed)
14:55:55 Feb 29 14:49:27.825: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m40.595388308s elapsed)
14:55:55 Feb 29 14:49:29.829: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m42.59903347s elapsed)
14:55:55 Feb 29 14:49:31.832: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m44.602018698s elapsed)
14:55:55 Feb 29 14:49:33.836: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m46.605876878s elapsed)
14:55:55 Feb 29 14:49:35.840: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m48.610046791s elapsed)
14:55:55 Feb 29 14:49:37.844: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m50.613781277s elapsed)
14:55:55 Feb 29 14:49:39.854: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m52.624094173s elapsed)
14:55:55 Feb 29 14:49:41.860: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m54.630550744s elapsed)
14:55:55 Feb 29 14:49:43.864: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m56.634172637s elapsed)
14:55:55 Feb 29 14:49:45.869: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m58.63948903s elapsed)
14:55:55 Feb 29 14:49:47.873: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m0.643135612s elapsed)
14:55:55 Feb 29 14:49:49.877: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m2.647470302s elapsed)
14:55:55 Feb 29 14:49:51.881: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m4.651137115s elapsed)
14:55:55 Feb 29 14:49:53.885: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m6.655102626s elapsed)
14:55:55 Feb 29 14:49:55.889: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m8.659097795s elapsed)
14:55:55 Feb 29 14:49:57.893: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m10.663495143s elapsed)
14:55:55 Feb 29 14:49:59.897: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m12.667322166s elapsed)
14:55:55 Feb 29 14:50:01.901: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m14.67117842s elapsed)
14:55:55 Feb 29 14:50:03.910: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m16.679818657s elapsed)
14:55:55 Feb 29 14:50:05.913: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m18.683340638s elapsed)
14:55:55 Feb 29 14:50:07.921: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m20.69096282s elapsed)
14:55:55 Feb 29 14:50:09.927: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m22.697135213s elapsed)
14:55:55 Feb 29 14:50:11.931: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m24.700827569s elapsed)
14:55:55 Feb 29 14:50:13.936: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m26.70621211s elapsed)
14:55:55 Feb 29 14:50:15.940: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m28.710144694s elapsed)
14:55:55 Feb 29 14:50:17.944: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m30.713880033s elapsed)
14:55:55 Feb 29 14:50:19.948: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m32.717650356s elapsed)
14:55:55 Feb 29 14:50:21.951: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m34.721059557s elapsed)
14:55:55 Feb 29 14:50:23.958: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m36.727608471s elapsed)
14:55:55 Feb 29 14:50:25.962: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m38.732520787s elapsed)
14:55:55 Feb 29 14:50:27.967: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m40.737490797s elapsed)
14:55:55 Feb 29 14:50:29.971: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m42.741214086s elapsed)
14:55:55 Feb 29 14:50:31.975: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m44.744627328s elapsed)
14:55:55 Feb 29 14:50:33.979: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m46.748988901s elapsed)
14:55:55 Feb 29 14:50:35.982: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m48.75244461s elapsed)
14:55:55 Feb 29 14:50:37.986: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m50.756274148s elapsed)
14:55:55 Feb 29 14:50:39.990: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m52.760202853s elapsed)
14:55:55 Feb 29 14:50:41.994: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m54.764189774s elapsed)
14:55:55 Feb 29 14:50:43.999: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m56.768787806s elapsed)
14:55:55 Feb 29 14:50:46.003: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m58.773103023s elapsed)
14:55:55 Feb 29 14:50:48.003: INFO: Waiting up to 5m0s for pod nginx-controller-h8inh status to be running
14:55:55 Feb 29 14:50:48.007: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4.211053ms elapsed)
14:55:55 Feb 29 14:50:50.012: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2.008983228s elapsed)
14:55:55 Feb 29 14:50:52.016: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4.013129998s elapsed)
14:55:55 Feb 29 14:50:54.021: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (6.017367481s elapsed)
14:55:55 Feb 29 14:50:56.024: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (8.02107039s elapsed)
14:55:55 Feb 29 14:50:58.029: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (10.02541349s elapsed)
14:55:55 Feb 29 14:51:00.033: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (12.029552985s elapsed)
14:55:55 Feb 29 14:51:02.039: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (14.036171006s elapsed)
14:55:55 Feb 29 14:51:04.043: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (16.040100224s elapsed)
14:55:55 Feb 29 14:51:06.047: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (18.044159343s elapsed)
14:55:55 Feb 29 14:51:08.052: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (20.048798763s elapsed)
14:55:55 Feb 29 14:51:10.058: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (22.054902853s elapsed)
14:55:55 Feb 29 14:51:12.062: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (24.058828936s elapsed)
14:55:55 Feb 29 14:51:14.066: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (26.062342652s elapsed)
14:55:55 Feb 29 14:51:16.070: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (28.066274694s elapsed)
14:55:55 Feb 29 14:51:18.073: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (30.070144825s elapsed)
14:55:55 Feb 29 14:51:20.086: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (32.082324045s elapsed)
14:55:55 Feb 29 14:51:22.090: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (34.086668007s elapsed)
14:55:55 Feb 29 14:51:24.094: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (36.090685907s elapsed)
14:55:55 Feb 29 14:51:26.098: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (38.094607125s elapsed)
14:55:55 Feb 29 14:51:28.102: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (40.09850594s elapsed)
14:55:55 Feb 29 14:51:30.105: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (42.102053212s elapsed)
14:55:55 Feb 29 14:51:32.109: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (44.105763653s elapsed)
14:55:55 Feb 29 14:51:34.113: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (46.109633313s elapsed)
14:55:55 Feb 29 14:51:36.117: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (48.113255528s elapsed)
14:55:55 Feb 29 14:51:38.120: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (50.117019458s elapsed)
14:55:55 Feb 29 14:51:40.124: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (52.121017084s elapsed)
14:55:55 Feb 29 14:51:42.131: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (54.128017746s elapsed)
14:55:55 Feb 29 14:51:44.136: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (56.133153645s elapsed)
14:55:55 Feb 29 14:51:46.142: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (58.138703473s elapsed)
14:55:55 Feb 29 14:51:48.146: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m0.142951356s elapsed)
14:55:55 Feb 29 14:51:50.151: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m2.147803745s elapsed)
14:55:55 Feb 29 14:51:52.155: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m4.151527539s elapsed)
14:55:55 Feb 29 14:51:54.159: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m6.155260186s elapsed)
14:55:55 Feb 29 14:51:56.162: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m8.158894618s elapsed)
14:55:55 Feb 29 14:51:58.166: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m10.16279543s elapsed)
14:55:55 Feb 29 14:52:00.170: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m12.166664238s elapsed)
14:55:55 Feb 29 14:52:02.174: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m14.170477718s elapsed)
14:55:55 Feb 29 14:52:04.177: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m16.173845691s elapsed)
14:55:55 Feb 29 14:52:06.181: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m18.1777854s elapsed)
14:55:55 Feb 29 14:52:08.190: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m20.186586746s elapsed)
14:55:55 Feb 29 14:52:10.193: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m22.190208344s elapsed)
14:55:55 Feb 29 14:52:12.198: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m24.194306234s elapsed)
14:55:55 Feb 29 14:52:14.201: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m26.198094678s elapsed)
14:55:55 Feb 29 14:52:16.205: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m28.201951293s elapsed)
14:55:55 Feb 29 14:52:18.209: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m30.205665439s elapsed)
14:55:55 Feb 29 14:52:20.213: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m32.210082255s elapsed)
14:55:55 Feb 29 14:52:22.217: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m34.213701417s elapsed)
14:55:55 Feb 29 14:52:24.221: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m36.217660413s elapsed)
14:55:55 Feb 29 14:52:26.225: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m38.221455902s elapsed)
14:55:55 Feb 29 14:52:28.234: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m40.231134268s elapsed)
14:55:55 Feb 29 14:52:30.242: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m42.238287924s elapsed)
14:55:55 Feb 29 14:52:32.254: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m44.251086385s elapsed)
14:55:55 Feb 29 14:52:34.260: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m46.256640711s elapsed)
14:55:55 Feb 29 14:52:36.267: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m48.263785131s elapsed)
14:55:55 Feb 29 14:52:38.277: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m50.274053874s elapsed)
14:55:55 Feb 29 14:52:40.285: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m52.281741783s elapsed)
14:55:55 Feb 29 14:52:42.289: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m54.285413431s elapsed)
14:55:55 Feb 29 14:52:44.293: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m56.28989383s elapsed)
14:55:55 Feb 29 14:52:46.298: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m58.294327581s elapsed)
14:55:55 Feb 29 14:52:48.302: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m0.298886666s elapsed)
14:55:55 Feb 29 14:52:50.307: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m2.303273246s elapsed)
14:55:55 Feb 29 14:52:52.311: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m4.30776469s elapsed)
14:55:55 Feb 29 14:52:54.317: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m6.313927782s elapsed)
14:55:55 Feb 29 14:52:56.321: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m8.318133634s elapsed)
14:55:55 Feb 29 14:52:58.326: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m10.322492904s elapsed)
14:55:55 Feb 29 14:53:00.330: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m12.327198643s elapsed)
14:55:55 Feb 29 14:53:02.334: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m14.330775062s elapsed)
14:55:55 Feb 29 14:53:04.338: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m16.334508684s elapsed)
14:55:55 Feb 29 14:53:06.344: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m18.340683374s elapsed)
14:55:55 Feb 29 14:53:08.348: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m20.344518882s elapsed)
14:55:55 Feb 29 14:53:10.351: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m22.347831402s elapsed)
14:55:55 Feb 29 14:53:12.355: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m24.351984641s elapsed)
14:55:55 Feb 29 14:53:14.359: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m26.355424285s elapsed)
14:55:55 Feb 29 14:53:16.362: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m28.359059512s elapsed)
14:55:55 Feb 29 14:53:18.369: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m30.365380933s elapsed)
14:55:55 Feb 29 14:53:20.377: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m32.373576478s elapsed)
14:55:55 Feb 29 14:53:22.381: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m34.37780978s elapsed)
14:55:55 Feb 29 14:53:24.385: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m36.38192322s elapsed)
14:55:55 Feb 29 14:53:26.389: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m38.385781905s elapsed)
14:55:55 Feb 29 14:53:28.393: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m40.389700512s elapsed)
14:55:55 Feb 29 14:53:30.412: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m42.408735121s elapsed)
14:55:55 Feb 29 14:53:32.416: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m44.412434592s elapsed)
14:55:55 Feb 29 14:53:34.420: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m46.416506436s elapsed)
14:55:55 Feb 29 14:53:36.424: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m48.421147451s elapsed)
14:55:55 Feb 29 14:53:38.428: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m50.424531085s elapsed)
14:55:55 Feb 29 14:53:40.431: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m52.42809243s elapsed)
14:55:55 Feb 29 14:53:42.441: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m54.437362409s elapsed)
14:55:55 Feb 29 14:53:44.444: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m56.440968638s elapsed)
14:55:55 Feb 29 14:53:46.448: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m58.445160653s elapsed)
14:55:55 Feb 29 14:53:48.452: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m0.449025955s elapsed)
14:55:55 Feb 29 14:53:50.456: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m2.452866072s elapsed)
14:55:55 Feb 29 14:53:52.464: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m4.460597301s elapsed)
14:55:55 Feb 29 14:53:54.476: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m6.473234019s elapsed)
14:55:55 Feb 29 14:53:56.480: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m8.476934413s elapsed)
14:55:55 Feb 29 14:53:58.484: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m10.480663504s elapsed)
14:55:55 Feb 29 14:54:00.488: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m12.484571084s elapsed)
14:55:55 Feb 29 14:54:02.491: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m14.488150476s elapsed)
14:55:55 Feb 29 14:54:04.495: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m16.491613464s elapsed)
14:55:55 Feb 29 14:54:06.498: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m18.495106773s elapsed)
14:55:55 Feb 29 14:54:08.502: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m20.49868058s elapsed)
14:55:55 Feb 29 14:54:10.506: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m22.502594883s elapsed)
14:55:55 Feb 29 14:54:12.511: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m24.507411779s elapsed)
14:55:55 Feb 29 14:54:14.514: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m26.510994085s elapsed)
14:55:55 Feb 29 14:54:16.519: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m28.5153312s elapsed)
14:55:55 Feb 29 14:54:18.523: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m30.519443897s elapsed)
14:55:55 Feb 29 14:54:20.527: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m32.523553138s elapsed)
14:55:55 Feb 29 14:54:22.535: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m34.531990009s elapsed)
14:55:55 Feb 29 14:54:24.539: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m36.535571536s elapsed)
14:55:55 Feb 29 14:54:26.543: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m38.539444702s elapsed)
14:55:55 Feb 29 14:54:28.546: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m40.542956695s elapsed)
14:55:55 Feb 29 14:54:30.552: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m42.549208619s elapsed)
14:55:55 Feb 29 14:54:32.558: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m44.555099434s elapsed)
14:55:55 Feb 29 14:54:34.563: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m46.559359149s elapsed)
14:55:55 Feb 29 14:54:36.567: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m48.563407504s elapsed)
14:55:55 Feb 29 14:54:38.570: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m50.567043354s elapsed)
14:55:55 Feb 29 14:54:40.574: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m52.570281177s elapsed)
14:55:55 Feb 29 14:54:42.577: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m54.573874763s elapsed)
14:55:55 Feb 29 14:54:44.581: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m56.577553334s elapsed)
14:55:55 Feb 29 14:54:46.586: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m58.583147738s elapsed)
14:55:55 Feb 29 14:54:48.590: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m0.586910288s elapsed)
14:55:55 Feb 29 14:54:50.594: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m2.590545622s elapsed)
14:55:55 Feb 29 14:54:52.597: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m4.594183648s elapsed)
14:55:55 Feb 29 14:54:54.601: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m6.597603705s elapsed)
14:55:55 Feb 29 14:54:56.610: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m8.606938741s elapsed)
14:55:55 Feb 29 14:54:58.614: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m10.610731259s elapsed)
14:55:55 Feb 29 14:55:00.617: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m12.613946615s elapsed)
14:55:55 Feb 29 14:55:02.621: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m14.617805149s elapsed)
14:55:55 Feb 29 14:55:04.629: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m16.625471868s elapsed)
14:55:55 Feb 29 14:55:06.633: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m18.629301384s elapsed)
14:55:55 Feb 29 14:55:08.641: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m20.637290114s elapsed)
14:55:55 Feb 29 14:55:10.644: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m22.64111968s elapsed)
14:55:55 Feb 29 14:55:12.648: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m24.645162778s elapsed)
14:55:55 Feb 29 14:55:14.652: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m26.648654287s elapsed)
14:55:55 Feb 29 14:55:16.655: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m28.652139493s elapsed)
14:55:55 Feb 29 14:55:18.660: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m30.656380224s elapsed)
14:55:55 Feb 29 14:55:20.663: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m32.660059881s elapsed)
14:55:55 Feb 29 14:55:22.667: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m34.663450733s elapsed)
14:55:55 Feb 29 14:55:24.671: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m36.667387975s elapsed)
14:55:55 Feb 29 14:55:26.674: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m38.671224595s elapsed)
14:55:55 Feb 29 14:55:28.678: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m40.675048194s elapsed)
14:55:55 Feb 29 14:55:30.682: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m42.678872766s elapsed)
14:55:55 Feb 29 14:55:32.686: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m44.682420629s elapsed)
14:55:55 Feb 29 14:55:34.689: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m46.686230956s elapsed)
14:55:55 Feb 29 14:55:36.693: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m48.689848937s elapsed)
14:55:55 Feb 29 14:55:38.697: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m50.693537391s elapsed)
14:55:55 Feb 29 14:55:40.706: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m52.702417132s elapsed)
14:55:55 Feb 29 14:55:42.709: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m54.706169746s elapsed)
14:55:55 Feb 29 14:55:44.714: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m56.710855695s elapsed)
14:55:55 Feb 29 14:55:46.718: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m58.714629186s elapsed)
14:55:55 Feb 29 14:55:48.718: INFO: Waiting up to 5m0s for pod nginx-controller-uub6h status to be running
14:55:55 Feb 29 14:55:48.721: INFO: Found pod 'nginx-controller-uub6h' on node 'gke-jenkins-e2e-f1ec7fe0-node-l2sf'
14:55:55 Feb 29 14:55:48.721: INFO: error in waiting for pods to come up: failed to wait for pods running: [gave up waiting for pod 'nginx-controller-cceta' to be 'running' after 5m0s gave up waiting for pod 'nginx-controller-h8inh' to be 'running' after 5m0s]
14:55:55 [AfterEach] Deployment
14:55:55 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework.go:84
14:55:55 STEP: Collecting events from namespace "e2e-tests-deployment-ezytf".
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:45:42 -0800 PST - event for nginx-controller: {replicaset-controller } SuccessfulCreate: Created pod: nginx-controller-h8inh
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:45:42 -0800 PST - event for nginx-controller: {replicaset-controller } SuccessfulCreate: Created pod: nginx-controller-cceta
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:45:42 -0800 PST - event for nginx-controller: {replicaset-controller } SuccessfulCreate: Created pod: nginx-controller-uub6h
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:45:42 -0800 PST - event for nginx-controller-cceta: {default-scheduler } Scheduled: Successfully assigned nginx-controller-cceta to gke-jenkins-e2e-f1ec7fe0-node-fkww
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:45:42 -0800 PST - event for nginx-controller-h8inh: {default-scheduler } Scheduled: Successfully assigned nginx-controller-h8inh to gke-jenkins-e2e-f1ec7fe0-node-8mwd
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:45:42 -0800 PST - event for nginx-controller-uub6h: {default-scheduler } Scheduled: Successfully assigned nginx-controller-uub6h to gke-jenkins-e2e-f1ec7fe0-node-l2sf
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:26 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} Pulling: pulling image "nginx"
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:28 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ErrImagePull: "Error pulling image (latest) from docker.io/library/nginx, Server error: 500 trying to fetch remote history for 6e36f46089ed3c0326d2f56d6282af5eab6000caaa04e44f327c37f13d13c933"
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:28 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} Failed: Failed to pull image "nginx": Error pulling image (latest) from docker.io/library/nginx, Server error: 500 trying to fetch remote history for 6e36f46089ed3c0326d2f56d6282af5eab6000caaa04e44f327c37f13d13c933
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:29 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} Pulling: pulling image "nginx"
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:30 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ImagePullBackOff: "Back-off pulling image \"nginx\""
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:30 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} BackOff: Back-off pulling image "nginx"
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:31 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ErrImagePull: "Could not reach any registry endpoint"
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:31 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} Failed: Failed to pull image "nginx": Could not reach any registry endpoint
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:34 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} BackOff: Back-off pulling image "nginx"
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:34 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ImagePullBackOff: "Back-off pulling image \"nginx\""
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:37 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} Pulling: pulling image "nginx"
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:39 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} BackOff: Back-off pulling image "nginx"
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:39 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ImagePullBackOff: "Back-off pulling image \"nginx\""
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:39 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ErrImagePull: "Could not reach any registry endpoint"
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:39 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} Failed: Failed to pull image "nginx": Could not reach any registry endpoint
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:47:05 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} Failed: Failed to pull image "nginx": Error pulling image (latest) from docker.io/library/nginx, Server error: Status 500 while fetching image layer (d4e3870153bf700b1df356ca654c38ac1767c55c51cecd0a2336278e3256a43b)
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:47:05 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ErrImagePull: "Error pulling image (latest) from docker.io/library/nginx, Server error: Status 500 while fetching image layer (d4e3870153bf700b1df356ca654c38ac1767c55c51cecd0a2336278e3256a43b)"
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:47:32 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ErrImagePull: "Could not reach any registry endpoint"
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:47:32 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} Failed: Failed to pull image "nginx": Could not reach any registry endpoint
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:47:55 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} Failed: Failed to pull image "nginx": Error pulling image (latest) from docker.io/library/nginx, Server error: 500 trying to fetch remote history for 6e36f46089ed3c0326d2f56d6282af5eab6000caaa04e44f327c37f13d13c933
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:47:55 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ErrImagePull: "Error pulling image (latest) from docker.io/library/nginx, Server error: 500 trying to fetch remote history for 6e36f46089ed3c0326d2f56d6282af5eab6000caaa04e44f327c37f13d13c933"
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:48:03 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} Failed: Failed to pull image "nginx": Error pulling image (latest) from docker.io/library/nginx, Server error: 500 trying to fetch remote history for 6e36f46089ed3c0326d2f56d6282af5eab6000caaa04e44f327c37f13d13c933
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:48:03 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ErrImagePull: "Error pulling image (latest) from docker.io/library/nginx, Server error: 500 trying to fetch remote history for 6e36f46089ed3c0326d2f56d6282af5eab6000caaa04e44f327c37f13d13c933"
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:49:03 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} Failed: Failed to pull image "nginx": Error pulling image (latest) from docker.io/library/nginx, HTTP code 500
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:49:03 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ErrImagePull: "Error pulling image (latest) from docker.io/library/nginx, HTTP code 500"
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:49:28 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} Pulled: Successfully pulled image "nginx"
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:49:28 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} Created: Created container with docker id 2682a301ced5
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:49:28 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} Started: Started container with docker id 2682a301ced5
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:53:18 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} Created: Created container with docker id 1cd3cd65e637
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:53:18 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} Started: Started container with docker id 1cd3cd65e637
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:53:18 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} Pulled: Successfully pulled image "nginx"
14:55:55 Feb 29 14:55:48.749: INFO: POD NODE PHASE GRACE CONDITIONS
14:55:55 Feb 29 14:55:48.749: INFO: nginx-controller-cceta gke-jenkins-e2e-f1ec7fe0-node-fkww Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:53:18 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: nginx-controller-h8inh gke-jenkins-e2e-f1ec7fe0-node-8mwd Pending [{Ready False 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:45:42 -0800 PST ContainersNotReady containers with unready status: [nginx]}]
14:55:55 Feb 29 14:55:48.749: INFO: nginx-controller-uub6h gke-jenkins-e2e-f1ec7fe0-node-l2sf Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:49:28 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: nginx-controller-10sxv gke-jenkins-e2e-f1ec7fe0-node-8mwd Pending [{Ready False 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:55:23 -0800 PST ContainersNotReady containers with unready status: [nginx]}]
14:55:55 Feb 29 14:55:48.749: INFO: nginx-controller-6tzlu gke-jenkins-e2e-f1ec7fe0-node-l2sf Pending [{Ready False 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:55:23 -0800 PST ContainersNotReady containers with unready status: [nginx]}]
14:55:55 Feb 29 14:55:48.749: INFO: nginx-controller-gq2aw gke-jenkins-e2e-f1ec7fe0-node-fkww Pending [{Ready False 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:55:23 -0800 PST ContainersNotReady containers with unready status: [nginx]}]
14:55:55 Feb 29 14:55:48.749: INFO: nginx-controller-ta510 gke-jenkins-e2e-f1ec7fe0-node-8mwd Pending [{Ready False 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:55:23 -0800 PST ContainersNotReady containers with unready status: [nginx]}]
14:55:55 Feb 29 14:55:48.749: INFO: mutability-test-0lbqb gke-jenkins-e2e-f1ec7fe0-node-fkww Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:49:22 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: mutability-test-0u77q gke-jenkins-e2e-f1ec7fe0-node-l2sf Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:49:27 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: fluentd-cloud-logging-gke-jenkins-e2e-f1ec7fe0-node-8mwd gke-jenkins-e2e-f1ec7fe0-node-8mwd Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:23 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: fluentd-cloud-logging-gke-jenkins-e2e-f1ec7fe0-node-fkww gke-jenkins-e2e-f1ec7fe0-node-fkww Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:07 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: fluentd-cloud-logging-gke-jenkins-e2e-f1ec7fe0-node-l2sf gke-jenkins-e2e-f1ec7fe0-node-l2sf Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:07 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: heapster-v14-762qq gke-jenkins-e2e-f1ec7fe0-node-l2sf Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:14 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: kube-dns-v10-87d3g gke-jenkins-e2e-f1ec7fe0-node-fkww Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:24 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: kube-proxy-gke-jenkins-e2e-f1ec7fe0-node-8mwd gke-jenkins-e2e-f1ec7fe0-node-8mwd Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:05 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: kube-proxy-gke-jenkins-e2e-f1ec7fe0-node-fkww gke-jenkins-e2e-f1ec7fe0-node-fkww Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:43:48 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: kube-proxy-gke-jenkins-e2e-f1ec7fe0-node-l2sf gke-jenkins-e2e-f1ec7fe0-node-l2sf Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:00 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: kubernetes-dashboard-v0.1.0-tm60w gke-jenkins-e2e-f1ec7fe0-node-fkww Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:08 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: l7-lb-controller-v0.5.2-h3ffn gke-jenkins-e2e-f1ec7fe0-node-l2sf Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:23 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO:
14:55:55 Feb 29 14:55:48.756: INFO:
14:55:55 Logging node info for node gke-jenkins-e2e-f1ec7fe0-node-8mwd
14:55:55 Feb 29 14:55:48.760: INFO: Node Info: &{{ } {gke-jenkins-e2e-f1ec7fe0-node-8mwd /api/v1/nodes/gke-jenkins-e2e-f1ec7fe0-node-8mwd d45ee606-df35-11e5-869e-42010af0000f 5164 0 2016-02-29 14:43:19 -0800 PST <nil> <nil> map[beta.kubernetes.io/instance-type:n1-standard-2 failure-domain.beta.kubernetes.io/region:us-central1 failure-domain.beta.kubernetes.io/zone:us-central1-f kubernetes.io/hostname:gke-jenkins-e2e-f1ec7fe0-node-8mwd] map[]} {10.180.2.0/24 7238163060162554869 gce://k8s-jkns-e2e-gke-ci/us-central1-f/gke-jenkins-e2e-f1ec7fe0-node-8mwd false} {map[cpu:{2.000 DecimalSI} memory:{7864139776.000 BinarySI} pods:{110.000 DecimalSI}] map[pods:{110.000 DecimalSI} cpu:{2.000 DecimalSI} memory:{7864139776.000 BinarySI}] [{OutOfDisk False 2016-02-29 14:55:47 -0800 PST 2016-02-29 14:43:19 -0800 PST KubeletHasSufficientDisk kubelet has sufficient disk space available} {Ready True 2016-02-29 14:55:47 -0800 PST 2016-02-29 14:44:00 -0800 PST KubeletReady kubelet is posting ready status}] [{InternalIP 10.240.0.2} {ExternalIP 173.255.112.50}] {{10250}} { DCE991B5-66DD-4ECF-7F5F-42F7B9B9B0C7 3b2b1188-dfdf-4a06-9073-e290f77082fd 3.16.0-4-amd64 Debian GNU/Linux 7 (wheezy) docker://1.9.1 v1.2.0-alpha.8.676+83d74dd9528e42 v1.2.0-alpha.8.676+83d74dd9528e42} [{[gcr.io/google_containers/kube-proxy:6952338e87c87e4bef2030a93d05b990] 165606485} {[redis:latest] 151331559} {[gcr.io/google_containers/nettest:1.8] 25178998} {[<none>:<none>] 142324527} {[nginx:latest] 134617408} {[<none>:<none>] 134617408} {[busybox:latest] 1113554} {[gcr.io/google_samples/gb-frontend:v4] 510252254} {[gcr.io/google_containers/nettest:1.7] 24051275} {[gcr.io/google_containers/fluentd-gcp:1.16] 406764477} {[gcr.io/google_containers/netexec:1.4] 7297019} {[gcr.io/google_containers/busybox:1.24] 1113554} {[gcr.io/google_containers/mounttest:0.6] 2084693} {[gcr.io/google_containers/hostexec:1.2] 13209617} {[gcr.io/google_containers/mounttest-user:0.3] 1718853} {[gcr.io/google_containers/mounttest:0.5] 1718853} {[gcr.io/google_containers/pause:2.0] 350164} {[gcr.io/google_containers/portforwardtester:1.0] 2296329} {[gcr.io/google_samples/gb-redisslave:v1] 109508753} {[gcr.io/google_containers/mounttest:0.2] 1752375} {[gcr.io/google_containers/pause:0.8.0] 241656} {[gcr.io/google_containers/eptest:0.1] 2970692} {[b.gcr.io/k8s_authenticated_test/serve_hostname:1.1] 4522409} {[gcr.io/google_containers/serve_hostname:1.1] 4522409} {[gcr.io/google_containers/update-demo:nautilus] 4555533} {[gcr.io/google_containers/update-demo:kitten] 4549069} {[gcr.io/google_containers/nginx:1.7.9] 91664166} {[gcr.io/google_containers/test-webserver:e2e] 4534272}]}}
14:55:55 Feb 29 14:55:48.760: INFO:
14:55:55 Logging kubelet events for node gke-jenkins-e2e-f1ec7fe0-node-8mwd
14:55:55 Feb 29 14:55:48.768: INFO:
14:55:55 Logging pods the kubelet thinks is on node gke-jenkins-e2e-f1ec7fe0-node-8mwd
14:55:55 Feb 29 14:55:48.796: INFO: nginx-controller-h8inh started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.796: INFO: fluentd-cloud-logging-gke-jenkins-e2e-f1ec7fe0-node-8mwd started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.796: INFO: kube-proxy-gke-jenkins-e2e-f1ec7fe0-node-8mwd started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.796: INFO: nginx-controller-10sxv started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.796: INFO: nginx-controller-ta510 started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="info"} => 1 @[0]
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="inspect_image"} => 80 @[0]
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="list_containers"} => 92 @[0]
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="list_images"} => 18 @[0]
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="logs"} => 1 @[0]
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="pull_image"} => 35 @[0]
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="start_container"} => 1 @[0]
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="stop_container"} => 52 @[0]
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="version"} => 72 @[0]
14:55:55 Feb 29 14:55:48.863: INFO:
14:55:55 Latency metrics for node gke-jenkins-e2e-f1ec7fe0-node-8mwd
14:55:55 Feb 29 14:55:48.863: INFO: {Operation: Method:pod_start_latency_microseconds Quantile:0.99 Latency:4m3.668161s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:create Method:pod_worker_latency_microseconds Quantile:0.99 Latency:1m18.437219s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:SyncPod Method:container_manager_latency_microseconds Quantile:0.99 Latency:1m18.080256s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:sync Method:pod_worker_latency_microseconds Quantile:0.99 Latency:1m9.755388s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation: Method:pod_start_latency_microseconds Quantile:0.9 Latency:1m5.820405s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:create Method:pod_worker_latency_microseconds Quantile:0.9 Latency:1m2.677207s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:pull_image Method:docker_operations_latency_microseconds Quantile:0.99 Latency:56.918473s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:update Method:pod_worker_latency_microseconds Quantile:0.99 Latency:30.560446s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:stop_container Method:docker_operations_latency_microseconds Quantile:0.99 Latency:30.09008s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:SyncPod Method:container_manager_latency_microseconds Quantile:0.9 Latency:29.353428s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:sync Method:pod_worker_latency_microseconds Quantile:0.9 Latency:11.242491s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:pull_image Method:docker_operations_latency_microseconds Quantile:0.9 Latency:10.805248s}
14:55:55 Feb 29 14:55:48.863: INFO:
14:55:55 Logging node info for node gke-jenkins-e2e-f1ec7fe0-node-fkww
14:55:55 Feb 29 14:55:48.868: INFO: Node Info: &{{ } {gke-jenkins-e2e-f1ec7fe0-node-fkww /api/v1/nodes/gke-jenkins-e2e-f1ec7fe0-node-fkww d359aed6-df35-11e5-869e-42010af0000f 5161 0 2016-02-29 14:43:17 -0800 PST <nil> <nil> map[beta.kubernetes.io/instance-type:n1-standard-2 failure-domain.beta.kubernetes.io/region:us-central1 failure-domain.beta.kubernetes.io/zone:us-central1-f kubernetes.io/hostname:gke-jenkins-e2e-f1ec7fe0-node-fkww] map[]} {10.180.0.0/24 11412112845876479197 gce://k8s-jkns-e2e-gke-ci/us-central1-f/gke-jenkins-e2e-f1ec7fe0-node-fkww false} {map[cpu:{2.000 DecimalSI} memory:{7864139776.000 BinarySI} pods:{110.000 DecimalSI}] map[cpu:{2.000 DecimalSI} memory:{7864139776.000 BinarySI} pods:{110.000 DecimalSI}] [{OutOfDisk False 2016-02-29 14:55:45 -0800 PST 2016-02-29 14:43:17 -0800 PST KubeletHasSufficientDisk kubelet has sufficient disk space available} {Ready True 2016-02-29 14:55:45 -0800 PST 2016-02-29 14:43:47 -0800 PST KubeletReady kubelet is posting ready status}] [{InternalIP 10.240.0.3} {ExternalIP 173.255.119.226}] {{10250}} { 4BAFA5DA-437E-C741-F5E4-5726E8EB84B7 55f15f24-cb47-418b-8e1a-286c4625a4a8 3.16.0-4-amd64 Debian GNU/Linux 7 (wheezy) docker://1.9.1 v1.2.0-alpha.8.676+83d74dd9528e42 v1.2.0-alpha.8.676+83d74dd9528e42} [{[gcr.io/google_containers/kube-proxy:6952338e87c87e4bef2030a93d05b990] 165606485} {[redis:latest] 151331559} {[gcr.io/google_containers/nettest:1.8] 25178998} {[nginx:latest] 134617408} {[<none>:<none>] 134617408} {[<none>:<none>] 1113554} {[busybox:latest] 1113554} {[gcr.io/google_samples/gb-frontend:v4] 510252254} {[gcr.io/google_containers/nettest:1.7] 24051275} {[gcr.io/google_containers/kubernetes-dashboard-amd64:v0.1.0] 35409795} {[gcr.io/google_containers/fluentd-gcp:1.16] 406764477} {[gcr.io/google_containers/netexec:1.4] 7297019} {[gcr.io/google_containers/etcd-amd64:2.2.1] 28192476} {[gcr.io/google_containers/fakegitserver:0.1] 5007469} {[gcr.io/google_containers/busybox:1.24] 1113554} {[gcr.io/google_containers/kube2sky:1.12] 24482187} {[gcr.io/google_containers/mounttest:0.6] 2084693} {[gcr.io/google_containers/mounttest-user:0.3] 1718853} {[gcr.io/google_containers/mounttest:0.5] 1718853} {[gcr.io/google_containers/skydns:2015-10-13-8c72f8c] 40551394} {[gcr.io/google_containers/pause:2.0] 350164} {[gcr.io/google_containers/exechealthz:1.0] 7099444} {[gcr.io/google_containers/jessie-dnsutils:e2e] 190148402} {[gcr.io/google_containers/mounttest:0.2] 1752375} {[gcr.io/google_containers/dnsutils:e2e] 141895666} {[gcr.io/google_containers/pause:0.8.0] 241656} {[gcr.io/google_containers/eptest:0.1] 2970692} {[gcr.io/google_containers/serve_hostname:1.1] 4522409} {[gcr.io/google_containers/update-demo:nautilus] 4555533} {[gcr.io/google_containers/nginx:1.7.9] 91664166} {[gcr.io/google_containers/test-webserver:e2e] 4534272} {[gcr.io/google_containers/liveness:e2e] 4387474}]}}
14:55:55 Feb 29 14:55:48.868: INFO:
14:55:55 Logging kubelet events for node gke-jenkins-e2e-f1ec7fe0-node-fkww
14:55:55 Feb 29 14:55:48.874: INFO:
14:55:55 Logging pods the kubelet thinks is on node gke-jenkins-e2e-f1ec7fe0-node-fkww
14:55:55 Feb 29 14:55:48.886: INFO: kube-proxy-gke-jenkins-e2e-f1ec7fe0-node-fkww started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.886: INFO: nginx-controller-gq2aw started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.886: INFO: nginx-controller-cceta started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.886: INFO: mutability-test-0lbqb started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.886: INFO: kube-dns-v10-87d3g started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.886: INFO: kubernetes-dashboard-v0.1.0-tm60w started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.886: INFO: fluentd-cloud-logging-gke-jenkins-e2e-f1ec7fe0-node-fkww started at <nil> (0 container statuses recorded)
14:55:55 W0229 14:55:48.941264 14311 request.go:627] Throttling request took 53.998802ms, request: https://104.197.114.165/api/v1/proxy/nodes/gke-jenkins-e2e-f1ec7fe0-node-fkww:10250/metrics
14:55:55 Feb 29 14:55:49.073: INFO: ERROR kubelet_docker_errors{operation_type="info"} => 1 @[0]
14:55:55 Feb 29 14:55:49.073: INFO: ERROR kubelet_docker_errors{operation_type="inspect_image"} => 64 @[0]
14:55:55 Feb 29 14:55:49.073: INFO: ERROR kubelet_docker_errors{operation_type="list_containers"} => 63 @[0]
14:55:55 Feb 29 14:55:49.073: INFO: ERROR kubelet_docker_errors{operation_type="list_images"} => 14 @[0]
14:55:55 Feb 29 14:55:49.073: INFO: ERROR kubelet_docker_errors{operation_type="pull_image"} => 21 @[0]
14:55:55 Feb 29 14:55:49.073: INFO: ERROR kubelet_docker_errors{operation_type="start_container"} => 1 @[0]
14:55:55 Feb 29 14:55:49.073: INFO: ERROR kubelet_docker_errors{operation_type="stop_container"} => 31 @[0]
14:55:55 Feb 29 14:55:49.073: INFO: ERROR kubelet_docker_errors{operation_type="version"} => 52 @[0]
14:55:55 Feb 29 14:55:49.074: INFO:
14:55:55 Latency metrics for node gke-jenkins-e2e-f1ec7fe0-node-fkww
14:55:55 Feb 29 14:55:49.074: INFO: {Operation: Method:pod_start_latency_microseconds Quantile:0.99 Latency:7m36.012229s}
14:55:55 Feb 29 14:55:49.074: INFO: {Operation:sync Method:pod_worker_latency_microseconds Quantile:0.99 Latency:36.084205s}
14:55:55 Feb 29 14:55:49.074: INFO: {Operation: Method:pod_start_latency_microseconds Quantile:0.9 Latency:36.006157s}
14:55:55 Feb 29 14:55:49.074: INFO: {Operation:stop_container Method:docker_operations_latency_microseconds Quantile:0.99 Latency:30.178837s}
14:55:55 Feb 29 14:55:49.074: INFO: {Operation:SyncPod Method:container_manager_latency_microseconds Quantile:0.99 Latency:26.024915s}
14:55:55 Feb 29 14:55:49.074: INFO: {Operation:pull_image Method:docker_operations_latency_microseconds Quantile:0.99 Latency:21.33357s}
14:55:55 Feb 29 14:55:49.074: INFO:
14:55:55 Logging node info for node gke-jenkins-e2e-f1ec7fe0-node-l2sf
14:55:55 W0229 14:55:49.141242 14311 request.go:627] Throttling request took 67.09998ms, request: https://104.197.114.165/api/v1/nodes/gke-jenkins-e2e-f1ec7fe0-node-l2sf
14:55:55 Feb 29 14:55:49.145: INFO: Node Info: &{{ } {gke-jenkins-e2e-f1ec7fe0-node-l2sf /api/v1/nodes/gke-jenkins-e2e-f1ec7fe0-node-l2sf d2bce91c-df35-11e5-869e-42010af0000f 5160 0 2016-02-29 14:43:16 -0800 PST <nil> <nil> map[beta.kubernetes.io/instance-type:n1-standard-2 failure-domain.beta.kubernetes.io/region:us-central1 failure-domain.beta.kubernetes.io/zone:us-central1-f kubernetes.io/hostname:gke-jenkins-e2e-f1ec7fe0-node-l2sf] map[]} {10.180.1.0/24 15180633427855587389 gce://k8s-jkns-e2e-gke-ci/us-central1-f/gke-jenkins-e2e-f1ec7fe0-node-l2sf false} {map[pods:{110.000 DecimalSI} cpu:{2.000 DecimalSI} memory:{7864139776.000 BinarySI}] map[cpu:{2.000 DecimalSI} memory:{7864139776.000 BinarySI} pods:{110.000 DecimalSI}] [{OutOfDisk False 2016-02-29 14:55:43 -0800 PST 2016-02-29 14:43:16 -0800 PST KubeletHasSufficientDisk kubelet has sufficient disk space available} {Ready True 2016-02-29 14:55:43 -0800 PST 2016-02-29 14:43:46 -0800 PST KubeletReady kubelet is posting ready status}] [{InternalIP 10.240.0.4} {ExternalIP 23.236.59.190}] {{10250}} { 19DD0CB4-8383-602F-3D1C-64B17EE856B9 4e5523c0-c6d5-42d4-8e11-efb76b337cc3 3.16.0-4-amd64 Debian GNU/Linux 7 (wheezy) docker://1.9.1 v1.2.0-alpha.8.676+83d74dd9528e42 v1.2.0-alpha.8.676+83d74dd9528e42} [{[gcr.io/google_containers/kube-proxy:6952338e87c87e4bef2030a93d05b990] 165606485} {[gcr.io/google_containers/heapster:v0.20.0-alpha8] 86984348} {[redis:latest] 151331559} {[gcr.io/google_containers/nettest:1.8] 25178998} {[nginx:latest] 134617408} {[gcr.io/google_samples/gb-frontend:v4] 510252254} {[gcr.io/google_containers/fluentd-gcp:1.16] 406764477} {[gcr.io/google_containers/netexec:1.4] 7297019} {[gcr.io/google_containers/glbc:0.5.2] 226351998} {[gcr.io/google_containers/busybox:1.24] 1113554} {[gcr.io/google_containers/mounttest:0.6] 2084693} {[gcr.io/google_containers/hostexec:1.2] 13209617} {[gcr.io/google_containers/mounttest-user:0.3] 1718853} {[gcr.io/google_containers/defaultbackend:1.0] 7513643} {[gcr.io/google_containers/mounttest:0.5] 1718853} {[gcr.io/google_containers/pause:2.0] 350164} {[gcr.io/google_containers/porter:cd5cb5791ebaa8641955f0e8c2a9bed669b1eaab] 5010921} {[gcr.io/google_containers/goproxy:0.1] 5489816} {[gcr.io/google_samples/gb-redisslave:v1] 109508753} {[gcr.io/google_containers/mounttest:0.2] 1752375} {[gcr.io/google_containers/pause:0.8.0] 241656} {[b.gcr.io/k8s_authenticated_test/serve_hostname:1.1] 4522409} {[gcr.io/google_containers/serve_hostname:1.1] 4522409} {[gcr.io/google_containers/update-demo:kitten] 4549069} {[gcr.io/google_containers/update-demo:nautilus] 4555533} {[gcr.io/google_containers/nginx:1.7.9] 91664166} {[gcr.io/google_containers/test-webserver:e2e] 4534272}]}}
14:55:55 Feb 29 14:55:49.146: INFO:
14:55:55 Logging kubelet events for node gke-jenkins-e2e-f1ec7fe0-node-l2sf
14:55:55 W0229 14:55:49.341238 14311 request.go:627] Throttling request took 194.963821ms, request: https://104.197.114.165/api/v1/namespaces/kube-system/events?fieldSelector=involvedObject.kind%3DNode%2CinvolvedObject.name%3Dgke-jenkins-e2e-f1ec7fe0-node-l2sf%2CinvolvedObject.namespace%3D%2Csource%3Dkubelet
14:55:55 Feb 29 14:55:49.346: INFO:
14:55:55 Logging pods the kubelet thinks is on node gke-jenkins-e2e-f1ec7fe0-node-l2sf
14:55:55 W0229 14:55:49.541273 14311 request.go:627] Throttling request took 193.107727ms, request: https://104.197.114.165/api/v1/proxy/nodes/gke-jenkins-e2e-f1ec7fe0-node-l2sf:10250/runningpods
14:55:55 Feb 29 14:55:49.564: INFO: fluentd-cloud-logging-gke-jenkins-e2e-f1ec7fe0-node-l2sf started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:49.564: INFO: kube-proxy-gke-jenkins-e2e-f1ec7fe0-node-l2sf started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:49.564: INFO: nginx-controller-6tzlu started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:49.564: INFO: nginx-controller-uub6h started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:49.564: INFO: mutability-test-0u77q started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:49.564: INFO: l7-lb-controller-v0.5.2-h3ffn started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:49.564: INFO: heapster-v14-762qq started at <nil> (0 container statuses recorded)
14:55:55 W0229 14:55:49.741215 14311 request.go:627] Throttling request took 175.733913ms, request: https://104.197.114.165/api/v1/proxy/nodes/gke-jenkins-e2e-f1ec7fe0-node-l2sf:10250/metrics
14:55:55 Feb 29 14:55:49.876: INFO: ERROR kubelet_docker_errors{operation_type="info"} => 1 @[0]
14:55:55 Feb 29 14:55:49.876: INFO: ERROR kubelet_docker_errors{operation_type="inspect_image"} => 48 @[0]
14:55:55 Feb 29 14:55:49.876: INFO: ERROR kubelet_docker_errors{operation_type="list_containers"} => 65 @[0]
14:55:55 Feb 29 14:55:49.876: INFO: ERROR kubelet_docker_errors{operation_type="list_images"} => 15 @[0]
14:55:55 Feb 29 14:55:49.876: INFO: ERROR kubelet_docker_errors{operation_type="pull_image"} => 26 @[0]
14:55:55 Feb 29 14:55:49.876: INFO: ERROR kubelet_docker_errors{operation_type="stop_container"} => 38 @[0]
14:55:55 Feb 29 14:55:49.876: INFO: ERROR kubelet_docker_errors{operation_type="version"} => 54 @[0]
14:55:55 Feb 29 14:55:49.876: INFO:
14:55:55 Latency metrics for node gke-jenkins-e2e-f1ec7fe0-node-l2sf
14:55:55 Feb 29 14:55:49.876: INFO: {Operation: Method:pod_start_latency_microseconds Quantile:0.99 Latency:3m46.644521s}
14:55:55 Feb 29 14:55:49.876: INFO: {Operation:create Method:pod_worker_latency_microseconds Quantile:0.99 Latency:55.520147s}
14:55:55 Feb 29 14:55:49.876: INFO: {Operation:sync Method:pod_worker_latency_microseconds Quantile:0.99 Latency:53.258439s}
14:55:55 Feb 29 14:55:49.876: INFO: {Operation:create Method:pod_worker_latency_microseconds Quantile:0.9 Latency:46.526966s}
14:55:55 Feb 29 14:55:49.876: INFO: {Operation:SyncPod Method:container_manager_latency_microseconds Quantile:0.99 Latency:38.112145s}
14:55:55 Feb 29 14:55:49.876: INFO: {Operation:update Method:pod_worker_latency_microseconds Quantile:0.99 Latency:30.421105s}
14:55:55 Feb 29 14:55:49.876: INFO: {Operation:stop_container Method:docker_operations_latency_microseconds Quantile:0.99 Latency:30.097734s}
14:55:55 Feb 29 14:55:49.876: INFO: {Operation: Method:pod_start_latency_microseconds Quantile:0.9 Latency:27.688855s}
14:55:55 Feb 29 14:55:49.876: INFO: {Operation:pull_image Method:docker_operations_latency_microseconds Quantile:0.99 Latency:20.27941s}
14:55:55 Feb 29 14:55:49.876: INFO: Waiting up to 1m0s for all nodes to be ready
14:55:55 W0229 14:55:49.941214 14311 request.go:627] Throttling request took 64.695123ms, request: https://104.197.114.165/api/v1/nodes
14:55:55 STEP: Destroying namespace "e2e-tests-deployment-ezytf" for this suite.
14:55:55 W0229 14:55:50.141262 14311 request.go:627] Throttling request took 194.312955ms, request: https://104.197.114.165/api/v1/namespaces/e2e-tests-deployment-ezytf
14:55:55 W0229 14:55:50.341242 14311 request.go:627] Throttling request took 194.516203ms, request: https://104.197.114.165/api/v1/namespaces/e2e-tests-deployment-ezytf
14:55:55
14:55:55
14:55:55 • Failure [615.210 seconds]
14:55:55 Deployment
14:55:55 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/deployment.go:72
14:55:55 deployment should label adopted RSs and pods [It]
14:55:55 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/deployment.go:71
14:55:55
14:55:55 Expected error:
14:55:55 <*errors.errorString | 0xc2084c0e60>: {
14:55:55 s: "failed to wait for pods running: [gave up waiting for pod 'nginx-controller-cceta' to be 'running' after 5m0s gave up waiting for pod 'nginx-controller-h8inh' to be 'running' after 5m0s]",
14:55:55 }
14:55:55 failed to wait for pods running: [gave up waiting for pod 'nginx-controller-cceta' to be 'running' after 5m0s gave up waiting for pod 'nginx-controller-h8inh' to be 'running' after 5m0s]
14:55:55 not to have occurred
14:55:55
14:55:55 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/deployment.go:837
14:55:55 ------------------------------"><pre class="notranslate"><code class="notranslate">14:55:55 ------------------------------
14:55:55 [BeforeEach] Deployment
14:55:55 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework.go:83
14:55:55 STEP: Creating a kubernetes client
14:55:55 Feb 29 14:45:40.139: INFO: >>> testContext.KubeConfig: /var/lib/jenkins/jobs/kubernetes-e2e-gke/workspace/.kube/config
14:55:55
14:55:55 STEP: Building a namespace api object
14:55:55 Feb 29 14:45:40.176: INFO: Waiting up to 2m0s for service account default to be provisioned in ns e2e-tests-deployment-ezytf
14:55:55 Feb 29 14:45:40.198: INFO: Service account default in ns e2e-tests-deployment-ezytf had 0 secrets, ignoring for 2s: <nil>
14:55:55 Feb 29 14:45:42.201: INFO: Service account default in ns e2e-tests-deployment-ezytf with secrets found. (2.025365056s)
14:55:55 STEP: Waiting for a default service account to be provisioned in namespace
14:55:55 Feb 29 14:45:42.201: INFO: Waiting up to 2m0s for service account default to be provisioned in ns e2e-tests-deployment-ezytf
14:55:55 Feb 29 14:45:42.206: INFO: Service account default in ns e2e-tests-deployment-ezytf with secrets found. (4.670131ms)
14:55:55 [It] deployment should label adopted RSs and pods
14:55:55 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/deployment.go:71
14:55:55 Feb 29 14:45:42.226: INFO: Pod name nginx: Found 0 pods out of 3
14:55:55 Feb 29 14:45:47.230: INFO: Pod name nginx: Found 3 pods out of 3
14:55:55 STEP: ensuring each pod is running
14:55:55 Feb 29 14:45:47.230: INFO: Waiting up to 5m0s for pod nginx-controller-cceta status to be running
14:55:55 Feb 29 14:45:47.233: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2.621607ms elapsed)
14:55:55 Feb 29 14:45:49.236: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2.006058051s elapsed)
14:55:55 Feb 29 14:45:51.240: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4.009845765s elapsed)
14:55:55 Feb 29 14:45:53.252: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (6.022251049s elapsed)
14:55:55 Feb 29 14:45:55.258: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (8.027621505s elapsed)
14:55:55 Feb 29 14:45:57.261: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (10.031267285s elapsed)
14:55:55 Feb 29 14:45:59.268: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (12.037638231s elapsed)
14:55:55 Feb 29 14:46:01.280: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (14.049614949s elapsed)
14:55:55 Feb 29 14:46:03.283: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (16.053400403s elapsed)
14:55:55 Feb 29 14:46:05.287: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (18.057109564s elapsed)
14:55:55 Feb 29 14:46:07.291: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (20.060796787s elapsed)
14:55:55 Feb 29 14:46:09.294: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (22.06440015s elapsed)
14:55:55 Feb 29 14:46:11.298: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (24.06814485s elapsed)
14:55:55 Feb 29 14:46:13.305: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (26.074887193s elapsed)
14:55:55 Feb 29 14:46:15.310: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (28.080003644s elapsed)
14:55:55 Feb 29 14:46:17.317: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (30.087297487s elapsed)
14:55:55 Feb 29 14:46:19.325: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (32.09515456s elapsed)
14:55:55 Feb 29 14:46:21.334: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (34.104373139s elapsed)
14:55:55 Feb 29 14:46:23.338: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (36.108334597s elapsed)
14:55:55 Feb 29 14:46:25.346: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (38.116396703s elapsed)
14:55:55 Feb 29 14:46:27.351: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (40.121301689s elapsed)
14:55:55 Feb 29 14:46:29.355: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (42.124813176s elapsed)
14:55:55 Feb 29 14:46:31.359: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (44.12878585s elapsed)
14:55:55 Feb 29 14:46:33.365: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (46.134847862s elapsed)
14:55:55 Feb 29 14:46:35.368: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (48.138450891s elapsed)
14:55:55 Feb 29 14:46:37.372: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (50.142281963s elapsed)
14:55:55 Feb 29 14:46:39.377: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (52.146617091s elapsed)
14:55:55 Feb 29 14:46:41.381: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (54.15122489s elapsed)
14:55:55 Feb 29 14:46:43.388: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (56.158110274s elapsed)
14:55:55 Feb 29 14:46:45.394: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (58.164077418s elapsed)
14:55:55 Feb 29 14:46:47.398: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m0.167681085s elapsed)
14:55:55 Feb 29 14:46:49.401: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m2.171265049s elapsed)
14:55:55 Feb 29 14:46:51.406: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m4.175785784s elapsed)
14:55:55 Feb 29 14:46:53.409: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m6.179102726s elapsed)
14:55:55 Feb 29 14:46:55.422: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m8.191951719s elapsed)
14:55:55 Feb 29 14:46:57.430: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m10.20031798s elapsed)
14:55:55 Feb 29 14:46:59.434: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m12.204123329s elapsed)
14:55:55 Feb 29 14:47:01.439: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m14.208718392s elapsed)
14:55:55 Feb 29 14:47:03.442: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m16.212384665s elapsed)
14:55:55 Feb 29 14:47:05.446: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m18.216284818s elapsed)
14:55:55 Feb 29 14:47:07.450: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m20.220123499s elapsed)
14:55:55 Feb 29 14:47:09.454: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m22.224276995s elapsed)
14:55:55 Feb 29 14:47:11.458: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m24.228176691s elapsed)
14:55:55 Feb 29 14:47:13.462: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m26.23182926s elapsed)
14:55:55 Feb 29 14:47:15.466: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m28.235893296s elapsed)
14:55:55 Feb 29 14:47:17.470: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m30.239743747s elapsed)
14:55:55 Feb 29 14:47:19.482: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m32.252012755s elapsed)
14:55:55 Feb 29 14:47:21.490: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m34.260158566s elapsed)
14:55:55 Feb 29 14:47:23.494: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m36.263887374s elapsed)
14:55:55 Feb 29 14:47:25.498: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m38.267798471s elapsed)
14:55:55 Feb 29 14:47:27.502: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m40.271853684s elapsed)
14:55:55 Feb 29 14:47:29.513: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m42.283292405s elapsed)
14:55:55 Feb 29 14:47:31.517: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m44.287226687s elapsed)
14:55:55 Feb 29 14:47:33.528: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m46.298316069s elapsed)
14:55:55 Feb 29 14:47:35.533: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m48.303456523s elapsed)
14:55:55 Feb 29 14:47:37.538: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m50.308487927s elapsed)
14:55:55 Feb 29 14:47:39.543: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m52.312857219s elapsed)
14:55:55 Feb 29 14:47:41.550: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m54.32038014s elapsed)
14:55:55 Feb 29 14:47:43.558: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m56.327828258s elapsed)
14:55:55 Feb 29 14:47:45.562: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m58.332220565s elapsed)
14:55:55 Feb 29 14:47:47.569: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m0.339085895s elapsed)
14:55:55 Feb 29 14:47:49.573: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m2.34325204s elapsed)
14:55:55 Feb 29 14:47:51.588: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m4.35849625s elapsed)
14:55:55 Feb 29 14:47:53.593: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m6.363014495s elapsed)
14:55:55 Feb 29 14:47:55.597: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m8.366630623s elapsed)
14:55:55 Feb 29 14:47:57.600: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m10.370008163s elapsed)
14:55:55 Feb 29 14:47:59.605: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m12.374953605s elapsed)
14:55:55 Feb 29 14:48:01.609: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m14.378645389s elapsed)
14:55:55 Feb 29 14:48:03.613: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m16.382921602s elapsed)
14:55:55 Feb 29 14:48:05.625: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m18.395545053s elapsed)
14:55:55 Feb 29 14:48:07.628: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m20.398245389s elapsed)
14:55:55 Feb 29 14:48:09.631: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m22.401138921s elapsed)
14:55:55 Feb 29 14:48:11.634: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m24.404201567s elapsed)
14:55:55 Feb 29 14:48:13.637: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m26.407054385s elapsed)
14:55:55 Feb 29 14:48:15.641: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m28.411133696s elapsed)
14:55:55 Feb 29 14:48:17.661: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m30.430611995s elapsed)
14:55:55 Feb 29 14:48:19.663: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m32.433593617s elapsed)
14:55:55 Feb 29 14:48:21.673: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m34.443030852s elapsed)
14:55:55 Feb 29 14:48:23.677: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m36.446616458s elapsed)
14:55:55 Feb 29 14:48:25.685: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m38.455065752s elapsed)
14:55:55 Feb 29 14:48:27.693: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m40.462934495s elapsed)
14:55:55 Feb 29 14:48:29.696: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m42.466417547s elapsed)
14:55:55 Feb 29 14:48:31.700: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m44.47012654s elapsed)
14:55:55 Feb 29 14:48:33.704: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m46.473811221s elapsed)
14:55:55 Feb 29 14:48:35.707: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m48.477187473s elapsed)
14:55:55 Feb 29 14:48:37.710: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m50.480501037s elapsed)
14:55:55 Feb 29 14:48:39.714: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m52.484141754s elapsed)
14:55:55 Feb 29 14:48:41.718: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m54.487909755s elapsed)
14:55:55 Feb 29 14:48:43.721: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m56.491389362s elapsed)
14:55:55 Feb 29 14:48:45.724: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m58.494594496s elapsed)
14:55:55 Feb 29 14:48:47.730: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m0.499728488s elapsed)
14:55:55 Feb 29 14:48:49.736: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m2.50568688s elapsed)
14:55:55 Feb 29 14:48:51.745: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m4.515130367s elapsed)
14:55:55 Feb 29 14:48:53.750: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m6.520567985s elapsed)
14:55:55 Feb 29 14:48:55.754: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m8.524143912s elapsed)
14:55:55 Feb 29 14:48:57.758: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m10.527734184s elapsed)
14:55:55 Feb 29 14:48:59.761: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m12.531404579s elapsed)
14:55:55 Feb 29 14:49:01.765: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m14.534863411s elapsed)
14:55:55 Feb 29 14:49:03.768: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m16.538322454s elapsed)
14:55:55 Feb 29 14:49:05.772: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m18.542429464s elapsed)
14:55:55 Feb 29 14:49:07.776: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m20.546443927s elapsed)
14:55:55 Feb 29 14:49:09.780: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m22.550147462s elapsed)
14:55:55 Feb 29 14:49:11.784: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m24.554155997s elapsed)
14:55:55 Feb 29 14:49:13.788: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m26.557779943s elapsed)
14:55:55 Feb 29 14:49:15.793: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m28.562762235s elapsed)
14:55:55 Feb 29 14:49:17.797: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m30.566609002s elapsed)
14:55:55 Feb 29 14:49:19.800: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m32.570050505s elapsed)
14:55:55 Feb 29 14:49:21.804: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m34.573663773s elapsed)
14:55:55 Feb 29 14:49:23.817: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m36.587287962s elapsed)
14:55:55 Feb 29 14:49:25.821: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m38.59092768s elapsed)
14:55:55 Feb 29 14:49:27.825: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m40.595388308s elapsed)
14:55:55 Feb 29 14:49:29.829: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m42.59903347s elapsed)
14:55:55 Feb 29 14:49:31.832: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m44.602018698s elapsed)
14:55:55 Feb 29 14:49:33.836: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m46.605876878s elapsed)
14:55:55 Feb 29 14:49:35.840: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m48.610046791s elapsed)
14:55:55 Feb 29 14:49:37.844: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m50.613781277s elapsed)
14:55:55 Feb 29 14:49:39.854: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m52.624094173s elapsed)
14:55:55 Feb 29 14:49:41.860: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m54.630550744s elapsed)
14:55:55 Feb 29 14:49:43.864: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m56.634172637s elapsed)
14:55:55 Feb 29 14:49:45.869: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m58.63948903s elapsed)
14:55:55 Feb 29 14:49:47.873: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m0.643135612s elapsed)
14:55:55 Feb 29 14:49:49.877: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m2.647470302s elapsed)
14:55:55 Feb 29 14:49:51.881: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m4.651137115s elapsed)
14:55:55 Feb 29 14:49:53.885: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m6.655102626s elapsed)
14:55:55 Feb 29 14:49:55.889: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m8.659097795s elapsed)
14:55:55 Feb 29 14:49:57.893: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m10.663495143s elapsed)
14:55:55 Feb 29 14:49:59.897: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m12.667322166s elapsed)
14:55:55 Feb 29 14:50:01.901: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m14.67117842s elapsed)
14:55:55 Feb 29 14:50:03.910: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m16.679818657s elapsed)
14:55:55 Feb 29 14:50:05.913: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m18.683340638s elapsed)
14:55:55 Feb 29 14:50:07.921: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m20.69096282s elapsed)
14:55:55 Feb 29 14:50:09.927: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m22.697135213s elapsed)
14:55:55 Feb 29 14:50:11.931: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m24.700827569s elapsed)
14:55:55 Feb 29 14:50:13.936: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m26.70621211s elapsed)
14:55:55 Feb 29 14:50:15.940: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m28.710144694s elapsed)
14:55:55 Feb 29 14:50:17.944: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m30.713880033s elapsed)
14:55:55 Feb 29 14:50:19.948: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m32.717650356s elapsed)
14:55:55 Feb 29 14:50:21.951: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m34.721059557s elapsed)
14:55:55 Feb 29 14:50:23.958: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m36.727608471s elapsed)
14:55:55 Feb 29 14:50:25.962: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m38.732520787s elapsed)
14:55:55 Feb 29 14:50:27.967: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m40.737490797s elapsed)
14:55:55 Feb 29 14:50:29.971: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m42.741214086s elapsed)
14:55:55 Feb 29 14:50:31.975: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m44.744627328s elapsed)
14:55:55 Feb 29 14:50:33.979: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m46.748988901s elapsed)
14:55:55 Feb 29 14:50:35.982: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m48.75244461s elapsed)
14:55:55 Feb 29 14:50:37.986: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m50.756274148s elapsed)
14:55:55 Feb 29 14:50:39.990: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m52.760202853s elapsed)
14:55:55 Feb 29 14:50:41.994: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m54.764189774s elapsed)
14:55:55 Feb 29 14:50:43.999: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m56.768787806s elapsed)
14:55:55 Feb 29 14:50:46.003: INFO: Waiting for pod nginx-controller-cceta in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m58.773103023s elapsed)
14:55:55 Feb 29 14:50:48.003: INFO: Waiting up to 5m0s for pod nginx-controller-h8inh status to be running
14:55:55 Feb 29 14:50:48.007: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4.211053ms elapsed)
14:55:55 Feb 29 14:50:50.012: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2.008983228s elapsed)
14:55:55 Feb 29 14:50:52.016: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4.013129998s elapsed)
14:55:55 Feb 29 14:50:54.021: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (6.017367481s elapsed)
14:55:55 Feb 29 14:50:56.024: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (8.02107039s elapsed)
14:55:55 Feb 29 14:50:58.029: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (10.02541349s elapsed)
14:55:55 Feb 29 14:51:00.033: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (12.029552985s elapsed)
14:55:55 Feb 29 14:51:02.039: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (14.036171006s elapsed)
14:55:55 Feb 29 14:51:04.043: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (16.040100224s elapsed)
14:55:55 Feb 29 14:51:06.047: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (18.044159343s elapsed)
14:55:55 Feb 29 14:51:08.052: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (20.048798763s elapsed)
14:55:55 Feb 29 14:51:10.058: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (22.054902853s elapsed)
14:55:55 Feb 29 14:51:12.062: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (24.058828936s elapsed)
14:55:55 Feb 29 14:51:14.066: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (26.062342652s elapsed)
14:55:55 Feb 29 14:51:16.070: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (28.066274694s elapsed)
14:55:55 Feb 29 14:51:18.073: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (30.070144825s elapsed)
14:55:55 Feb 29 14:51:20.086: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (32.082324045s elapsed)
14:55:55 Feb 29 14:51:22.090: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (34.086668007s elapsed)
14:55:55 Feb 29 14:51:24.094: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (36.090685907s elapsed)
14:55:55 Feb 29 14:51:26.098: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (38.094607125s elapsed)
14:55:55 Feb 29 14:51:28.102: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (40.09850594s elapsed)
14:55:55 Feb 29 14:51:30.105: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (42.102053212s elapsed)
14:55:55 Feb 29 14:51:32.109: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (44.105763653s elapsed)
14:55:55 Feb 29 14:51:34.113: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (46.109633313s elapsed)
14:55:55 Feb 29 14:51:36.117: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (48.113255528s elapsed)
14:55:55 Feb 29 14:51:38.120: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (50.117019458s elapsed)
14:55:55 Feb 29 14:51:40.124: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (52.121017084s elapsed)
14:55:55 Feb 29 14:51:42.131: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (54.128017746s elapsed)
14:55:55 Feb 29 14:51:44.136: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (56.133153645s elapsed)
14:55:55 Feb 29 14:51:46.142: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (58.138703473s elapsed)
14:55:55 Feb 29 14:51:48.146: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m0.142951356s elapsed)
14:55:55 Feb 29 14:51:50.151: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m2.147803745s elapsed)
14:55:55 Feb 29 14:51:52.155: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m4.151527539s elapsed)
14:55:55 Feb 29 14:51:54.159: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m6.155260186s elapsed)
14:55:55 Feb 29 14:51:56.162: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m8.158894618s elapsed)
14:55:55 Feb 29 14:51:58.166: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m10.16279543s elapsed)
14:55:55 Feb 29 14:52:00.170: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m12.166664238s elapsed)
14:55:55 Feb 29 14:52:02.174: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m14.170477718s elapsed)
14:55:55 Feb 29 14:52:04.177: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m16.173845691s elapsed)
14:55:55 Feb 29 14:52:06.181: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m18.1777854s elapsed)
14:55:55 Feb 29 14:52:08.190: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m20.186586746s elapsed)
14:55:55 Feb 29 14:52:10.193: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m22.190208344s elapsed)
14:55:55 Feb 29 14:52:12.198: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m24.194306234s elapsed)
14:55:55 Feb 29 14:52:14.201: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m26.198094678s elapsed)
14:55:55 Feb 29 14:52:16.205: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m28.201951293s elapsed)
14:55:55 Feb 29 14:52:18.209: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m30.205665439s elapsed)
14:55:55 Feb 29 14:52:20.213: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m32.210082255s elapsed)
14:55:55 Feb 29 14:52:22.217: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m34.213701417s elapsed)
14:55:55 Feb 29 14:52:24.221: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m36.217660413s elapsed)
14:55:55 Feb 29 14:52:26.225: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m38.221455902s elapsed)
14:55:55 Feb 29 14:52:28.234: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m40.231134268s elapsed)
14:55:55 Feb 29 14:52:30.242: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m42.238287924s elapsed)
14:55:55 Feb 29 14:52:32.254: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m44.251086385s elapsed)
14:55:55 Feb 29 14:52:34.260: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m46.256640711s elapsed)
14:55:55 Feb 29 14:52:36.267: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m48.263785131s elapsed)
14:55:55 Feb 29 14:52:38.277: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m50.274053874s elapsed)
14:55:55 Feb 29 14:52:40.285: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m52.281741783s elapsed)
14:55:55 Feb 29 14:52:42.289: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m54.285413431s elapsed)
14:55:55 Feb 29 14:52:44.293: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m56.28989383s elapsed)
14:55:55 Feb 29 14:52:46.298: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (1m58.294327581s elapsed)
14:55:55 Feb 29 14:52:48.302: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m0.298886666s elapsed)
14:55:55 Feb 29 14:52:50.307: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m2.303273246s elapsed)
14:55:55 Feb 29 14:52:52.311: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m4.30776469s elapsed)
14:55:55 Feb 29 14:52:54.317: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m6.313927782s elapsed)
14:55:55 Feb 29 14:52:56.321: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m8.318133634s elapsed)
14:55:55 Feb 29 14:52:58.326: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m10.322492904s elapsed)
14:55:55 Feb 29 14:53:00.330: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m12.327198643s elapsed)
14:55:55 Feb 29 14:53:02.334: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m14.330775062s elapsed)
14:55:55 Feb 29 14:53:04.338: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m16.334508684s elapsed)
14:55:55 Feb 29 14:53:06.344: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m18.340683374s elapsed)
14:55:55 Feb 29 14:53:08.348: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m20.344518882s elapsed)
14:55:55 Feb 29 14:53:10.351: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m22.347831402s elapsed)
14:55:55 Feb 29 14:53:12.355: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m24.351984641s elapsed)
14:55:55 Feb 29 14:53:14.359: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m26.355424285s elapsed)
14:55:55 Feb 29 14:53:16.362: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m28.359059512s elapsed)
14:55:55 Feb 29 14:53:18.369: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m30.365380933s elapsed)
14:55:55 Feb 29 14:53:20.377: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m32.373576478s elapsed)
14:55:55 Feb 29 14:53:22.381: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m34.37780978s elapsed)
14:55:55 Feb 29 14:53:24.385: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m36.38192322s elapsed)
14:55:55 Feb 29 14:53:26.389: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m38.385781905s elapsed)
14:55:55 Feb 29 14:53:28.393: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m40.389700512s elapsed)
14:55:55 Feb 29 14:53:30.412: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m42.408735121s elapsed)
14:55:55 Feb 29 14:53:32.416: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m44.412434592s elapsed)
14:55:55 Feb 29 14:53:34.420: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m46.416506436s elapsed)
14:55:55 Feb 29 14:53:36.424: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m48.421147451s elapsed)
14:55:55 Feb 29 14:53:38.428: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m50.424531085s elapsed)
14:55:55 Feb 29 14:53:40.431: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m52.42809243s elapsed)
14:55:55 Feb 29 14:53:42.441: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m54.437362409s elapsed)
14:55:55 Feb 29 14:53:44.444: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m56.440968638s elapsed)
14:55:55 Feb 29 14:53:46.448: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (2m58.445160653s elapsed)
14:55:55 Feb 29 14:53:48.452: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m0.449025955s elapsed)
14:55:55 Feb 29 14:53:50.456: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m2.452866072s elapsed)
14:55:55 Feb 29 14:53:52.464: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m4.460597301s elapsed)
14:55:55 Feb 29 14:53:54.476: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m6.473234019s elapsed)
14:55:55 Feb 29 14:53:56.480: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m8.476934413s elapsed)
14:55:55 Feb 29 14:53:58.484: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m10.480663504s elapsed)
14:55:55 Feb 29 14:54:00.488: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m12.484571084s elapsed)
14:55:55 Feb 29 14:54:02.491: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m14.488150476s elapsed)
14:55:55 Feb 29 14:54:04.495: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m16.491613464s elapsed)
14:55:55 Feb 29 14:54:06.498: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m18.495106773s elapsed)
14:55:55 Feb 29 14:54:08.502: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m20.49868058s elapsed)
14:55:55 Feb 29 14:54:10.506: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m22.502594883s elapsed)
14:55:55 Feb 29 14:54:12.511: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m24.507411779s elapsed)
14:55:55 Feb 29 14:54:14.514: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m26.510994085s elapsed)
14:55:55 Feb 29 14:54:16.519: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m28.5153312s elapsed)
14:55:55 Feb 29 14:54:18.523: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m30.519443897s elapsed)
14:55:55 Feb 29 14:54:20.527: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m32.523553138s elapsed)
14:55:55 Feb 29 14:54:22.535: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m34.531990009s elapsed)
14:55:55 Feb 29 14:54:24.539: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m36.535571536s elapsed)
14:55:55 Feb 29 14:54:26.543: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m38.539444702s elapsed)
14:55:55 Feb 29 14:54:28.546: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m40.542956695s elapsed)
14:55:55 Feb 29 14:54:30.552: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m42.549208619s elapsed)
14:55:55 Feb 29 14:54:32.558: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m44.555099434s elapsed)
14:55:55 Feb 29 14:54:34.563: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m46.559359149s elapsed)
14:55:55 Feb 29 14:54:36.567: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m48.563407504s elapsed)
14:55:55 Feb 29 14:54:38.570: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m50.567043354s elapsed)
14:55:55 Feb 29 14:54:40.574: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m52.570281177s elapsed)
14:55:55 Feb 29 14:54:42.577: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m54.573874763s elapsed)
14:55:55 Feb 29 14:54:44.581: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m56.577553334s elapsed)
14:55:55 Feb 29 14:54:46.586: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (3m58.583147738s elapsed)
14:55:55 Feb 29 14:54:48.590: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m0.586910288s elapsed)
14:55:55 Feb 29 14:54:50.594: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m2.590545622s elapsed)
14:55:55 Feb 29 14:54:52.597: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m4.594183648s elapsed)
14:55:55 Feb 29 14:54:54.601: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m6.597603705s elapsed)
14:55:55 Feb 29 14:54:56.610: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m8.606938741s elapsed)
14:55:55 Feb 29 14:54:58.614: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m10.610731259s elapsed)
14:55:55 Feb 29 14:55:00.617: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m12.613946615s elapsed)
14:55:55 Feb 29 14:55:02.621: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m14.617805149s elapsed)
14:55:55 Feb 29 14:55:04.629: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m16.625471868s elapsed)
14:55:55 Feb 29 14:55:06.633: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m18.629301384s elapsed)
14:55:55 Feb 29 14:55:08.641: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m20.637290114s elapsed)
14:55:55 Feb 29 14:55:10.644: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m22.64111968s elapsed)
14:55:55 Feb 29 14:55:12.648: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m24.645162778s elapsed)
14:55:55 Feb 29 14:55:14.652: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m26.648654287s elapsed)
14:55:55 Feb 29 14:55:16.655: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m28.652139493s elapsed)
14:55:55 Feb 29 14:55:18.660: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m30.656380224s elapsed)
14:55:55 Feb 29 14:55:20.663: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m32.660059881s elapsed)
14:55:55 Feb 29 14:55:22.667: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m34.663450733s elapsed)
14:55:55 Feb 29 14:55:24.671: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m36.667387975s elapsed)
14:55:55 Feb 29 14:55:26.674: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m38.671224595s elapsed)
14:55:55 Feb 29 14:55:28.678: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m40.675048194s elapsed)
14:55:55 Feb 29 14:55:30.682: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m42.678872766s elapsed)
14:55:55 Feb 29 14:55:32.686: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m44.682420629s elapsed)
14:55:55 Feb 29 14:55:34.689: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m46.686230956s elapsed)
14:55:55 Feb 29 14:55:36.693: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m48.689848937s elapsed)
14:55:55 Feb 29 14:55:38.697: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m50.693537391s elapsed)
14:55:55 Feb 29 14:55:40.706: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m52.702417132s elapsed)
14:55:55 Feb 29 14:55:42.709: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m54.706169746s elapsed)
14:55:55 Feb 29 14:55:44.714: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m56.710855695s elapsed)
14:55:55 Feb 29 14:55:46.718: INFO: Waiting for pod nginx-controller-h8inh in namespace 'e2e-tests-deployment-ezytf' status to be 'running'(found phase: "Pending", readiness: false) (4m58.714629186s elapsed)
14:55:55 Feb 29 14:55:48.718: INFO: Waiting up to 5m0s for pod nginx-controller-uub6h status to be running
14:55:55 Feb 29 14:55:48.721: INFO: Found pod 'nginx-controller-uub6h' on node 'gke-jenkins-e2e-f1ec7fe0-node-l2sf'
14:55:55 Feb 29 14:55:48.721: INFO: error in waiting for pods to come up: failed to wait for pods running: [gave up waiting for pod 'nginx-controller-cceta' to be 'running' after 5m0s gave up waiting for pod 'nginx-controller-h8inh' to be 'running' after 5m0s]
14:55:55 [AfterEach] Deployment
14:55:55 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework.go:84
14:55:55 STEP: Collecting events from namespace "e2e-tests-deployment-ezytf".
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:45:42 -0800 PST - event for nginx-controller: {replicaset-controller } SuccessfulCreate: Created pod: nginx-controller-h8inh
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:45:42 -0800 PST - event for nginx-controller: {replicaset-controller } SuccessfulCreate: Created pod: nginx-controller-cceta
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:45:42 -0800 PST - event for nginx-controller: {replicaset-controller } SuccessfulCreate: Created pod: nginx-controller-uub6h
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:45:42 -0800 PST - event for nginx-controller-cceta: {default-scheduler } Scheduled: Successfully assigned nginx-controller-cceta to gke-jenkins-e2e-f1ec7fe0-node-fkww
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:45:42 -0800 PST - event for nginx-controller-h8inh: {default-scheduler } Scheduled: Successfully assigned nginx-controller-h8inh to gke-jenkins-e2e-f1ec7fe0-node-8mwd
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:45:42 -0800 PST - event for nginx-controller-uub6h: {default-scheduler } Scheduled: Successfully assigned nginx-controller-uub6h to gke-jenkins-e2e-f1ec7fe0-node-l2sf
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:26 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} Pulling: pulling image "nginx"
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:28 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ErrImagePull: "Error pulling image (latest) from docker.io/library/nginx, Server error: 500 trying to fetch remote history for 6e36f46089ed3c0326d2f56d6282af5eab6000caaa04e44f327c37f13d13c933"
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:28 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} Failed: Failed to pull image "nginx": Error pulling image (latest) from docker.io/library/nginx, Server error: 500 trying to fetch remote history for 6e36f46089ed3c0326d2f56d6282af5eab6000caaa04e44f327c37f13d13c933
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:29 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} Pulling: pulling image "nginx"
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:30 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ImagePullBackOff: "Back-off pulling image \"nginx\""
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:30 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} BackOff: Back-off pulling image "nginx"
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:31 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ErrImagePull: "Could not reach any registry endpoint"
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:31 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} Failed: Failed to pull image "nginx": Could not reach any registry endpoint
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:34 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} BackOff: Back-off pulling image "nginx"
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:34 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ImagePullBackOff: "Back-off pulling image \"nginx\""
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:37 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} Pulling: pulling image "nginx"
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:39 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} BackOff: Back-off pulling image "nginx"
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:39 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ImagePullBackOff: "Back-off pulling image \"nginx\""
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:39 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ErrImagePull: "Could not reach any registry endpoint"
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:46:39 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} Failed: Failed to pull image "nginx": Could not reach any registry endpoint
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:47:05 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} Failed: Failed to pull image "nginx": Error pulling image (latest) from docker.io/library/nginx, Server error: Status 500 while fetching image layer (d4e3870153bf700b1df356ca654c38ac1767c55c51cecd0a2336278e3256a43b)
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:47:05 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ErrImagePull: "Error pulling image (latest) from docker.io/library/nginx, Server error: Status 500 while fetching image layer (d4e3870153bf700b1df356ca654c38ac1767c55c51cecd0a2336278e3256a43b)"
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:47:32 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ErrImagePull: "Could not reach any registry endpoint"
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:47:32 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} Failed: Failed to pull image "nginx": Could not reach any registry endpoint
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:47:55 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} Failed: Failed to pull image "nginx": Error pulling image (latest) from docker.io/library/nginx, Server error: 500 trying to fetch remote history for 6e36f46089ed3c0326d2f56d6282af5eab6000caaa04e44f327c37f13d13c933
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:47:55 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ErrImagePull: "Error pulling image (latest) from docker.io/library/nginx, Server error: 500 trying to fetch remote history for 6e36f46089ed3c0326d2f56d6282af5eab6000caaa04e44f327c37f13d13c933"
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:48:03 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} Failed: Failed to pull image "nginx": Error pulling image (latest) from docker.io/library/nginx, Server error: 500 trying to fetch remote history for 6e36f46089ed3c0326d2f56d6282af5eab6000caaa04e44f327c37f13d13c933
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:48:03 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ErrImagePull: "Error pulling image (latest) from docker.io/library/nginx, Server error: 500 trying to fetch remote history for 6e36f46089ed3c0326d2f56d6282af5eab6000caaa04e44f327c37f13d13c933"
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:49:03 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} Failed: Failed to pull image "nginx": Error pulling image (latest) from docker.io/library/nginx, HTTP code 500
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:49:03 -0800 PST - event for nginx-controller-h8inh: {kubelet gke-jenkins-e2e-f1ec7fe0-node-8mwd} FailedSync: Error syncing pod, skipping: failed to "StartContainer" for "nginx" with ErrImagePull: "Error pulling image (latest) from docker.io/library/nginx, HTTP code 500"
14:55:55
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:49:28 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} Pulled: Successfully pulled image "nginx"
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:49:28 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} Created: Created container with docker id 2682a301ced5
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:49:28 -0800 PST - event for nginx-controller-uub6h: {kubelet gke-jenkins-e2e-f1ec7fe0-node-l2sf} Started: Started container with docker id 2682a301ced5
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:53:18 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} Created: Created container with docker id 1cd3cd65e637
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:53:18 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} Started: Started container with docker id 1cd3cd65e637
14:55:55 Feb 29 14:55:48.735: INFO: At 2016-02-29 14:53:18 -0800 PST - event for nginx-controller-cceta: {kubelet gke-jenkins-e2e-f1ec7fe0-node-fkww} Pulled: Successfully pulled image "nginx"
14:55:55 Feb 29 14:55:48.749: INFO: POD NODE PHASE GRACE CONDITIONS
14:55:55 Feb 29 14:55:48.749: INFO: nginx-controller-cceta gke-jenkins-e2e-f1ec7fe0-node-fkww Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:53:18 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: nginx-controller-h8inh gke-jenkins-e2e-f1ec7fe0-node-8mwd Pending [{Ready False 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:45:42 -0800 PST ContainersNotReady containers with unready status: [nginx]}]
14:55:55 Feb 29 14:55:48.749: INFO: nginx-controller-uub6h gke-jenkins-e2e-f1ec7fe0-node-l2sf Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:49:28 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: nginx-controller-10sxv gke-jenkins-e2e-f1ec7fe0-node-8mwd Pending [{Ready False 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:55:23 -0800 PST ContainersNotReady containers with unready status: [nginx]}]
14:55:55 Feb 29 14:55:48.749: INFO: nginx-controller-6tzlu gke-jenkins-e2e-f1ec7fe0-node-l2sf Pending [{Ready False 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:55:23 -0800 PST ContainersNotReady containers with unready status: [nginx]}]
14:55:55 Feb 29 14:55:48.749: INFO: nginx-controller-gq2aw gke-jenkins-e2e-f1ec7fe0-node-fkww Pending [{Ready False 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:55:23 -0800 PST ContainersNotReady containers with unready status: [nginx]}]
14:55:55 Feb 29 14:55:48.749: INFO: nginx-controller-ta510 gke-jenkins-e2e-f1ec7fe0-node-8mwd Pending [{Ready False 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:55:23 -0800 PST ContainersNotReady containers with unready status: [nginx]}]
14:55:55 Feb 29 14:55:48.749: INFO: mutability-test-0lbqb gke-jenkins-e2e-f1ec7fe0-node-fkww Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:49:22 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: mutability-test-0u77q gke-jenkins-e2e-f1ec7fe0-node-l2sf Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:49:27 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: fluentd-cloud-logging-gke-jenkins-e2e-f1ec7fe0-node-8mwd gke-jenkins-e2e-f1ec7fe0-node-8mwd Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:23 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: fluentd-cloud-logging-gke-jenkins-e2e-f1ec7fe0-node-fkww gke-jenkins-e2e-f1ec7fe0-node-fkww Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:07 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: fluentd-cloud-logging-gke-jenkins-e2e-f1ec7fe0-node-l2sf gke-jenkins-e2e-f1ec7fe0-node-l2sf Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:07 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: heapster-v14-762qq gke-jenkins-e2e-f1ec7fe0-node-l2sf Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:14 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: kube-dns-v10-87d3g gke-jenkins-e2e-f1ec7fe0-node-fkww Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:24 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: kube-proxy-gke-jenkins-e2e-f1ec7fe0-node-8mwd gke-jenkins-e2e-f1ec7fe0-node-8mwd Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:05 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: kube-proxy-gke-jenkins-e2e-f1ec7fe0-node-fkww gke-jenkins-e2e-f1ec7fe0-node-fkww Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:43:48 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: kube-proxy-gke-jenkins-e2e-f1ec7fe0-node-l2sf gke-jenkins-e2e-f1ec7fe0-node-l2sf Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:00 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: kubernetes-dashboard-v0.1.0-tm60w gke-jenkins-e2e-f1ec7fe0-node-fkww Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:08 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO: l7-lb-controller-v0.5.2-h3ffn gke-jenkins-e2e-f1ec7fe0-node-l2sf Running [{Ready True 0001-01-01 00:00:00 +0000 UTC 2016-02-29 14:44:23 -0800 PST }]
14:55:55 Feb 29 14:55:48.749: INFO:
14:55:55 Feb 29 14:55:48.756: INFO:
14:55:55 Logging node info for node gke-jenkins-e2e-f1ec7fe0-node-8mwd
14:55:55 Feb 29 14:55:48.760: INFO: Node Info: &{{ } {gke-jenkins-e2e-f1ec7fe0-node-8mwd /api/v1/nodes/gke-jenkins-e2e-f1ec7fe0-node-8mwd d45ee606-df35-11e5-869e-42010af0000f 5164 0 2016-02-29 14:43:19 -0800 PST <nil> <nil> map[beta.kubernetes.io/instance-type:n1-standard-2 failure-domain.beta.kubernetes.io/region:us-central1 failure-domain.beta.kubernetes.io/zone:us-central1-f kubernetes.io/hostname:gke-jenkins-e2e-f1ec7fe0-node-8mwd] map[]} {10.180.2.0/24 7238163060162554869 gce://k8s-jkns-e2e-gke-ci/us-central1-f/gke-jenkins-e2e-f1ec7fe0-node-8mwd false} {map[cpu:{2.000 DecimalSI} memory:{7864139776.000 BinarySI} pods:{110.000 DecimalSI}] map[pods:{110.000 DecimalSI} cpu:{2.000 DecimalSI} memory:{7864139776.000 BinarySI}] [{OutOfDisk False 2016-02-29 14:55:47 -0800 PST 2016-02-29 14:43:19 -0800 PST KubeletHasSufficientDisk kubelet has sufficient disk space available} {Ready True 2016-02-29 14:55:47 -0800 PST 2016-02-29 14:44:00 -0800 PST KubeletReady kubelet is posting ready status}] [{InternalIP 10.240.0.2} {ExternalIP 173.255.112.50}] {{10250}} { DCE991B5-66DD-4ECF-7F5F-42F7B9B9B0C7 3b2b1188-dfdf-4a06-9073-e290f77082fd 3.16.0-4-amd64 Debian GNU/Linux 7 (wheezy) docker://1.9.1 v1.2.0-alpha.8.676+83d74dd9528e42 v1.2.0-alpha.8.676+83d74dd9528e42} [{[gcr.io/google_containers/kube-proxy:6952338e87c87e4bef2030a93d05b990] 165606485} {[redis:latest] 151331559} {[gcr.io/google_containers/nettest:1.8] 25178998} {[<none>:<none>] 142324527} {[nginx:latest] 134617408} {[<none>:<none>] 134617408} {[busybox:latest] 1113554} {[gcr.io/google_samples/gb-frontend:v4] 510252254} {[gcr.io/google_containers/nettest:1.7] 24051275} {[gcr.io/google_containers/fluentd-gcp:1.16] 406764477} {[gcr.io/google_containers/netexec:1.4] 7297019} {[gcr.io/google_containers/busybox:1.24] 1113554} {[gcr.io/google_containers/mounttest:0.6] 2084693} {[gcr.io/google_containers/hostexec:1.2] 13209617} {[gcr.io/google_containers/mounttest-user:0.3] 1718853} {[gcr.io/google_containers/mounttest:0.5] 1718853} {[gcr.io/google_containers/pause:2.0] 350164} {[gcr.io/google_containers/portforwardtester:1.0] 2296329} {[gcr.io/google_samples/gb-redisslave:v1] 109508753} {[gcr.io/google_containers/mounttest:0.2] 1752375} {[gcr.io/google_containers/pause:0.8.0] 241656} {[gcr.io/google_containers/eptest:0.1] 2970692} {[b.gcr.io/k8s_authenticated_test/serve_hostname:1.1] 4522409} {[gcr.io/google_containers/serve_hostname:1.1] 4522409} {[gcr.io/google_containers/update-demo:nautilus] 4555533} {[gcr.io/google_containers/update-demo:kitten] 4549069} {[gcr.io/google_containers/nginx:1.7.9] 91664166} {[gcr.io/google_containers/test-webserver:e2e] 4534272}]}}
14:55:55 Feb 29 14:55:48.760: INFO:
14:55:55 Logging kubelet events for node gke-jenkins-e2e-f1ec7fe0-node-8mwd
14:55:55 Feb 29 14:55:48.768: INFO:
14:55:55 Logging pods the kubelet thinks is on node gke-jenkins-e2e-f1ec7fe0-node-8mwd
14:55:55 Feb 29 14:55:48.796: INFO: nginx-controller-h8inh started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.796: INFO: fluentd-cloud-logging-gke-jenkins-e2e-f1ec7fe0-node-8mwd started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.796: INFO: kube-proxy-gke-jenkins-e2e-f1ec7fe0-node-8mwd started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.796: INFO: nginx-controller-10sxv started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.796: INFO: nginx-controller-ta510 started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="info"} => 1 @[0]
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="inspect_image"} => 80 @[0]
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="list_containers"} => 92 @[0]
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="list_images"} => 18 @[0]
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="logs"} => 1 @[0]
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="pull_image"} => 35 @[0]
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="start_container"} => 1 @[0]
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="stop_container"} => 52 @[0]
14:55:55 Feb 29 14:55:48.863: INFO: ERROR kubelet_docker_errors{operation_type="version"} => 72 @[0]
14:55:55 Feb 29 14:55:48.863: INFO:
14:55:55 Latency metrics for node gke-jenkins-e2e-f1ec7fe0-node-8mwd
14:55:55 Feb 29 14:55:48.863: INFO: {Operation: Method:pod_start_latency_microseconds Quantile:0.99 Latency:4m3.668161s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:create Method:pod_worker_latency_microseconds Quantile:0.99 Latency:1m18.437219s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:SyncPod Method:container_manager_latency_microseconds Quantile:0.99 Latency:1m18.080256s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:sync Method:pod_worker_latency_microseconds Quantile:0.99 Latency:1m9.755388s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation: Method:pod_start_latency_microseconds Quantile:0.9 Latency:1m5.820405s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:create Method:pod_worker_latency_microseconds Quantile:0.9 Latency:1m2.677207s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:pull_image Method:docker_operations_latency_microseconds Quantile:0.99 Latency:56.918473s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:update Method:pod_worker_latency_microseconds Quantile:0.99 Latency:30.560446s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:stop_container Method:docker_operations_latency_microseconds Quantile:0.99 Latency:30.09008s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:SyncPod Method:container_manager_latency_microseconds Quantile:0.9 Latency:29.353428s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:sync Method:pod_worker_latency_microseconds Quantile:0.9 Latency:11.242491s}
14:55:55 Feb 29 14:55:48.863: INFO: {Operation:pull_image Method:docker_operations_latency_microseconds Quantile:0.9 Latency:10.805248s}
14:55:55 Feb 29 14:55:48.863: INFO:
14:55:55 Logging node info for node gke-jenkins-e2e-f1ec7fe0-node-fkww
14:55:55 Feb 29 14:55:48.868: INFO: Node Info: &{{ } {gke-jenkins-e2e-f1ec7fe0-node-fkww /api/v1/nodes/gke-jenkins-e2e-f1ec7fe0-node-fkww d359aed6-df35-11e5-869e-42010af0000f 5161 0 2016-02-29 14:43:17 -0800 PST <nil> <nil> map[beta.kubernetes.io/instance-type:n1-standard-2 failure-domain.beta.kubernetes.io/region:us-central1 failure-domain.beta.kubernetes.io/zone:us-central1-f kubernetes.io/hostname:gke-jenkins-e2e-f1ec7fe0-node-fkww] map[]} {10.180.0.0/24 11412112845876479197 gce://k8s-jkns-e2e-gke-ci/us-central1-f/gke-jenkins-e2e-f1ec7fe0-node-fkww false} {map[cpu:{2.000 DecimalSI} memory:{7864139776.000 BinarySI} pods:{110.000 DecimalSI}] map[cpu:{2.000 DecimalSI} memory:{7864139776.000 BinarySI} pods:{110.000 DecimalSI}] [{OutOfDisk False 2016-02-29 14:55:45 -0800 PST 2016-02-29 14:43:17 -0800 PST KubeletHasSufficientDisk kubelet has sufficient disk space available} {Ready True 2016-02-29 14:55:45 -0800 PST 2016-02-29 14:43:47 -0800 PST KubeletReady kubelet is posting ready status}] [{InternalIP 10.240.0.3} {ExternalIP 173.255.119.226}] {{10250}} { 4BAFA5DA-437E-C741-F5E4-5726E8EB84B7 55f15f24-cb47-418b-8e1a-286c4625a4a8 3.16.0-4-amd64 Debian GNU/Linux 7 (wheezy) docker://1.9.1 v1.2.0-alpha.8.676+83d74dd9528e42 v1.2.0-alpha.8.676+83d74dd9528e42} [{[gcr.io/google_containers/kube-proxy:6952338e87c87e4bef2030a93d05b990] 165606485} {[redis:latest] 151331559} {[gcr.io/google_containers/nettest:1.8] 25178998} {[nginx:latest] 134617408} {[<none>:<none>] 134617408} {[<none>:<none>] 1113554} {[busybox:latest] 1113554} {[gcr.io/google_samples/gb-frontend:v4] 510252254} {[gcr.io/google_containers/nettest:1.7] 24051275} {[gcr.io/google_containers/kubernetes-dashboard-amd64:v0.1.0] 35409795} {[gcr.io/google_containers/fluentd-gcp:1.16] 406764477} {[gcr.io/google_containers/netexec:1.4] 7297019} {[gcr.io/google_containers/etcd-amd64:2.2.1] 28192476} {[gcr.io/google_containers/fakegitserver:0.1] 5007469} {[gcr.io/google_containers/busybox:1.24] 1113554} {[gcr.io/google_containers/kube2sky:1.12] 24482187} {[gcr.io/google_containers/mounttest:0.6] 2084693} {[gcr.io/google_containers/mounttest-user:0.3] 1718853} {[gcr.io/google_containers/mounttest:0.5] 1718853} {[gcr.io/google_containers/skydns:2015-10-13-8c72f8c] 40551394} {[gcr.io/google_containers/pause:2.0] 350164} {[gcr.io/google_containers/exechealthz:1.0] 7099444} {[gcr.io/google_containers/jessie-dnsutils:e2e] 190148402} {[gcr.io/google_containers/mounttest:0.2] 1752375} {[gcr.io/google_containers/dnsutils:e2e] 141895666} {[gcr.io/google_containers/pause:0.8.0] 241656} {[gcr.io/google_containers/eptest:0.1] 2970692} {[gcr.io/google_containers/serve_hostname:1.1] 4522409} {[gcr.io/google_containers/update-demo:nautilus] 4555533} {[gcr.io/google_containers/nginx:1.7.9] 91664166} {[gcr.io/google_containers/test-webserver:e2e] 4534272} {[gcr.io/google_containers/liveness:e2e] 4387474}]}}
14:55:55 Feb 29 14:55:48.868: INFO:
14:55:55 Logging kubelet events for node gke-jenkins-e2e-f1ec7fe0-node-fkww
14:55:55 Feb 29 14:55:48.874: INFO:
14:55:55 Logging pods the kubelet thinks is on node gke-jenkins-e2e-f1ec7fe0-node-fkww
14:55:55 Feb 29 14:55:48.886: INFO: kube-proxy-gke-jenkins-e2e-f1ec7fe0-node-fkww started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.886: INFO: nginx-controller-gq2aw started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.886: INFO: nginx-controller-cceta started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.886: INFO: mutability-test-0lbqb started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.886: INFO: kube-dns-v10-87d3g started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.886: INFO: kubernetes-dashboard-v0.1.0-tm60w started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:48.886: INFO: fluentd-cloud-logging-gke-jenkins-e2e-f1ec7fe0-node-fkww started at <nil> (0 container statuses recorded)
14:55:55 W0229 14:55:48.941264 14311 request.go:627] Throttling request took 53.998802ms, request: https://104.197.114.165/api/v1/proxy/nodes/gke-jenkins-e2e-f1ec7fe0-node-fkww:10250/metrics
14:55:55 Feb 29 14:55:49.073: INFO: ERROR kubelet_docker_errors{operation_type="info"} => 1 @[0]
14:55:55 Feb 29 14:55:49.073: INFO: ERROR kubelet_docker_errors{operation_type="inspect_image"} => 64 @[0]
14:55:55 Feb 29 14:55:49.073: INFO: ERROR kubelet_docker_errors{operation_type="list_containers"} => 63 @[0]
14:55:55 Feb 29 14:55:49.073: INFO: ERROR kubelet_docker_errors{operation_type="list_images"} => 14 @[0]
14:55:55 Feb 29 14:55:49.073: INFO: ERROR kubelet_docker_errors{operation_type="pull_image"} => 21 @[0]
14:55:55 Feb 29 14:55:49.073: INFO: ERROR kubelet_docker_errors{operation_type="start_container"} => 1 @[0]
14:55:55 Feb 29 14:55:49.073: INFO: ERROR kubelet_docker_errors{operation_type="stop_container"} => 31 @[0]
14:55:55 Feb 29 14:55:49.073: INFO: ERROR kubelet_docker_errors{operation_type="version"} => 52 @[0]
14:55:55 Feb 29 14:55:49.074: INFO:
14:55:55 Latency metrics for node gke-jenkins-e2e-f1ec7fe0-node-fkww
14:55:55 Feb 29 14:55:49.074: INFO: {Operation: Method:pod_start_latency_microseconds Quantile:0.99 Latency:7m36.012229s}
14:55:55 Feb 29 14:55:49.074: INFO: {Operation:sync Method:pod_worker_latency_microseconds Quantile:0.99 Latency:36.084205s}
14:55:55 Feb 29 14:55:49.074: INFO: {Operation: Method:pod_start_latency_microseconds Quantile:0.9 Latency:36.006157s}
14:55:55 Feb 29 14:55:49.074: INFO: {Operation:stop_container Method:docker_operations_latency_microseconds Quantile:0.99 Latency:30.178837s}
14:55:55 Feb 29 14:55:49.074: INFO: {Operation:SyncPod Method:container_manager_latency_microseconds Quantile:0.99 Latency:26.024915s}
14:55:55 Feb 29 14:55:49.074: INFO: {Operation:pull_image Method:docker_operations_latency_microseconds Quantile:0.99 Latency:21.33357s}
14:55:55 Feb 29 14:55:49.074: INFO:
14:55:55 Logging node info for node gke-jenkins-e2e-f1ec7fe0-node-l2sf
14:55:55 W0229 14:55:49.141242 14311 request.go:627] Throttling request took 67.09998ms, request: https://104.197.114.165/api/v1/nodes/gke-jenkins-e2e-f1ec7fe0-node-l2sf
14:55:55 Feb 29 14:55:49.145: INFO: Node Info: &{{ } {gke-jenkins-e2e-f1ec7fe0-node-l2sf /api/v1/nodes/gke-jenkins-e2e-f1ec7fe0-node-l2sf d2bce91c-df35-11e5-869e-42010af0000f 5160 0 2016-02-29 14:43:16 -0800 PST <nil> <nil> map[beta.kubernetes.io/instance-type:n1-standard-2 failure-domain.beta.kubernetes.io/region:us-central1 failure-domain.beta.kubernetes.io/zone:us-central1-f kubernetes.io/hostname:gke-jenkins-e2e-f1ec7fe0-node-l2sf] map[]} {10.180.1.0/24 15180633427855587389 gce://k8s-jkns-e2e-gke-ci/us-central1-f/gke-jenkins-e2e-f1ec7fe0-node-l2sf false} {map[pods:{110.000 DecimalSI} cpu:{2.000 DecimalSI} memory:{7864139776.000 BinarySI}] map[cpu:{2.000 DecimalSI} memory:{7864139776.000 BinarySI} pods:{110.000 DecimalSI}] [{OutOfDisk False 2016-02-29 14:55:43 -0800 PST 2016-02-29 14:43:16 -0800 PST KubeletHasSufficientDisk kubelet has sufficient disk space available} {Ready True 2016-02-29 14:55:43 -0800 PST 2016-02-29 14:43:46 -0800 PST KubeletReady kubelet is posting ready status}] [{InternalIP 10.240.0.4} {ExternalIP 23.236.59.190}] {{10250}} { 19DD0CB4-8383-602F-3D1C-64B17EE856B9 4e5523c0-c6d5-42d4-8e11-efb76b337cc3 3.16.0-4-amd64 Debian GNU/Linux 7 (wheezy) docker://1.9.1 v1.2.0-alpha.8.676+83d74dd9528e42 v1.2.0-alpha.8.676+83d74dd9528e42} [{[gcr.io/google_containers/kube-proxy:6952338e87c87e4bef2030a93d05b990] 165606485} {[gcr.io/google_containers/heapster:v0.20.0-alpha8] 86984348} {[redis:latest] 151331559} {[gcr.io/google_containers/nettest:1.8] 25178998} {[nginx:latest] 134617408} {[gcr.io/google_samples/gb-frontend:v4] 510252254} {[gcr.io/google_containers/fluentd-gcp:1.16] 406764477} {[gcr.io/google_containers/netexec:1.4] 7297019} {[gcr.io/google_containers/glbc:0.5.2] 226351998} {[gcr.io/google_containers/busybox:1.24] 1113554} {[gcr.io/google_containers/mounttest:0.6] 2084693} {[gcr.io/google_containers/hostexec:1.2] 13209617} {[gcr.io/google_containers/mounttest-user:0.3] 1718853} {[gcr.io/google_containers/defaultbackend:1.0] 7513643} {[gcr.io/google_containers/mounttest:0.5] 1718853} {[gcr.io/google_containers/pause:2.0] 350164} {[gcr.io/google_containers/porter:cd5cb5791ebaa8641955f0e8c2a9bed669b1eaab] 5010921} {[gcr.io/google_containers/goproxy:0.1] 5489816} {[gcr.io/google_samples/gb-redisslave:v1] 109508753} {[gcr.io/google_containers/mounttest:0.2] 1752375} {[gcr.io/google_containers/pause:0.8.0] 241656} {[b.gcr.io/k8s_authenticated_test/serve_hostname:1.1] 4522409} {[gcr.io/google_containers/serve_hostname:1.1] 4522409} {[gcr.io/google_containers/update-demo:kitten] 4549069} {[gcr.io/google_containers/update-demo:nautilus] 4555533} {[gcr.io/google_containers/nginx:1.7.9] 91664166} {[gcr.io/google_containers/test-webserver:e2e] 4534272}]}}
14:55:55 Feb 29 14:55:49.146: INFO:
14:55:55 Logging kubelet events for node gke-jenkins-e2e-f1ec7fe0-node-l2sf
14:55:55 W0229 14:55:49.341238 14311 request.go:627] Throttling request took 194.963821ms, request: https://104.197.114.165/api/v1/namespaces/kube-system/events?fieldSelector=involvedObject.kind%3DNode%2CinvolvedObject.name%3Dgke-jenkins-e2e-f1ec7fe0-node-l2sf%2CinvolvedObject.namespace%3D%2Csource%3Dkubelet
14:55:55 Feb 29 14:55:49.346: INFO:
14:55:55 Logging pods the kubelet thinks is on node gke-jenkins-e2e-f1ec7fe0-node-l2sf
14:55:55 W0229 14:55:49.541273 14311 request.go:627] Throttling request took 193.107727ms, request: https://104.197.114.165/api/v1/proxy/nodes/gke-jenkins-e2e-f1ec7fe0-node-l2sf:10250/runningpods
14:55:55 Feb 29 14:55:49.564: INFO: fluentd-cloud-logging-gke-jenkins-e2e-f1ec7fe0-node-l2sf started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:49.564: INFO: kube-proxy-gke-jenkins-e2e-f1ec7fe0-node-l2sf started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:49.564: INFO: nginx-controller-6tzlu started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:49.564: INFO: nginx-controller-uub6h started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:49.564: INFO: mutability-test-0u77q started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:49.564: INFO: l7-lb-controller-v0.5.2-h3ffn started at <nil> (0 container statuses recorded)
14:55:55 Feb 29 14:55:49.564: INFO: heapster-v14-762qq started at <nil> (0 container statuses recorded)
14:55:55 W0229 14:55:49.741215 14311 request.go:627] Throttling request took 175.733913ms, request: https://104.197.114.165/api/v1/proxy/nodes/gke-jenkins-e2e-f1ec7fe0-node-l2sf:10250/metrics
14:55:55 Feb 29 14:55:49.876: INFO: ERROR kubelet_docker_errors{operation_type="info"} => 1 @[0]
14:55:55 Feb 29 14:55:49.876: INFO: ERROR kubelet_docker_errors{operation_type="inspect_image"} => 48 @[0]
14:55:55 Feb 29 14:55:49.876: INFO: ERROR kubelet_docker_errors{operation_type="list_containers"} => 65 @[0]
14:55:55 Feb 29 14:55:49.876: INFO: ERROR kubelet_docker_errors{operation_type="list_images"} => 15 @[0]
14:55:55 Feb 29 14:55:49.876: INFO: ERROR kubelet_docker_errors{operation_type="pull_image"} => 26 @[0]
14:55:55 Feb 29 14:55:49.876: INFO: ERROR kubelet_docker_errors{operation_type="stop_container"} => 38 @[0]
14:55:55 Feb 29 14:55:49.876: INFO: ERROR kubelet_docker_errors{operation_type="version"} => 54 @[0]
14:55:55 Feb 29 14:55:49.876: INFO:
14:55:55 Latency metrics for node gke-jenkins-e2e-f1ec7fe0-node-l2sf
14:55:55 Feb 29 14:55:49.876: INFO: {Operation: Method:pod_start_latency_microseconds Quantile:0.99 Latency:3m46.644521s}
14:55:55 Feb 29 14:55:49.876: INFO: {Operation:create Method:pod_worker_latency_microseconds Quantile:0.99 Latency:55.520147s}
14:55:55 Feb 29 14:55:49.876: INFO: {Operation:sync Method:pod_worker_latency_microseconds Quantile:0.99 Latency:53.258439s}
14:55:55 Feb 29 14:55:49.876: INFO: {Operation:create Method:pod_worker_latency_microseconds Quantile:0.9 Latency:46.526966s}
14:55:55 Feb 29 14:55:49.876: INFO: {Operation:SyncPod Method:container_manager_latency_microseconds Quantile:0.99 Latency:38.112145s}
14:55:55 Feb 29 14:55:49.876: INFO: {Operation:update Method:pod_worker_latency_microseconds Quantile:0.99 Latency:30.421105s}
14:55:55 Feb 29 14:55:49.876: INFO: {Operation:stop_container Method:docker_operations_latency_microseconds Quantile:0.99 Latency:30.097734s}
14:55:55 Feb 29 14:55:49.876: INFO: {Operation: Method:pod_start_latency_microseconds Quantile:0.9 Latency:27.688855s}
14:55:55 Feb 29 14:55:49.876: INFO: {Operation:pull_image Method:docker_operations_latency_microseconds Quantile:0.99 Latency:20.27941s}
14:55:55 Feb 29 14:55:49.876: INFO: Waiting up to 1m0s for all nodes to be ready
14:55:55 W0229 14:55:49.941214 14311 request.go:627] Throttling request took 64.695123ms, request: https://104.197.114.165/api/v1/nodes
14:55:55 STEP: Destroying namespace "e2e-tests-deployment-ezytf" for this suite.
14:55:55 W0229 14:55:50.141262 14311 request.go:627] Throttling request took 194.312955ms, request: https://104.197.114.165/api/v1/namespaces/e2e-tests-deployment-ezytf
14:55:55 W0229 14:55:50.341242 14311 request.go:627] Throttling request took 194.516203ms, request: https://104.197.114.165/api/v1/namespaces/e2e-tests-deployment-ezytf
14:55:55
14:55:55
14:55:55 • Failure [615.210 seconds]
14:55:55 Deployment
14:55:55 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/deployment.go:72
14:55:55 deployment should label adopted RSs and pods [It]
14:55:55 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/deployment.go:71
14:55:55
14:55:55 Expected error:
14:55:55 <*errors.errorString | 0xc2084c0e60>: {
14:55:55 s: "failed to wait for pods running: [gave up waiting for pod 'nginx-controller-cceta' to be 'running' after 5m0s gave up waiting for pod 'nginx-controller-h8inh' to be 'running' after 5m0s]",
14:55:55 }
14:55:55 failed to wait for pods running: [gave up waiting for pod 'nginx-controller-cceta' to be 'running' after 5m0s gave up waiting for pod 'nginx-controller-h8inh' to be 'running' after 5m0s]
14:55:55 not to have occurred
14:55:55
14:55:55 /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/deployment.go:837
14:55:55 ------------------------------
</code></pre></div>
<p dir="auto">Assigning to <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/bgrant0607/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/bgrant0607">@bgrant0607</a> for triage.</p> | <p dir="auto">e2e-gce/8241 failed a kubectl test which timed out getting the redis pod into running:<br>
scheduler assigned redis-master-k6hgb to node .<em>-dhly (.</em>-128.243)<br>
logs from kubelet on .*-dhly</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="I0827 20:12:52.079928 4057 server.go:705] GET /stats/e2e-tests-kubectl-8myat/redis-master-k6hgb/feaaa022-4cf7-11e5-87c2-42010af0a6de/redis-master: (130.42µs) 404 [[Go 1.1 package http] 10.245.1.8:44772]
I0827 20:12:52.178302 4057 server.go:638] Event(api.ObjectReference{Kind:"Pod", Namespace:"e2e-tests-kubectl-8myat", Name:"redis-master-k6hgb", UID:"feaaa022-4cf7-11e5-87c2-42010af0a6de", APIVersion:"v1", ResourceVersion:"11824", FieldPath:"implicitly required container POD"}): reason: 'Started' Started with docker id e1387779aa97
I0827 20:12:52.182865 4057 manager.go:797] Added container: "/e1387779aa9731ad436d359162493b8712f95994eefa6bbffd904c3dadef8665" (aliases: [k8s_POD.84e5fe25_redis-master-k6hgb_e2e-tests-kubectl-8myat_feaaa022-4cf7-11e5-87c2-42010af0a6de_a08de96f e1387779aa9731ad436d359162493b8712f95994eefa6bbffd904c3dadef8665], namespace: "docker")
I0827 20:12:52.189654 4057 manager.go:1723] Creating container &{Name:redis-master Image:redis Command:[] Args:[] WorkingDir: Ports:[{Name:redis-server HostPort:0 ContainerPort:6379 Protocol:TCP HostIP:}] Env:[] Resources:{Limits:map[] Requests:map[]} VolumeMounts:[{Name:default-token-twa2p ReadOnly:true MountPath:/var/run/secrets/kubernetes.io/serviceaccount}] LivenessProbe:<nil> ReadinessProbe:<nil> Lifecycle:<nil> TerminationMessagePath:/dev/termination-log ImagePullPolicy:IfNotPresent SecurityContext:<nil> Stdin:false TTY:false} in pod redis-master-k6hgb_e2e-tests-kubectl-8myat
I0827 20:12:52.191004 4057 server.go:638] Event(api.ObjectReference{Kind:"Pod", Namespace:"e2e-tests-kubectl-8myat", Name:"redis-master-k6hgb", UID:"feaaa022-4cf7-11e5-87c2-42010af0a6de", APIVersion:"v1", ResourceVersion:"11824", FieldPath:"spec.containers{redis-master}"}): reason: 'pulling' Pulling image "redis"
...
I0827 20:14:56.286208 4057 server.go:705] GET /stats/e2e-tests-kubectl-8myat/redis-master-k6hgb/feaaa022-4cf7-11e5-87c2-42010af0a6de/redis-master: (9.484031ms) 404 [[Go 1.1 package http] 10.245.1.8:45044]
I0827 20:15:06.594298 4057 server.go:705] GET /stats/e2e-tests-kubectl-8myat/redis-master-k6hgb/feaaa022-4cf7-11e5-87c2-42010af0a6de/redis-master: (34.694138ms) 404 [[Go 1.1 package http] 10.245.1.8:45177]
I0827 20:15:16.972562 4057 server.go:705] GET /stats/e2e-tests-kubectl-8myat/redis-master-k6hgb/feaaa022-4cf7-11e5-87c2-42010af0a6de/redis-master: (141.647µs) 404 [[Go 1.1 package http] 10.245.1.8:45198]
I0827 20:15:27.314306 4057 server.go:705] GET /stats/e2e-tests-kubectl-8myat/redis-master-k6hgb/feaaa022-4cf7-11e5-87c2-42010af0a6de/redis-master: (7.672739ms) 404 [[Go 1.1 package http] 10.245.1.8:45083]
I0827 20:15:37.573487 4057 server.go:705] GET /stats/e2e-tests-kubectl-8myat/redis-master-k6hgb/feaaa022-4cf7-11e5-87c2-42010af0a6de/redis-master: (1.932497ms) 404 [[Go 1.1 package http] 10.245.1.8:45177]
I0827 20:15:48.369530 4057 server.go:705] GET /stats/e2e-tests-kubectl-8myat/redis-master-k6hgb/feaaa022-4cf7-11e5-87c2-42010af0a6de/redis-master: (5.908228ms) 404 [[Go 1.1 package http] 10.245.1.8:45256]
I0827 20:15:58.694011 4057 server.go:705] GET /stats/e2e-tests-kubectl-8myat/redis-master-k6hgb/feaaa022-4cf7-11e5-87c2-42010af0a6de/redis-master: (6.382311ms) 404 [[Go 1.1 package http] 10.245.1.8:45274]
...
I0827 20:17:52.554839 4057 kubelet.go:1777] SyncLoop (REMOVE): "redis-master-k6hgb_e2e-tests-kubectl-8myat"
I0827 20:17:52.554949 4057 kubelet.go:1615] Killing unwanted pod "redis-master-k6hgb"
"><pre class="notranslate"><code class="notranslate">I0827 20:12:52.079928 4057 server.go:705] GET /stats/e2e-tests-kubectl-8myat/redis-master-k6hgb/feaaa022-4cf7-11e5-87c2-42010af0a6de/redis-master: (130.42µs) 404 [[Go 1.1 package http] 10.245.1.8:44772]
I0827 20:12:52.178302 4057 server.go:638] Event(api.ObjectReference{Kind:"Pod", Namespace:"e2e-tests-kubectl-8myat", Name:"redis-master-k6hgb", UID:"feaaa022-4cf7-11e5-87c2-42010af0a6de", APIVersion:"v1", ResourceVersion:"11824", FieldPath:"implicitly required container POD"}): reason: 'Started' Started with docker id e1387779aa97
I0827 20:12:52.182865 4057 manager.go:797] Added container: "/e1387779aa9731ad436d359162493b8712f95994eefa6bbffd904c3dadef8665" (aliases: [k8s_POD.84e5fe25_redis-master-k6hgb_e2e-tests-kubectl-8myat_feaaa022-4cf7-11e5-87c2-42010af0a6de_a08de96f e1387779aa9731ad436d359162493b8712f95994eefa6bbffd904c3dadef8665], namespace: "docker")
I0827 20:12:52.189654 4057 manager.go:1723] Creating container &{Name:redis-master Image:redis Command:[] Args:[] WorkingDir: Ports:[{Name:redis-server HostPort:0 ContainerPort:6379 Protocol:TCP HostIP:}] Env:[] Resources:{Limits:map[] Requests:map[]} VolumeMounts:[{Name:default-token-twa2p ReadOnly:true MountPath:/var/run/secrets/kubernetes.io/serviceaccount}] LivenessProbe:<nil> ReadinessProbe:<nil> Lifecycle:<nil> TerminationMessagePath:/dev/termination-log ImagePullPolicy:IfNotPresent SecurityContext:<nil> Stdin:false TTY:false} in pod redis-master-k6hgb_e2e-tests-kubectl-8myat
I0827 20:12:52.191004 4057 server.go:638] Event(api.ObjectReference{Kind:"Pod", Namespace:"e2e-tests-kubectl-8myat", Name:"redis-master-k6hgb", UID:"feaaa022-4cf7-11e5-87c2-42010af0a6de", APIVersion:"v1", ResourceVersion:"11824", FieldPath:"spec.containers{redis-master}"}): reason: 'pulling' Pulling image "redis"
...
I0827 20:14:56.286208 4057 server.go:705] GET /stats/e2e-tests-kubectl-8myat/redis-master-k6hgb/feaaa022-4cf7-11e5-87c2-42010af0a6de/redis-master: (9.484031ms) 404 [[Go 1.1 package http] 10.245.1.8:45044]
I0827 20:15:06.594298 4057 server.go:705] GET /stats/e2e-tests-kubectl-8myat/redis-master-k6hgb/feaaa022-4cf7-11e5-87c2-42010af0a6de/redis-master: (34.694138ms) 404 [[Go 1.1 package http] 10.245.1.8:45177]
I0827 20:15:16.972562 4057 server.go:705] GET /stats/e2e-tests-kubectl-8myat/redis-master-k6hgb/feaaa022-4cf7-11e5-87c2-42010af0a6de/redis-master: (141.647µs) 404 [[Go 1.1 package http] 10.245.1.8:45198]
I0827 20:15:27.314306 4057 server.go:705] GET /stats/e2e-tests-kubectl-8myat/redis-master-k6hgb/feaaa022-4cf7-11e5-87c2-42010af0a6de/redis-master: (7.672739ms) 404 [[Go 1.1 package http] 10.245.1.8:45083]
I0827 20:15:37.573487 4057 server.go:705] GET /stats/e2e-tests-kubectl-8myat/redis-master-k6hgb/feaaa022-4cf7-11e5-87c2-42010af0a6de/redis-master: (1.932497ms) 404 [[Go 1.1 package http] 10.245.1.8:45177]
I0827 20:15:48.369530 4057 server.go:705] GET /stats/e2e-tests-kubectl-8myat/redis-master-k6hgb/feaaa022-4cf7-11e5-87c2-42010af0a6de/redis-master: (5.908228ms) 404 [[Go 1.1 package http] 10.245.1.8:45256]
I0827 20:15:58.694011 4057 server.go:705] GET /stats/e2e-tests-kubectl-8myat/redis-master-k6hgb/feaaa022-4cf7-11e5-87c2-42010af0a6de/redis-master: (6.382311ms) 404 [[Go 1.1 package http] 10.245.1.8:45274]
...
I0827 20:17:52.554839 4057 kubelet.go:1777] SyncLoop (REMOVE): "redis-master-k6hgb_e2e-tests-kubectl-8myat"
I0827 20:17:52.554949 4057 kubelet.go:1615] Killing unwanted pod "redis-master-k6hgb"
</code></pre></div>
<p dir="auto">The test uses the guestbook example which has the stock redis image.</p>
<p dir="auto"><a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/mwielgus/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/mwielgus">@mwielgus</a></p> | 1 |
<h2 dir="auto">🐛 Bug 🐛</h2>
<p dir="auto">When loading a saved sparse tensor an error occurs: <code class="notranslate">RuntimeError: size is inconsistent with indices: for dim 0, size is 3 but found index 140115406052656</code><br>
Clearly, that is impossible (can't even fit such a large tensor on my memory)</p>
<h2 dir="auto">To Reproduce</h2>
<p dir="auto">Steps to reproduce the behavior:</p>
<ol dir="auto">
<li>Install PyTorch and torch-sparse</li>
<li>Run the below code (which produces the above error for me)</li>
</ol>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="import torch
x=torch.zeros(3,3)
x[1][1]=1
x=x.to_sparse()
torch.save(x, './scratch/test4e')
y = torch.load('./scratch/test4e')"><pre class="notranslate"><code class="notranslate">import torch
x=torch.zeros(3,3)
x[1][1]=1
x=x.to_sparse()
torch.save(x, './scratch/test4e')
y = torch.load('./scratch/test4e')
</code></pre></div>
<h2 dir="auto">Expected behavior</h2>
<p dir="auto">Loads <code class="notranslate">y</code> normally ...</p>
<h2 dir="auto">Environment</h2>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="
PyTorch version: 1.4.0
Is debug build: No
CUDA used to build PyTorch: 10.1
OS: Ubuntu 18.04.4 LTS
GCC version: (Ubuntu 7.4.0-1ubuntu1~18.04.1) 7.4.0
CMake version: version 3.10.2
Python version: 3.8
Is CUDA available: Yes
CUDA runtime version: 10.1.243
GPU models and configuration: GPU 0: GeForce GTX 1070
Nvidia driver version: 440.59
cuDNN version: Could not collect
Versions of relevant libraries:
[pip3] numpy==1.18.1
[pip3] torch==1.4.0
[pip3] torchvision==0.5.0
[conda] torch 1.4.0 <pip>
[conda] torch-cluster 1.4.5 <pip>
[conda] torch-geometric 1.4.1 <pip>
[conda] torch-scatter 2.0.3 <pip>
[conda] torch-sparse 0.5.1 <pip>"><pre class="notranslate"><code class="notranslate">
PyTorch version: 1.4.0
Is debug build: No
CUDA used to build PyTorch: 10.1
OS: Ubuntu 18.04.4 LTS
GCC version: (Ubuntu 7.4.0-1ubuntu1~18.04.1) 7.4.0
CMake version: version 3.10.2
Python version: 3.8
Is CUDA available: Yes
CUDA runtime version: 10.1.243
GPU models and configuration: GPU 0: GeForce GTX 1070
Nvidia driver version: 440.59
cuDNN version: Could not collect
Versions of relevant libraries:
[pip3] numpy==1.18.1
[pip3] torch==1.4.0
[pip3] torchvision==0.5.0
[conda] torch 1.4.0 <pip>
[conda] torch-cluster 1.4.5 <pip>
[conda] torch-geometric 1.4.1 <pip>
[conda] torch-scatter 2.0.3 <pip>
[conda] torch-sparse 0.5.1 <pip>
</code></pre></div>
<p dir="auto">cc <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/ezyang/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/ezyang">@ezyang</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/gchanan/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/gchanan">@gchanan</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/zou3519/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/zou3519">@zou3519</a> <a class="user-mention notranslate" data-hovercard-type="user" data-hovercard-url="/users/vincentqb/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="https://github.com/vincentqb">@vincentqb</a></p> | <p dir="auto"><a href="https://github.com/pytorch/pytorch/blob/04c0971679374148aa4105e2e998de6478eca1eb/torch/nn/modules/rnn.py#L723-L727">Current implementation</a> of LSTMCell calls internal C function which makes it infeasible to write a layer normalization of LSTM cell by wrapping it with <code class="notranslate">nn.LayerNorm</code>.</p>
<p dir="auto">It would be great to add this feature.</p>
<p dir="auto">We could add trainable parameters inside <code class="notranslate">RNNCellBase</code> as follows</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="class RNNCellBase(nn.Module):
def __init__(self, input_size, hidden_size, bias, num_chunks):
super(RNNCellBase, self).__init__()
self.input_size = input_size
self.hidden_size = hidden_size
self.bias = bias
self.weight_ih = Parameter(torch.Tensor(num_chunks * hidden_size, input_size))
self.weight_hh = Parameter(torch.Tensor(num_chunks * hidden_size, hidden_size))
self.ln_ih = nn.LayerNorm(num_chunks * hidden_size)
self.ln_hh = nn.LayerNorm(num_chunks * hidden_size)
self.ln_c = nn.LayerNorm(hidden_size)"><pre class="notranslate"><span class="pl-k">class</span> <span class="pl-v">RNNCellBase</span>(<span class="pl-s1">nn</span>.<span class="pl-v">Module</span>):
<span class="pl-k">def</span> <span class="pl-en">__init__</span>(<span class="pl-s1">self</span>, <span class="pl-s1">input_size</span>, <span class="pl-s1">hidden_size</span>, <span class="pl-s1">bias</span>, <span class="pl-s1">num_chunks</span>):
<span class="pl-en">super</span>(<span class="pl-v">RNNCellBase</span>, <span class="pl-s1">self</span>).<span class="pl-en">__init__</span>()
<span class="pl-s1">self</span>.<span class="pl-s1">input_size</span> <span class="pl-c1">=</span> <span class="pl-s1">input_size</span>
<span class="pl-s1">self</span>.<span class="pl-s1">hidden_size</span> <span class="pl-c1">=</span> <span class="pl-s1">hidden_size</span>
<span class="pl-s1">self</span>.<span class="pl-s1">bias</span> <span class="pl-c1">=</span> <span class="pl-s1">bias</span>
<span class="pl-s1">self</span>.<span class="pl-s1">weight_ih</span> <span class="pl-c1">=</span> <span class="pl-v">Parameter</span>(<span class="pl-s1">torch</span>.<span class="pl-v">Tensor</span>(<span class="pl-s1">num_chunks</span> <span class="pl-c1">*</span> <span class="pl-s1">hidden_size</span>, <span class="pl-s1">input_size</span>))
<span class="pl-s1">self</span>.<span class="pl-s1">weight_hh</span> <span class="pl-c1">=</span> <span class="pl-v">Parameter</span>(<span class="pl-s1">torch</span>.<span class="pl-v">Tensor</span>(<span class="pl-s1">num_chunks</span> <span class="pl-c1">*</span> <span class="pl-s1">hidden_size</span>, <span class="pl-s1">hidden_size</span>))
<span class="pl-s1">self</span>.<span class="pl-s1">ln_ih</span> <span class="pl-c1">=</span> <span class="pl-s1">nn</span>.<span class="pl-v">LayerNorm</span>(<span class="pl-s1">num_chunks</span> <span class="pl-c1">*</span> <span class="pl-s1">hidden_size</span>)
<span class="pl-s1">self</span>.<span class="pl-s1">ln_hh</span> <span class="pl-c1">=</span> <span class="pl-s1">nn</span>.<span class="pl-v">LayerNorm</span>(<span class="pl-s1">num_chunks</span> <span class="pl-c1">*</span> <span class="pl-s1">hidden_size</span>)
<span class="pl-s1">self</span>.<span class="pl-s1">ln_c</span> <span class="pl-c1">=</span> <span class="pl-s1">nn</span>.<span class="pl-v">LayerNorm</span>(<span class="pl-s1">hidden_size</span>)</pre></div>
<p dir="auto">However, the it is impossible to use it for the computation</p>
<div class="highlight highlight-source-python notranslate position-relative overflow-auto" dir="auto" data-snippet-clipboard-copy-content="return _VF.lstm_cell(
input, hx,
self.weight_ih, self.weight_hh,
self.bias_ih, self.bias_hh,
)"><pre class="notranslate"><span class="pl-k">return</span> <span class="pl-s1">_VF</span>.<span class="pl-en">lstm_cell</span>(
<span class="pl-s1">input</span>, <span class="pl-s1">hx</span>,
<span class="pl-s1">self</span>.<span class="pl-s1">weight_ih</span>, <span class="pl-s1">self</span>.<span class="pl-s1">weight_hh</span>,
<span class="pl-s1">self</span>.<span class="pl-s1">bias_ih</span>, <span class="pl-s1">self</span>.<span class="pl-s1">bias_hh</span>,
)</pre></div>
<p dir="auto">as it calls C functions.</p>
<p dir="auto">And writing it purely in Python, the speed would be too slow.</p> | 0 |
<h5 dir="auto">System information (version)</h5>
<ul dir="auto">
<li>OpenCV => Git Master</li>
<li>Operating System / Platform => Raspbain Buster - 32 Bit</li>
<li>Compiler => gcc 8.3.0</li>
</ul>
<h5 dir="auto">Detailed description</h5>
<p dir="auto">Cmake is running fine, but make fails with internal compiler error: Segmentation fault</p>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="[ 1%] Linking C static library ../lib/libquirc.a
during GIMPLE pass: slp
/home/pi/opencv/3rdparty/protobuf/src/google/protobuf/arena.cc: In member function 'google::protobuf::internal::ArenaImpl::Block* google::protobuf::internal::ArenaImpl::GetBlockSlow(void*, google::protobuf::internal::ArenaImpl::Block*, size_t)':
/home/pi/opencv/3rdparty/protobuf/src/google/protobuf/arena.cc:230:19: internal compiler error: Segmentation fault
ArenaImpl::Block* ArenaImpl::GetBlockSlow(void* me, Block* my_full_block,
^~~~~~~~~
[ 1%] Built target quir"><pre class="notranslate"><code class="notranslate">[ 1%] Linking C static library ../lib/libquirc.a
during GIMPLE pass: slp
/home/pi/opencv/3rdparty/protobuf/src/google/protobuf/arena.cc: In member function 'google::protobuf::internal::ArenaImpl::Block* google::protobuf::internal::ArenaImpl::GetBlockSlow(void*, google::protobuf::internal::ArenaImpl::Block*, size_t)':
/home/pi/opencv/3rdparty/protobuf/src/google/protobuf/arena.cc:230:19: internal compiler error: Segmentation fault
ArenaImpl::Block* ArenaImpl::GetBlockSlow(void* me, Block* my_full_block,
^~~~~~~~~
[ 1%] Built target quir
</code></pre></div>
<h5 dir="auto">Steps to reproduce</h5>
<ul dir="auto">
<li><code class="notranslate">cd ~</code></li>
<li><code class="notranslate">git clone https://github.com/opencv/opencv</code></li>
<li><code class="notranslate">git clone https://github.com/opencv/opencv_contrib</code></li>
<li><code class="notranslate">cd opencv</code></li>
<li><code class="notranslate">mkdir build && cd build</code></li>
</ul>
<div class="snippet-clipboard-content notranslate position-relative overflow-auto" data-snippet-clipboard-copy-content="cmake -D CMAKE_BUILD_TYPE=RELEASE \
-D CMAKE_INSTALL_PREFIX=/usr/local \
-D OPENCV_EXTRA_MODULES_PATH=~/opencv_contrib/modules \
-D ENABLE_NEON=ON \
-D ENABLE_VFPV3=ON \
-D BUILD_TESTS=OFF \
-D OPENCV_ENABLE_NONFREE=ON \
-D INSTALL_PYTHON_EXAMPLES=OFF \
-D CMAKE_CXX_FLAGS=-latomic \
-D OPENCV_EXTRA_EXE_LINKER_FLAGS=-latomic \
-D BUILD_EXAMPLES=OFF .."><pre class="notranslate"><code class="notranslate">cmake -D CMAKE_BUILD_TYPE=RELEASE \
-D CMAKE_INSTALL_PREFIX=/usr/local \
-D OPENCV_EXTRA_MODULES_PATH=~/opencv_contrib/modules \
-D ENABLE_NEON=ON \
-D ENABLE_VFPV3=ON \
-D BUILD_TESTS=OFF \
-D OPENCV_ENABLE_NONFREE=ON \
-D INSTALL_PYTHON_EXAMPLES=OFF \
-D CMAKE_CXX_FLAGS=-latomic \
-D OPENCV_EXTRA_EXE_LINKER_FLAGS=-latomic \
-D BUILD_EXAMPLES=OFF ..
</code></pre></div>
<ul dir="auto">
<li><code class="notranslate">make -j4</code></li>
</ul> | <p dir="auto">Can anyone help me with this compilation error? Thanks</p>
<p dir="auto">Raspberry Pi 3 B+<br>
OpenCV 4.1.0</p>
<p dir="auto"><code class="notranslate">cmake -D CMAKE_BUILD_TYPE=RELEASE \ -D CMAKE_INSTALL_PREFIX=/usr/local \ -D OPENCV_EXTRA_MODULES_PATH=~/opencv_contrib/modules \ -D ENABLE_NEON=ON \ -D ENABLE_VFPV3=ON \ -D BUILD_TESTS=OFF \ -D INSTALL_PYTHON_EXAMPLES=OFF \ -D CMAKE_SIZEOF_VOID_P=4 \ -D BUILD_opencv_python3=TRUE \ -D PYTHON3_LIBRARY=/usr/lib/arm-linux-gnueabihf/libpython3.7m.so \ -D BUILD_EXAMPLES=OFF ..</code></p>
<p dir="auto">Error during compilation:</p>
<p dir="auto"><code class="notranslate">[ 98%] Built target gen_opencv_python_source [ 98%] Building CXX object modules/python3/CMakeFiles/opencv_python3.dir/__/src2/cv2.cpp.o during RTL pass: expand In file included from /home/pi/opencv/modules/python/src2/cv2.cpp:1722: /home/pi/opencv/build/modules/python_bindings_generator/pyopencv_generated_types.h: In function ‘PyObject* pyopencv_cv_detail_detail_BlocksGainCompensator_apply(PyObject*, PyObject*, PyObject*)’: /home/pi/opencv/build/modules/python_bindings_generator/pyopencv_generated_types.h:79645:17: internal compiler error: Segmentation fault const char* keywords[] = { "index", "corner", "image", "mask", NULL }; ^~~~~~~~ 0x76a7311f ??? ../sysdeps/unix/sysv/linux/arm/sigrestorer.S:64 0x76a5d717 __libc_start_main /build/glibc-FUvrFr/glibc-2.28/csu/libc-start.c:308 Please submit a full bug report, with preprocessed source if appropriate. Please include the complete backtrace with any bug report. See <file:///usr/share/doc/gcc-8/README.Bugs> for instructions. make[2]: *** [modules/python3/CMakeFiles/opencv_python3.dir/build.make:63: modules/python3/CMakeFiles/opencv_python3.dir/__/src2/cv2.cpp.o] Error 1 make[1]: *** [CMakeFiles/Makefile2:11793: modules/python3/CMakeFiles/opencv_python3.dir/all] Error 2 make: *** [Makefile:163: all] Error 2 </code></p>
<p dir="auto"><a href="https://pastebin.com/5FnMjAQe" rel="nofollow">https://pastebin.com/5FnMjAQe</a><br>
Complete compilation: <a href="https://pastebin.com/ujKu1mRT" rel="nofollow">https://pastebin.com/ujKu1mRT</a></p> | 1 |
<h5 dir="auto">Issue Type:</h5>
<p dir="auto">Feature Idea</p>
<h5 dir="auto">Ansible Version:</h5>
<p dir="auto">ansible 1.5.4</p>
<h5 dir="auto">Environment:</h5>
<p dir="auto">N/A</p>
<h5 dir="auto">Summary:</h5>
<p dir="auto">Related to <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="15690632" data-permission-text="Title is private" data-url="https://github.com/ansible/ansible/issues/3245" data-hovercard-type="issue" data-hovercard-url="/ansible/ansible/issues/3245/hovercard" href="https://github.com/ansible/ansible/issues/3245">#3245</a> I'd like to request the ability to use vars_prompt in a role, such as:</p>
<ul dir="auto">
<li>an action "include_vars_prompt" similar to "include_vars" but which functions as vars_prompt does</li>
</ul>
<p dir="auto">OR</p>
<ul dir="auto">
<li>have the ability to have vars_prompt function in role vars files (perhaps a ROLES/vars/prompt.yml file in addition to main.yml)</li>
</ul>
<p dir="auto">My use case:<br>
My deployment system uses a role to handle the actual deployment. Specifically, hosts are grouped in inventory into application groups, each group has a group_vars file that sets the required vars (git repos, pip files, application name, etc..) for the role to perform the deployment with.</p>
<p dir="auto">Our main playbook then has a play for each application, targeted at the application group, that calls the deployment role.</p>
<p dir="auto">What I would like to do now is easily add vars_prompt: functionality into the deployment role so that I may prompt the user for the desired release to deploy. For automated invocation, the release would need to be supplied as an extra var, which is fine.</p>
<p dir="auto">Having this feature would allow me to avoid having to duplicate via copy-n-paste the vars_prompt bits for each application deployment play. Since I have multiple prompts per application and would like to use Jinja templates to define the prompt and default vars, the repetition of this code gets unwieldy quickly. Moreso when I end up needing to split applications up into separate playbooks.</p>
<p dir="auto">Thanks!</p>
<p dir="auto">-n</p> | <p dir="auto">Neither of the following obvious approaches work for storing <code class="notranslate">/etc/ssh/ssh_host_key</code> in an Ansible vault, i.e. a vars file:</p>
<ul dir="auto">
<li>YAML 1.1's <code class="notranslate">!binary</code> (<a href="http://yaml.org/spec/1.1/" rel="nofollow">spec</a>, <a href="http://yaml.org/type/binary.html" rel="nofollow">tag draft</a>), because the parser fails</li>
<li><code class="notranslate">copy</code> with <code class="notranslate">content</code> set to <code class="notranslate">"{{ ssh_host_key|b64decode }}"</code>, because the Jinja2 template engine <a href="https://groups.google.com/forum/#!topic/ansible-devel/lR0k2S0GIQw/discussion" rel="nofollow">corrupts the result</a></li>
</ul>
<p dir="auto">All the workarounds I've found involve writing to temporary files and decoding them, plus use of <code class="notranslate">register</code> and <code class="notranslate">changed_when</code> to detect whether the target really changed or not.</p>
<p dir="auto">Could we add the <code class="notranslate">!binary</code> parser to avoid the need to use Jinja2, or fix our use of Jinja2, or fix Jinja2?</p>
<p dir="auto">See also: <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="102672021" data-permission-text="Title is private" data-url="https://github.com/ansible/ansible-modules-core/issues/1961" data-hovercard-type="issue" data-hovercard-url="/ansible/ansible-modules-core/issues/1961/hovercard" href="https://github.com/ansible/ansible-modules-core/issues/1961">ansible/ansible-modules-core#1961</a></p> | 0 |
<p dir="auto"><strong><a href="https://jira.spring.io/secure/ViewProfile.jspa?name=shaug" rel="nofollow">Scott Haug</a></strong> opened <strong><a href="https://jira.spring.io/browse/SPR-2850?redirect=false" rel="nofollow">SPR-2850</a></strong> and commented</p>
<p dir="auto">Similar to <a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="398073279" data-permission-text="Title is private" data-url="https://github.com/spring-projects/spring-framework/issues/7571" data-hovercard-type="issue" data-hovercard-url="/spring-projects/spring-framework/issues/7571/hovercard" href="https://github.com/spring-projects/spring-framework/issues/7571">#7571</a>, a change in the Quartz 1.6 API has made some of the nice Spring-related features incompatible with that release. Namely, the changes related to the following issue have made SimpleTriggerBean and CronTriggerBean useless with Quartz 1.6</p>
<p dir="auto"><a href="http://jira.opensymphony.com/browse/QUARTZ-417" rel="nofollow">http://jira.opensymphony.com/browse/QUARTZ-417</a></p>
<p dir="auto">This change resulted in org.quartz.impl.jdbcjobstore.JobStoreSupport#storeTrigger using '==' rather than instanceof to determine if the trigger is a SimpleTrigger or CronTrigger. Ostensibly, this is to allow custom triggers to extend either of these classes, but it prevents Spring from providing extended configuration functionality through subclassing. Personally, I think a better solution on Quartz's part would have been to introduce a marker interface like CustomTrigger, but I'm not sure how receptive the Quartz team will be to change this yet again to accomodate Spring code.</p>
<p dir="auto">My workaround has been to turn SimpleTriggerBean and CronTriggerBean into FactoryBeans. This is much clunkier, as it requires a lot more copying of data, but it gets the job done and is backwards-compatible.</p>
<hr>
<p dir="auto"><strong>Affects:</strong> 2.0 final</p>
<p dir="auto"><strong>Issue Links:</strong></p>
<ul dir="auto">
<li><a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="398074401" data-permission-text="Title is private" data-url="https://github.com/spring-projects/spring-framework/issues/7705" data-hovercard-type="issue" data-hovercard-url="/spring-projects/spring-framework/issues/7705/hovercard" href="https://github.com/spring-projects/spring-framework/issues/7705">#7705</a> Quartz 1.6/CronTriggerBean persistence problem (<em><strong>"is duplicated by"</strong></em>)</li>
<li><a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="398073196" data-permission-text="Title is private" data-url="https://github.com/spring-projects/spring-framework/issues/7564" data-hovercard-type="issue" data-hovercard-url="/spring-projects/spring-framework/issues/7564/hovercard" href="https://github.com/spring-projects/spring-framework/issues/7564">#7564</a> Spring Quartz support, persistent jobs. (<em><strong>"is duplicated by"</strong></em>)</li>
</ul>
<p dir="auto">19 votes, 22 watchers</p> | <p dir="auto"><strong><a href="https://jira.spring.io/secure/ViewProfile.jspa?name=cae29" rel="nofollow">Caesar Loh</a></strong> opened <strong><a href="https://jira.spring.io/browse/SPR-4597?redirect=false" rel="nofollow">SPR-4597</a></strong> and commented</p>
<p dir="auto">In line 108 of NativeJdbcExtractorAdapter, it will return a null for connection.</p>
<p dir="auto">This occur when the datasource is being retrieved from another server rather than the holding server. The metadata.getConnection() will return a null, and obviously by comparing it with the original connection, it will be different. Thus, this method will return a null.</p>
<p dir="auto">How about doing a simple check for null before returning the connection? In any case, metadataConnection will simply return the same connection.</p>
<hr>
<p dir="auto"><strong>Affects:</strong> 2.0.6, 2.0.7, 2.0.8, 2.5.1, 2.5.2</p>
<p dir="auto"><strong>Issue Links:</strong></p>
<ul dir="auto">
<li><a class="issue-link js-issue-link" data-error-text="Failed to load title" data-id="398086493" data-permission-text="Title is private" data-url="https://github.com/spring-projects/spring-framework/issues/9273" data-hovercard-type="issue" data-hovercard-url="/spring-projects/spring-framework/issues/9273/hovercard" href="https://github.com/spring-projects/spring-framework/issues/9273">#9273</a> NativeJdbcExtractorAdapter .getNativeConnectionConnection (<em><strong>"is duplicated by"</strong></em>)</li>
</ul>
<p dir="auto"><strong>Backported to:</strong> <a href="https://github.com/spring-projects/spring-framework/milestone/46?closed=1">2.0.9</a></p> | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.