repo_name
stringlengths
9
75
topic
stringclasses
30 values
issue_number
int64
1
203k
title
stringlengths
1
976
body
stringlengths
0
254k
state
stringclasses
2 values
created_at
stringlengths
20
20
updated_at
stringlengths
20
20
url
stringlengths
38
105
labels
sequencelengths
0
9
user_login
stringlengths
1
39
comments_count
int64
0
452
HumanSignal/labelImg
deep-learning
255
Bug in Verify?
Looks like in in prebuild labelImg_v1.6.0 for windows `Verify Image` not working properly when you press verify and then press next green background still present in next image.
closed
2018-03-21T14:23:18Z
2018-05-22T05:53:29Z
https://github.com/HumanSignal/labelImg/issues/255
[]
mrgloom
1
scikit-hep/awkward
numpy
2,462
Running over zero-length arrays, time spent mostly in function overhead
### Version of Awkward Array 2.2.0 ### Description and code to reproduce N.B.: This is lower priority, but it scales with analysis size/complexity! When running column optimization in dask awkward a typetracer is passed through the whole analysis code. This triggers all relevant function calls for the analysis and produces an accurate accounting of data needs which is great, but as analyses grow larger (especially with systematics overhead!) the tracing begins to take quite some time for processing no data. I understand this is only so reducible, but responsiveness is definitely part of user experience and it would be sad if we lost users because of really excellent core functionality taking too long. Here is a pyinstrument trace showing what functions stacks are grabbing the most time in a mostly realistic analysis with full systematics: ``` _ ._ __/__ _ _ _ _ _/_ Recorded: 08:25:40 Samples: 3840 /_//_/// /_\ / //_// / //_'/ // Duration: 3.983 CPU time: 3.988 / _/ v4.3.0 Program: yimu_test5.py 3.983 <module> yimu_test5.py:1 โ””โ”€ 3.983 necessary_columns dask_awkward/lib/inspect.py:11 โ””โ”€ 3.983 _necessary_columns dask_awkward/lib/optimize.py:392 โ””โ”€ 3.979 _get_column_reports dask_awkward/lib/optimize.py:335 โ”œโ”€ 3.900 get_sync dask/local.py:551 โ”‚ โ””โ”€ 3.898 get_async dask/local.py:350 โ”‚ โ”œโ”€ 3.382 fire_tasks dask/local.py:452 โ”‚ โ”‚ โ””โ”€ 3.331 submit dask/local.py:539 โ”‚ โ”‚ โ””โ”€ 3.292 batch_execute_tasks dask/local.py:234 โ”‚ โ”‚ โ””โ”€ 3.292 <listcomp> dask/local.py:238 โ”‚ โ”‚ โ””โ”€ 3.287 execute_task dask/local.py:214 โ”‚ โ”‚ โ””โ”€ 3.279 _execute_task dask/core.py:82 โ”‚ โ”‚ โ””โ”€ 3.254 __call__ dask/optimization.py:987 โ”‚ โ”‚ โ””โ”€ 3.246 get dask/core.py:128 โ”‚ โ”‚ โ””โ”€ 3.211 _execute_task dask/core.py:82 โ”‚ โ”‚ โ”œโ”€ 1.186 __array_ufunc__ awkward/highlevel.py:1289 โ”‚ โ”‚ โ”‚ โ””โ”€ 1.138 array_ufunc awkward/_connect/numpy.py:219 โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.496 _array_ufunc_custom_cast awkward/_connect/numpy.py:147 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.449 find_custom_cast awkward/_behavior.py:60 โ”‚ โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.271 __iter__ _collections_abc.py:742 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.146 [self] โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.106 __getitem__ collections/__init__.py:892 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.121 [self] โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.382 broadcast_and_apply awkward/_broadcasting.py:1078 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.356 apply_step awkward/_broadcasting.py:370 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.336 continuation awkward/_broadcasting.py:1023 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.332 broadcast_any_list awkward/_broadcasting.py:500 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.299 apply_step awkward/_broadcasting.py:370 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.275 continuation awkward/_broadcasting.py:1023 โ”‚ โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.222 broadcast_any_list awkward/_broadcasting.py:500 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.189 apply_step awkward/_broadcasting.py:370 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.126 continuation awkward/_broadcasting.py:1023 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.105 broadcast_any_list awkward/_broadcasting.py:500 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.085 apply_step awkward/_broadcasting.py:370 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.066 action awkward/_connect/numpy.py:228 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.055 find_ufunc awkward/_behavior.py:109 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.055 action awkward/_connect/numpy.py:228 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.043 find_ufunc awkward/_behavior.py:109 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.049 broadcast_any_indexed awkward/_broadcasting.py:979 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.213 recursively_apply awkward/_do.py:20 โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.148 _recursively_apply awkward/contents/listoffsetarray.py:2069 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.144 continuation awkward/contents/listoffsetarray.py:2084 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.106 _recursively_apply awkward/contents/numpyarray.py:1261 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.105 unary_action awkward/_connect/numpy.py:309 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.105 action awkward/_connect/numpy.py:228 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.085 find_ufunc awkward/_behavior.py:109 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.043 __iter__ _collections_abc.py:742 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.045 _recursively_apply awkward/contents/indexedarray.py:1030 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.043 continuation awkward/contents/indexedarray.py:1055 โ”‚ โ”‚ โ”œโ”€ 0.558 __call__ dask_awkward/lib/core.py:1894 โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.407 pt coffea/nanoevents/methods/vector.py:121 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.407 r coffea/nanoevents/methods/vector.py:85 โ”‚ โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.361 r2 coffea/nanoevents/methods/vector.py:111 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.336 func numpy/lib/mixins.py:18 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.335 __array_ufunc__ awkward/highlevel.py:1289 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.323 array_ufunc awkward/_connect/numpy.py:219 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.222 broadcast_and_apply awkward/_broadcasting.py:1078 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.213 apply_step awkward/_broadcasting.py:370 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.205 continuation awkward/_broadcasting.py:1023 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.203 broadcast_any_list awkward/_broadcasting.py:500 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.191 apply_step awkward/_broadcasting.py:370 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.181 continuation awkward/_broadcasting.py:1023 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.113 broadcast_any_list awkward/_broadcasting.py:500 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.099 apply_step awkward/_broadcasting.py:370 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.087 continuation awkward/_broadcasting.py:1023 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.067 broadcast_any_list awkward/_broadcasting.py:500 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.057 apply_step awkward/_broadcasting.py:370 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.067 broadcast_any_indexed awkward/_broadcasting.py:979 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.051 apply_step awkward/_broadcasting.py:370 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.050 continuation awkward/_broadcasting.py:1023 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.050 broadcast_any_list awkward/_broadcasting.py:500 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.048 apply_step awkward/_broadcasting.py:370 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.048 continuation awkward/_broadcasting.py:1023 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.048 broadcast_any_indexed awkward/_broadcasting.py:979 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.092 _array_ufunc_custom_cast awkward/_connect/numpy.py:147 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.088 find_custom_cast awkward/_behavior.py:60 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.049 __iter__ _collections_abc.py:742 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.047 __array_ufunc__ awkward/highlevel.py:1289 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.044 array_ufunc awkward/_connect/numpy.py:219 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.140 eta coffea/nanoevents/methods/vector.py:512 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.100 r coffea/nanoevents/methods/vector.py:85 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.088 r2 coffea/nanoevents/methods/vector.py:111 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.075 func numpy/lib/mixins.py:18 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.075 __array_ufunc__ awkward/highlevel.py:1289 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.074 array_ufunc awkward/_connect/numpy.py:219 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.047 broadcast_and_apply awkward/_broadcasting.py:1078 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.045 apply_step awkward/_broadcasting.py:370 โ”‚ โ”‚ โ”œโ”€ 0.408 apply dask/utils.py:41 โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.201 sum awkward/operations/ak_sum.py:13 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.180 _impl awkward/operations/ak_sum.py:267 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.174 reduce awkward/_do.py:259 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.126 _reduce_next awkward/contents/listoffsetarray.py:1464 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.040 _reduce_next awkward/contents/listoffsetarray.py:1464 โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.061 count awkward/operations/ak_count.py:12 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.052 _impl awkward/operations/ak_count.py:110 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.049 reduce awkward/_do.py:259 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.057 mean awkward/operations/ak_mean.py:13 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.057 _impl awkward/operations/ak_mean.py:152 โ”‚ โ”‚ โ”œโ”€ 0.399 __call__ dask_awkward/lib/structure.py:861 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.398 with_field awkward/operations/ak_with_field.py:19 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.381 _impl awkward/operations/ak_with_field.py:55 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.360 broadcast_and_apply awkward/_broadcasting.py:1078 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.353 apply_step awkward/_broadcasting.py:370 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.351 continuation awkward/_broadcasting.py:1023 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.350 broadcast_any_list awkward/_broadcasting.py:500 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.338 apply_step awkward/_broadcasting.py:370 โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.276 continuation awkward/_broadcasting.py:1023 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.260 broadcast_any_list awkward/_broadcasting.py:500 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.246 apply_step awkward/_broadcasting.py:370 โ”‚ โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.151 continuation awkward/_broadcasting.py:1023 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.092 broadcast_any_indexed awkward/_broadcasting.py:979 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.082 <listcomp> awkward/_broadcasting.py:980 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.082 project awkward/contents/indexedarray.py:408 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.080 _carry awkward/contents/recordarray.py:496 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.079 <listcomp> awkward/contents/recordarray.py:517 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.071 _carry awkward/contents/numpyarray.py:315 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.046 __getitem__ awkward/_nplikes/typetracer.py:378 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.059 broadcast_any_list awkward/_broadcasting.py:500 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.056 apply_step awkward/_broadcasting.py:370 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.053 continuation awkward/_broadcasting.py:1023 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.052 broadcast_any_indexed awkward/_broadcasting.py:979 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.047 <listcomp> awkward/_broadcasting.py:980 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.047 project awkward/contents/indexedarray.py:408 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.045 _carry awkward/contents/recordarray.py:496 โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.042 <listcomp> awkward/contents/recordarray.py:517 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.093 action awkward/operations/ak_with_field.py:123 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.081 <listcomp> awkward/operations/ak_with_field.py:156 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.079 __getitem__ awkward/contents/content.py:535 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.076 _getitem awkward/contents/content.py:538 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.051 _getitem_field awkward/contents/recordarray.py:458 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.048 content awkward/contents/recordarray.py:396 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.058 action awkward/operations/ak_with_field.py:123 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.052 <listcomp> awkward/operations/ak_with_field.py:156 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.052 __getitem__ awkward/contents/content.py:535 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.051 _getitem awkward/contents/content.py:538 โ”‚ โ”‚ โ”œโ”€ 0.324 __getitem__ awkward/highlevel.py:518 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.256 __getitem__ awkward/contents/content.py:535 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.255 _getitem awkward/contents/content.py:538 โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.156 _getitem awkward/contents/content.py:538 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.142 _getitem awkward/contents/content.py:538 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.094 _getitem_next awkward/contents/regulararray.py:451 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.080 _getitem_next_jagged awkward/contents/listoffsetarray.py:416 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.071 _getitem_next_jagged awkward/contents/listarray.py:389 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.058 _getitem_next awkward/contents/regulararray.py:451 โ”‚ โ”‚ โ”œโ”€ 0.187 __call__ dask_awkward/lib/core.py:1886 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.187 delta_r coffea/nanoevents/methods/vector.py:602 โ”‚ โ”‚ โ”‚ โ”œโ”€ 0.095 eta coffea/nanoevents/methods/vector.py:512 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.068 r coffea/nanoevents/methods/vector.py:85 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.061 r2 coffea/nanoevents/methods/vector.py:111 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.056 func numpy/lib/mixins.py:18 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.056 __array_ufunc__ awkward/highlevel.py:1289 โ”‚ โ”‚ โ”‚ โ”‚ โ””โ”€ 0.053 array_ufunc awkward/_connect/numpy.py:219 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.050 delta_phi coffea/nanoevents/methods/vector.py:199 โ”‚ โ”‚ โ””โ”€ 0.058 __call__ dask_awkward/lib/structure.py:807 โ”‚ โ”‚ โ””โ”€ 0.058 where awkward/operations/ak_where.py:14 โ”‚ โ”‚ โ””โ”€ 0.056 _impl3 awkward/operations/ak_where.py:97 โ”‚ โ”‚ โ””โ”€ 0.052 broadcast_and_apply awkward/_broadcasting.py:1078 โ”‚ โ”‚ โ””โ”€ 0.049 apply_step awkward/_broadcasting.py:370 โ”‚ โ”‚ โ””โ”€ 0.048 continuation awkward/_broadcasting.py:1023 โ”‚ โ”‚ โ””โ”€ 0.048 broadcast_any_list awkward/_broadcasting.py:500 โ”‚ โ”‚ โ””โ”€ 0.042 apply_step awkward/_broadcasting.py:370 โ”‚ โ”œโ”€ 0.252 keys dask/highlevelgraph.py:530 โ”‚ โ”‚ โ””โ”€ 0.252 to_dict dask/highlevelgraph.py:522 โ”‚ โ”‚ โ””โ”€ 0.252 ensure_dict dask/utils.py:1236 โ”‚ โ”‚ โ””โ”€ 0.234 __iter__ _collections_abc.py:719 โ”‚ โ”‚ โ””โ”€ 0.233 __iter__ dask/blockwise.py:493 โ”‚ โ”‚ โ””โ”€ 0.232 _dict dask/blockwise.py:452 โ”‚ โ”‚ โ”œโ”€ 0.066 fuse dask/optimization.py:450 โ”‚ โ”‚ โ”œโ”€ 0.060 dims dask/blockwise.py:440 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.059 _make_dims dask/blockwise.py:1480 โ”‚ โ”‚ โ”‚ โ””โ”€ 0.058 broadcast_dimensions dask/blockwise.py:1420 โ”‚ โ”‚ โ””โ”€ 0.051 make_blockwise_graph dask/blockwise.py:759 โ”‚ โ”œโ”€ 0.129 start_state_from_dask dask/local.py:140 โ”‚ โ”‚ โ””โ”€ 0.113 reverse_dict dask/core.py:307 โ”‚ โ”‚ โ””โ”€ 0.112 [self] โ”‚ โ””โ”€ 0.052 order dask/order.py:84 โ””โ”€ 0.042 mock dask_awkward/layers/layers.py:94 ``` I am happy to provide this test code if it is needed! (this cannot really be made minimal as it's ~emergent behavior)
closed
2023-05-16T13:32:38Z
2023-06-16T14:06:59Z
https://github.com/scikit-hep/awkward/issues/2462
[ "performance" ]
lgray
11
proplot-dev/proplot
data-visualization
390
Proplot not compatible with matplotlib 3.6
Looks like the current version of proplot does not work with matplotlib==3.6 (though it works fine with 3.5). The import: > import proplot as pplt Throws: > AttributeError: module 'matplotlib.cm' has no attribute 'cmap_d' From: proplot/colors.py:3111, > _cmap_database = _init_cmap_database() This might be fixable by changing the reference to cmap_d to point instead to matplotlib.colormaps.
open
2022-09-20T18:36:31Z
2023-08-02T12:52:33Z
https://github.com/proplot-dev/proplot/issues/390
[ "dependencies" ]
swamidass
14
dask/dask
scikit-learn
10,986
CI is printing tracebacks for all xfailed tests which can be very confusing
See #10984 for more context. We can definitely disable this, we would have to add the setting to the pytest call if we want to do this
closed
2024-03-07T15:33:01Z
2024-03-27T14:23:59Z
https://github.com/dask/dask/issues/10986
[ "needs triage" ]
phofl
0
huggingface/peft
pytorch
2,015
XLoRA: training issues, Gradients will be None
I installed PEFT from source. And use the latest versions of Transformers and TRL. I passed the XLoRA model to TRL but the training doesn't seem to work (training loss doesn't decrease and validation loss remains constant). I get this warning: UserWarning: None of the inputs have requires_grad=True. Gradients will be None I load Llama 3.1 (without quantization) and then run this code: ``` adapters = dict() adapters["0"] = './adapter1/' adapters["1"] = './adapter2/' peft_config = XLoraConfig( task_type=TaskType.CAUSAL_LM, peft_type=PeftType.XLORA, hidden_size=model.config.hidden_size, xlora_depth=8, adapters=adapters, xlora_size=2048, layerwise_scalings=True, xlora_dropout_p=0.2 ) xlora_model = get_peft_model(model, peft_config) training_arguments = SFTConfig( output_dir="./output/", optim="paged_adamw_8bit", per_device_train_batch_size=2, gradient_accumulation_steps=16, save_strategy="epoch", log_level="debug", logging_steps=1, learning_rate=1e-5, bf16 = True, num_train_epochs=1, warmup_ratio=0.1, lr_scheduler_type="linear", dataset_text_field="text", max_seq_length=512, ) trainer = SFTTrainer( model=xlora_model, train_dataset=ds, tokenizer=tokenizer, args=training_arguments, ) trainer.train() ``` I also observed another bug: The adapters must be named "0", "1", etc in the adapters dict() otherwise training won't start and will say that the adapters don't exist. Maybe @EricLBuehler can help with this?
closed
2024-08-18T10:30:50Z
2024-11-17T15:03:26Z
https://github.com/huggingface/peft/issues/2015
[]
benjamin-marie
10
jschneier/django-storages
django
997
parameters are ignored if s3 custom URL is provided
https://github.com/jschneier/django-storages/blob/770332b598712da27ecdba75c9e202ad6a1a8722/storages/backends/s3boto3.py#L554-L578 If additional parameters are passed in to the url method when a custom_url is in use (like when S3 is behind CloudFront), they're ignored on the generated URL. They're already passed in as a dict, so they could easily be passed in to `urllib.parse.urlunparse` instead of using a format string when the URL is built (which might be a good fix anyway).
closed
2021-04-02T00:31:39Z
2021-09-11T18:10:43Z
https://github.com/jschneier/django-storages/issues/997
[]
dannysauer
1
docarray/docarray
pydantic
1,444
docs: explain type coercion
A Document parses data an coerces its type if needed. For example: ```python from docarray import BaseDoc from docarray.typing import TorchTensor import numpy as np class Doc(BaseDoc): embedding: TorchTensor[5] d = Doc(embedding=np.random.rand(5)) print(d.embedding) print(type(d.embedding)) ``` ```bash TorchTensor([0.5059, 0.1513, 0.5784, 0.9147, 0.6029], dtype=torch.float64) <class 'docarray.typing.tensor.torch_tensor.TorchTensor'> ``` For everyone coming from pydantic this is obvious, but everyone else would probably manually convert the np array to torch before passing it. So I think this is a feature worth highlighting
closed
2023-04-24T16:19:30Z
2023-05-31T09:22:44Z
https://github.com/docarray/docarray/issues/1444
[ "area/docs" ]
JohannesMessner
3
dynaconf/dynaconf
flask
723
[bug] Get Vault Secret via dynaconf throws `permission denied` error
**Describe the bug** Accessing Vault secrets via dynaconf throws a` permission denied` error even though the user/role has access to the secrets and its path. **To Reproduce** Steps to reproduce the behaviour: 1. Having the following folder structure in the Vault <details> <summary> Vault Store </summary> ``` /apps/mna-satqe-hashicorp/robottelo ``` The secrets are stored in the above path in Corporate Vault. </details> 2. Having the following app code: <details> <summary> Code </summary> **config/__init__.py** ```python from dynaconf import LazySettings os.environ["VAULT_KV_VERSION_FOR_DYNACONF"] = "1" os.environ['VAULT_ROLE_ID_FOR_DYNACONF'] = 'mna-satqe-hashicorp' os.environ["VAULT_SECRET_ID_FOR_DYNACONF"] = "abcd-323333-23232323-sdsfsf" os.environ["VAULT_MOUNT_POINT_FOR_DYNACONF"] = "apps" os.environ["VAULT_PATH_FOR_DYNACONF"] = "mna-satqe-hashicorp/robottelo" settingsd = LazySettings(environments=True, vault_enabled=True, vault={'url': 'https://vault.corp.compony.com:8200', 'verify': '/etc/ipa/ipa.crt'}) print(settings.ad_password) ``` </details> 4. Executing under the following environment <details> <summary> Execution </summary> ```bash $ pytest -v -s tests/foreman/api/test_product.py -k test_positive_create_with_name Traceback (most recent call last): . . . . File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/dynaconf/base.py", line 113, in __getattr__ self._setup() File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/dynaconf/base.py", line 163, in _setup self._wrapped = Settings( File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/dynaconf/base.py", line 233, in __init__ self.execute_loaders() File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/dynaconf/base.py", line 977, in execute_loaders loader.load(self, env, silent=silent, key=key) File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/dynaconf/loaders/vault_loader.py", line 74, in load dirs = client.secrets.kv.list_secrets( File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/hvac/api/secrets_engines/kv_v1.py", line 58, in list_secrets return self._adapter.list( File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/hvac/adapters.py", line 165, in list return self.request("list", url, **kwargs) File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/hvac/adapters.py", line 364, in request response = super(JSONAdapter, self).request(*args, **kwargs) File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/hvac/adapters.py", line 330, in request utils.raise_for_error( File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/hvac/utils.py", line 41, in raise_for_error raise exceptions.Forbidden(message, errors=errors, method=method, url=url) hvac.exceptions.Forbidden: 1 error occurred: * permission denied , on list https://vault.corp.company.com:8200/v1/apps/mna-satqe-hashicorp/robottelo ``` </details> **Expected behaviour** DynaConf should be able to access the secret successfully and print in my case above. **Environment (please complete the following information):** - OS: [MacOS 12.1] - Dynaconf Version [3.1.7] - Frameworks in use (pytest) **Additional context** The curl command to access the secrets from the same path works as expected. ``` # curl --header "X-Vault-Token: s.saswewetergfd3333dddd" https://vault.corp.company.com:8200/v1/apps/data/mna-satqe-hashicorp/robottelo {"request_id":"a75e5b8a-c958-8608-dad0-85f516c62c69","lease_id":"","renewable":false,"lease_duration":0,"data":{"data":{"ad_password":"SecretPAss3456"}, "metadata":{"created_time":"2022-02-25T11:51:47.258977787Z","deletion_time":"","destroyed":false,"version":1}},"wrap_info":null,"warnings":null,"auth":null} ```
closed
2022-02-28T14:02:16Z
2022-03-30T20:34:43Z
https://github.com/dynaconf/dynaconf/issues/723
[ "bug" ]
jyejare
0
long2ice/fastapi-cache
fastapi
31
How to set the redis db in the cache?
closed
2021-08-02T09:02:43Z
2021-08-02T09:18:48Z
https://github.com/long2ice/fastapi-cache/issues/31
[]
liushuangdan
2
gee-community/geemap
streamlit
455
Add GeoPandas/Fiona support
This will allow users to open any vector dataset supported by GeoPandas/Fiona and display it on the map. It can also be converted to ee.FeatureCollection if needed. Fiona supported drivers: https://github.com/Toblerity/Fiona/blob/master/fiona/drvsupport.py geopandas.read_file https://geopandas.org/docs/reference/api/geopandas.read_file.html#geopandas.read_file
closed
2021-05-04T01:20:46Z
2021-05-05T02:19:02Z
https://github.com/gee-community/geemap/issues/455
[ "Feature Request" ]
giswqs
1
Johnserf-Seed/TikTokDownload
api
405
[Feature]่ƒฝๅฆ็ป™ๅธฆๆœ‰้•ฟๆ–‡ๆœฌๆ ‡้ข˜็š„่ง†้ข‘ ๅ•็‹ฌๅŠ ไธ€ไธชๆ”พ่ฏฅ้•ฟๆ–‡ๆœฌ็š„txtๆ–‡ๆกฃ
ๆœ‰ไบ›้•ฟๆ–‡ๆœฌๆœฌ่บซ็š„ๅ†…ๅฎนไนŸๆ˜ฏๅพˆไธ้”™็š„
open
2023-04-21T00:34:29Z
2023-08-04T12:37:34Z
https://github.com/Johnserf-Seed/TikTokDownload/issues/405
[ "้œ€ๆฑ‚ๅปบ่ฎฎ(enhancement)" ]
NightMoonIsland
1
unit8co/darts
data-science
1,901
RuntimeError: Lightning can't create new processes if CUDA is already initialized.
**Describe the bug** I encounter a problem that says `RuntimeError: Lightning can't create new processes if CUDA is already initialized. ` **To Reproduce** ```python import numpy as np import pandas as pd import torch import matplotlib.pyplot as plt from darts.dataprocessing.transformers import Scaler from darts.models import RNNModel from darts.metrics import mape from darts.datasets import AirPassengersDataset # Read data: series = AirPassengersDataset().load() series = series.astype(np.float32) # Create training and validation sets: train, val = series.split_after(pd.Timestamp("19590101")) # Normalize the time series (note: we avoid fitting the transformer on the validation set) transformer = Scaler() train_transformed = transformer.fit_transform(train) val_transformed = transformer.transform(val) series_transformed = transformer.transform(series) torch.cuda.set_device(0) my_model = RNNModel( model="RNN", hidden_dim=20, dropout=0, batch_size=16, n_epochs=300, optimizer_kwargs={"lr": 1e-3}, model_name="Air_RNN", log_tensorboard=True, random_state=42, training_length=20, input_chunk_length=14, force_reset=True, ) my_model.fit(train_transformed, val_series=val_transformed) ``` **Expected behavior** I want it works properly. **System (please complete the following information):** - Python version: 3.9 - darts version: 0.24.0 - torch version: 1.12.0 **Additional context** Detail error: ```shell GPU available: True (cuda), used: True TPU available: False, using: 0 TPU cores IPU available: False, using: 0 IPUs HPU available: False, using: 0 HPUs --------------------------------------------------------------------------- RuntimeError Traceback (most recent call last) Cell In[9], line 40 24 torch.cuda.set_device(0) 26 my_model = RNNModel( 27 model="RNN", 28 hidden_dim=20, (...) 38 force_reset=True, 39 ) ---> 40 my_model.fit(train_transformed, val_series=val_transformed) File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/darts/utils/torch.py:112, in random_method.<locals>.decorator(self, *args, **kwargs) 110 with fork_rng(): 111 manual_seed(self._random_instance.randint(0, high=MAX_TORCH_SEED_VALUE)) --> 112 return decorated(self, *args, **kwargs) File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/darts/models/forecasting/torch_forecasting_model.py:705, in TorchForecastingModel.fit(self, series, past_covariates, future_covariates, val_series, val_past_covariates, val_future_covariates, trainer, verbose, epochs, max_samples_per_ts, num_loader_workers) 699 # call super fit only if user is actually fitting the model 700 super().fit( 701 series=seq2series(series), 702 past_covariates=seq2series(past_covariates), 703 future_covariates=seq2series(future_covariates), 704 ) --> 705 return self.fit_from_dataset(*params) File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/darts/utils/torch.py:112, in random_method.<locals>.decorator(self, *args, **kwargs) 110 with fork_rng(): 111 manual_seed(self._random_instance.randint(0, high=MAX_TORCH_SEED_VALUE)) --> 112 return decorated(self, *args, **kwargs) File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/darts/models/forecasting/torch_forecasting_model.py:897, in TorchForecastingModel.fit_from_dataset(self, train_dataset, val_dataset, trainer, verbose, epochs, num_loader_workers) 846 @random_method 847 def fit_from_dataset( 848 self, (...) 854 num_loader_workers: int = 0, 855 ) -> "TorchForecastingModel": 856 """ 857 Train the model with a specific :class:`darts.utils.data.TrainingDataset` instance. 858 These datasets implement a PyTorch ``Dataset``, and specify how the target and covariates are sliced (...) 895 Fitted model. 896 """ --> 897 self._train( 898 *self._setup_for_train( 899 train_dataset=train_dataset, 900 val_dataset=val_dataset, 901 trainer=trainer, 902 verbose=verbose, 903 epochs=epochs, 904 num_loader_workers=num_loader_workers, 905 ) 906 ) 907 return self File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/darts/models/forecasting/torch_forecasting_model.py:1041, in TorchForecastingModel._train(self, trainer, model, train_loader, val_loader) 1038 ckpt_path = self.load_ckpt_path 1039 self.load_ckpt_path = None -> 1041 trainer.fit( 1042 model, 1043 train_dataloaders=train_loader, 1044 val_dataloaders=val_loader, 1045 ckpt_path=ckpt_path, 1046 ) 1047 self.model = model 1048 self.trainer = trainer File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/pytorch_lightning/trainer/trainer.py:531, in Trainer.fit(self, model, train_dataloaders, val_dataloaders, datamodule, ckpt_path) 529 model = _maybe_unwrap_optimized(model) 530 self.strategy._lightning_module = model --> 531 call._call_and_handle_interrupt( 532 self, self._fit_impl, model, train_dataloaders, val_dataloaders, datamodule, ckpt_path 533 ) File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/pytorch_lightning/trainer/call.py:41, in _call_and_handle_interrupt(trainer, trainer_fn, *args, **kwargs) 39 try: 40 if trainer.strategy.launcher is not None: ---> 41 return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) 42 return trainer_fn(*args, **kwargs) 44 except _TunerExitException: File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/pytorch_lightning/strategies/launchers/multiprocessing.py:99, in _MultiProcessingLauncher.launch(self, function, trainer, *args, **kwargs) 97 self._check_torchdistx_support() 98 if self._start_method in ("fork", "forkserver"): ---> 99 _check_bad_cuda_fork() 101 # The default cluster environment in Lightning chooses a random free port number 102 # This needs to be done in the main process here before starting processes to ensure each rank will connect 103 # through the same port 104 assert self._strategy.cluster_environment is not None File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/lightning_fabric/strategies/launchers/multiprocessing.py:189, in _check_bad_cuda_fork() 187 if _IS_INTERACTIVE: 188 message += " You will have to restart the Python kernel." --> 189 raise RuntimeError(message) RuntimeError: Lightning can't create new processes if CUDA is already initialized. Did you manually call `torch.cuda.*` functions, have moved the model to the device, or allocated memory on the GPU any other way? Please remove any such calls, or change the selected strategy. You will have to restart the Python kernel. ```
closed
2023-07-17T07:24:45Z
2023-07-18T15:40:08Z
https://github.com/unit8co/darts/issues/1901
[ "bug", "triage" ]
iDestro
2
quasarstream/python-ffmpeg-video-streaming
dash
43
Setting custom ffmpeg/ffprobe binary path
I'd like to use python-ffmpeg-video-streaming and set a custom path to the ffmpeg and ffprobe binaries (as opposed to using the value found in the path), as I have multiple custom ffmpeg builds available on my system. I see the run_command call in _media.py: ` run_command: bool = True, ffmpeg_bin: str = 'ffmpeg', monitor: callable = None, **options):` And in _command_builder.py: `def command_builder(ffmpeg_bin: str, media):` How can I override this value without having to modify the code myself & recompile?
closed
2020-10-30T15:22:42Z
2020-11-01T13:59:16Z
https://github.com/quasarstream/python-ffmpeg-video-streaming/issues/43
[ "question" ]
CraigSample
1
ijl/orjson
numpy
341
Problem installing orjson package
When I am installing orjson, I get this error. It was working fine an hour back and I see that something got updated here in the meanwhile. Can anybody help me please? ``` Collecting orjson Downloading orjson-3.8.6.tar.gz (655 kB) โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” 655.1/655.1 kB 47.3 MB/s eta 0:00:00 Installing build dependencies: started Installing build dependencies: finished with status 'done' Getting requirements to build wheel: started Getting requirements to build wheel: finished with status 'done' Preparing metadata (pyproject.toml): started Preparing metadata (pyproject.toml): finished with status 'error' error: subprocess-exited-with-error ร— Preparing metadata (pyproject.toml) did not run successfully. โ”‚ exit code: 1 โ•ฐโ”€> [6 lines of output] Cargo, the Rust package manager, is not installed or is not on PATH. This package requires Rust and Cargo to compile extensions. Install it through the system's package manager or via https://rustup.rs/ Checking for Rust toolchain.... [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. error: metadata-generation-failed ร— Encountered error while generating package metadata. โ•ฐโ”€> See above for output. ```
closed
2023-02-09T15:27:54Z
2023-02-10T09:40:59Z
https://github.com/ijl/orjson/issues/341
[]
saicharan0112
7
modelscope/data-juicer
data-visualization
81
[Bug]: NameError: name 'fingerprint_warnings' is not defined TypeError: cannot pickle 'OpenCC' object
### Before Reporting ๆŠฅๅ‘Šไน‹ๅ‰ - [X] I have pulled the latest code of main branch to run again and the bug still existed. ๆˆ‘ๅทฒ็ปๆ‹‰ๅ–ไบ†ไธปๅˆ†ๆ”ฏไธŠๆœ€ๆ–ฐ็š„ไปฃ็ ๏ผŒ้‡ๆ–ฐ่ฟ่กŒไน‹ๅŽ๏ผŒ้—ฎ้ข˜ไปไธ่ƒฝ่งฃๅ†ณใ€‚ - [X] I have read the [README](https://github.com/alibaba/data-juicer/blob/main/README.md) carefully and no error occurred during the installation process. (Otherwise, we recommend that you can ask a question using the Question template) ๆˆ‘ๅทฒ็ปไป”็ป†้˜…่ฏปไบ† [README](https://github.com/alibaba/data-juicer/blob/main/README_ZH.md) ไธŠ็š„ๆ“ไฝœๆŒ‡ๅผ•๏ผŒๅนถไธ”ๅœจๅฎ‰่ฃ…่ฟ‡็จ‹ไธญๆฒกๆœ‰้”™่ฏฏๅ‘็”Ÿใ€‚๏ผˆๅฆๅˆ™๏ผŒๆˆ‘ไปฌๅปบ่ฎฎๆ‚จไฝฟ็”จQuestionๆจกๆฟๅ‘ๆˆ‘ไปฌ่ฟ›่กŒๆ้—ฎ๏ผ‰ ### Search before reporting ๅ…ˆๆœ็ดข๏ผŒๅ†ๆŠฅๅ‘Š - [X] I have searched the Data-Juicer [issues](https://github.com/alibaba/data-juicer/issues) and found no similar bugs. ๆˆ‘ๅทฒ็ปๅœจ [issueๅˆ—่กจ](https://github.com/alibaba/data-juicer/issues) ไธญๆœ็ดขไฝ†ๆ˜ฏๆฒกๆœ‰ๅ‘็Žฐ็ฑปไผผ็š„bugๆŠฅๅ‘Šใ€‚ ### OS ็ณป็ปŸ ubuntu ### Installation Method ๅฎ‰่ฃ…ๆ–นๅผ pip ### Data-Juicer Version Data-Juicer็‰ˆๆœฌ v0.1.2 ### Python Version Python็‰ˆๆœฌ 3.8 ### Describe the bug ๆ่ฟฐ่ฟ™ไธชbug ![image](https://github.com/alibaba/data-juicer/assets/66729924/2f12490e-efe4-462f-aed1-390a44ac7275) ### To Reproduce ๅฆ‚ไฝ•ๅค็Žฐ ``` # configs/demo/process.yaml # global parameters project_name: 'demo-process' dataset_path: './demos/data/demo-dataset.jsonl' # path to your dataset directory or file np: 4 # number of subprocess to process your dataset export_path: './outputs/demo-process/demo-processed.jsonl' # process schedule # a list of several process operators with their arguments process: # - language_id_score_filter: # lang: 'zh' - chinese_convert_mapper: mode: 's2t' ``` ``` python tools/process_data.py --config configs/demo/process.yaml ``` ๆŠฅ้”™ๅฆ‚ไธŠ้ขๆˆชๅ›พ ไฝฟ็”จไปฅไธ‹ๅ•็‹ฌ่„šๆœฌๆต‹่ฏ•ๆ˜ฏๆญฃๅธธ็š„๏ผš ``` from data_juicer.ops.mapper.chinese_convert_mapper import ChineseConvertMapper text = {"text": "ๆˆ‘ๅœจ้ฆฌ่ทฏ้‚Šๅซฒ็ท‘่ฎŠ"} op = ChineseConvertMapper(mode = 't2s') aa = op.process(text) print(aa) ``` ### Configs ้…็ฝฎไฟกๆฏ _No response_ ### Logs ๆŠฅ้”™ๆ—ฅๅฟ— _No response_ ### Screenshots ๆˆชๅ›พ _No response_ ### Additional ้ขๅค–ไฟกๆฏ _No response_
closed
2023-11-16T09:39:32Z
2023-11-17T08:28:45Z
https://github.com/modelscope/data-juicer/issues/81
[ "bug" ]
simplew2011
3
seleniumbase/SeleniumBase
web-scraping
2,759
When using ThreadPoolExecutor, a minor problem occurs.
Now, I'm using like below. ``` python def setup_driver() -> Driver: driver = Driver(browser="chrome", headless=True, no_sandbox=True) return driver ``` ``` python # def crawling(driver): driver = setup_driver() driver.get(url) ... with ThreadPoolExecutor(max_workers=4) as executor: results[A] = executor.submit( crawling ) results[B] = executor.submit( crawling ) results[C] = executor.submit( crawling ) results[D] = executor.submit( crawling ) ``` This doesn't work well in parallel. How can I fix it? It worked fine when using Selenium's Webdriver, but I want to use it(Seleniumbase) because of chromedriver's automatic download feature. ``` python def setup_driver() -> webdriver.Chrome: options = Options() options.add_argument("--headless") options.add_argument("--no-sandbox") options.add_argument("--disable-dev-shm-usage") driver = webdriver.Chrome(service=Service(), options=options) return driver ``` In above, it worked fine in local, but when running in environments like AWS Lambda, the Chrome WebDriver was not automatically installed properly. Thank you.
closed
2024-05-07T06:30:09Z
2024-05-07T13:18:00Z
https://github.com/seleniumbase/SeleniumBase/issues/2759
[ "invalid usage" ]
hwk06023
1
explosion/spaCy
nlp
13,351
Incorrect detection of sentence boundaries, if last sentence missing eos symbol for trf model
## How to reproduce the behaviour ``` In [69]: len(list(spacy.load("en_core_web_trf")("The first sentence. The second sentence. The last one").sents)) Out[69]: 1 <<<<<<<<<<<<<<<<<<<<<< WRONG In [70]: len(list(spacy.load("en_core_web_trf")("The first sentence. The second sentence. The last one.").sents)) Out[70]: 3 In [71]: len(list(spacy.load("en_core_web_sm")("The first sentence. The second sentence. The last one").sents)) Out[71]: 3 In [72]: len(list(spacy.load("en_core_web_sm")("The first sentence. The second sentence. The last one.").sents)) Out[72]: 3 ``` ## Your Environment * Operating System: max os x 10.3 * Python Version Used: 3.11 * spaCy Version Used: 3.7.4 * Environment Information: ``` en_core_web_trf.__version__ >> '3.7.3' en_core_web_sm.__version__ >> '3.7.1' ```
closed
2024-02-25T20:50:50Z
2024-02-27T12:58:36Z
https://github.com/explosion/spaCy/issues/13351
[ "feat / parser", "perf / accuracy" ]
koder-ua
0
widgetti/solara
jupyter
563
Functions run very slow when they are called under use_thread
**Problem** A function runs very slow when it is run under `solara.use_thread`. **How much slow:** On my Mac, it is about 20 times. On Hugging Face (which uses Linux), 26 times slower. **How to reproduce:** I deployed the [code](https://huggingface.co/spaces/hkayabilisim/test_solara_use_thread/blob/main/test_thread.py) to https://huggingface.co/spaces/hkayabilisim/test_solara_use_thread for you to repeat the case. For demonstration, I used the prime checking code here: https://github.com/widgetti/solara/blob/master/solara/website/pages/api/use_thread.py To test the code, use large primes such as 160423, 203789, 364289, 991961, 1203793, 1667321, 3704053. **Screenshot** <img width="770" alt="image" src="https://github.com/widgetti/solara/assets/2515171/a93a6131-2081-4dc0-bc3f-4e89c9265b72">
closed
2024-03-17T19:55:39Z
2024-03-27T05:59:38Z
https://github.com/widgetti/solara/issues/563
[]
hkayabilisim
3
indico/indico
flask
6,735
When using the scroll bar with the mouse in select items in a regform, scrolling does not work
**Describe the bug** If you use a select item in a regform that has enough items to enable scrolling (like country), the user may scroll with the wheel / touchpad, but if the scroll bar is clicked on to scroll "the old way", the options closes instead of allowing to scroll by clicking. **To Reproduce** 1. create an event and a regform, add the country field 2. try to scroll via clicking **Expected behavior** scrolling should be possible via scrollbar **Additional context** [sandbox event.](https://sandbox.getindico.io/event/4588/registrations/1570/)
closed
2025-02-06T18:36:22Z
2025-02-07T10:33:40Z
https://github.com/indico/indico/issues/6735
[ "bug" ]
knusperman
1
joerick/pyinstrument
django
33
Maximum recursion depth with python 3.4
Hi, Thanks for a great tool! I get a maximum recursion depth when running pyinstrument with python 3.4.3, but it completes fine with 2.7.9. Full error log below. I have these versions for both python 3.4.3 and 2.7.9 ``` pyinstrument (2.0.0) pyinstrument-cext (0.1.2) ``` ``` Traceback (most recent call last): File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/runpy.py", line 170, in _run_module_as_main "__main__", mod_spec) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/runpy.py", line 85, in _run_code exec(code, run_globals) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/__main__.py", line 138, in <module> main() File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/__main__.py", line 101, in main f.write(profiler.output(renderer=renderer, **renderer_kwargs)) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/profiler.py", line 93, in output return renderer.render(self.starting_frame(root=root)) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/profiler.py", line 82, in starting_frame return self.first_interesting_frame() File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/profiler.py", line 69, in first_interesting_frame while len(frame.children) <= 1: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time for child in self.children: File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True) RuntimeError: maximum recursion depth exceeded while calling a Python object ```
closed
2018-01-10T04:49:21Z
2018-02-11T16:24:43Z
https://github.com/joerick/pyinstrument/issues/33
[]
ksahlin
3
deezer/spleeter
deep-learning
258
Tensorflow/stream_executor/cuda/cuda_dnn.cc:329 Could not create cudnn handle
I have been getting the above error when running the train command. Spleeter is running on a Windows 10 laptop core i7 gen 9. With a nvidia geforce gtx 1660 ti. A friend has lent me their gaming laptop to do these experiments. I have installed CUDA toolset 10.0 (i have also tried with 10.1 and 10.2). I have copied the cuDNN files as described by the nvidia documentation. I installed Anaconda3 spleeter was installed using conda (conda install -c conda-forge spleeter-gpu) The separation works fine, and runs in a few seconds. This is something to do with the CUDA installation, but I am not sure where to look. Anyone out there who has this working any pointers would be grateful.
closed
2020-02-03T22:22:19Z
2020-02-06T20:30:14Z
https://github.com/deezer/spleeter/issues/258
[ "question" ]
stickyninja3
3
Avaiga/taipy
automation
1,683
[๐Ÿ› BUG] Table width issue with persistent pane and d-flex
### What went wrong? ๐Ÿค” The table takes more than the screen width when having a persistent pane. I am using *with* to create the content of the pane. If I use another page to put it inside the pane, the persistent pane will not even work. ### Expected Behavior This shouldn't change the layout of the application where there is no pane. And if there is a pane, the layout should just change so that the pane appears by the side of the app. ### Steps to Reproduce Issue Create a folder data and put this CSV in it. [modified_supermarkt_sales_plus.csv](https://github.com/user-attachments/files/16664692/modified_supermarkt_sales_plus.csv) Add this code and run it: ```python from taipy.gui import Gui, notify import pandas as pd import taipy.gui.builder as tgb import json # Load and prepare data data = pd.read_csv("data/modified_supermarkt_sales_plus.csv") data["Date"] = pd.to_datetime(data["Date"]) data["Review"] = ["[Review](Review)" for _ in range(len(data))] data["Total ($)"] = data["Total"] data["Total (โ‚ฌ)"] = data["Total"] * 1.2 displayed_data = data.copy() # Initialize state variables with default values show_city_info_pane = True selected_view = "Simple view" selected_currency = "USD" selected_dates = [data["Date"].min().date(), data["Date"].max().date()] selected_prices = [0, 5000] selected_city = "All" selected_product_line = "All" selected_branch = "All" rate_info = "Good" rate_price = "Good" open_dialog_review = False selected_row_for_review = None # Load city information from a JSON file city_info_dict = {} # Function to filter the data based on selected criteria def filter(state): filtered_data = state.data if state.selected_city != "All": filtered_data = filtered_data[filtered_data["City"] == state.selected_city] if state.selected_product_line != "All": filtered_data = filtered_data[ filtered_data["Product_line"] == state.selected_product_line ] if state.selected_branch != "All": filtered_data = filtered_data[filtered_data["Branch"] == state.selected_branch] filtered_data = filtered_data[ (filtered_data["Date"].dt.date >= state.selected_dates[0]) & (filtered_data["Total"] >= state.selected_prices[0]) & (filtered_data["Total"] <= state.selected_prices[1]) ] state.displayed_data = filtered_data state.city_info_partial.update_content(state, build_city_info(state.displayed_data)) # Function to convert the total values based on the selected currency def convert(state): if state.selected_currency == "USD": state.displayed_data["Total"] = state.displayed_data["Total ($)"] elif state.selected_currency == "EUR": state.displayed_data["Total"] = state.displayed_data["Total (โ‚ฌ)"] state.refresh("displayed_data") # Function to handle the review submission def send_review(state, id, payload): state.open_dialog_review = False # Build basic filters section def build_basic_filters(): tgb.text("### Basic Filters", mode="md") tgb.selector( value="{selected_product_line}", lov=["All"] + data["Product_line"].unique().tolist(), dropdown=True, filter=True, label="Product Line", on_change=filter, class_name="fullwidth", ) tgb.selector( value="{selected_city}", lov=["All"] + data["City"].unique().tolist(), dropdown=True, filter=True, label="City", on_change=filter, class_name="fullwidth", ) tgb.selector( value="{selected_branch}", lov=["All"] + data["Branch"].unique().tolist(), dropdown=True, filter=True, label="Branch", on_change=filter, class_name="fullwidth", ) # Build conversion section def build_conversion(): tgb.text("### Conversion", mode="md") tgb.selector( value="{selected_currency}", lov=["USD", "EUR"], dropdown=True, label="Currency", on_change=convert, class_name="fullwidth", ) tgb.text("Date Range") tgb.date_range( "{selected_dates}", label_start="Start", label_end="End", on_change=filter ) tgb.text("Price Range") tgb.slider( "{selected_prices}", min=0, max=5000, on_change=filter, continuous=False, width="100%", ) # Function to handle the review process def open_review(state: State, var_name: str, payload: dict): index = payload["index"] data = getattr(state, var_name).copy() state.selected_row_for_review = data.iloc[index].to_frame().T state.open_dialog_review = True with tgb.Page() as review_page: tgb.text("Rate info", mode="md") tgb.table("{selected_row_for_review}") tgb.selector( value="{rate_info}", lov=["Good", "Bad"], dropdown=True, label="Rate info", class_name="fullwidth", # native in 4.0 ) tgb.text("Rate price", mode="md") tgb.selector( value="{rate_price}", lov=["Good", "Bad"], dropdown=True, label="Rate price", class_name="fullwidth", # native in 4.0 ) # Build city information pane def build_city_info(displayed_data): with tgb.Page() as page: tgb.text("### City Information", mode="md") for city in displayed_data["City"].unique(): with tgb.expandable(title=city, expanded=False): tgb.text( city_info_dict.get(city, "No information available."), mode="md" ) return page # Build the main GUI page with tgb.Page() as page: with tgb.part(class_name="container d-flex"): with tgb.part(): tgb.text("Sales Insights", class_name="h1 text-center") with tgb.layout("1 1 1", gap="20px", columns__mobile="1"): with tgb.part(): build_basic_filters() with tgb.part(): build_conversion() tgb.html("hr") tgb.toggle( value="{selected_view}", lov=["Simple view", "Advanced view", "Raw view"], ) with tgb.part(render="{selected_view=='Raw view'}"): tgb.table( "{data}", on_action=open_review, filter=True, ) with tgb.part(render="{selected_view=='Simple view'}"): tgb.table( "{displayed_data}", columns=["Date", "City", "Product_line", "Total", "Review"], group_by__City=True, group_by__Product_line=True, apply_Total="mean", filter=True, on_action=open_review, ) with tgb.part(render="{selected_view=='Advanced view'}"): tgb.table( "{displayed_data}", columns=[ "City", "Product_line", "Total", "Quantity", "Tax_5%", "Total", "Date", "Review", ], filter=True, on_action=open_review, ) def open_info_pane(state): state.show_city_info_pane = True tgb.button( "City info", on_action=open_info_pane, id="open_pane", ) with tgb.pane( open="{show_city_info_pane}", width="300px", persistent=True, anchor="right", ): tgb.part(partial="{city_info_partial}") tgb.dialog( page="review_page", open="{open_dialog_review}", on_action=send_review, labels=["Cancel", "Send"], width="500px", title="Review the selected row", ) # Define pages for the GUI pages = {"page": page, "review_page": review_page} # Run the GUI application if __name__ == "__main__": gui = Gui(pages=pages) city_info_partial = gui.add_partial(build_city_info(displayed_data)) gui.run(title="Sales", port=2452) ``` ### Version of Taipy develop - 4.0.0.dev0 ### Acceptance Criteria - [ ] Ensure new code is unit tested, and check code coverage is at least 90%. - [ ] Create related issue in taipy-doc for documentation and Release Notes. ### Code of Conduct - [X] I have checked the [existing issues](https://github.com/Avaiga/taipy/issues?q=is%3Aissue+). - [ ] I am willing to work on this issue (optional)
closed
2024-08-19T19:51:35Z
2024-08-23T14:37:33Z
https://github.com/Avaiga/taipy/issues/1683
[ "๐Ÿ–ฐ GUI", "๐Ÿ’ฅMalfunction", "๐ŸŸจ Priority: Medium" ]
FlorianJacta
1
hpcaitech/ColossalAI
deep-learning
6,102
[BUG]: assert grad_chunk.l2_norm is not None
### Is there an existing issue for this bug? - [X] I have searched the existing issues ### ๐Ÿ› Describe the bug Modify the code to adapt to qwen2vl(transformers.Qwen2VLForConditionalGeneration) and find that the loss can be calculated, but partial chunk : grad_chunk.l2_norm is None.........(LLM is ok) ![image](https://github.com/user-attachments/assets/b21cedda-30d3-4aad-9112-4c1ab0597211) modified source code to print more information๏ผš ![image](https://github.com/user-attachments/assets/927e9fd6-7a14-4cac-a3cb-2bf7a098ceaa) result: ![image](https://github.com/user-attachments/assets/c0517e1e-1ab1-498b-9d43-7229270126d1) ### Environment _No response_
closed
2024-10-25T09:55:40Z
2024-11-11T13:22:28Z
https://github.com/hpcaitech/ColossalAI/issues/6102
[ "bug" ]
liangzz1991
12
quantmind/pulsar
asyncio
217
How to communicate between two applications in MutiApp?
i create two application, flask wsgi server and Echo server in MutiApp, can i send data to Echo server from browser? How to communicate between two applications in MutiApp?
closed
2016-03-16T02:58:08Z
2016-03-16T22:54:07Z
https://github.com/quantmind/pulsar/issues/217
[ "question", "moved" ]
bright-pan
2
iterative/dvc
data-science
10,431
dvc repro: pygit2 update breaking changes
# Bug Report <!-- ## Issue name Issue names must follow the pattern `command: description` where the command is the dvc command that you are trying to run. The description should describe the consequence of the bug. Example: `repro: doesn't detect input changes` --> ## Description <!-- A clear and concise description of what the bug is. --> pygit2 renamed GIT_OBJ_COMMIT to GIT_OBJECT_COMMIT (I think) because we get error GIT_OBJ_COMMIT not found. ### Reproduce <!-- Step list of how to reproduce the bug --> install newest dvc , run dvc repro <!-- Example: 1. dvc init 2. Copy dataset.zip to the directory 3. dvc add dataset.zip 4. dvc run -d dataset.zip -o model ./train.sh 5. modify dataset.zip 6. dvc repro --> ### Expected <!-- A clear and concise description of what you expect to happen. --> no crash ### Environment information <!-- This is required to ensure that we can reproduce the bug. --> **Output of `dvc doctor`:** ```console $ dvc doctor ``` **Additional Information (if any):** <!-- Please check https://github.com/iterative/dvc/wiki/Debugging-DVC on ways to gather more information regarding the issue. If applicable, please also provide a `--verbose` output of the command, eg: `dvc add --verbose`. If the issue is regarding the performance, please attach the profiling information and the benchmark comparisons. --> you can fix it by using pip install pygit2==1.14.1
closed
2024-05-19T08:50:29Z
2024-05-22T15:20:48Z
https://github.com/iterative/dvc/issues/10431
[ "bug", "p1-important", "dependencies" ]
SimiPro
3
laurentS/slowapi
fastapi
156
slowapi uses limits==2.8.0 which contains usage of deprecated pkg_resources
**Describe the bug** Our application uses `slowapi==0.1.7` which makes use of `limits==2.8.0` The problem is `limits==2.8.0` uses `pkg_resources` which is deprecated and when we run tests they fail with: ```python src/common/limiter.py:1: in <module> from slowapi.extension import Limiter .nox/test-3-10-item/lib/python3.10/site-packages/slowapi/__init__.py:1: in <module> from .extension import Limiter, _rate_limit_exceeded_handler .nox/test-3-10-item/lib/python3.10/site-packages/slowapi/extension.py:25: in <module> from limits import RateLimitItem # type: ignore .nox/test-3-10-item/lib/python3.10/site-packages/limits/__init__.py:5: in <module> from . import _version, aio, storage, strategies .nox/test-3-10-item/lib/python3.10/site-packages/limits/aio/__init__.py:1: in <module> from . import storage, strategies .nox/test-3-10-item/lib/python3.10/site-packages/limits/aio/storage/__init__.py:6: in <module> from .base import MovingWindowSupport, Storage .nox/test-3-10-item/lib/python3.10/site-packages/limits/aio/storage/base.py:5: in <module> from limits.storage.registry import StorageRegistry .nox/test-3-10-item/lib/python3.10/site-packages/limits/storage/__init__.py:12: in <module> from .base import MovingWindowSupport, Storage .nox/test-3-10-item/lib/python3.10/site-packages/limits/storage/base.py:6: in <module> from limits.util import LazyDependency .nox/test-3-10-item/lib/python3.10/site-packages/limits/util.py:11: in <module> import pkg_resources .nox/test-3-10-item/lib/python3.10/site-packages/pkg_resources/__init__.py:121: in <module> warnings.warn("pkg_resources is deprecated as an API", DeprecationWarning) E DeprecationWarning: pkg_resources is deprecated as an API ``` I tried to manually edit `.nox/test-3-10-item/lib/python3.10/site-packages/limits/util.py` and apply the fix (Remove deprecated use of pkg_resources and switch to importlib_resource) and we don't get the warning anymore. **To Reproduce** On Python 3.10 we get this warning when running tests through nox **Expected behavior** `limits` fixed the issue in 3.3.0 https://github.com/alisaifee/limits/blob/master/HISTORY.rst#v330 `slowapi` should use a more updated version of `limits` **Screenshots** (none) **Your app (please complete the following information):** Deps used: ```python fastapi==0.95.0 limits==2.8.0 slowapi==0.1.7 ``` **Python**: Python 3.10.9 **Additional context** The best would be if `slowapi` could move to `limits>=3.3.0` so this problem would be solved. In alternative, a way to suppress `E DeprecationWarning: pkg_resources is deprecated as an API` could work as well. Please note that this error is not happening on all our devs machines so I suspect other people may have some particular setting which is suppressing these warnings. The problem is that we can't find what it is. Thanks
closed
2023-04-06T16:14:34Z
2023-04-07T11:34:54Z
https://github.com/laurentS/slowapi/issues/156
[]
andreagrandi
0
alteryx/featuretools
scikit-learn
2,119
Adding spark to the latest dependency checker lowered the checker's reported pandas version
Making an issue to track this for now. Possibly pyspark 3.3 will allow us to use pandas 1.4 with pyspark
closed
2022-06-16T22:19:38Z
2022-07-11T18:58:21Z
https://github.com/alteryx/featuretools/issues/2119
[]
rwedge
1
yihong0618/running_page
data-visualization
450
่ฏทๆ•™ไธ€ไธ‹yihong็š„ๅทฅไฝœๆต
ๆˆ‘ๆณจๆ„ๅˆฐ๏ผŒyihong็š„running_page้ƒจ็ฝฒๅœจvercelไธŠใ€‚github actionๆฏๅคฉไผš่ท‘๏ผŒRUN_TYPEๆ˜ฏpassใ€‚็œ‹ๆ—ฅๅฟ—ไนŸๆฒกๆœ‰ๆŠ“ๅ–ๆ•ฐๆฎใ€‚ scriptsไธญ็š„data.dbไนŸๆฒกๆœ‰ๆ›ดๆ–ฐใ€‚ ๆ‰€ไปฅyihongๆ˜ฏๅฆ‚ไฝ•ๆ›ดๆ–ฐๆ•ฐๆฎ็š„ใ€‚ๆœ‰ไธ€ไธช็งๆœ‰ไป“ๅบ“๏ผŸ
closed
2023-07-15T07:48:10Z
2023-07-16T02:38:23Z
https://github.com/yihong0618/running_page/issues/450
[]
agassiyzh
1
davidsandberg/facenet
tensorflow
1,010
็”จ่‡ชๅทฑ็š„ๆ•ฐๆฎๅ’Œ้ข„่ฎญ็ปƒๆจกๅž‹้‡ๆ–ฐ่ฎญ็ปƒtrain_softmax.pyๅ‡บ้”™
TypeError: Cannot create initializer for non-floating point type.
open
2019-04-23T08:00:29Z
2021-07-26T08:02:33Z
https://github.com/davidsandberg/facenet/issues/1010
[]
Galaxy-yyyyy
4
DistrictDataLabs/yellowbrick
scikit-learn
552
Add FAQ section to docs
**Describe the solution you'd like** It would be great to add an FAQ section to the docs with answers to commonly asked questions (e.g., how to set the title; change the size; change the colormap). **Is your feature request related to a problem? Please describe.** I saw in [this issue](https://github.com/DistrictDataLabs/yellowbrick/issues/551) a suggestion for an FAQ; it sounds like it would be helpful for new users to have an FAQ section in the docs.
closed
2018-08-09T16:31:18Z
2018-08-15T02:12:45Z
https://github.com/DistrictDataLabs/yellowbrick/issues/552
[ "type: documentation" ]
peterespinosa
2
JaidedAI/EasyOCR
machine-learning
1,136
1.7.1, Macbook M1 silicon, default Reader cannot read text, have to disable GPU to fix issue
After upgarded to 1.7.1, previously working code not able to work. `reader = easyocr.Reader(['ch_sim','en'])` `result = reader.readtext(file_path)` result will be empty. Must specify `gpu=False` to make code work.
open
2023-09-08T06:02:14Z
2023-09-08T06:02:36Z
https://github.com/JaidedAI/EasyOCR/issues/1136
[]
webdkt
0
hyperspy/hyperspy
data-visualization
3,344
Dark spot locator algorithm
#### Describe the functionality you would like to see. A function that automatically detects the positions and radii of dark spots in an image, and the ability to perform manual corrections by the clicking in a figure to add or remove positions. #### Describe the context I've developed a function in matplotlib that uses the skimage [Laplacian of Gaussian blob detection algorithm](https://scikit-image.org/docs/stable/api/skimage.feature.html#skimage.feature.blob_log) for locating and counting threading dislocations (dark spots) in cathodoluminescence panchromatic images ) I would then use these results to calculate a threading dislocation density in InGaN samples. #### Additional information I've managed to do this with HyperSpy Markers, where the results and marker positions are saved to Signal2D metadata for panchromatic image. However, I'm still working on how to do the manual adjustments with a hyperspy figures. ``` def dark_spot_locator(image, max_sigma=30, min_sigma=12, num_sigma=30, threshold = 0.1, permanent = False, color = 'r', linewidth = 2, invert_img = True): import numpy as np import matplotlib.pyplot as plt import hyperspy.api as hs import lumispy as lum from skimage.util import invert from skimage.feature import blob_log # Assigns image intensity values of Signal2D image_array = image.data # Inverts color as blob_log detects white blobs on black background if invert_img == True: image_array = invert(image_array) # Laplacian of Gaussian algorithm, outputting array of shape [..., 3] # containing dark spot offsets and radii blobs = blob_log(image_array, max_sigma=max_sigma, num_sigma=num_sigma, threshold=threshold) # Scales pixel pos to real space, and multiplies by reccommended factor # of sqrt(2) for radii y = blobs[:, 0] * image.axes_manager[0].scale x = blobs[:, 1]* image.axes_manager[0].scale r = blobs[:, 2] * image.axes_manager[0].scale * np.sqrt(2) # Declares markers object, and takes in offsets and radii found by LoG alogorithm markers = hs.plot.markers.Circles( sizes=r, offsets=np.stack((x, y), axis=1), edgecolor=color, linewidth=linewidth, units = 'width') # Adds markers to Signal2D image.add_marker(markers, permanent = permanent) image = hs.load('Data\M1910-3_pmt_map05.dm4') dark_spot_locator(image, permanent = True, max_sigma=5, min_sigma=0.5, num_sigma=5, threshold = 0.01, invert_img = True) ``` Raw panchromatic image, showing dark spots (threading dislocations) and dark lines (misfit dislocations): ![planar_dis_panchromatic](https://github.com/hyperspy/hyperspy/assets/41931869/87c689de-a4e9-4ddd-88e1-0a5b525ed194) LoG algorithm found positions in matplotlib function and figure: ![planar_dis_log_counting](https://github.com/hyperspy/hyperspy/assets/41931869/ffa39411-31ee-4863-8f40-4791ce8e4479) LoG algorithm found positions (green) + manual corrections make by clicking on incorrect positions to remove and adding other positions (red) ![planar_dis_manual_counting](https://github.com/hyperspy/hyperspy/assets/41931869/108a7bdb-2cb1-4c16-a78d-f74da336d61c) This has definitely been personally useful for my mentioned cathodolumescence data but we (@jlaehne) wonder if there are other potential applications for which it would be worth integrating a feature like this into LumiSpy/HyperSpy. When I load the .dm4 CL image into hyperspy it stores it as a Signal2D. It doesn't look like LumiSpy currently has a signal for storing monochromatic/panchromatic images. We were wondering if a feature like this would be best placed in LumiSpy in which case we would consider how to best implement it with a signal for CL pan/mono images, or if on the otherhand this may be more suited to be upstreamed to HyperSpy.
closed
2024-03-28T17:54:44Z
2024-09-17T15:39:40Z
https://github.com/hyperspy/hyperspy/issues/3344
[]
aidanc151
8
miguelgrinberg/python-socketio
asyncio
868
"KeyError: 0" while calling emit to a room from a celery + gevent task
**Describe the bug** Howdy, I am seeing the following error `KeyError: 0` while calling emit to a room from a celery + gevent task. Thanks in advance! **To Reproduce** Steps to reproduce the behavior: 1. Setup with Celery with gevent and Kombu 2. Send event to room from celery task with gevent 3. receive error on emit Python 3.9 Django / Celery app **Expected behavior** A clear and concise description of what you expected to happen. **Logs** ``` trunk-player-worker_1 | [2022-02-06 17:55:07,941: INFO/MainProcess] Task radio.tasks.broadcast_transmission[fc227c55-2ddc-4fae-86a7-8bc83c9faba9] succeeded in 0.028156599029898643s: None trunk-player_1 | Traceback (most recent call last): trunk-player_1 | File "src/gevent/greenlet.py", line 906, in gevent._gevent_cgreenlet.Greenlet.run trunk-player_1 | File "/usr/local/lib/python3.9/site-packages/socketio/pubsub_manager.py", line 168, in _thread trunk-player_1 | self._handle_emit(data) trunk-player_1 | File "/usr/local/lib/python3.9/site-packages/socketio/pubsub_manager.py", line 116, in _handle_emit trunk-player_1 | super(PubSubManager, self).emit(message['event'], message['data'], trunk-player_1 | File "/usr/local/lib/python3.9/site-packages/socketio/base_manager.py", line 165, in emit trunk-player_1 | for sid, eio_sid in self.get_participants(namespace, room): trunk-player_1 | File "/usr/local/lib/python3.9/site-packages/socketio/base_manager.py", line 43, in get_participants trunk-player_1 | participants = ns[room[0]]._fwdm.copy() if room[0] in ns else {} trunk-player_1 | KeyError: 0 trunk-player_1 | 2022-02-06T23:55:07Z <Thread at 0x7ff4b49339d0: <bound method PubSubManager._thread of <socketio.kombu_manager.KombuManager object at 0x7ff4b5fbcb20>>> failed with KeyError ``` **Additional context** The sio.emit is running in a celery task This also seems to stall the rabbitMQ Queue This also appears to break the UWSGI socket io as it stops responding but stays connected code: ```python @shared_task def broadcast_transmission(event: str, room: str, data: dict): try: mgr = socketio.KombuManager( os.getenv("CELERY_BROKER_URL", "ampq://user:pass@127.0.0.1/") ) sio = socketio.Server( async_mode="gevent", client_manager=mgr, logger=False, engineio_logger=False ) sio.emit(event, data, room=room) except Exception as e: if settings.SEND_TELEMETRY: capture_exception(e) raise(e) ```
closed
2022-02-07T02:01:00Z
2022-02-07T02:15:11Z
https://github.com/miguelgrinberg/python-socketio/issues/868
[]
MaxwellDPS
1
xorbitsai/xorbits
numpy
56
using arrow dtype bugfix
closed
2022-12-08T11:24:38Z
2022-12-08T11:24:59Z
https://github.com/xorbitsai/xorbits/issues/56
[]
UranusSeven
0
AUTOMATIC1111/stable-diffusion-webui
deep-learning
15,879
[Bug]:
### Checklist - [ ] The issue exists after disabling all extensions - [X] The issue exists on a clean installation of webui - [ ] The issue is caused by an extension, but I believe it is caused by a bug in the webui - [X] The issue exists in the current version of the webui - [X] The issue has not been reported before recently - [ ] The issue has been reported before but has not been fixed yet ### What happened? Fresh Installation of automatic 1111. I had been stuck on this error for a while. I swapped through a few versions of pythons(from 3.10.6 to 3.10.11 to 3.12.3.) However, I noticed that at the top it still says "Python 3.10.6 (tags/v3.10.6:9c7b4bd, Aug 1 2022, 21:53:49) [MSC v.1932 64 bit (AMD64)] Version: v1.9.3" even though I swapped to 3.10.11 and 3.12.3. The main error is that it cannot import one file into python. I understand its probably something wrong with my Python or something, could anybody tell me how to fix this? ### Steps to reproduce the problem 1. Download a fresh install Double Click webui-user.bat ### What should have happened? WebUi should have started ### What browsers do you use to access the UI ? _No response_ ### Sysinfo [System.txt](https://github.com/AUTOMATIC1111/stable-diffusion-webui/files/15438075/System.txt) ### Console logs ```Shell Python 3.10.6 (tags/v3.10.6:9c7b4bd, Aug 1 2022, 21:53:49) [MSC v.1932 64 bit (AMD64)] Version: v1.9.3 Commit hash: 1c0a0c4c26f78c32095ebc7f8af82f5c04fca8c0 Launching Web UI with arguments: no module 'xformers'. Processing without... no module 'xformers'. Processing without... No module 'xformers'. Proceeding without it. Traceback (most recent call last): File "C:\Users\Admin\Downloads\sd.webui\webui\launch.py", line 48, in <module> main() File "C:\Users\Admin\Downloads\sd.webui\webui\launch.py", line 44, in main start() File "C:\Users\Admin\Downloads\sd.webui\webui\modules\launch_utils.py", line 465, in start import webui File "C:\Users\Admin\Downloads\sd.webui\webui\webui.py", line 13, in <module> initialize.imports() File "C:\Users\Admin\Downloads\sd.webui\webui\modules\initialize.py", line 39, in imports from modules import processing, gradio_extensons, ui # noqa: F401 File "C:\Users\Admin\Downloads\sd.webui\webui\modules\processing.py", line 18, in <module> import modules.sd_hijack File "C:\Users\Admin\Downloads\sd.webui\webui\modules\sd_hijack.py", line 5, in <module> from modules import devices, sd_hijack_optimizations, shared, script_callbacks, errors, sd_unet, patches File "C:\Users\Admin\Downloads\sd.webui\webui\modules\sd_hijack_optimizations.py", line 13, in <module> from modules.hypernetworks import hypernetwork File "C:\Users\Admin\Downloads\sd.webui\webui\modules\hypernetworks\hypernetwork.py", line 8, in <module> import modules.textual_inversion.dataset File "C:\Users\Admin\Downloads\sd.webui\webui\modules\textual_inversion\dataset.py", line 12, in <module> from modules import devices, shared, images File "C:\Users\Admin\Downloads\sd.webui\webui\modules\images.py", line 22, in <module> from modules import sd_samplers, shared, script_callbacks, errors File "C:\Users\Admin\Downloads\sd.webui\webui\modules\sd_samplers.py", line 5, in <module> from modules import sd_samplers_kdiffusion, sd_samplers_timesteps, sd_samplers_lcm, shared, sd_samplers_common, sd_schedulers File "C:\Users\Admin\Downloads\sd.webui\webui\modules\sd_samplers_kdiffusion.py", line 3, in <module> import k_diffusion.sampling File "C:\Users\Admin\Downloads\sd.webui\webui\repositories\k-diffusion\k_diffusion\__init__.py", line 1, in <module> from . import augmentation, config, evaluation, external, gns, layers, models, sampling, utils File "C:\Users\Admin\Downloads\sd.webui\webui\repositories\k-diffusion\k_diffusion\evaluation.py", line 6, in <module> import clip File "C:\Users\Admin\Downloads\sd.webui\system\python\lib\site-packages\clip\__init__.py", line 1, in <module> from .clip import * File "C:\Users\Admin\Downloads\sd.webui\system\python\lib\site-packages\clip\clip.py", line 6, in <module> from pkg_resources import packaging ImportError: cannot import name 'packaging' from 'pkg_resources' (C:\Users\Admin\Downloads\sd.webui\system\python\lib\site-packages\pkg_resources\__init__.py) Press any key to continue . . . ``` ### Additional information _No response_
open
2024-05-24T17:27:42Z
2024-05-28T18:46:11Z
https://github.com/AUTOMATIC1111/stable-diffusion-webui/issues/15879
[ "bug-report" ]
CrashedPlatz
2
rasbt/watermark
jupyter
8
%install_ext is deprecated...
IPython deprecated the `%install_ext` magic (https://github.com/ipython/ipython/issues/8634). The replacement story is using a package to distribute the extension and then use `%load_ext package.extension` to load the extension. Ah, and shamless plug: if you don't want to work with your own package, there is also now https://github.com/ipython-contrib/IPython-extensions watermark would be a great addition there!
closed
2015-10-09T11:01:48Z
2016-05-21T18:35:32Z
https://github.com/rasbt/watermark/issues/8
[]
jankatins
1
TencentARC/GFPGAN
pytorch
466
T
https://github.com/kubernetes-client/python/issues/991
open
2023-11-10T09:46:22Z
2023-11-10T09:46:22Z
https://github.com/TencentARC/GFPGAN/issues/466
[]
AbdullahAlForman
0
tflearn/tflearn
data-science
703
tflearn DNN examples segmentation fault
I tried out the titanic example, but running gives a segmentation fault. I reduced the number of lines and it will segfault with the following: import tflearn net = tflearn.input_data(shape=[None, 6]) model = tflearn.DNN(net) It does come back and I can print the "model" variable so it seems to occur on exit of this simple script. On the "titanic" example, it will segfault on: model.fit(data, labels, n_epoch=10, batch_size=16, show_metric=True) I traced this to: self.summ_writer = writer_summary( self.tensorboard_dir + run_id, self.session.graph) in helpers/trainers.py. "wrtier_summary" prints as follows: <class 'tensorflow.python.summary.writer.writer.FileWriter'> I have also tried finetuning.py and it segfaults as well. I am running: OS: Fedora 19 (old, but everything else works) Python v2.7.5 tensorflow: v1.0.1 numpy: 1.12.1 scipy: 0.19.0 I am running this in a virtualenv environment and everything was installed with pip in this environment. Is this a known issue with this environment?
open
2017-04-08T02:09:23Z
2017-04-18T02:24:32Z
https://github.com/tflearn/tflearn/issues/703
[]
remkonoteboom
3
Evil0ctal/Douyin_TikTok_Download_API
fastapi
347
[Feature request] Can I get the "paid partnership" information of the tiktok video?
**Is your feature request related to a problem? Please describe.** I found the API can't return the "paid partnership" information of the video. For example, https://www.tiktok.com/@bobbyparrish/video/7249131513927781675, it shows it is a "paid partnership" video, but i can't find this information returned by API. Can you help check this, thanks! ![Screenshot 2024-04-06 122047](https://github.com/Evil0ctal/Douyin_TikTok_Download_API/assets/104165305/3f278328-1ba2-4f00-aaf0-450fef330919)
closed
2024-04-06T17:23:57Z
2024-04-15T01:08:05Z
https://github.com/Evil0ctal/Douyin_TikTok_Download_API/issues/347
[ "enhancement" ]
YepJin
3
pydantic/pydantic
pydantic
11,211
A strange new bug with HttpUrl in v2.10.4
### Initial Checks - [X] I confirm that I'm using Pydantic V2 ### Description I get `'{"a":"a=\'a\'"}'` which is not a correct dump. Everything works fine in v2.9. ### Example Code ```Python from pydantic import HttpUrl, BaseModel class A(BaseModel): a: str class B(BaseModel): a: HttpUrl | A b = B(a=A(a="a")) b.model_dump_json() ``` ### Python, Pydantic & OS Version ```Text pydantic version: 2.10.4 pydantic-core version: 2.27.2 pydantic-core build: profile=release pgo=false install path: D:\Conda\envs\personal\Lib\site-packages\pydantic python version: 3.10.14 | packaged by Anaconda, Inc. | (main, Mar 21 2024, 16:20:14) [MSC v.1916 64 bit (AMD64)] platform: Windows-10-10.0.19045-SP0 related packages: fastapi-0.115.0 pydantic-settings-2.5.2 typing_extensions-4.12.2 commit: unknown ```
closed
2025-01-03T08:38:50Z
2025-01-08T15:56:30Z
https://github.com/pydantic/pydantic/issues/11211
[ "bug V2" ]
atodniAr
2
mwaskom/seaborn
matplotlib
3,046
Support "layer labels" with legend
It should be possible to represent a a _layer_ in the legend, in addition to a _mapping_. An example usecase would be ```python ( so.Plot(data, "x", "y") .add(so.Line(color="r"), so.PolyFit(1), label="1st degree") .add(so.Line(color="b"), so.Polyfit(2), label="2nd degree") .add(so.Dots()) ) ``` We'd like this plot to have a legend with two entries: a red line with the label "1st degree" and a blue line with the label "2nd degree". Complications! - How does this interact with mapped properties? Should this only be allowed if there's no mapping for that layer? Or should we get a legend with the mapped properties set to a "neutral" value (similar to when you have, e.g. two separate legends for color and linewidth)? - How should this legend appear by default? Perhaps as a separate legend panel from the variable legends? Should it always be first? If we want it to have a title, where does that happen? - How would we refer to this legend in TBD legend customization interface? - Implementation-wise, should this expand the functionality of `Plot._update_legend_contents`, or should we separate out methods for "labeled layer" legends and mapping legends? - Is there anything else we might want to (or users might expect us to) use the concept of a "layer label" for?
closed
2022-09-26T01:15:22Z
2023-08-31T13:01:22Z
https://github.com/mwaskom/seaborn/issues/3046
[ "objects-plot", "feature" ]
mwaskom
3
robotframework/robotframework
automation
4,669
Importing custom library not working if module name differ class name
Hello, the documentation says: > Python classes are always inside a module. If the name of a class implementing a library is the same as the name of the module, Robot Framework allows dropping the class name when importing the library. [...] **If the module name and class name are different**, libraries must be taken into use using both module and class names, such as **mymodule.MyLibrary** or parent.submodule.MyLib. Simple reproduction code: **CustomLibrary.py** ```python from robot.api.deco import library, keyword @library class Lib: @keyword def add(self, a, b): return a+b def some_method(a, b): return a+b ``` **test.robot** ``` *** Settings *** Library CustomLibrary.Lib *** Test Cases *** Example Test ${result} add ${1} ${2} Log to Console result = ${result} ``` Both files are on the same level (for simplify demonstration purpose) I'm getting `No keyword with name 'add' found.` error. However changing `Library CustomLibrary.Lib` into `Library CustomLibrary.py` and `add` into `Some Method` works perfectly. I found [some old examples](https://www.youtube.com/watch?v=Y8pD5-SP_0c) that shows that my way should work. Python: 3.11 64bit Windows: 10 64bit robotframework: 6.0.2
closed
2023-03-02T13:29:20Z
2023-03-06T08:29:25Z
https://github.com/robotframework/robotframework/issues/4669
[]
Redysz
10
jupyter/nbgrader
jupyter
938
Accessing scores per question from assignments
First, a question: Is there a way to export grades, or interact with the gradebook so as to get the grades per assignment, per question? As far as I can tell, there is not (although perhaps possible with a custom Exporter?). Unless I've missed something that would currently make this possible, I think this might be a useful feature: - It would allow for better auditing of assignments, seeing where students are having trouble, tracking problem topics, and individual outlier questions (that suggests a problem with the Q). - It would also be useful to deal grading quirks: it would be easier to recalculate total assignment scores including or excluding certain questions, if issues arise (I have this issue right now in which a bad test case can sometimes fail, and I can't find a clear way to exclude this question from the overall grades). Perhaps there is something in the underlying grading that makes this type of logging / exporting difficult, but if anyone knows a suggested way to do this, I would like to try it.
closed
2018-03-01T01:54:15Z
2022-12-05T06:58:28Z
https://github.com/jupyter/nbgrader/issues/938
[ "duplicate" ]
TomDonoghue
9
davidsandberg/facenet
tensorflow
895
Poor finetuning accuracy
I excluded the Logits layer and finetuned the model with 40 identities. The resulting model has got poor accuracy. It detects even unknown person as 100% known person. why so?
closed
2018-10-10T14:18:11Z
2019-04-15T13:51:59Z
https://github.com/davidsandberg/facenet/issues/895
[]
Zumbalamambo
0
ARM-DOE/pyart
data-visualization
682
Add nearest neighbor and most common in ROI in map_gates_to_grid
It would be useful to have a nearest neighbor and most common value within the radius of influence in the gridding routines. I have some idea how to do this in the python implementation (and have tested it, but it is sloooooow), but in the fast c-code code written by @jjhelmus - ummm, yeah. Use case would be gridding hydrometeor identification from polar data (you don't do HID on gridded data - shame! shame!), or for tags related to use of particular rainfall algorithms applied at the polar level, etc.
closed
2017-08-28T18:54:29Z
2018-05-22T13:26:51Z
https://github.com/ARM-DOE/pyart/issues/682
[]
swnesbitt
2
modin-project/modin
data-science
6,828
BUG: The csv file cannot be read if there are square brackets in the csv file path or full path.
### Modin version checks - [X] I have checked that this issue has not already been reported. - [X] I have confirmed this bug exists on the latest released version of Modin. - [ ] I have confirmed this bug exists on the main branch of Modin. (In order to do this you can follow [this guide](https://modin.readthedocs.io/en/stable/getting_started/installation.html#installing-from-the-github-master-branch).) ### Reproducible Example ```python df = pd.read_csv("/home/user/[DE]/[PROJECT]/total_data_231127.csv", names=['seq', 'A', 'B'], usecols=['A','B'], sep="\t", dtype=str, na_values=['\\N'], ) ``` ### Issue Description I thought it was a Korean (UTF-8) problem at first, so I tested it several times to find the error. (Because, I am using the Korean version of Ubuntu 22.04 and Korean is included in the file path.) After removing Korean, I tried to read_csv using the file full path, but the same error occurred. Below is the test screen using square brackets. (Using "MODIN_ENGINE" = "dask") ![modin1](https://github.com/modin-project/modin/assets/43741821/77edf677-efb3-41fb-8790-b52845b9d10e) ![modin2](https://github.com/modin-project/modin/assets/43741821/21589227-5b98-41c7-b492-b992e065a5ac) I tried these tests just in case, but it was no use. ![modin3](https://github.com/modin-project/modin/assets/43741821/264e3b28-0232-4b7d-a48c-66a27a7ff010) ![modin4](https://github.com/modin-project/modin/assets/43741821/71998b92-7d08-4500-a4d4-461cd8fa121c) Currently, I'm using it very well after removing square brackets. However I think it would be better to leave such a report, so I'm writing in a bug report. Thank you. ### Expected Behavior FileNotFoundError: [Errno 2] No such file or directory: '/home/user/[DE]/[PROJECT]/total_data_231127.csv' ### Error Logs <details> ```python-traceback Replace this line with the error backtrace (if applicable). ``` </details> ### Installed Versions <details> - Ubuntu 22.04.2 LTS (Korean) - conda 22.9.0 - jupyter notebook 6.5.3 - pip list print ... dask 2023.5.0 modin 0.23.1.post0 modin-spreadsheet 0.1.2 ... </details>
open
2023-12-15T01:49:32Z
2024-06-28T22:35:39Z
https://github.com/modin-project/modin/issues/6828
[ "bug ๐Ÿฆ—", "External", "P3" ]
JacobKwon
4
darrenburns/posting
automation
50
Support for navigating by direction and custom keybindings
Hello! First off, I love the look and feel of the interface. I'm using the galaxy theme and it's just so beautiful it almost makes me cry. Just one thing about the navigation, I'm personally very fond of having ctrl+hjkl for moving between splits/panes etc. This interface would be complete for me if there was a possibility to set directional navigation. I looked into the codebase, to start with just for setting some keybinds since I couldn't find a way to set it in config, but I also wanted to see how the directional nav could be done, and I'm a little puzzled (disclaimer, I'm not very experienced with Python or Textual). It seems to be quite hard with the structure of the screen -> widgets, to get which widget exists in a certain direction, instead of next/prev. I wanted to just do a hacky solution with `get_widget_by_id` (and finding the ids by `self.walk_children(Widget)`), but it turns out that the `get_widget_by_id` only goes downward in the tree. Do you have any pointers regarding how such an implementation might be done? I could probably spend some time trying to implement it, not guaranteeing that it will be a great implementation though. Perhaps just point me to relevant documentation, and maybe where you'd see such an addition belonging. And if you don't feel like having such an option, then I'd appreciate if you could help me find a hacky solution that could work so that I can have a local fork and use that. Like just having a fixed layout and focusing by ID:s based on where I am located. Another thing, I really think it would be nice to have keymaps as an option in the config file. It'd have to be something like: ```yaml keymaps: main_screen: send_request: ctrl+enter url_bar: focus_next: down ``` But I see that it would require quite a bit of work. Same goes there though, I could try implementing that. But a few pointers would probably help, if you've thought about it perhaps. Thanks for creating a fantastic program!
closed
2024-07-19T21:16:46Z
2024-08-02T22:58:43Z
https://github.com/darrenburns/posting/issues/50
[]
albinjon
0
PokeAPI/pokeapi
api
307
Wrong evolution chain for Wurmple?
[JSON-Response](https://pokeapi.co/api/v2/evolution-chain/135/) To me it looks like the chain shows: wurmple -> silcoon -> ( beautifly | cascoon - > dustox ) I hope this visualization makes it clear.
closed
2017-10-08T00:03:11Z
2017-10-08T16:16:30Z
https://github.com/PokeAPI/pokeapi/issues/307
[ "question" ]
HansWurst90
2
davidsandberg/facenet
tensorflow
964
VGG19 Model
Can some one share VGG19 architecture, modified to run with facenet code?
open
2019-01-29T09:44:52Z
2019-03-20T12:21:00Z
https://github.com/davidsandberg/facenet/issues/964
[]
Shahnawazgrewal
1
cvat-ai/cvat
pytorch
9,219
Bug Report: Missing Label Numbers in CVAT Prevents Setting Default Labels Beyond 10
### Actions before raising this issue - [x] I searched the existing issues and did not find anything similar. - [x] I read/searched [the docs](https://docs.cvat.ai/docs/) ### Steps to Reproduce _No response_ ### Expected Behavior _No response_ ### Possible Solution _No response_ ### Context In CVAT, the Labels section does not display assigned numbers for labels. Because of this, I cannot set a default label using hotkeys (Ctrl + [number]) when I have more than 10 labels. Since only the first 10 labels are accessible via keyboard shortcuts, selecting labels beyond that requires manual selection, which slows down the annotation process. docs: ![Image](https://github.com/user-attachments/assets/be427e60-6109-411b-9133-61a2e0393c28) my env: ![Image](https://github.com/user-attachments/assets/150cf955-4e62-444f-97b0-b3487023bab5) ### Environment ```Markdown ```
open
2025-03-17T08:39:50Z
2025-03-17T08:39:50Z
https://github.com/cvat-ai/cvat/issues/9219
[ "bug" ]
ntl-y
0
Buuntu/fastapi-react
fastapi
44
Add posts to react-admin and db migrations as example
Can serve as an example of how to handle relations
open
2020-05-26T17:40:23Z
2020-05-26T17:40:23Z
https://github.com/Buuntu/fastapi-react/issues/44
[ "enhancement" ]
Buuntu
0
automagica/automagica
automation
30
Could not install in vscode
I got this error while instalilng the pip Could not find a version that satisfies the requirement sty==1.0.0b2 (from Automagica==0.3.1) (from versions: ) No matching distribution found for sty==1.0.0b2 (from Automagica==0.3.1) Can you help me?
closed
2018-09-04T19:28:27Z
2018-11-20T07:56:21Z
https://github.com/automagica/automagica/issues/30
[]
Cemoo
4
polakowo/vectorbt
data-visualization
693
Doc Bug report
https://github.com/polakowo/vectorbt/blob/master/examples/MACDVolume.ipynb <img width="1407" alt="image" src="https://github.com/polakowo/vectorbt/assets/22385710/369bc694-034b-4df6-a8f3-7d0ceb7e9034"> It is disappear
open
2024-03-05T11:45:59Z
2024-03-16T10:52:21Z
https://github.com/polakowo/vectorbt/issues/693
[]
GithubJaimeLee
1
aidlearning/AidLearning-FrameWork
jupyter
188
other tablet archecture @aidlearning . please is there the posibility of having aidlearning apk that can be installed on other phone or tablets with archecture other than aarch64. I have a 32bit tablet and it does not install on it. thanks
closed
2021-09-26T10:03:31Z
2021-10-01T16:45:54Z
https://github.com/aidlearning/AidLearning-FrameWork/issues/188
[]
ichit
1
Asabeneh/30-Days-Of-Python
pandas
157
V
V
closed
2021-04-28T17:10:08Z
2021-07-05T21:58:51Z
https://github.com/Asabeneh/30-Days-Of-Python/issues/157
[]
eknatx
0
manrajgrover/halo
jupyter
46
Disable spinners when not outputting to a TTY
I am using halo for the spinners and it works absolutely fine while running the management command from the command line, however I do not want the spinners while running my test suite. This is cluttering my tests and I want to disable it. I don't see an option to disable it. The only other option is to write the output to a file. Is it possible to disable the spinners when not outputting to a TTY ? ## Description - Disable spinners when not outputting to a TTY ### System settings - Operating System: Ubuntu 16.04.3 LTS - Python version: 3.6.3 - Halo version: 0.0.10
closed
2018-01-24T05:10:31Z
2018-02-20T19:24:14Z
https://github.com/manrajgrover/halo/issues/46
[ "bug" ]
vidyarani-dg
9
xuebinqin/U-2-Net
computer-vision
119
What does u2net_train.py expect in train_data/DUTS/DUTS-TR/DUTS-TR/gt_aug/?
Obviously I'm a complete newbie to ML and all this. I put my jpgs to train_data/DUTS/DUTS-TR/DUTS-TR/im_aug/ and they're recognized, but I'm supposed to put some label png files to train_data/DUTS/DUTS-TR/DUTS-TR/gt_aug/ as this error message says: FileNotFoundError: No such file: '/home/user/Desktop/U-2-Net/train_data/DUTS/DUTS-TR/DUTS-TR/gt_aug/0500.png' What does it expect and how do I create these label files?
open
2020-12-07T14:21:54Z
2021-01-02T19:10:10Z
https://github.com/xuebinqin/U-2-Net/issues/119
[]
coveritytest
4
OpenInterpreter/open-interpreter
python
659
Cryptic error when now API key set in a system
### Describe the bug Interpreter fails if there's no API key set in system variable, even though I self-host the model and don't need any key. ``` Open Interpreter will require approval before running code. Use interpreter -y to bypass this. Press CTRL-C to exit. > hello Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new LiteLLM.Info: If you need to debug this error, use `litellm.set_verbose=True'. Traceback (most recent call last): File "E:\Python\3.10\lib\site-packages\interpreter\core\respond.py", line 44, in respond for chunk in interpreter._llm(messages_for_llm): File "E:\Python\3.10\lib\site-packages\interpreter\llm\convert_to_coding_llm.py", line 19, in coding_llm for chunk in text_llm(messages): File "E:\Python\3.10\lib\site-packages\interpreter\llm\setup_text_llm.py", line 117, in base_llm return litellm.completion(**params) File "E:\Python\3.10\lib\site-packages\litellm\utils.py", line 798, in wrapper raise e File "E:\Python\3.10\lib\site-packages\litellm\utils.py", line 757, in wrapper result = original_function(*args, **kwargs) File "E:\Python\3.10\lib\site-packages\litellm\timeout.py", line 53, in wrapper result = future.result(timeout=local_timeout_duration) File "E:\Python\3.10\lib\concurrent\futures\_base.py", line 458, in result return self.__get_result() File "E:\Python\3.10\lib\concurrent\futures\_base.py", line 403, in __get_result raise self._exception File "E:\Python\3.10\lib\site-packages\litellm\timeout.py", line 42, in async_func return func(*args, **kwargs) File "E:\Python\3.10\lib\site-packages\litellm\main.py", line 1195, in completion raise exception_type( File "E:\Python\3.10\lib\site-packages\litellm\utils.py", line 3173, in exception_type raise e File "E:\Python\3.10\lib\site-packages\litellm\utils.py", line 2530, in exception_type raise original_exception File "E:\Python\3.10\lib\site-packages\litellm\main.py", line 447, in completion raise e File "E:\Python\3.10\lib\site-packages\litellm\main.py", line 429, in completion response = openai.ChatCompletion.create( File "E:\Python\3.10\lib\site-packages\openai\api_resources\chat_completion.py", line 25, in create return super().create(*args, **kwargs) File "E:\Python\3.10\lib\site-packages\openai\api_resources\abstract\engine_api_resource.py", line 151, in create ) = cls.__prepare_create_request( File "E:\Python\3.10\lib\site-packages\openai\api_resources\abstract\engine_api_resource.py", line 108, in __prepare_create_request requestor = api_requestor.APIRequestor( File "E:\Python\3.10\lib\site-packages\openai\api_requestor.py", line 139, in __init__ self.api_key = key or util.default_api_key() File "E:\Python\3.10\lib\site-packages\openai\util.py", line 186, in default_api_key raise openai.error.AuthenticationError( openai.error.AuthenticationError: No API key provided. You can set your API key in code using 'openai.api_key = <API-KEY>', or you can set the environment variable OPENAI_API_KEY=<API-KEY>). If your API key is stored in a file, you can point the openai module at it with 'openai.api_key_path = <PATH>'. You can generate API keys in the OpenAI web interface. See https://platform.openai.com/account/api-keys for details. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "E:\Python\3.10\lib\runpy.py", line 196, in _run_module_as_main return _run_code(code, main_globals, None, File "E:\Python\3.10\lib\runpy.py", line 86, in _run_code exec(code, run_globals) File "E:\Python\3.10\Scripts\interpreter.exe\__main__.py", line 7, in <module> File "E:\Python\3.10\lib\site-packages\interpreter\core\core.py", line 26, in cli cli(self) File "E:\Python\3.10\lib\site-packages\interpreter\cli\cli.py", line 254, in cli interpreter.chat() File "E:\Python\3.10\lib\site-packages\interpreter\core\core.py", line 88, in chat for _ in self._streaming_chat(message=message, display=display): File "E:\Python\3.10\lib\site-packages\interpreter\core\core.py", line 109, in _streaming_chat yield from terminal_interface(self, message) File "E:\Python\3.10\lib\site-packages\interpreter\terminal_interface\terminal_interface.py", line 62, in terminal_interface for chunk in interpreter.chat(message, display=False, stream=True): File "E:\Python\3.10\lib\site-packages\interpreter\core\core.py", line 117, in _streaming_chat yield from self._respond() File "E:\Python\3.10\lib\site-packages\interpreter\core\core.py", line 143, in _respond yield from respond(self) File "E:\Python\3.10\lib\site-packages\interpreter\core\respond.py", line 91, in respond raise Exception(f"{output}\n\nThere might be an issue with your API key(s).\n\nTo reset your API key (we'll use OPENAI_API_KEY for this example, but you may need to reset your ANTHROPIC_API_KEY, HUGGINGFACE_API_KEY, etc):\n Mac/Linux: 'export OPENAI_API_KEY=your-key-here',\n Windows: 'setx OPENAI_API_KEY your-key-here' then restart terminal.\n\n") Exception: Traceback (most recent call last): File "E:\Python\3.10\lib\site-packages\interpreter\core\respond.py", line 44, in respond for chunk in interpreter._llm(messages_for_llm): File "E:\Python\3.10\lib\site-packages\interpreter\llm\convert_to_coding_llm.py", line 19, in coding_llm for chunk in text_llm(messages): File "E:\Python\3.10\lib\site-packages\interpreter\llm\setup_text_llm.py", line 117, in base_llm return litellm.completion(**params) File "E:\Python\3.10\lib\site-packages\litellm\utils.py", line 798, in wrapper raise e File "E:\Python\3.10\lib\site-packages\litellm\utils.py", line 757, in wrapper result = original_function(*args, **kwargs) File "E:\Python\3.10\lib\site-packages\litellm\timeout.py", line 53, in wrapper result = future.result(timeout=local_timeout_duration) File "E:\Python\3.10\lib\concurrent\futures\_base.py", line 458, in result return self.__get_result() File "E:\Python\3.10\lib\concurrent\futures\_base.py", line 403, in __get_result raise self._exception File "E:\Python\3.10\lib\site-packages\litellm\timeout.py", line 42, in async_func return func(*args, **kwargs) File "E:\Python\3.10\lib\site-packages\litellm\main.py", line 1195, in completion raise exception_type( File "E:\Python\3.10\lib\site-packages\litellm\utils.py", line 3173, in exception_type raise e File "E:\Python\3.10\lib\site-packages\litellm\utils.py", line 2530, in exception_type raise original_exception File "E:\Python\3.10\lib\site-packages\litellm\main.py", line 447, in completion raise e File "E:\Python\3.10\lib\site-packages\litellm\main.py", line 429, in completion response = openai.ChatCompletion.create( File "E:\Python\3.10\lib\site-packages\openai\api_resources\chat_completion.py", line 25, in create return super().create(*args, **kwargs) File "E:\Python\3.10\lib\site-packages\openai\api_resources\abstract\engine_api_resource.py", line 151, in create ) = cls.__prepare_create_request( File "E:\Python\3.10\lib\site-packages\openai\api_resources\abstract\engine_api_resource.py", line 108, in __prepare_create_request requestor = api_requestor.APIRequestor( File "E:\Python\3.10\lib\site-packages\openai\api_requestor.py", line 139, in __init__ self.api_key = key or util.default_api_key() File "E:\Python\3.10\lib\site-packages\openai\util.py", line 186, in default_api_key raise openai.error.AuthenticationError( openai.error.AuthenticationError: No API key provided. You can set your API key in code using 'openai.api_key = <API-KEY>', or you can set the environment variable OPENAI_API_KEY=<API-KEY>). If your API key is stored in a file, you can point the openai module at it with 'openai.api_key_path = <PATH>'. You can generate API keys in the OpenAI web interface. See https://platform.openai.com/account/api-keys for details. There might be an issue with your API key(s). To reset your API key (we'll use OPENAI_API_KEY for this example, but you may need to reset your ANTHROPIC_API_KEY, HUGGINGFACE_API_KEY, etc): Mac/Linux: 'export OPENAI_API_KEY=your-key-here', Windo ``` ### Reproduce #### This works `interpreter -ab "http://localhost:1234/v1" -m openai/TheBloke/dolphin-2.1-mistral-7B-GGUF --context_window 4000 --max_tokens 2000` **--api_key NONE** #### This doesn't `interpreter -ab "http://localhost:1234/v1" -m openai/TheBloke/dolphin-2.1-mistral-7B-GGUF --context_window 4000 --max_tokens 2000` ### Expected behavior Open Interpreter should start with default / empty key and connect to the local server, as the key is not required. ### Screenshots _No response_ ### Open Interpreter version 0.1.10 ### Python version 3.10.10 ### Operating System name and version Windows 11 ### Additional context LLM is self-hosted in LM Studio.
closed
2023-10-19T10:13:11Z
2024-02-12T17:40:25Z
https://github.com/OpenInterpreter/open-interpreter/issues/659
[ "Bug", "External" ]
DjWarmonger
6
MaxHalford/prince
scikit-learn
112
ca.transform() for columns: compute principal coordinates for the columns
Hi, How to compute principal coordinates for the columns, where columns are new ones. Indeed, the function ca.transform() is only for rows, not for columns. Thanks, Clรฉlia
closed
2021-05-18T15:36:49Z
2021-05-18T15:50:19Z
https://github.com/MaxHalford/prince/issues/112
[]
clelialopez
0
deepset-ai/haystack
machine-learning
8,239
Update tests in core integration: Amazon bedrock
closed
2024-08-15T14:26:07Z
2024-08-19T13:33:49Z
https://github.com/deepset-ai/haystack/issues/8239
[ "P1" ]
shadeMe
0
koxudaxi/datamodel-code-generator
fastapi
1,941
CLI options `use-standard-collections` and `use-union-operator` are not being applied for input file type `graphql`
**Describe the bug** CLI options `use-standard-collections` and `use-union-operator` are not being applied for input file type `graphql` **To Reproduce** Example schema: ```graphql type DummyType { id: ID! label: String! value: Int } type DummyType2 { list_types: [DummyType] } ``` Used commandline: ``` $ datamodel-codegen --input test.graphql --input-file-type=graphql --output-model-type pydantic_v2.BaseModel --use-standard-collections --use-union-operator --output test_models.py ``` **Expected behavior** `Optional` should be replaced by union operator `|`, e.g. `Optional[str]` should be replaced by `str | None`. `List` should be replaced by standard type `list` **Version:** - OS: MAC OS Sonoma 14.4.1 - Python version: 3.11.9 - datamodel-code-generator version: 0.25.6 **Additional context** Add any other context about the problem here.
closed
2024-04-30T09:01:15Z
2024-07-04T15:21:09Z
https://github.com/koxudaxi/datamodel-code-generator/issues/1941
[]
paultop6
2
kornia/kornia
computer-vision
2,597
Add an alias `average_enpoint_error = aepe` (or viceversa)
i'd also add an alias `average_enpoint_error = aepe` (or viceversa) to have more explicit name functions _Originally posted by @edgarriba in https://github.com/kornia/kornia/pull/2500#discussion_r1335526211_
closed
2023-10-03T08:49:33Z
2023-10-10T22:36:21Z
https://github.com/kornia/kornia/issues/2597
[ "enhancement :rocket:" ]
edgarriba
2
dsdanielpark/Bard-API
api
91
Suggest new chat feature
* A new chat option for each new chat a person wants to do. * The chats should be saved so that users can review them at any time. * The chat threads could be organized by date, topic, or user. * The chat threads could also be tagged with keywords so that users can easily search for them.
closed
2023-07-04T15:30:07Z
2024-02-06T05:54:24Z
https://github.com/dsdanielpark/Bard-API/issues/91
[]
itsmeshashank13
4
psf/black
python
4,396
INTERNAL ERROR while formatting (log included)
<!-- Please make sure that the bug is not already fixed either in newer versions or the current development version. To confirm this, you have three options: 1. Update Black's version if a newer release exists: `pip install -U black` 2. Use the online formatter at <https://black.vercel.app/?version=main>, which will use the latest main branch. Note that the online formatter currently runs on an older version of Python and may not support newer syntax, such as the extended f-string syntax added in Python 3.12. 3. Or run _Black_ on your machine: - create a new virtualenv (make sure it's the same Python version); - clone this repository; - run `pip install -e .[d]`; - run `pip install -r test_requirements.txt` - make sure it's sane by running `python -m pytest`; and - run `black` like you did last time. --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> I was formatting my code when an error broke. The console said that I needed to report it, so here I am **To Reproduce** <!-- Minimal steps to reproduce the behavior with source code and Black's configuration. --> [blk_zse1vd0s.log](https://github.com/user-attachments/files/16110113/blk_zse1vd0s.log) I have no idea how to reproduce it, but I'll attach a log file. The resulting error is: > INTERNAL ERROR: Black produced code that is not equivalent to the source. Please report a bug on https://github.com/psf/black/issues. This diff might be helpful: C:\Users\XXX\AppData\Local\Temp\blk_zse1vd0s.log **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> Not this... **Environment** <!-- Please complete the following information: --> - Black's version: <!-- e.g. [main] --> 24.4.2 - OS and Python version: <!-- e.g. [Linux/Python 3.7.4rc1] --> Win 10, Python 3.10 **Additional context** <!-- Add any other context about the problem here. --> Good luck and thank you :)
closed
2024-07-05T13:15:02Z
2024-07-05T23:36:12Z
https://github.com/psf/black/issues/4396
[ "T: bug", "F: fmtskip" ]
teverus
2
unytics/bigfunctions
data-visualization
33
[new]: `xml2json(xml)`
### Check the idea has not already been suggested - [X] I could not find my idea in [existing issues](https://github.com/unytics/bigfunctions/issues?q=is%3Aissue+is%3Aopen+label%3Anew-bigfunction) ### Edit the title above with self-explanatory function name and argument names - [X] The function name and the argument names I entered in the title above seems self explanatory to me. ### BigFunction Description as it would appear in the documentation convert xml to json string using a js library Inspired from https://github.com/salrashid123/bq-udf-xml ### Examples of (arguments, expected output) as they would appear in the documentation `<a><b>foo</b></a>` --> `{"a": {"b": "foo"}}`
closed
2022-12-07T22:36:44Z
2023-04-08T15:01:56Z
https://github.com/unytics/bigfunctions/issues/33
[ "good first issue", "new-bigfunction" ]
unytics
0
Johnserf-Seed/TikTokDownload
api
249
[Feature] ๅธŒๆœ›ๅŠ ๅ…ฅ้‡ๅ‘ฝๅไธ‹่ฝฝ็š„ๆ–‡ไปถๅŠŸ่ƒฝ
**ๆ่ฟฐๆ‚จๆƒณ่ฆ็š„ๅŠŸ่ƒฝ** TikTokDownload ๅ‘ฝไปค่กŒ่ฟ่กŒๆ—ถๅŽ้ขๅŠ ไธชๅฏ้€‰้€‰้กน๏ผŒ่‡ชๅฎšไน‰ไธ‹่ฝฝๆ–‡ไปถ็š„ๆ–‡ไปถๅ **ๆ่ฟฐไฝ ๆƒณ่ฆๆ›ฟๆข็š„ๅŠŸ่ƒฝๆˆ–้œ€่ฆไฟฎๆ”น็š„่งฃๅ†ณๆ–นๆกˆ**
closed
2022-11-14T10:30:48Z
2022-11-27T11:45:34Z
https://github.com/Johnserf-Seed/TikTokDownload/issues/249
[ "้œ€ๆฑ‚ๅปบ่ฎฎ(enhancement)" ]
SageJustus
0
piskvorky/gensim
nlp
3,367
Phrases.analyse_sentence() performs a greedy search for phrases
<!-- **IMPORTANT**: - Use the [Gensim mailing list](https://groups.google.com/forum/#!forum/gensim) to ask general or usage questions. Github issues are only for bug reports. - Check [Recipes&FAQ](https://github.com/RaRe-Technologies/gensim/wiki/Recipes-&-FAQ) first for common answers. Github bug reports that do not include relevant information and context will be closed without an answer. Thanks! --> #### Problem description When computing phrases, it is desirable that `Phrases.analyze_sentence()` implement a certain look-ahead and return phrases with higher scores. Let's say we have bigrams `('a_b', 0.25)` and `('b_c', 0.45)` and a `s = ['a', 'b', 'c']`. At present, the return value of `phrases.analyze_sentence(s)` is going to be: ``` [('a_b', 0.25), ('c'), None] ``` although, since the second bigram has a higher score, it makes sense for the return value to be: ``` [('a', None), ('b_c'), 0.45] ``` Is there value in implementing this, optimised version of `analyze_sentence()`? _[Update]_ PS: Would it be okay to (eventually) raise a PR for this? #### Versions ``` Linux-4.19.0-21-cloud-amd64-x86_64-with-debian-10.12 Python 3.7.12 | packaged by conda-forge | (default, Oct 26 2021, 06:08:53) [GCC 9.4.0] Bits 64 NumPy 1.19.5 SciPy 1.7.3 gensim 4.2.0 FAST_VERSION 0 ```
open
2022-07-21T01:36:17Z
2022-07-25T18:04:27Z
https://github.com/piskvorky/gensim/issues/3367
[]
chaturv3di
1
ultralytics/ultralytics
machine-learning
19,519
model.train() returns None when training on multiple GPUs.
### Search before asking - [x] I have searched the Ultralytics YOLO [issues](https://github.com/ultralytics/ultralytics/issues) and found no similar bug report. ### Ultralytics YOLO Component Train ### Bug When training the YOLO model (both v11 and v12) on multiple GPUs (n_gpu > 1), the results returned from `model.train()` is always `None`. Hereโ€™s an example code block I used to test this issue: `results = model.train( data = "./dataset.yaml", epochs=10, fraction=0.2, batch=64, imgsz=128, conf=0.3, iou=0.2, device = [0,1], )` Here is a screenshot showing the type of `results` output: ![Image](https://github.com/user-attachments/assets/ff8d1e22-c328-4acd-896c-c8844dd658ce) ### Environment Ultralytics 8.3.82 ๐Ÿš€ Python-3.11.10 torch-2.5.1+cu124 CUDA:0 (Tesla T4, 14931MiB) Setup complete โœ… (4 CPUs, 14.6 GB RAM, 34.2/94.3 GB disk) OS Linux-6.1.123+-x86_64-with-glibc2.35 Environment Linux Python 3.11.10 Install pip Path /opt/conda/lib/python3.11/site-packages/ultralytics RAM 14.64 GB Disk 34.2/94.3 GB CPU Intel Xeon 2.30GHz CPU count 4 GPU Tesla T4, 14931MiB GPU count 2 CUDA 12.4 numpy โœ… 1.26.4<=2.1.1,>=1.23.0 matplotlib โœ… 3.9.2>=3.3.0 opencv-python โœ… 4.11.0.86>=4.6.0 pillow โœ… 10.4.0>=7.1.2 pyyaml โœ… 6.0.2>=5.3.1 requests โœ… 2.32.3>=2.23.0 scipy โœ… 1.14.1>=1.4.1 torch โœ… 2.5.1+cu124>=1.8.0 torch โœ… 2.5.1+cu124!=2.4.0,>=1.8.0; sys_platform == "win32" torchvision โœ… 0.20.1+cu124>=0.9.0 tqdm โœ… 4.67.1>=4.64.0 psutil โœ… 6.1.0 py-cpuinfo โœ… 9.0.0 pandas โœ… 2.1.4>=1.1.4 seaborn โœ… 0.13.2>=0.11.0 ultralytics-thop โœ… 2.0.14>=2.0.0 {'OS': 'Linux-6.1.123+-x86_64-with-glibc2.35', 'Environment': 'Linux', 'Python': '3.11.10', 'Install': 'pip', 'Path': '/opt/conda/lib/python3.11/site-packages/ultralytics', 'RAM': '14.64 GB', 'Disk': '34.2/94.3 GB', 'CPU': 'Intel Xeon 2.30GHz', 'CPU count': 4, 'GPU': 'Tesla T4, 14931MiB', 'GPU count': 2, 'CUDA': '12.4', 'Package Info': {'numpy': 'โœ… 1.26.4<=2.1.1,>=1.23.0', 'matplotlib': 'โœ… 3.9.2>=3.3.0', 'opencv-python': 'โœ… 4.11.0.86>=4.6.0', 'pillow': 'โœ… 10.4.0>=7.1.2', 'pyyaml': 'โœ… 6.0.2>=5.3.1', 'requests': 'โœ… 2.32.3>=2.23.0', 'scipy': 'โœ… 1.14.1>=1.4.1', 'torch': 'โœ… 2.5.1+cu124!=2.4.0,>=1.8.0; sys_platform == "win32"', 'torchvision': 'โœ… 0.20.1+cu124>=0.9.0', 'tqdm': 'โœ… 4.67.1>=4.64.0', 'psutil': 'โœ… 6.1.0', 'py-cpuinfo': 'โœ… 9.0.0', 'pandas': 'โœ… 2.1.4>=1.1.4', 'seaborn': 'โœ… 0.13.2>=0.11.0', 'ultralytics-thop': 'โœ… 2.0.14>=2.0.0'}} ### Minimal Reproducible Example `results = model.train( data = "./dataset.yaml", epochs=10, fraction=0.2, batch=64, imgsz=128, conf=0.3, iou=0.2, device = [0,1], )` `print(type(results))` `print(results.save_dir)` ### Additional _No response_ ### Are you willing to submit a PR? - [ ] Yes I'd like to help by submitting a PR!
closed
2025-03-04T20:59:43Z
2025-03-18T00:13:22Z
https://github.com/ultralytics/ultralytics/issues/19519
[ "bug", "devops" ]
xiaoyu-wen-1118
4
pydantic/FastUI
fastapi
318
How to use a custom Oauth provider (e.g. with demo app)
Would it be possible to add an example custom oauth provider? Iโ€™m having an issue getting it to work just right with fastui.
open
2024-05-22T14:18:33Z
2024-06-27T19:11:51Z
https://github.com/pydantic/FastUI/issues/318
[]
jimkring
2
benlubas/molten-nvim
jupyter
105
[Help] Questions regarding demo & overall usage experience
Hello, I tried molten first time today. I have never had experience with notebooks in nvim before. First impressions are very good and promising. However, I can't understand how can I do some basic things, which were demonstrated in the video attached to the readme. - Currently my cell outputs are only shown when the cursor hovers chunk of code being executed. I don't want it to disappear after I move cursor away. How can I configure it this way? - There is an options to execute line of code, and operator like `<MoltenEvaluateOperator>j20`, which is usable, but clearly not the smoothest workflow for cell execution. Is there an option to execute entire cell? I noticed that in demo it seems to be working this way.
closed
2023-12-27T08:55:17Z
2023-12-27T16:44:11Z
https://github.com/benlubas/molten-nvim/issues/105
[]
V3RGANz
5
benbusby/whoogle-search
flask
280
[QUESTION] How to update herokuapp to new whoogle version via web?
See the title so i am asking how to do it via herokuweb without delete the app
closed
2021-04-11T05:57:37Z
2021-04-12T13:58:14Z
https://github.com/benbusby/whoogle-search/issues/280
[ "question" ]
itsnotsaved
1
scikit-image/scikit-image
computer-vision
7,423
Rolling ball algorithm scales badly with radius
### Description: I was playing with rolling ball background subtraction, and naively tried to invoke `skimage.restoration.rolling_ball(img, radius=50)` on a single-plane 582x200 RGB image. I noticed it eating CPU like crazy, so I did some further testing. Here is what I found: radius | time (s) | s/r -- | -- | -- 5 | 0.627 | 0.1254 10 | 3.83 | 0.383 15 | 10 | 0.6666666667 20 | 20.7 | 1.035 25 | 39.4 | 1.576 30 | 67 | 2.233333333 ![image](https://github.com/scikit-image/scikit-image/assets/556626/8c907401-e9f3-4609-a25f-e77f7d1ad283) Conversely, the following Groovy script in Fiji takes 17 ms to complete on my system with radius=50 on the same image: ```groovy #@ ImagePlus imp #@ Double(value=50) radius import ij.plugin.filter.BackgroundSubtracter start = System.currentTimeMillis() BackgroundSubtracter bs = new BackgroundSubtracter() bs.rollingBallBackground(imp.getProcessor(), radius, false, false, false, false, false) end = System.currentTimeMillis() println(end - start) imp.updateAndDraw() ``` Is this a known limitation of the algorithm as currently implemented in skimage? ### Way to reproduce: ```python img = np.random.random([200, 582, 3]) from skimage.restoration import rolling_ball rolling_ball(img, radius=50) ``` ### Version information: ```Shell 3.11.6 | packaged by conda-forge | (main, Oct 3 2023, 10:40:35) [GCC 12.3.0] Linux-6.5.0-28-generic-x86_64-with-glibc2.38 scikit-image version: 0.22.0 numpy version: 1.26.2 ``` I also experienced the same issue using skimage within Pyodide (not sure which versions, sorry). Given that the default radius for the `rolling_ball` routine as currently written is 100, I feel like I must be doing something wrong?
open
2024-05-15T19:03:07Z
2024-11-14T02:39:16Z
https://github.com/scikit-image/scikit-image/issues/7423
[ ":sleeping: Dormant", ":bug: Bug" ]
ctrueden
11
streamlit/streamlit
deep-learning
10,281
Full page reload when opening dialog
### Checklist - [x] I have searched the [existing issues](https://github.com/streamlit/streamlit/issues) for similar issues. - [x] I added a very descriptive title to this issue. - [x] I have provided sufficient information below to help reproduce this issue. ### Summary Hello, I am currently on the nightly version. I want to show a dialog on my page. The weird thing is that my full page is reloading when I try to show a dialog. It takes a lot of time before the dialog shows up. How can I fix that? It is just a dialog with a single button. No large datasets involved. ### Reproducible Code Example ```Python ``` ### Steps To Reproduce _No response_ ### Expected Behavior _No response_ ### Current Behavior _No response_ ### Is this a regression? - [ ] Yes, this used to work in a previous version. ### Debug info - Streamlit version: - Python version: - Operating System: - Browser: ### Additional Information _No response_
closed
2025-01-29T12:44:13Z
2025-01-30T18:28:13Z
https://github.com/streamlit/streamlit/issues/10281
[ "type:bug", "status:awaiting-user-response" ]
Martijn3161
6
jupyterhub/repo2docker
jupyter
449
Docs page has lost its sidebar
https://repo2docker.readthedocs.io/en/latest/ now doesn't show a menu/sidebar anymore and seems to miss the title below the logo. <img width="843" alt="screen shot 2018-10-19 at 18 01 12" src="https://user-images.githubusercontent.com/1448859/47229855-0b85b780-d3c9-11e8-8ff8-ece8a1b3a872.png"> cc @choldgraf the sphinx templates level 11 wizard ๐Ÿง™โ€โ™‚๏ธ
closed
2018-10-19T16:02:35Z
2018-10-20T06:41:17Z
https://github.com/jupyterhub/repo2docker/issues/449
[]
betatim
1
jina-ai/serve
fastapi
5,966
Endless Waiting executor0 but the service works normally.
the code to run from #5959 ``` DEBUG executor0/rep-0@39476 Setting signal handlers [07/14/23 08:58:52] DEBUG executor0/rep-0@39476 Signal handlers already set DEBUG executor0-replica-set@39272 Waiting for ReplicaSet to start successfully [07/14/23 08:58:52] โ ‹ Waiting ... โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” 0/2 -:--:--DEBUG gateway/rep-0@39477 Setting signal handlers [07/14/23 08:58:52] DEBUG gateway-replica-set@39272 Waiting for ReplicaSet to start successfully [07/14/23 08:58:52] โ ‹ Waiting ... โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” 0/2 -:--:--DEBUG gateway/rep-0@39477 Signal handlers already set DEBUG gateway/rep-0@39477 adding connection for deployment executor0/heads/0 to grpc://0.0.0.0:62917 [07/14/23 08:58:52] DEBUG gateway/rep-0@39477 create_connection connection for executor0 to grpc://0.0.0.0:62917 DEBUG gateway/rep-0@39477 create_connection connection for executor0 to grpc://0.0.0.0:62917 DEBUG gateway/rep-0@39477 connection for deployment executor0/heads/0 to grpc://0.0.0.0:62917 added DEBUG gateway/rep-0@39477 Setting up GRPC server DEBUG gateway/rep-0@39477 Get all endpoints from TopologyGraph DEBUG gateway/rep-0@39477 Running GatewayRuntime warmup DEBUG gateway/rep-0@39477 Getting Endpoints data from executor0 DEBUG gateway/rep-0@39477 starting warmup task for deployment executor0 DEBUG gateway/rep-0@39477 gRPC call to executor0 for EndpointDiscovery errored, with error <AioRpcError of RPC that terminated with: status = StatusCode.UNAVAILABLE details = "failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused" debug_error_string = "UNKNOWN:Failed to pick subchannel {created_time:"2023-07-14T08:58:52.240257381+08:00", children:[UNKNOWN:failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused {grpc_status:14, created_time:"2023-07-14T08:58:52.240253323+08:00"}]}" > and for the 1th time. DEBUG gateway/rep-0@39477 resetting connection for executor0 to 0.0.0.0:62917 DEBUG gateway/rep-0@39477 create_connection connection for executor0 to 0.0.0.0:62917 โ น Waiting executor0 gateway... โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” 0/2 0:00:01DEBUG gateway/rep-0@39477 gRPC call to executor0 for EndpointDiscovery errored, with error <AioRpcError of RPC that terminated with: [07/14/23 08:58:53] status = StatusCode.UNAVAILABLE details = "failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused" debug_error_string = "UNKNOWN:Failed to pick subchannel {created_time:"2023-07-14T08:58:53.241945446+08:00", children:[UNKNOWN:failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused {grpc_status:14, created_time:"2023-07-14T08:58:53.241944256+08:00"}]}" > and for the 2th time. DEBUG gateway/rep-0@39477 resetting connection for executor0 to 0.0.0.0:62917 DEBUG gateway/rep-0@39477 create_connection connection for executor0 to 0.0.0.0:62917 โ ฆ Waiting executor0 gateway... โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” 0/2 0:00:02DEBUG gateway/rep-0@39477 gRPC call to executor0 for EndpointDiscovery errored, with error <AioRpcError of RPC that terminated with: [07/14/23 08:58:55] status = StatusCode.UNAVAILABLE details = "failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused" debug_error_string = "UNKNOWN:Failed to pick subchannel {created_time:"2023-07-14T08:58:55.150302259+08:00", children:[UNKNOWN:failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused {created_time:"2023-07-14T08:58:55.150299709+08:00", grpc_status:14}]}" > and for the 3th time. DEBUG gateway/rep-0@39477 resetting connection for executor0 to 0.0.0.0:62917 DEBUG gateway/rep-0@39477 create_connection connection for executor0 to 0.0.0.0:62917 โ ง Waiting executor0 gateway... โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” 0/2 0:00:05DEBUG gateway/rep-0@39477 gRPC call to executor0 for EndpointDiscovery errored, with error <AioRpcError of RPC that terminated with: [07/14/23 08:58:57] status = StatusCode.UNAVAILABLE details = "failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused" debug_error_string = "UNKNOWN:Failed to pick subchannel {created_time:"2023-07-14T08:58:57.65715233+08:00", children:[UNKNOWN:failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused {created_time:"2023-07-14T08:58:57.657149449+08:00", grpc_status:14}]}" > and for the 4th time. DEBUG gateway/rep-0@39477 gRPC call for executor0 failed, retries exhausted DEBUG gateway/rep-0@39477 resetting connection for executor0 to 0.0.0.0:62917 DEBUG gateway/rep-0@39477 create_connection connection for executor0 to 0.0.0.0:62917 WARNIโ€ฆ gateway/rep-0@39477 Getting endpoints failed: failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused. Waiting for another trial โ  Waiting executor0 gateway... โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” 0/2 0:00:06DEBUG gateway/rep-0@39477 Getting Endpoints data from executor0 [07/14/23 08:58:58] โ ™ Waiting executor0 gateway... โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” 0/2 0:00:09DEBUG gateway/rep-0@39477 gRPC call to executor0 for EndpointDiscovery errored, with error <AioRpcError of RPC that terminated with: [07/14/23 08:59:01] status = StatusCode.UNAVAILABLE details = "failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused" debug_error_string = "UNKNOWN:Failed to pick subchannel {created_time:"2023-07-14T08:59:01.975496864+08:00", children:[UNKNOWN:failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused {grpc_status:14, created_time:"2023-07-14T08:59:01.975494113+08:00"}]}" > and for the 1th time. DEBUG gateway/rep-0@39477 resetting connection for executor0 to 0.0.0.0:62917 DEBUG gateway/rep-0@39477 create_connection connection for executor0 to 0.0.0.0:62917 Loading checkpoint shards: 100%|โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆ| 7/7 [00:11<00:00, 1.67s/it] โ ฆ Waiting executor0 gateway... โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” 0/2 0:00:17DEBUG gateway/rep-0@39477 gRPC call to executor0 for EndpointDiscovery errored, with error <AioRpcError of RPC that terminated with: [07/14/23 08:59:09] status = StatusCode.UNAVAILABLE details = "failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused" debug_error_string = "UNKNOWN:Failed to pick subchannel {created_time:"2023-07-14T08:59:09.637254313+08:00", children:[UNKNOWN:failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused {grpc_status:14, created_time:"2023-07-14T08:59:09.637251379+08:00"}]}" > and for the 2th time. โ ‡ Waiting executor0 gateway... โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” 0/2 0:00:17DEBUG gateway/rep-0@39477 resetting connection for executor0 to 0.0.0.0:62917 DEBUG gateway/rep-0@39477 create_connection connection for executor0 to 0.0.0.0:62917 โ ‹ Waiting executor0 gateway... โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” 0/2 0:00:19DEBUG executor0/rep-0@39476 <__main__.GLMInference object at 0x7f129e307f40> is successfully loaded! [07/14/23 08:59:11] DEBUG executor0/rep-0@39476 Setting up GRPC server INFO executor0/rep-0@39476 start server bound to 0.0.0.0:62917 DEBUG executor0/rep-0@39476 server bound to 0.0.0.0:62917 started DEBUG executor0/rep-0@39476 GRPC server setup successful DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC [07/14/23 08:59:11] DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:12] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:12.556621633+08:00", grpc_status:4}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:13] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:13.666544143+08:00", grpc_status:4}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:14] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:14.774546914+08:00", grpc_status:4}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:15] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {grpc_status:4, created_time:"2023-07-14T08:59:15.882533002+08:00"}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:16] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {grpc_status:4, created_time:"2023-07-14T08:59:16.990531376+08:00"}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC [07/14/23 08:59:17] DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:18] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:18.098532957+08:00", grpc_status:4}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:19] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {grpc_status:4, created_time:"2023-07-14T08:59:19.206557554+08:00"}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:20] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:20.313535613+08:00", grpc_status:4}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC โ ผ Waiting executor0 gateway... โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” 0/2 0:00:29DEBUG executor0/rep-0@39476 got an endpoint discovery request [07/14/23 08:59:21] DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:21] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:21.421539286+08:00", grpc_status:4}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready โ ผ Waiting executor0 gateway... โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” 0/2 0:00:29DEBUG gateway/rep-0@39477 Got all endpoints from TopologyGraph {'/chat', '/chat_manas', '_jina_dry_run_'} [07/14/23 08:59:21] INFO gateway/rep-0@39477 start server bound to 0.0.0.0:19002 DEBUG gateway/rep-0@39477 server bound to 0.0.0.0:19002 started DEBUG gateway/rep-0@39477 GRPC server setup successful DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC DEBUG gateway/rep-0@39272 ready and listening [07/14/23 08:59:21] DEBUG gateway-replica-set@39272 ReplicaSet started successfully [07/14/23 08:59:21] DEBUG gateway@39272 Deployment started successfully [07/14/23 08:59:21] โ ฆ Waiting executor0 gateway... โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ”โ” 0/2 0:00:29DEBUG executor0/rep-0@39476 recv _status request DEBUG gateway/rep-0@39477 completed warmup task in 29.353113174438477s. DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:22] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {grpc_status:4, created_time:"2023-07-14T08:59:22.529546379+08:00"}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:23] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:23.636531418+08:00", grpc_status:4}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:25] status = StatusCode.UNAVAILABLE details = "failed to connect to all addresses; last error: UNKNOWN: HTTP proxy returned response code 407" debug_error_string = "UNKNOWN:Failed to pick subchannel {created_time:"2023-07-14T08:59:25.646676539+08:00", children:[UNKNOWN:failed to connect to all addresses; last error: UNKNOWN: HTTP proxy returned response code 407 {created_time:"2023-07-14T08:59:25.646673591+08:00", grpc_status:14}]}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:26] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {grpc_status:4, created_time:"2023-07-14T08:59:26.754537615+08:00"}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:27] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {grpc_status:4, created_time:"2023-07-14T08:59:27.863538749+08:00"}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:28] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {grpc_status:4, created_time:"2023-07-14T08:59:28.971538289+08:00"}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC [07/14/23 08:59:29] DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:30] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:30.078544313+08:00", grpc_status:4}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:31] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:31.185541927+08:00", grpc_status:4}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:32] status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:32.29353679+08:00", grpc_status:4}" > DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready ```
closed
2023-07-14T01:11:22Z
2023-07-19T06:51:13Z
https://github.com/jina-ai/serve/issues/5966
[]
wqh17101
41
holoviz/panel
plotly
7,015
pyodide-worker raising error with pyodide 0.26.1
I was trying to upgrade the pyodide and wasm code + documentation to latest version. I found that `--pyodide-worker` will not work with latest version of pyodide 0.26.1 and panel 1.5.0b2 ## Reproduce - `pip install panel==1.5.0-b.2` - Follow the instructions in the reference example https://panel.holoviz.org/how_to/wasm/convert.html - after the conversion manually update 0.25.0 to 0.26.1 in the script.js file. See the error in the browser console ![image](https://github.com/user-attachments/assets/d7a99e4a-8767-42ed-aeed-02219ad68e9a) ```bash PythonError: Traceback (most recent call last): File "/lib/python312.zip/_pyodide/_base.py", line 574, in eval_code_async await CodeRunner( File "/lib/python312.zip/_pyodide/_base.py", line 394, in run_async coroutine = eval(self.code, globals, locals) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/lib/python312.zip/contextlib.py", line 158, in __exit__ self.gen.throw(value) File "/lib/python3.12/site-packages/panel/util/parameters.py", line 51, in edit_readonly yield File "<exec>", line 8, in <module> File "/lib/python3.12/site-packages/param/parameterized.py", line 2319, in update restore = dict(self_._update(arg, **kwargs)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/lib/python3.12/site-packages/param/parameterized.py", line 2352, in _update self_._batch_call_watchers() File "/lib/python312.zip/contextlib.py", line 158, in __exit__ self.gen.throw(value) File "/lib/python3.12/site-packages/param/parameterized.py", line 284, in _batch_call_watchers yield File "/lib/python3.12/site-packages/param/parameterized.py", line 2546, in _batch_call_watchers self_._execute_watcher(watcher, events) File "/lib/python3.12/site-packages/param/parameterized.py", line 2506, in _execute_watcher watcher.fn(*args, **kwargs) File "/lib/python3.12/site-packages/panel/reactive.py", line 412, in _param_change self._apply_update(named_events, properties, model, ref) File "/lib/python312.zip/contextlib.py", line 158, in __exit__ self.gen.throw(value) File "/lib/python3.12/site-packages/panel/io/document.py", line 375, in unlocked yield File "/lib/python3.12/site-packages/panel/reactive.py", line 316, in _apply_update self._update_model(events, msg, root, model, doc, comm) File "/lib/python3.12/site-packages/panel/reactive.py", line 344, in _update_model model.update(**msg) File "/lib/python3.12/site-packages/bokeh/core/has_props.py", line 483, in update setattr(self, k, v) File "/lib/python3.12/site-packages/bokeh/core/has_props.py", line 336, in __setattr__ return super().__setattr__(name, value) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/lib/python3.12/site-packages/bokeh/core/property/descriptors.py", line 332, in __set__ self._set(obj, old, value, setter=setter) File "/lib/python3.12/site-packages/bokeh/core/property/descriptors.py", line 620, in _set self._trigger(obj, old, value, hint=hint, setter=setter) File "/lib/python3.12/site-packages/bokeh/core/property/descriptors.py", line 698, in _trigger obj.trigger(self.name, old, value, hint, setter) File "/lib/python3.12/site-packages/bokeh/model/model.py", line 571, in trigger super().trigger(descriptor.name, old, new, hint=hint, setter=setter) File "/lib/python3.12/site-packages/bokeh/util/callback_manager.py", line 186, in trigger self.document.callbacks.notify_change(cast(Model, self), attr, old, new, hint, setter, invoke) File "/lib/python3.12/site-packages/bokeh/document/callbacks.py", line 251, in notify_change self.trigger_on_change(event) File "/lib/python3.12/site-packages/bokeh/document/callbacks.py", line 423, in trigger_on_change invoke_with_curdoc(doc, invoke_callbacks) File "/lib/python312.zip/contextlib.py", line 158, in __exit__ self.gen.throw(value) File "/lib/python3.12/site-packages/bokeh/io/doc.py", line 82, in patch_curdoc yield File "/lib/python3.12/site-packages/bokeh/document/callbacks.py", line 453, in invoke_with_curdoc return f() ^^^ File "/lib/python3.12/site-packages/bokeh/document/callbacks.py", line 422, in invoke_callbacks cb(event) File "/lib/python3.12/site-packages/panel/io/pyodide.py", line 346, in pysync dispatch_fn(json_patch, pyodide.ffi.to_js(buffer_map), msg_id) File "https://mnr-jupyterhub.de-prod.dk/mt-uk/user/masma/vscode/proxy/8000/pyodide/script.js", line 4, in sendPatch pyodide.ffi.JsException: DataCloneError: Failed to execute 'postMessage' on 'DedicatedWorkerGlobalScope': [object Map] could not be cloned. at new_error (pyodide.asm.js:10:9965) at pyodide.asm.wasm:0x16dbfd at pyodide.asm.wasm:0x177431 at _PyEM_TrampolineCall_JS (pyodide.asm.js:10:125894) at pyodide.asm.wasm:0x1c2eaf at pyodide.asm.wasm:0x2c7c0f at pyodide.asm.wasm:0x20a884 at pyodide.asm.wasm:0x1c359c at pyodide.asm.wasm:0x1c38ab at pyodide.asm.wasm:0x1c3929 at pyodide.asm.wasm:0x29e95d at pyodide.asm.wasm:0x2a4f54 at pyodide.asm.wasm:0x1c3a69 at pyodide.asm.wasm:0x1c36d2 at pyodide.asm.wasm:0x176a95 at callPyObjectKwargs (pyodide.asm.js:10:64068) at Module.callPyObjectMaybePromising (pyodide.asm.js:10:65316) at wrapper (pyodide.asm.js:10:27006) at Cn.e.port1.onmessage (pyodide.asm.js:10:102154) ```
closed
2024-07-26T06:02:39Z
2024-08-02T14:43:57Z
https://github.com/holoviz/panel/issues/7015
[]
MarcSkovMadsen
15
PaddlePaddle/models
nlp
5,254
PaddleCVๆจกๅž‹ๅบ“ไธ‹็š„3Dๆฃ€ๆต‹้‡Œ็š„M3D-RPNๆจกๅž‹ๆ–‡ๆกฃๆœ‰้—ฎ้ข˜
่ฟ™ไธช็›ฎๅฝ•ไธ‹็š„ๆ–‡ๆกฃ ๅฟซ้€Ÿๅผ€ๅง‹็›ฎๅฝ•ไธ‹็š„cd M3D-RPNๅŽ ่ฟ่กŒln -s /path/to/kitti dataset/kittiๆ็คบ๏ผš ln:failed to create symbolic link "dataset/kitti":ๆฒกๆœ‰่ฟ™ไธช็›ฎๅฝ• ๅฆๅค–๏ผšๆœฌๆจกๅž‹ๆ˜ฏๅœจ1.8ไธ‹ๅผ€ๅ‘็š„๏ผŒไป€ไนˆๆ—ถๅ€™่ƒฝๆไพ›ไธช2.0็š„็‰ˆๆœฌๅ‘ข
open
2021-02-01T03:26:28Z
2024-02-26T05:09:17Z
https://github.com/PaddlePaddle/models/issues/5254
[]
Sqhttwl
2
K3D-tools/K3D-jupyter
jupyter
59
better check/warning on data inconsistency
`import k3d import numpy as np plot = k3d.plot() plot += k3d.points(np.array([[0,0,0], [0,0,2.5]]), colors = [0xff0000]) plot.display() ` Produces an empty plot. The user mistake is that one color is given for 2 points. It should checked and error displayed.
closed
2017-06-13T12:26:27Z
2017-10-22T06:30:35Z
https://github.com/K3D-tools/K3D-jupyter/issues/59
[]
marcinofulus
3
ultralytics/yolov5
deep-learning
12,504
fine tune problem
> @joryoker you can quickly fine-tune your trained custom dataset model by using transfer learning. Transfer learning allows you to build upon your existing model by leveraging the knowledge it has gained from the initial training. > > To perform fine-tuning with your 10 new images, follow these steps: > > 1. Load your previously trained model and weights. > 2. Freeze the layers of the model except for the last few layers. > 3. Train the model using your 10 new images for a few epochs, allowing it to adapt to the new data. > 4. Optionally, you can gradually unfreeze more layers and continue training if needed. > > By fine-tuning with a smaller dataset, you can save considerable training time while still improving the model's performance on the new images. Remember to use a learning rate appropriate for fine-tuning. > > If you need more detailed guidance, the [YOLOv5 official tutorials](https://docs.ultralytics.com/yolov5/tutorials/) provide step-by-step instructions on various topics including transfer learning. > > Let me know if you have any further questions or need additional assistance! I'm a little confused. My initial data set has 30,000 pieces of data, and then I added dozens of new data for fine tune (based on the model trained on the previous initial data set). The new data is similarity as the initial data set. , and the final accuracy of the training is very high, but for new data, the detection results are not as good as the retrained model, and there are always a few error detections. Moreover, when the model is trained, the accuracy has high fluctuations, which is not as smooth as retraining. Is it because there is too little data and the impact on loss is relatively small? If that's the case, I still need to retrain. The base model already use "strip_optimizer(weights)" make requires_grad = False,and use lower learning rate(0.0001) to fine tune. ![0](https://github.com/ultralytics/yolov5/assets/52022528/805c13ca-6e85-4a4f-a741-a3be336df853) _Originally posted by @BossCrab-jyj in https://github.com/ultralytics/yolov5/issues/12069#issuecomment-1855075732_
closed
2023-12-14T05:38:34Z
2023-12-18T01:28:06Z
https://github.com/ultralytics/yolov5/issues/12504
[]
BossCrab-jyj
0
cupy/cupy
numpy
8,718
`import cupyx.scipy.signal` emits FutureWarning
### Description The warning should not be emitted. ### To Reproduce ```py >>> import cupyx.scipy.signal /home/maehashi/Development/cupy/cupyx/jit/_interface.py:173: FutureWarning: cupyx.jit.rawkernel is experimental. The interface can change in the future. cupy._util.experimental('cupyx.jit.rawkernel') ``` ### Installation Built from GitHub source ### Environment ``` # Paste the output here ``` ### Additional Information _No response_
open
2024-11-02T04:03:47Z
2024-11-05T04:44:35Z
https://github.com/cupy/cupy/issues/8718
[ "cat:bug", "prio:high" ]
kmaehashi
0
ultralytics/ultralytics
computer-vision
19,832
Is there an error in the padding of the validation set? Is it allowed not to pad to imgsz?
### Search before asking - [x] I have searched the Ultralytics YOLO [issues](https://github.com/ultralytics/ultralytics/issues) and [discussions](https://github.com/orgs/ultralytics/discussions) and found no similar questions. ### Question I found that the training set was resized directly, but the validation set was scaled to keep the aspect ratio. The training set is easy to understand, but the validation set is not padded to the imgsz size after being scaled proportionally, but padded to 24 in length and 32 in width. Is this a bug or intentional? I did not find the code for the validation set to do this step.I did it very simply, and accidentally discovered that the anchor was not 80*80. I further discovered that the image size of the validation set was not transformed according to imgsz, for example, imgsz (320, 320), the image size (720, 1280), the actual size was (224, 352), it was not filled to 320, but 32 more, and then 224 did not make up 320. I don't know if it's intentional, like allowing not to pad to the imgsz size, but it's important to me where the code is doing the extra padding and why. ![Image](https://github.com/user-attachments/assets/f3956b12-d733-4ce4-ab82-59c1e8252a31) ### Additional _No response_
open
2025-03-23T10:34:15Z
2025-03-23T15:57:33Z
https://github.com/ultralytics/ultralytics/issues/19832
[ "question", "detect" ]
Hua2544
10
tqdm/tqdm
pandas
1,231
Using generator with process_map crashes, despite passing `total` parameter to tqdm_class
When using process_map on a generator, it crashes if `len` can't be called. For example: ```python โ†ช cd /var/www/cl/docker/nginx/ && sudo docker-compose exec cl-python /opt/courtlistener/manage.py dump_anon_docket_html --offset 74850 --processes 4 ; bush "Done making html" 0%| | 0/2703724 [0Traceback (most recent call last): File "/opt/courtlistener/manage.py", line 15, in <module> main() File "/opt/courtlistener/manage.py", line 11, in main execute_from_command_line(sys.argv) File "/usr/local/lib/python3.8/site-packages/django/core/management/__init__.py", line 401, in execute_from_command_line utility.execute() File "/usr/local/lib/python3.8/site-packages/django/core/management/__init__.py", line 395, in execute self.fetch_command(subcommand).run_from_argv(self.argv) File "/usr/local/lib/python3.8/site-packages/django/core/management/base.py", line 330, in run_from_argv self.execute(*args, **cmd_options) File "/usr/local/lib/python3.8/site-packages/django/core/management/base.py", line 371, in execute output = self.handle(*args, **options) File "/opt/courtlistener/cl/corpus_importer/management/commands/dump_anon_docket_html.py", line 68, in handle make_html(options) File "/opt/courtlistener/cl/corpus_importer/management/commands/dump_anon_docket_html.py", line 40, in make_html process_map( File "/usr/local/lib/python3.8/site-packages/tqdm/contrib/concurrent.py", line 130, in process_map return _executor_map(ProcessPoolExecutor, fn, *iterables, **tqdm_kwargs) File "/usr/local/lib/python3.8/site-packages/tqdm/contrib/concurrent.py", line 61, in _executor_map kwargs["total"] = len(iterables[0]) TypeError: object of type 'generator' has no len() Sentry is attempting to send 1 pending error messages Waiting up to 2 seconds Press Ctrl-C to quit ``` That's a lot, here's the important part: ```python kwargs["total"] = len(iterables[0]) TypeError: object of type 'generator' has no len() ``` What's funny about that is that I *passed* the total parameter to my `tqdm_class` parameter: ```python progress_bar = tqdm( total=total, ) process_map( _write_item_to_disk, objects, max_workers=options["processes"], tqdm_class=progress_bar, total=total, ) ``` It makes sense when you look in the code though. There, you find that the `process_map` calls `_executor_map` with whatever tqdm_kwargs it gets: ```python return _executor_map(ProcessPoolExecutor, fn, *iterables, **tqdm_kwargs) ``` And `_executor_map` does: ```python if "total" not in kwargs: kwargs["total"] = len(iterables[0]) ``` Which crashes. The fix is probably either: 1. If there's a `tqdm_class` parameter, `_executor_map` should use the `total` parameter it provides, instead of using the kwarg provided to `process_map`; or 2. The documentation should be updated to note that the `total` parameter is a supported argument in the `process_map` function. I think number 2 is easier (no code change), but number 1 makes more sense because it provides a simpler API for how to set the `total` parameter. You could do a combination of the two also: 3. Update the docs to note that total can be provided to the `process_map`, but prefer the `total` parameter that is passed to the `tqdm_class` class, if there is one. Version info: 4.59.0 3.8.8 (default, Mar 12 2021, 19:44:18) [GCC 8.3.0] linux
closed
2021-08-19T21:48:24Z
2021-08-22T21:32:50Z
https://github.com/tqdm/tqdm/issues/1231
[ "p0-bug-critical โ˜ข", "to-merge โ†ฐ", "c1-quick ๐Ÿ•" ]
mlissner
1
huggingface/datasets
pytorch
7,066
One subset per file in repo ?
Right now we consider all the files of a dataset to be the same data, e.g. ``` single_subset_dataset/ โ”œโ”€โ”€ train0.jsonl โ”œโ”€โ”€ train1.jsonl โ””โ”€โ”€ train2.jsonl ``` but in cases like this, each file is actually a different subset of the dataset and should be loaded separately ``` many_subsets_dataset/ โ”œโ”€โ”€ animals.jsonl โ”œโ”€โ”€ trees.jsonl โ””โ”€โ”€ metadata.jsonl ``` It would be nice to detect those subsets automatically using a simple heuristic. For example we can group files together if their paths names are the same except some digits ?
open
2024-07-23T12:43:59Z
2024-07-23T12:43:59Z
https://github.com/huggingface/datasets/issues/7066
[]
lhoestq
0
google/seq2seq
tensorflow
155
InvalidArgumentError while training on GPU
Hi After 1000 steps, when evaluation starts, I get the following error when training on GPU. I did the exact same training with the same data (200.000 training sentences) and parameters (example_config_nmt_small) on CPU without this error. Could anyone point me in the right direction as to what I did wrong exactly? > File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/seq2seq/seq2seq/contrib/seq2seq/decoder.py", line 287, in dynamic_decode swap_memory=swap_memory) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/ops/control_flow_ops.py", line 2605, in while_loop result = context.BuildLoop(cond, body, loop_vars, shape_invariants) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/ops/control_flow_ops.py", line 2438, in BuildLoop pred, body, original_loop_vars, loop_vars, shape_invariants) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/ops/control_flow_ops.py", line 2388, in _BuildLoop body_result = body(*packed_vars_for_body) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/seq2seq/seq2seq/contrib/seq2seq/decoder.py", line 240, in body decoder_finished) = decoder.step(time, inputs, state) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/seq2seq/seq2seq/decoders/attention_decoder.py", line 187, in step time=time_, outputs=outputs, state=cell_state, sample_ids=sample_ids) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/seq2seq/seq2seq/contrib/seq2seq/helper.py", line 134, in next_inputs time=time, outputs=outputs, state=state, sample_ids=sample_ids) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/seq2seq/seq2seq/decoders/attention_decoder.py", line 155, in att_next_inputs name=name) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/seq2seq/seq2seq/contrib/seq2seq/helper.py", line 204, in next_inputs lambda: nest.map_structure(read_from_ta, self._input_tas)) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/ops/control_flow_ops.py", line 1745, in cond _, res_f = context_f.BuildCondBranch(fn2) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/ops/control_flow_ops.py", line 1639, in BuildCondBranch r = fn() File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/seq2seq/seq2seq/contrib/seq2seq/helper.py", line 204, in <lambda> lambda: nest.map_structure(read_from_ta, self._input_tas)) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/util/nest.py", line 302, in map_structure structure[0], [func(*x) for x in entries]) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/util/nest.py", line 302, in <listcomp> structure[0], [func(*x) for x in entries]) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/seq2seq/seq2seq/contrib/seq2seq/helper.py", line 201, in read_from_ta return inp.read(next_time) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/ops/tensor_array_ops.py", line 250, in read name=name) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/ops/gen_data_flow_ops.py", line 2421, in _tensor_array_read_v3 name=name) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/framework/op_def_library.py", line 763, in apply_op op_def=op_def) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/framework/ops.py", line 2327, in create_op original_op=self._default_original_op, op_def=op_def) File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/framework/ops.py", line 1226, in __init__ self._traceback = _extract_stack() > InvalidArgumentError (see above for traceback): Tried to read from index 49 but array size is: 49 > [[Node: model/att_seq2seq/decode/attention_decoder_1/decoder/while/CustomHelperNextInputs/TrainingHelperNextInputs/cond/TensorArrayReadV3 = TensorArrayReadV3[_class=["loc:@model/att_seq2seq/decode/TrainingHelper/TensorArray"], dtype=DT_FLOAT, _device="/job:localhost/replica:0/task:0/cpu:0"](model/att_seq2seq/decode/attention_decoder_1/decoder/while/CustomHelperNextInputs/TrainingHelperNextInputs/cond/TensorArrayReadV3/Switch, model/att_seq2seq/decode/attention_decoder_1/decoder/while/CustomHelperNextInputs/TrainingHelperNextInputs/cond/TensorArrayReadV3/Switch_1/_437, model/att_seq2seq/decode/attention_decoder_1/decoder/while/CustomHelperNextInputs/TrainingHelperNextInputs/cond/TensorArrayReadV3/Switch_2)]] > [[Node: model/att_seq2seq/decode/attention_decoder_1/decoder/while/CustomHelperNextInputs/TrainingHelperNextInputs/cond/TensorArrayReadV3/_423 = _Recv[client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_1011_model/att_seq2seq/decode/attention_decoder_1/decoder/while/CustomHelperNextInputs/TrainingHelperNextInputs/cond/TensorArrayReadV3", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/gpu:0"](^_cloopmodel/att_seq2seq/decode/attention_decoder_1/decoder/while/NextIteration_2/_250)]] Thanks in advance.
closed
2017-04-12T11:13:59Z
2017-04-12T15:09:59Z
https://github.com/google/seq2seq/issues/155
[]
ghost
1
Lightning-AI/pytorch-lightning
deep-learning
20,451
Error: Invalid value for '--accelerator': 'auto' is not one of 'cpu', 'gpu', 'cuda', 'mps', 'tpu'.
### Bug description when run this codes ``` fabric run ./path/to/train.py \ --devices=auto \ --accelerator=auto ``` get problem below ``` Error: Invalid value for '--accelerator': 'auto' is not one of 'cpu', 'gpu', 'cuda', 'mps', 'tpu'. ``` ### What version are you seeing the problem on? v2.4 ### How to reproduce the bug _No response_ ### Error messages and logs ``` # Error messages and logs here please ``` ### Environment <details> <summary>Current environment</summary> ``` #- PyTorch Lightning Version (e.g., 2.4.0): #- PyTorch Version (e.g., 2.4): #- Python version (e.g., 3.12): #- OS (e.g., Linux): #- CUDA/cuDNN version: #- GPU models and configuration: #- How you installed Lightning(`conda`, `pip`, source): ``` </details> ### More info _No response_ cc @borda @justusschock @awaelchli
open
2024-11-26T12:55:57Z
2024-12-07T06:33:03Z
https://github.com/Lightning-AI/pytorch-lightning/issues/20451
[ "feature", "help wanted", "fabric", "ver: 2.4.x" ]
wkd88
2
profusion/sgqlc
graphql
108
sessions support for requests
Can you publish the latest changes to the pypi please? for local development I can use direct github link to the latest commit as dependency I'd like to refer to it from my published package (setup.py doesn't seem to support direct links)
closed
2020-07-10T15:47:09Z
2020-07-10T16:45:52Z
https://github.com/profusion/sgqlc/issues/108
[]
vavalomi
2
slackapi/bolt-python
fastapi
742
Testing slack bot events to our chatbot in Python
Issue: We are currently unable to receive the payloads from the Slackbot. We reached out to Slack Support via Feedback and Aubrey from the support desk executed a test and they notified us about below: `Your configured Event Subscriptions Request URL endpoint is not successfully acknowledging the Slack event callbacks meaning that even after receiving the event callback, your endpoint is not sending back the expected HTTP 200 status code, but rather it's sending back HTTP 403 which indicates that the callback requests were "Forbidden" so it seems that your endpoint is rejecting our event callbacks.` Goal: Our goal is for the slack events to be registered properly in the chatbot, currently we are not receiving it so we are also exploring other ways to be able to test this and identify what is wrong. #### The `slack_bolt` version slack==0.0.2 slack-bolt==1.6.0 slack-sdk==3.18.4 slackblocks==0.2.5 slackclient==2.7.2 slackeventsapi==3.0.1 #### Python runtime version Python 3.7.11 #### OS info #1 SMP Sat Aug 27 02:44:20 EDT 2022 #### Problem: The chatbot does not seem to receive the payload from slack. Current situation: Once we attempt to trigger an event, the chatbot remains the same saying "โšกBolt App is Running" #### What we have tried: โ€ข Reinstalling the application to ensure that all my changes were saved and were reflected properly in my Slack Environment โ€ข Testing the communication by curling through to the URL (it responded with challenge parameter). โ€ข Testing the communication by entering some text via DM / channel communication and opening the homepage. Since we wanted to test the below events, we tried executing appropriate actions for these: - app_home_opened - message.channels - message.groups - message.im ##### For the items that the team asked us for earlier, it is included below: โ€ข Slack API ID: A043X1Y2MK6 โ€ข URL of Slack Workspace: https://docomo-data-strategy.slack.com/ โ€ข What event type are you expecting to get? We subscribed to below events. We wanted to try app_home_opened, message.im, message.channels, etc. ## Requirements Please read the [Contributing guidelines](https://github.com/slackapi/bolt-python/blob/main/.github/contributing.md) and [Code of Conduct](https://slackhq.github.io/code-of-conduct) before creating this issue or pull request. By submitting, you are agreeing to those rules.
closed
2022-10-19T06:36:18Z
2022-10-20T05:55:44Z
https://github.com/slackapi/bolt-python/issues/742
[ "question", "need info" ]
jdimapilis02
2
roboflow/supervision
computer-vision
953
TRACKING AND CENTER ID IN LINE COUNTER SOLUTION
### Search before asking - [X] I have searched the Supervision [issues](https://github.com/roboflow/supervision/issues) and found no similar feature requests. ### Question Hello Supervision, I'm working with your line crossing counting solution I created this .py script, it works in common scenarios and with the camera in the ideal position (which is great), but I'm working in a scenario where the camera is positioned on the ceiling and pointing towards the floor... it can detect it, but the tracking does not remain constant because it is apparently looking for a center ID, which I think is hard to happen as the image is not three dimensional.... how could it adapt the tracking or this center ID for this type of scenario... The model that Iยดm using here itยดs alredy a custom model ![frame_sv](https://github.com/roboflow/supervision/assets/76708662/65509628-00c5-44a0-804e-dd6373a246b1) ### Additional Iยดm using this code to run the inference/counting: import os import numpy as np import supervision as sv from ultralytics import YOLO # Define o caminho do vรญdeo fonte e do vรญdeo alvo #HOME = os.getcwd() SOURCE_VIDEO_PATH = os.path.join( "Identificacao-vendedor.mp4") TARGET_VIDEO_PATH = SOURCE_VIDEO_PATH.replace(".mp4", "_output.mp4") # Inicializaรงรฃo do modelo YOLO MODEL = "best (5).pt" model = YOLO(MODEL) model.fuse() # Define as classes de interesse selected_classes = [0,1] # Exemplo: ID 0 para 'pessoa' # Configuraรงรฃo da zona de linha e anotadores LINE_START = sv.Point(300, 600) LINE_END = sv.Point(292, 50) line_zone = sv.LineZone(start=LINE_START, end=LINE_END) line_zone_annotator = sv.LineZoneAnnotator(thickness=4, text_thickness=4, text_scale=1) # Configuraรงรฃo do rastreador ByteTrack byte_tracker = sv.ByteTrack(track_thresh=0.05, track_buffer=30, match_thresh=0.8, frame_rate=24) # Anotadores adicionais box_annotator = sv.BoxAnnotator(thickness=3, text_thickness=0, text_scale=0.5) trace_annotator = sv.TraceAnnotator(thickness=4, trace_length=50) # Funรงรฃo de callback para processamento de cada frame def callback(frame: np.ndarray, index: int) -> np.ndarray: results = model(frame, verbose=True, device='0', conf=0.015, iou=0.02, imgsz=1280)[0] detections = sv.Detections.from_ultralytics(results) detections = detections[np.isin(detections.class_id, selected_classes)] detections = byte_tracker.update_with_detections(detections) line_zone.trigger(detections) labels = [f"#{tracker_id} {model.model.names[class_id]} {confidence:0.2f}" for _, _, confidence, class_id, tracker_id, _ in detections] annotated_frame = trace_annotator.annotate(scene=frame.copy(), detections=detections) annotated_frame = box_annotator.annotate(scene=annotated_frame, detections=detections, labels=labels) annotated_frame = line_zone_annotator.annotate(annotated_frame, line_counter=line_zone) return annotated_frame # Funรงรฃo para processar o vรญdeo def process_video(source_path: str, target_path: str, callback, debug: bool = False) -> None: source_video_info = sv.VideoInfo.from_video_path(video_path=source_path) with sv.VideoSink(target_path=target_path, video_info=source_video_info) as sink: for index, frame in enumerate(sv.get_video_frames_generator(source_path=source_path)): result_frame = callback(frame, index) sink.write_frame(frame=result_frame) if debug and index == 0: # Exemplo de condiรงรฃo para depuraรงรฃo break # Chamada da funรงรฃo para processar o vรญdeo process_video(source_path=SOURCE_VIDEO_PATH, target_path=TARGET_VIDEO_PATH, callback=callback, debug=False)
closed
2024-02-28T12:03:05Z
2024-02-28T12:12:23Z
https://github.com/roboflow/supervision/issues/953
[ "question" ]
Rasantis
1
OFA-Sys/Chinese-CLIP
nlp
240
ไฝฟ็”จ่‡ชๅทฑๆž„ๅปบ็š„ๆ•ฐๆฎ้›†๏ผŒๅพฎ่ฐƒlossไธไธ‹้™
ๆ‚จๅฅฝ๏ผŒ่ฏทๆ•™ไธ€ไธ‹๏ผŒๆˆ‘ๆƒณ็”จclipๆฅๅพฎ่ฐƒไธ€ไธชๅˆ†็ฑปไปปๅŠกใ€‚ ๅฎž้ชŒไฝฟ็”จไบ†cifar-10ๆž„ๅปบไบ†ไธ€ไธชๆ•ฐๆฎ้›†ใ€‚textไธบ็ฑปๅˆซ็š„ไธญๆ–‡ๅ็งฐ๏ผŒไธ€ๅ…ฑๆœ‰5w็š„่ฎญ็ปƒ้›†ๅ’Œ1w็š„้ชŒ่ฏ้›†ใ€‚ jsonไธญ็š„ๆ ทไพ‹ๅฆ‚ไธ‹ {"text_id":2, "text":"้ธŸ","image_ids":[0]} {"text_id":1, "text":"ๆฑฝ่ฝฆ","image_ids":[4001]} ๆ•ฐๆฎ่ฝฌๆข่ฟ‡็จ‹ๅ’Œ่ฎญ็ปƒ่ฟ‡็จ‹ๆฒกๆœ‰ๆŠฅ้”™ใ€‚ ่ฎญ็ปƒ่ฟ‡็จ‹ไธญ๏ผŒlossๅชๆœ‰็ฌฌไธ€ไธชepochไธ‹้™๏ผŒๅŽ็ปญไธ€็›ด้œ‡่กใ€‚ไธ”ๅœจ่ฎญ็ปƒ้›†ๅ’Œ้ชŒ่ฏ้›†ไธŠๅ‡†็กฎ็އ้ƒฝๅพˆไฝŽ๏ผŒไฝŽไบŽzero-shot็š„ๆ•ˆๆžœใ€‚ ่ฏท้—ฎไธ‹่ฟ™ๅฏ่ƒฝๆ˜ฏไป€ไนˆ้—ฎ้ข˜๏ผŸ
closed
2023-12-27T06:35:20Z
2024-09-20T02:45:16Z
https://github.com/OFA-Sys/Chinese-CLIP/issues/240
[]
LiqunW
2
csurfer/pyheat
matplotlib
6
use pyheat for cgi
I watch the code,find pyheat can't be uesd for cgi
closed
2017-02-13T03:13:29Z
2017-02-21T02:37:58Z
https://github.com/csurfer/pyheat/issues/6
[]
wheat-attack
2
Neoteroi/BlackSheep
asyncio
272
Router blueprint
##### _Note: consider using [Discussions](https://github.com/Neoteroi/BlackSheep/discussions) to open a conversation about new featuresโ€ฆ_ **๐Ÿš€ Feature Request** i don't know, though have tried to check, but i don't know if there is anything like blueprint like of flask or Router like of Fastapi, will really make sense, rather than creating new app instances and mounting it can be really costly. Thanks for blacksheep
closed
2022-08-02T17:30:40Z
2023-07-15T16:39:41Z
https://github.com/Neoteroi/BlackSheep/issues/272
[ "document" ]
Kokoserver
4
healthchecks/healthchecks
django
834
Export and import support
I have a self hosted healthchecks instance and I had to change my VPS last month. I exported all the database and tried to import back but it's not being easy. Maybe if there is a option to export and import in some known format like json or yaml. Is there something like that in the roadmap for this project? If not, maybe I can give a shot.
open
2023-05-30T12:23:29Z
2024-07-29T07:46:45Z
https://github.com/healthchecks/healthchecks/issues/834
[ "feature" ]
pfsmorigo
1
plotly/plotly.py
plotly
4,986
make Kaleido 1.0 (new architecture) the default for static image generation
open
2025-01-28T15:42:59Z
2025-02-13T21:14:59Z
https://github.com/plotly/plotly.py/issues/4986
[ "feature", "P1" ]
gvwilson
0
Hironsan/BossSensor
computer-vision
32
Problems about './store/model.h5'
When I read โ€œboss_train.pyโ€,I found a constant named "FILE_PATH".I can't find its value "./store/model.h5" in any folders.Is it a forein file?
open
2020-01-07T10:55:39Z
2020-01-07T11:08:43Z
https://github.com/Hironsan/BossSensor/issues/32
[]
SCYtelils
1
junyanz/pytorch-CycleGAN-and-pix2pix
deep-learning
1,319
Is it possible to change shape of origin image using cyclegan?
Thanks for your contribution. I am trying to convert the 3D model to anime style, it works well so far. In fake B image, the model outlines the lines on the edges to make it more anime. Also color have change a little. Besides lines and color, I wanna a little bit shape changes. Is it possible?
closed
2021-09-23T04:03:56Z
2021-12-02T21:07:28Z
https://github.com/junyanz/pytorch-CycleGAN-and-pix2pix/issues/1319
[]
codexq123
1
coqui-ai/TTS
pytorch
3,989
[Bug] Install on clean Ubuntu 22.04 fails in SudachiPy build....
### Describe the bug I'm trying to install TTS on an ARM server with a clean version of Ubuntu 22.04 after "apt install python3-pip". Here's the bug in the log section. TIA ### To Reproduce Start with a clean version of 22.04 on an arm server. Run "apt install python3-pip" Run "pip install TTS" ### Expected behavior _No response_ ### Logs ```shell pip install TTS Collecting TTS Using cached TTS-0.22.0.tar.gz (1.7 MB) Installing build dependencies ... done Getting requirements to build wheel ... done Preparing metadata (pyproject.toml) ... done Collecting unidecode>=1.3.2 Using cached Unidecode-1.3.8-py3-none-any.whl (235 kB) Collecting umap-learn>=0.5.1 Using cached umap_learn-0.5.6-py3-none-any.whl (85 kB) Collecting coqpit>=0.0.16 Using cached coqpit-0.0.17-py3-none-any.whl (13 kB) Collecting torch>=2.1 Using cached torch-2.4.1-cp310-cp310-manylinux2014_aarch64.whl (89.7 MB) Collecting bangla Using cached bangla-0.0.2-py2.py3-none-any.whl (6.2 kB) Collecting spacy[ja]>=3 Using cached spacy-3.7.6.tar.gz (1.3 MB) Installing build dependencies ... done Getting requirements to build wheel ... done Installing backend dependencies ... done Preparing metadata (pyproject.toml) ... done Collecting bnnumerizer Using cached bnnumerizer-0.0.2.tar.gz (4.7 kB) Preparing metadata (setup.py) ... done Collecting einops>=0.6.0 Using cached einops-0.8.0-py3-none-any.whl (43 kB) Collecting inflect>=5.6.0 Using cached inflect-7.3.1-py3-none-any.whl (34 kB) Collecting trainer>=0.0.32 Using cached trainer-0.0.36-py3-none-any.whl (51 kB) Collecting flask>=2.0.1 Using cached flask-3.0.3-py3-none-any.whl (101 kB) Collecting scipy>=1.11.2 Using cached scipy-1.14.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (35.6 MB) Collecting packaging>=23.1 Using cached packaging-24.1-py3-none-any.whl (53 kB) Collecting fsspec>=2023.6.0 Using cached fsspec-2024.9.0-py3-none-any.whl (179 kB) Collecting nltk Using cached nltk-3.9.1-py3-none-any.whl (1.5 MB) Collecting gruut[de,es,fr]==2.2.3 Using cached gruut-2.2.3.tar.gz (73 kB) Preparing metadata (setup.py) ... done Collecting anyascii>=0.3.0 Using cached anyascii-0.3.2-py3-none-any.whl (289 kB) Collecting tqdm>=4.64.1 Using cached tqdm-4.66.5-py3-none-any.whl (78 kB) Collecting cython>=0.29.30 Using cached Cython-3.0.11-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (3.5 MB) Collecting transformers>=4.33.0 Using cached transformers-4.44.2-py3-none-any.whl (9.5 MB) Collecting numba>=0.57.0 Using cached numba-0.60.0-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl (3.4 MB) Collecting g2pkk>=0.1.1 Using cached g2pkk-0.1.2-py3-none-any.whl (25 kB) Collecting encodec>=0.1.1 Using cached encodec-0.1.1.tar.gz (3.7 MB) Preparing metadata (setup.py) ... done Collecting pyyaml>=6.0 Using cached PyYAML-6.0.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (718 kB) Collecting scikit-learn>=1.3.0 Using cached scikit_learn-1.5.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (12.5 MB) Collecting aiohttp>=3.8.1 Using cached aiohttp-3.10.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (1.2 MB) Collecting torchaudio Using cached torchaudio-2.4.1-cp310-cp310-manylinux2014_aarch64.whl (1.7 MB) Collecting jamo Using cached jamo-0.4.1-py3-none-any.whl (9.5 kB) Collecting jieba Using cached jieba-0.42.1.tar.gz (19.2 MB) Preparing metadata (setup.py) ... done Collecting pandas<2.0,>=1.4 Using cached pandas-1.5.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (11.4 MB) Collecting bnunicodenormalizer Using cached bnunicodenormalizer-0.1.7-py3-none-any.whl (23 kB) Collecting pypinyin Using cached pypinyin-0.52.0-py2.py3-none-any.whl (833 kB) Collecting numpy==1.22.0 Using cached numpy-1.22.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (13.4 MB) Collecting soundfile>=0.12.0 Using cached soundfile-0.12.1-py2.py3-none-any.whl (24 kB) Collecting pysbd>=0.3.4 Using cached pysbd-0.3.4-py3-none-any.whl (71 kB) Collecting hangul-romanize Using cached hangul_romanize-0.1.0-py3-none-any.whl (4.6 kB) Collecting librosa>=0.10.0 Using cached librosa-0.10.2.post1-py3-none-any.whl (260 kB) Collecting matplotlib>=3.7.0 Using cached matplotlib-3.9.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (8.2 MB) Collecting num2words Using cached num2words-0.5.13-py3-none-any.whl (143 kB) Requirement already satisfied: Babel<3.0.0,>=2.8.0 in /usr/lib/python3/dist-packages (from gruut[de,es,fr]==2.2.3->TTS) (2.8.0) Collecting dateparser~=1.1.0 Using cached dateparser-1.1.8-py2.py3-none-any.whl (293 kB) Collecting gruut-ipa<1.0,>=0.12.0 Using cached gruut-ipa-0.13.0.tar.gz (101 kB) Preparing metadata (setup.py) ... done Collecting gruut_lang_en~=2.0.0 Using cached gruut_lang_en-2.0.1.tar.gz (15.3 MB) Preparing metadata (setup.py) ... done Collecting jsonlines~=1.2.0 Using cached jsonlines-1.2.0-py2.py3-none-any.whl (7.6 kB) Collecting networkx<3.0.0,>=2.5.0 Using cached networkx-2.8.8-py3-none-any.whl (2.0 MB) Collecting python-crfsuite~=0.9.7 Using cached python-crfsuite-0.9.10.tar.gz (478 kB) Preparing metadata (setup.py) ... done Collecting gruut_lang_fr~=2.0.0 Using cached gruut_lang_fr-2.0.2.tar.gz (10.9 MB) Preparing metadata (setup.py) ... done Collecting gruut_lang_de~=2.0.0 Using cached gruut_lang_de-2.0.1.tar.gz (18.1 MB) Preparing metadata (setup.py) ... done Collecting gruut_lang_es~=2.0.0 Using cached gruut_lang_es-2.0.1.tar.gz (31.4 MB) Preparing metadata (setup.py) ... done Collecting async-timeout<5.0,>=4.0 Using cached async_timeout-4.0.3-py3-none-any.whl (5.7 kB) Collecting multidict<7.0,>=4.5 Using cached multidict-6.0.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (126 kB) Collecting frozenlist>=1.1.1 Using cached frozenlist-1.4.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (238 kB) Collecting aiohappyeyeballs>=2.3.0 Using cached aiohappyeyeballs-2.4.0-py3-none-any.whl (12 kB) Requirement already satisfied: attrs>=17.3.0 in /usr/lib/python3/dist-packages (from aiohttp>=3.8.1->TTS) (21.2.0) Collecting yarl<2.0,>=1.0 Using cached yarl-1.9.11-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (462 kB) Collecting aiosignal>=1.1.2 Using cached aiosignal-1.3.1-py3-none-any.whl (7.6 kB) Collecting Jinja2>=3.1.2 Using cached jinja2-3.1.4-py3-none-any.whl (133 kB) Collecting blinker>=1.6.2 Using cached blinker-1.8.2-py3-none-any.whl (9.5 kB) Collecting click>=8.1.3 Using cached click-8.1.7-py3-none-any.whl (97 kB) Collecting itsdangerous>=2.1.2 Using cached itsdangerous-2.2.0-py3-none-any.whl (16 kB) Collecting Werkzeug>=3.0.0 Using cached werkzeug-3.0.4-py3-none-any.whl (227 kB) Requirement already satisfied: more-itertools>=8.5.0 in /usr/lib/python3/dist-packages (from inflect>=5.6.0->TTS) (8.10.0) Collecting typeguard>=4.0.1 Using cached typeguard-4.3.0-py3-none-any.whl (35 kB) Collecting pooch>=1.1 Using cached pooch-1.8.2-py3-none-any.whl (64 kB) Collecting typing-extensions>=4.1.1 Using cached typing_extensions-4.12.2-py3-none-any.whl (37 kB) Collecting soxr>=0.3.2 Using cached soxr-0.5.0.post1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (221 kB) Collecting audioread>=2.1.9 Using cached audioread-3.0.1-py3-none-any.whl (23 kB) Collecting lazy-loader>=0.1 Using cached lazy_loader-0.4-py3-none-any.whl (12 kB) Collecting decorator>=4.3.0 Using cached decorator-5.1.1-py3-none-any.whl (9.1 kB) Collecting librosa>=0.10.0 Using cached librosa-0.10.2-py3-none-any.whl (260 kB) Using cached librosa-0.10.1-py3-none-any.whl (253 kB) Using cached librosa-0.10.0.post2-py3-none-any.whl (253 kB) Collecting pooch<1.7,>=1.0 Using cached pooch-1.6.0-py3-none-any.whl (56 kB) Collecting librosa>=0.10.0 Using cached librosa-0.10.0.post1-py3-none-any.whl (252 kB) Using cached librosa-0.10.0-py3-none-any.whl (252 kB) Collecting joblib>=0.14 Using cached joblib-1.4.2-py3-none-any.whl (301 kB) Collecting msgpack>=1.0 Using cached msgpack-1.0.8-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (376 kB) Collecting matplotlib>=3.7.0 Using cached matplotlib-3.9.1.post1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (8.2 MB) Using cached matplotlib-3.9.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (8.2 MB) Using cached matplotlib-3.8.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (11.4 MB) Collecting kiwisolver>=1.3.1 Using cached kiwisolver-1.4.7-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (1.4 MB) Collecting fonttools>=4.22.0 Using cached fonttools-4.53.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (4.6 MB) Collecting python-dateutil>=2.7 Using cached python_dateutil-2.9.0.post0-py2.py3-none-any.whl (229 kB) Collecting pillow>=8 Using cached pillow-10.4.0-cp310-cp310-manylinux_2_28_aarch64.whl (4.4 MB) Collecting cycler>=0.10 Using cached cycler-0.12.1-py3-none-any.whl (8.3 kB) Collecting contourpy>=1.0.1 Using cached contourpy-1.3.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (308 kB) Requirement already satisfied: pyparsing>=2.3.1 in /usr/lib/python3/dist-packages (from matplotlib>=3.7.0->TTS) (2.4.7) Collecting docopt>=0.6.2 Using cached docopt-0.6.2.tar.gz (25 kB) Preparing metadata (setup.py) ... done Collecting llvmlite<0.44,>=0.43.0dev0 Using cached llvmlite-0.43.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (42.9 MB) Requirement already satisfied: pytz>=2020.1 in /usr/lib/python3/dist-packages (from pandas<2.0,>=1.4->TTS) (2022.1) Collecting threadpoolctl>=3.1.0 Using cached threadpoolctl-3.5.0-py3-none-any.whl (18 kB) Collecting scipy>=1.11.2 Using cached scipy-1.14.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (35.6 MB) Using cached scipy-1.13.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (33.7 MB) Using cached scipy-1.13.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (33.7 MB) Using cached scipy-1.12.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (34.8 MB) Using cached scipy-1.11.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (32.9 MB) Collecting cffi>=1.0 Using cached cffi-1.17.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (448 kB) Collecting cymem<2.1.0,>=2.0.2 Using cached cymem-2.0.8-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (44 kB) Collecting typer<1.0.0,>=0.3.0 Using cached typer-0.12.5-py3-none-any.whl (47 kB) Requirement already satisfied: setuptools in /usr/lib/python3/dist-packages (from spacy[ja]>=3->TTS) (59.6.0) Collecting spacy-loggers<2.0.0,>=1.0.0 Using cached spacy_loggers-1.0.5-py3-none-any.whl (22 kB) Collecting wasabi<1.2.0,>=0.9.1 Using cached wasabi-1.1.3-py3-none-any.whl (27 kB) Requirement already satisfied: requests<3.0.0,>=2.13.0 in /usr/lib/python3/dist-packages (from spacy[ja]>=3->TTS) (2.25.1) Collecting murmurhash<1.1.0,>=0.28.0 Using cached murmurhash-1.0.10-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (29 kB) Collecting pydantic!=1.8,!=1.8.1,<3.0.0,>=1.7.4 Using cached pydantic-2.9.0-py3-none-any.whl (434 kB) Collecting catalogue<2.1.0,>=2.0.6 Using cached catalogue-2.0.10-py3-none-any.whl (17 kB) Collecting spacy-legacy<3.1.0,>=3.0.11 Using cached spacy_legacy-3.0.12-py2.py3-none-any.whl (29 kB) Collecting langcodes<4.0.0,>=3.2.0 Using cached langcodes-3.4.0-py3-none-any.whl (182 kB) Collecting srsly<3.0.0,>=2.4.3 Using cached srsly-2.4.8-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (488 kB) Collecting thinc<8.3.0,>=8.2.2 Using cached thinc-8.2.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (868 kB) Collecting preshed<3.1.0,>=3.0.2 Using cached preshed-3.0.9-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (150 kB) Collecting weasel<0.5.0,>=0.1.0 Using cached weasel-0.4.1-py3-none-any.whl (50 kB) Collecting sudachipy!=0.6.1,>=0.5.2 Using cached SudachiPy-0.6.8.tar.gz (161 kB) Installing build dependencies ... done Getting requirements to build wheel ... error error: subprocess-exited-with-error ร— Getting requirements to build wheel did not run successfully. โ”‚ exit code: 1 โ•ฐโ”€> [36 lines of output] Traceback (most recent call last): File "/usr/lib/python3/dist-packages/pip/_vendor/pep517/in_process/_in_process.py", line 363, in <module> main() File "/usr/lib/python3/dist-packages/pip/_vendor/pep517/in_process/_in_process.py", line 345, in main json_out['return_val'] = hook(**hook_input['kwargs']) File "/usr/lib/python3/dist-packages/pip/_vendor/pep517/in_process/_in_process.py", line 130, in get_requires_for_build_wheel return hook(config_settings) File "/usr/lib/python3/dist-packages/setuptools/build_meta.py", line 162, in get_requires_for_build_wheel return self._get_build_requires( File "/usr/lib/python3/dist-packages/setuptools/build_meta.py", line 143, in _get_build_requires self.run_setup() File "/usr/lib/python3/dist-packages/setuptools/build_meta.py", line 267, in run_setup super(_BuildMetaLegacyBackend, File "/usr/lib/python3/dist-packages/setuptools/build_meta.py", line 158, in run_setup exec(compile(code, __file__, 'exec'), locals()) File "setup.py", line 18, in <module> setup( File "/usr/lib/python3/dist-packages/setuptools/__init__.py", line 153, in setup return distutils.core.setup(**attrs) File "/usr/lib/python3/dist-packages/setuptools/_distutils/core.py", line 109, in setup _setup_distribution = dist = klass(attrs) File "/usr/lib/python3/dist-packages/setuptools/dist.py", line 459, in __init__ _Distribution.__init__( File "/usr/lib/python3/dist-packages/setuptools/_distutils/dist.py", line 293, in __init__ self.finalize_options() File "/usr/lib/python3/dist-packages/setuptools/dist.py", line 836, in finalize_options for ep in sorted(loaded, key=by_order): File "/usr/lib/python3/dist-packages/setuptools/dist.py", line 835, in <lambda> loaded = map(lambda e: e.load(), filtered) File "/usr/lib/python3/dist-packages/pkg_resources/__init__.py", line 2464, in load self.require(*args, **kwargs) File "/usr/lib/python3/dist-packages/pkg_resources/__init__.py", line 2487, in require items = working_set.resolve(reqs, env, installer, extras=self.extras) File "/usr/lib/python3/dist-packages/pkg_resources/__init__.py", line 782, in resolve raise VersionConflict(dist, req).with_context(dependent_req) pkg_resources.VersionConflict: (setuptools 59.6.0 (/usr/lib/python3/dist-packages), Requirement.parse('setuptools>=62.4')) [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. error: subprocess-exited-with-error ร— Getting requirements to build wheel did not run successfully. โ”‚ exit code: 1 โ•ฐโ”€> See above for output. note: This error originates from a subprocess, and is likely not a problem with pip. ``` ### Environment ```shell Not finished. ``` ### Additional context _No response_
closed
2024-09-06T12:47:00Z
2024-12-28T11:58:05Z
https://github.com/coqui-ai/TTS/issues/3989
[ "bug", "wontfix" ]
pwayner
8
miguelgrinberg/microblog
flask
314
Part XVI: Full-Text Search / Constant Redirects
I am enjoying your tutorial, thank you. I just wanted to reference this Issue #219 I am using your files from the Git Repo, and I ran into the same issue #219. I found this via Google Fu. [https://stackoverflow.com/questions/61237524/validating-get-params-with-wtforms-in-flask](https://stackoverflow.com/questions/61237524/validating-get-params-with-wtforms-in-flask) Lead me to update: app/main/forms.py YOURS: ``` if 'csrf_enabled' not in kwargs: kwargs['csrf_enabled'] = False ``` I added: ``` if 'meta' not in kwargs: kwargs['meta'] = {'csrf': False} ``` By doing so allowed "g.search_form.validate" (app/main/routes.py) to return True when I put data into the 'q' field. Until I had that, I would always be redirected. ``` 192.168.1.188 - - [22/Jan/2022 16:57:15] "GET /search?q=crap HTTP/1.1" 302 - 192.168.1.188 - - [22/Jan/2022 16:57:15] "GET /explore HTTP/1.1" 200 - ```
closed
2022-01-22T22:08:02Z
2022-01-23T00:16:18Z
https://github.com/miguelgrinberg/microblog/issues/314
[ "bug" ]
nuonce
2
recommenders-team/recommenders
machine-learning
1,918
[FEATURE] Find a way to run the tests for external contributors after approval from core devs
### Description <!--- Describe your expected feature in detail --> Currently, external contributors can do PRs, but after a core dev approves and executes to run the tests, there is an error: ``` Run azure/login@v1 Error: Az CLI Login failed. Please check the credentials and make sure az is installed on the runner. For more information refer https://aka.ms/create-secrets-for-GitHub-workflows ``` See example #1916 ### Expected behavior with the suggested feature <!--- For example: --> <!--- *Adding algorithm xxx will help people understand more about xxx use case scenarios. --> ### Other Comments Related to https://github.com/microsoft/recommenders/pull/1840
closed
2023-04-11T10:47:26Z
2023-04-11T10:49:12Z
https://github.com/recommenders-team/recommenders/issues/1918
[ "enhancement" ]
miguelgfierro
1
keras-team/keras
data-science
20,686
Error after upgrade from 3.6 to 3.7
I have a model, that use combined input. Using version prior to 3.7 it works very fine, when using this to predict ![image](https://github.com/user-attachments/assets/43bda454-74fd-49a0-b564-06003c4f0507) Dropping the tf.function and using a normal predict also works fine Output is typical something like this ![image](https://github.com/user-attachments/assets/94c7add5-6c06-400a-b071-97bd7b4434e2) Upgrading to keras 3.7, using the same model result in this output ![image](https://github.com/user-attachments/assets/018fe9d2-82c9-40d2-93a0-bd95681c9abf) Do you see the difference? shape[0] for x and b is now None But the real problem is the prediction, that is no longer correct Just to be sure, I print the shape of x and b before clalling the tf.function (1238, 42) (1238, 15) It looks like there is a problem finding the batch size for combined input, so for now I will have to stay at 3.6
open
2024-12-24T16:30:40Z
2024-12-31T13:56:13Z
https://github.com/keras-team/keras/issues/20686
[ "type:Bug" ]
ThorvaldAagaard
4
pallets-eco/flask-sqlalchemy
flask
350
http://flask-sqlalchemy.pocoo.org/2.1/ down?
http://flask-sqlalchemy.pocoo.org/2.1/ down?
closed
2015-11-19T06:08:33Z
2020-12-05T21:31:20Z
https://github.com/pallets-eco/flask-sqlalchemy/issues/350
[]
rlam3
3