repo_name
stringlengths 9
75
| topic
stringclasses 30
values | issue_number
int64 1
203k
| title
stringlengths 1
976
| body
stringlengths 0
254k
| state
stringclasses 2
values | created_at
stringlengths 20
20
| updated_at
stringlengths 20
20
| url
stringlengths 38
105
| labels
sequencelengths 0
9
| user_login
stringlengths 1
39
| comments_count
int64 0
452
|
---|---|---|---|---|---|---|---|---|---|---|---|
HumanSignal/labelImg | deep-learning | 255 | Bug in Verify? | Looks like in in prebuild labelImg_v1.6.0 for windows `Verify Image` not working properly when you press verify and then press next green background still present in next image. | closed | 2018-03-21T14:23:18Z | 2018-05-22T05:53:29Z | https://github.com/HumanSignal/labelImg/issues/255 | [] | mrgloom | 1 |
scikit-hep/awkward | numpy | 2,462 | Running over zero-length arrays, time spent mostly in function overhead | ### Version of Awkward Array
2.2.0
### Description and code to reproduce
N.B.: This is lower priority, but it scales with analysis size/complexity!
When running column optimization in dask awkward a typetracer is passed through the whole analysis code.
This triggers all relevant function calls for the analysis and produces an accurate accounting of data needs which is great, but as analyses grow larger (especially with systematics overhead!) the tracing begins to take quite some time for processing no data.
I understand this is only so reducible, but responsiveness is definitely part of user experience and it would be sad if we lost users because of really excellent core functionality taking too long.
Here is a pyinstrument trace showing what functions stacks are grabbing the most time in a mostly realistic analysis with full systematics:
```
_ ._ __/__ _ _ _ _ _/_ Recorded: 08:25:40 Samples: 3840
/_//_/// /_\ / //_// / //_'/ // Duration: 3.983 CPU time: 3.988
/ _/ v4.3.0
Program: yimu_test5.py
3.983 <module> yimu_test5.py:1
โโ 3.983 necessary_columns dask_awkward/lib/inspect.py:11
โโ 3.983 _necessary_columns dask_awkward/lib/optimize.py:392
โโ 3.979 _get_column_reports dask_awkward/lib/optimize.py:335
โโ 3.900 get_sync dask/local.py:551
โ โโ 3.898 get_async dask/local.py:350
โ โโ 3.382 fire_tasks dask/local.py:452
โ โ โโ 3.331 submit dask/local.py:539
โ โ โโ 3.292 batch_execute_tasks dask/local.py:234
โ โ โโ 3.292 <listcomp> dask/local.py:238
โ โ โโ 3.287 execute_task dask/local.py:214
โ โ โโ 3.279 _execute_task dask/core.py:82
โ โ โโ 3.254 __call__ dask/optimization.py:987
โ โ โโ 3.246 get dask/core.py:128
โ โ โโ 3.211 _execute_task dask/core.py:82
โ โ โโ 1.186 __array_ufunc__ awkward/highlevel.py:1289
โ โ โ โโ 1.138 array_ufunc awkward/_connect/numpy.py:219
โ โ โ โโ 0.496 _array_ufunc_custom_cast awkward/_connect/numpy.py:147
โ โ โ โ โโ 0.449 find_custom_cast awkward/_behavior.py:60
โ โ โ โ โโ 0.271 __iter__ _collections_abc.py:742
โ โ โ โ โ โโ 0.146 [self]
โ โ โ โ โ โโ 0.106 __getitem__ collections/__init__.py:892
โ โ โ โ โโ 0.121 [self]
โ โ โ โโ 0.382 broadcast_and_apply awkward/_broadcasting.py:1078
โ โ โ โ โโ 0.356 apply_step awkward/_broadcasting.py:370
โ โ โ โ โโ 0.336 continuation awkward/_broadcasting.py:1023
โ โ โ โ โโ 0.332 broadcast_any_list awkward/_broadcasting.py:500
โ โ โ โ โโ 0.299 apply_step awkward/_broadcasting.py:370
โ โ โ โ โโ 0.275 continuation awkward/_broadcasting.py:1023
โ โ โ โ โโ 0.222 broadcast_any_list awkward/_broadcasting.py:500
โ โ โ โ โ โโ 0.189 apply_step awkward/_broadcasting.py:370
โ โ โ โ โ โโ 0.126 continuation awkward/_broadcasting.py:1023
โ โ โ โ โ โ โโ 0.105 broadcast_any_list awkward/_broadcasting.py:500
โ โ โ โ โ โ โโ 0.085 apply_step awkward/_broadcasting.py:370
โ โ โ โ โ โ โโ 0.066 action awkward/_connect/numpy.py:228
โ โ โ โ โ โ โโ 0.055 find_ufunc awkward/_behavior.py:109
โ โ โ โ โ โโ 0.055 action awkward/_connect/numpy.py:228
โ โ โ โ โ โโ 0.043 find_ufunc awkward/_behavior.py:109
โ โ โ โ โโ 0.049 broadcast_any_indexed awkward/_broadcasting.py:979
โ โ โ โโ 0.213 recursively_apply awkward/_do.py:20
โ โ โ โโ 0.148 _recursively_apply awkward/contents/listoffsetarray.py:2069
โ โ โ โ โโ 0.144 continuation awkward/contents/listoffsetarray.py:2084
โ โ โ โ โโ 0.106 _recursively_apply awkward/contents/numpyarray.py:1261
โ โ โ โ โโ 0.105 unary_action awkward/_connect/numpy.py:309
โ โ โ โ โโ 0.105 action awkward/_connect/numpy.py:228
โ โ โ โ โโ 0.085 find_ufunc awkward/_behavior.py:109
โ โ โ โ โโ 0.043 __iter__ _collections_abc.py:742
โ โ โ โโ 0.045 _recursively_apply awkward/contents/indexedarray.py:1030
โ โ โ โโ 0.043 continuation awkward/contents/indexedarray.py:1055
โ โ โโ 0.558 __call__ dask_awkward/lib/core.py:1894
โ โ โ โโ 0.407 pt coffea/nanoevents/methods/vector.py:121
โ โ โ โ โโ 0.407 r coffea/nanoevents/methods/vector.py:85
โ โ โ โ โโ 0.361 r2 coffea/nanoevents/methods/vector.py:111
โ โ โ โ โ โโ 0.336 func numpy/lib/mixins.py:18
โ โ โ โ โ โโ 0.335 __array_ufunc__ awkward/highlevel.py:1289
โ โ โ โ โ โโ 0.323 array_ufunc awkward/_connect/numpy.py:219
โ โ โ โ โ โโ 0.222 broadcast_and_apply awkward/_broadcasting.py:1078
โ โ โ โ โ โ โโ 0.213 apply_step awkward/_broadcasting.py:370
โ โ โ โ โ โ โโ 0.205 continuation awkward/_broadcasting.py:1023
โ โ โ โ โ โ โโ 0.203 broadcast_any_list awkward/_broadcasting.py:500
โ โ โ โ โ โ โโ 0.191 apply_step awkward/_broadcasting.py:370
โ โ โ โ โ โ โโ 0.181 continuation awkward/_broadcasting.py:1023
โ โ โ โ โ โ โโ 0.113 broadcast_any_list awkward/_broadcasting.py:500
โ โ โ โ โ โ โ โโ 0.099 apply_step awkward/_broadcasting.py:370
โ โ โ โ โ โ โ โโ 0.087 continuation awkward/_broadcasting.py:1023
โ โ โ โ โ โ โ โโ 0.067 broadcast_any_list awkward/_broadcasting.py:500
โ โ โ โ โ โ โ โโ 0.057 apply_step awkward/_broadcasting.py:370
โ โ โ โ โ โ โโ 0.067 broadcast_any_indexed awkward/_broadcasting.py:979
โ โ โ โ โ โ โโ 0.051 apply_step awkward/_broadcasting.py:370
โ โ โ โ โ โ โโ 0.050 continuation awkward/_broadcasting.py:1023
โ โ โ โ โ โ โโ 0.050 broadcast_any_list awkward/_broadcasting.py:500
โ โ โ โ โ โ โโ 0.048 apply_step awkward/_broadcasting.py:370
โ โ โ โ โ โ โโ 0.048 continuation awkward/_broadcasting.py:1023
โ โ โ โ โ โ โโ 0.048 broadcast_any_indexed awkward/_broadcasting.py:979
โ โ โ โ โ โโ 0.092 _array_ufunc_custom_cast awkward/_connect/numpy.py:147
โ โ โ โ โ โโ 0.088 find_custom_cast awkward/_behavior.py:60
โ โ โ โ โ โโ 0.049 __iter__ _collections_abc.py:742
โ โ โ โ โโ 0.047 __array_ufunc__ awkward/highlevel.py:1289
โ โ โ โ โโ 0.044 array_ufunc awkward/_connect/numpy.py:219
โ โ โ โโ 0.140 eta coffea/nanoevents/methods/vector.py:512
โ โ โ โโ 0.100 r coffea/nanoevents/methods/vector.py:85
โ โ โ โโ 0.088 r2 coffea/nanoevents/methods/vector.py:111
โ โ โ โโ 0.075 func numpy/lib/mixins.py:18
โ โ โ โโ 0.075 __array_ufunc__ awkward/highlevel.py:1289
โ โ โ โโ 0.074 array_ufunc awkward/_connect/numpy.py:219
โ โ โ โโ 0.047 broadcast_and_apply awkward/_broadcasting.py:1078
โ โ โ โโ 0.045 apply_step awkward/_broadcasting.py:370
โ โ โโ 0.408 apply dask/utils.py:41
โ โ โ โโ 0.201 sum awkward/operations/ak_sum.py:13
โ โ โ โ โโ 0.180 _impl awkward/operations/ak_sum.py:267
โ โ โ โ โโ 0.174 reduce awkward/_do.py:259
โ โ โ โ โโ 0.126 _reduce_next awkward/contents/listoffsetarray.py:1464
โ โ โ โ โโ 0.040 _reduce_next awkward/contents/listoffsetarray.py:1464
โ โ โ โโ 0.061 count awkward/operations/ak_count.py:12
โ โ โ โ โโ 0.052 _impl awkward/operations/ak_count.py:110
โ โ โ โ โโ 0.049 reduce awkward/_do.py:259
โ โ โ โโ 0.057 mean awkward/operations/ak_mean.py:13
โ โ โ โโ 0.057 _impl awkward/operations/ak_mean.py:152
โ โ โโ 0.399 __call__ dask_awkward/lib/structure.py:861
โ โ โ โโ 0.398 with_field awkward/operations/ak_with_field.py:19
โ โ โ โโ 0.381 _impl awkward/operations/ak_with_field.py:55
โ โ โ โโ 0.360 broadcast_and_apply awkward/_broadcasting.py:1078
โ โ โ โโ 0.353 apply_step awkward/_broadcasting.py:370
โ โ โ โโ 0.351 continuation awkward/_broadcasting.py:1023
โ โ โ โโ 0.350 broadcast_any_list awkward/_broadcasting.py:500
โ โ โ โโ 0.338 apply_step awkward/_broadcasting.py:370
โ โ โ โโ 0.276 continuation awkward/_broadcasting.py:1023
โ โ โ โ โโ 0.260 broadcast_any_list awkward/_broadcasting.py:500
โ โ โ โ โโ 0.246 apply_step awkward/_broadcasting.py:370
โ โ โ โ โโ 0.151 continuation awkward/_broadcasting.py:1023
โ โ โ โ โ โโ 0.092 broadcast_any_indexed awkward/_broadcasting.py:979
โ โ โ โ โ โ โโ 0.082 <listcomp> awkward/_broadcasting.py:980
โ โ โ โ โ โ โโ 0.082 project awkward/contents/indexedarray.py:408
โ โ โ โ โ โ โโ 0.080 _carry awkward/contents/recordarray.py:496
โ โ โ โ โ โ โโ 0.079 <listcomp> awkward/contents/recordarray.py:517
โ โ โ โ โ โ โโ 0.071 _carry awkward/contents/numpyarray.py:315
โ โ โ โ โ โ โโ 0.046 __getitem__ awkward/_nplikes/typetracer.py:378
โ โ โ โ โ โโ 0.059 broadcast_any_list awkward/_broadcasting.py:500
โ โ โ โ โ โโ 0.056 apply_step awkward/_broadcasting.py:370
โ โ โ โ โ โโ 0.053 continuation awkward/_broadcasting.py:1023
โ โ โ โ โ โโ 0.052 broadcast_any_indexed awkward/_broadcasting.py:979
โ โ โ โ โ โโ 0.047 <listcomp> awkward/_broadcasting.py:980
โ โ โ โ โ โโ 0.047 project awkward/contents/indexedarray.py:408
โ โ โ โ โ โโ 0.045 _carry awkward/contents/recordarray.py:496
โ โ โ โ โ โโ 0.042 <listcomp> awkward/contents/recordarray.py:517
โ โ โ โ โโ 0.093 action awkward/operations/ak_with_field.py:123
โ โ โ โ โโ 0.081 <listcomp> awkward/operations/ak_with_field.py:156
โ โ โ โ โโ 0.079 __getitem__ awkward/contents/content.py:535
โ โ โ โ โโ 0.076 _getitem awkward/contents/content.py:538
โ โ โ โ โโ 0.051 _getitem_field awkward/contents/recordarray.py:458
โ โ โ โ โโ 0.048 content awkward/contents/recordarray.py:396
โ โ โ โโ 0.058 action awkward/operations/ak_with_field.py:123
โ โ โ โโ 0.052 <listcomp> awkward/operations/ak_with_field.py:156
โ โ โ โโ 0.052 __getitem__ awkward/contents/content.py:535
โ โ โ โโ 0.051 _getitem awkward/contents/content.py:538
โ โ โโ 0.324 __getitem__ awkward/highlevel.py:518
โ โ โ โโ 0.256 __getitem__ awkward/contents/content.py:535
โ โ โ โโ 0.255 _getitem awkward/contents/content.py:538
โ โ โ โโ 0.156 _getitem awkward/contents/content.py:538
โ โ โ โ โโ 0.142 _getitem awkward/contents/content.py:538
โ โ โ โ โโ 0.094 _getitem_next awkward/contents/regulararray.py:451
โ โ โ โ โโ 0.080 _getitem_next_jagged awkward/contents/listoffsetarray.py:416
โ โ โ โ โโ 0.071 _getitem_next_jagged awkward/contents/listarray.py:389
โ โ โ โโ 0.058 _getitem_next awkward/contents/regulararray.py:451
โ โ โโ 0.187 __call__ dask_awkward/lib/core.py:1886
โ โ โ โโ 0.187 delta_r coffea/nanoevents/methods/vector.py:602
โ โ โ โโ 0.095 eta coffea/nanoevents/methods/vector.py:512
โ โ โ โ โโ 0.068 r coffea/nanoevents/methods/vector.py:85
โ โ โ โ โโ 0.061 r2 coffea/nanoevents/methods/vector.py:111
โ โ โ โ โโ 0.056 func numpy/lib/mixins.py:18
โ โ โ โ โโ 0.056 __array_ufunc__ awkward/highlevel.py:1289
โ โ โ โ โโ 0.053 array_ufunc awkward/_connect/numpy.py:219
โ โ โ โโ 0.050 delta_phi coffea/nanoevents/methods/vector.py:199
โ โ โโ 0.058 __call__ dask_awkward/lib/structure.py:807
โ โ โโ 0.058 where awkward/operations/ak_where.py:14
โ โ โโ 0.056 _impl3 awkward/operations/ak_where.py:97
โ โ โโ 0.052 broadcast_and_apply awkward/_broadcasting.py:1078
โ โ โโ 0.049 apply_step awkward/_broadcasting.py:370
โ โ โโ 0.048 continuation awkward/_broadcasting.py:1023
โ โ โโ 0.048 broadcast_any_list awkward/_broadcasting.py:500
โ โ โโ 0.042 apply_step awkward/_broadcasting.py:370
โ โโ 0.252 keys dask/highlevelgraph.py:530
โ โ โโ 0.252 to_dict dask/highlevelgraph.py:522
โ โ โโ 0.252 ensure_dict dask/utils.py:1236
โ โ โโ 0.234 __iter__ _collections_abc.py:719
โ โ โโ 0.233 __iter__ dask/blockwise.py:493
โ โ โโ 0.232 _dict dask/blockwise.py:452
โ โ โโ 0.066 fuse dask/optimization.py:450
โ โ โโ 0.060 dims dask/blockwise.py:440
โ โ โ โโ 0.059 _make_dims dask/blockwise.py:1480
โ โ โ โโ 0.058 broadcast_dimensions dask/blockwise.py:1420
โ โ โโ 0.051 make_blockwise_graph dask/blockwise.py:759
โ โโ 0.129 start_state_from_dask dask/local.py:140
โ โ โโ 0.113 reverse_dict dask/core.py:307
โ โ โโ 0.112 [self]
โ โโ 0.052 order dask/order.py:84
โโ 0.042 mock dask_awkward/layers/layers.py:94
```
I am happy to provide this test code if it is needed! (this cannot really be made minimal as it's ~emergent behavior) | closed | 2023-05-16T13:32:38Z | 2023-06-16T14:06:59Z | https://github.com/scikit-hep/awkward/issues/2462 | [
"performance"
] | lgray | 11 |
proplot-dev/proplot | data-visualization | 390 | Proplot not compatible with matplotlib 3.6 | Looks like the current version of proplot does not work with matplotlib==3.6 (though it works fine with 3.5).
The import:
> import proplot as pplt
Throws:
> AttributeError: module 'matplotlib.cm' has no attribute 'cmap_d'
From: proplot/colors.py:3111,
> _cmap_database = _init_cmap_database()
This might be fixable by changing the reference to cmap_d to point instead to matplotlib.colormaps.
| open | 2022-09-20T18:36:31Z | 2023-08-02T12:52:33Z | https://github.com/proplot-dev/proplot/issues/390 | [
"dependencies"
] | swamidass | 14 |
dask/dask | scikit-learn | 10,986 | CI is printing tracebacks for all xfailed tests which can be very confusing | See #10984 for more context. We can definitely disable this, we would have to add the setting to the pytest call if we want to do this | closed | 2024-03-07T15:33:01Z | 2024-03-27T14:23:59Z | https://github.com/dask/dask/issues/10986 | [
"needs triage"
] | phofl | 0 |
huggingface/peft | pytorch | 2,015 | XLoRA: training issues, Gradients will be None | I installed PEFT from source.
And use the latest versions of Transformers and TRL.
I passed the XLoRA model to TRL but the training doesn't seem to work (training loss doesn't decrease and validation loss remains constant). I get this warning:
UserWarning: None of the inputs have requires_grad=True. Gradients will be None
I load Llama 3.1 (without quantization) and then run this code:
```
adapters = dict()
adapters["0"] = './adapter1/'
adapters["1"] = './adapter2/'
peft_config = XLoraConfig(
task_type=TaskType.CAUSAL_LM,
peft_type=PeftType.XLORA,
hidden_size=model.config.hidden_size,
xlora_depth=8,
adapters=adapters,
xlora_size=2048,
layerwise_scalings=True,
xlora_dropout_p=0.2
)
xlora_model = get_peft_model(model, peft_config)
training_arguments = SFTConfig(
output_dir="./output/",
optim="paged_adamw_8bit",
per_device_train_batch_size=2,
gradient_accumulation_steps=16,
save_strategy="epoch",
log_level="debug",
logging_steps=1,
learning_rate=1e-5,
bf16 = True,
num_train_epochs=1,
warmup_ratio=0.1,
lr_scheduler_type="linear",
dataset_text_field="text",
max_seq_length=512,
)
trainer = SFTTrainer(
model=xlora_model,
train_dataset=ds,
tokenizer=tokenizer,
args=training_arguments,
)
trainer.train()
```
I also observed another bug: The adapters must be named "0", "1", etc in the adapters dict() otherwise training won't start and will say that the adapters don't exist.
Maybe @EricLBuehler can help with this?
| closed | 2024-08-18T10:30:50Z | 2024-11-17T15:03:26Z | https://github.com/huggingface/peft/issues/2015 | [] | benjamin-marie | 10 |
jschneier/django-storages | django | 997 | parameters are ignored if s3 custom URL is provided | https://github.com/jschneier/django-storages/blob/770332b598712da27ecdba75c9e202ad6a1a8722/storages/backends/s3boto3.py#L554-L578
If additional parameters are passed in to the url method when a custom_url is in use (like when S3 is behind CloudFront), they're ignored on the generated URL. They're already passed in as a dict, so they could easily be passed in to `urllib.parse.urlunparse` instead of using a format string when the URL is built (which might be a good fix anyway). | closed | 2021-04-02T00:31:39Z | 2021-09-11T18:10:43Z | https://github.com/jschneier/django-storages/issues/997 | [] | dannysauer | 1 |
docarray/docarray | pydantic | 1,444 | docs: explain type coercion | A Document parses data an coerces its type if needed. For example:
```python
from docarray import BaseDoc
from docarray.typing import TorchTensor
import numpy as np
class Doc(BaseDoc):
embedding: TorchTensor[5]
d = Doc(embedding=np.random.rand(5))
print(d.embedding)
print(type(d.embedding))
```
```bash
TorchTensor([0.5059, 0.1513, 0.5784, 0.9147, 0.6029], dtype=torch.float64)
<class 'docarray.typing.tensor.torch_tensor.TorchTensor'>
```
For everyone coming from pydantic this is obvious, but everyone else would probably manually convert the np array to torch before passing it.
So I think this is a feature worth highlighting | closed | 2023-04-24T16:19:30Z | 2023-05-31T09:22:44Z | https://github.com/docarray/docarray/issues/1444 | [
"area/docs"
] | JohannesMessner | 3 |
dynaconf/dynaconf | flask | 723 | [bug] Get Vault Secret via dynaconf throws `permission denied` error | **Describe the bug**
Accessing Vault secrets via dynaconf throws a` permission denied` error even though the user/role has access to the secrets and its path.
**To Reproduce**
Steps to reproduce the behaviour:
1. Having the following folder structure in the Vault
<details>
<summary> Vault Store </summary>
```
/apps/mna-satqe-hashicorp/robottelo
```
The secrets are stored in the above path in Corporate Vault.
</details>
2. Having the following app code:
<details>
<summary> Code </summary>
**config/__init__.py**
```python
from dynaconf import LazySettings
os.environ["VAULT_KV_VERSION_FOR_DYNACONF"] = "1"
os.environ['VAULT_ROLE_ID_FOR_DYNACONF'] = 'mna-satqe-hashicorp'
os.environ["VAULT_SECRET_ID_FOR_DYNACONF"] = "abcd-323333-23232323-sdsfsf"
os.environ["VAULT_MOUNT_POINT_FOR_DYNACONF"] = "apps"
os.environ["VAULT_PATH_FOR_DYNACONF"] = "mna-satqe-hashicorp/robottelo"
settingsd = LazySettings(environments=True, vault_enabled=True, vault={'url': 'https://vault.corp.compony.com:8200', 'verify': '/etc/ipa/ipa.crt'})
print(settings.ad_password)
```
</details>
4. Executing under the following environment
<details>
<summary> Execution </summary>
```bash
$ pytest -v -s tests/foreman/api/test_product.py -k test_positive_create_with_name
Traceback (most recent call last):
.
.
.
.
File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/dynaconf/base.py", line 113, in __getattr__
self._setup()
File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/dynaconf/base.py", line 163, in _setup
self._wrapped = Settings(
File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/dynaconf/base.py", line 233, in __init__
self.execute_loaders()
File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/dynaconf/base.py", line 977, in execute_loaders
loader.load(self, env, silent=silent, key=key)
File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/dynaconf/loaders/vault_loader.py", line 74, in load
dirs = client.secrets.kv.list_secrets(
File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/hvac/api/secrets_engines/kv_v1.py", line 58, in list_secrets
return self._adapter.list(
File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/hvac/adapters.py", line 165, in list
return self.request("list", url, **kwargs)
File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/hvac/adapters.py", line 364, in request
response = super(JSONAdapter, self).request(*args, **kwargs)
File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/hvac/adapters.py", line 330, in request
utils.raise_for_error(
File "/Users/jitendrayejare/JWorkspace/Envs/rtello/lib/python3.9/site-packages/hvac/utils.py", line 41, in raise_for_error
raise exceptions.Forbidden(message, errors=errors, method=method, url=url)
hvac.exceptions.Forbidden: 1 error occurred:
* permission denied
, on list https://vault.corp.company.com:8200/v1/apps/mna-satqe-hashicorp/robottelo
```
</details>
**Expected behaviour**
DynaConf should be able to access the secret successfully and print in my case above.
**Environment (please complete the following information):**
- OS: [MacOS 12.1]
- Dynaconf Version [3.1.7]
- Frameworks in use (pytest)
**Additional context**
The curl command to access the secrets from the same path works as expected.
```
# curl --header "X-Vault-Token: s.saswewetergfd3333dddd" https://vault.corp.company.com:8200/v1/apps/data/mna-satqe-hashicorp/robottelo
{"request_id":"a75e5b8a-c958-8608-dad0-85f516c62c69","lease_id":"","renewable":false,"lease_duration":0,"data":{"data":{"ad_password":"SecretPAss3456"}, "metadata":{"created_time":"2022-02-25T11:51:47.258977787Z","deletion_time":"","destroyed":false,"version":1}},"wrap_info":null,"warnings":null,"auth":null}
```
| closed | 2022-02-28T14:02:16Z | 2022-03-30T20:34:43Z | https://github.com/dynaconf/dynaconf/issues/723 | [
"bug"
] | jyejare | 0 |
long2ice/fastapi-cache | fastapi | 31 | How to set the redis db in the cache? | closed | 2021-08-02T09:02:43Z | 2021-08-02T09:18:48Z | https://github.com/long2ice/fastapi-cache/issues/31 | [] | liushuangdan | 2 |
|
gee-community/geemap | streamlit | 455 | Add GeoPandas/Fiona support | This will allow users to open any vector dataset supported by GeoPandas/Fiona and display it on the map. It can also be converted to ee.FeatureCollection if needed.
Fiona supported drivers:
https://github.com/Toblerity/Fiona/blob/master/fiona/drvsupport.py
geopandas.read_file
https://geopandas.org/docs/reference/api/geopandas.read_file.html#geopandas.read_file | closed | 2021-05-04T01:20:46Z | 2021-05-05T02:19:02Z | https://github.com/gee-community/geemap/issues/455 | [
"Feature Request"
] | giswqs | 1 |
Johnserf-Seed/TikTokDownload | api | 405 | [Feature]่ฝๅฆ็ปๅธฆๆ้ฟๆๆฌๆ ้ข็่ง้ข ๅ็ฌๅ ไธไธชๆพ่ฏฅ้ฟๆๆฌ็txtๆๆกฃ | ๆไบ้ฟๆๆฌๆฌ่บซ็ๅ
ๅฎนไนๆฏๅพไธ้็ | open | 2023-04-21T00:34:29Z | 2023-08-04T12:37:34Z | https://github.com/Johnserf-Seed/TikTokDownload/issues/405 | [
"้ๆฑๅปบ่ฎฎ(enhancement)"
] | NightMoonIsland | 1 |
unit8co/darts | data-science | 1,901 | RuntimeError: Lightning can't create new processes if CUDA is already initialized. | **Describe the bug**
I encounter a problem that says `RuntimeError: Lightning can't create new processes if CUDA is already initialized. `
**To Reproduce**
```python
import numpy as np
import pandas as pd
import torch
import matplotlib.pyplot as plt
from darts.dataprocessing.transformers import Scaler
from darts.models import RNNModel
from darts.metrics import mape
from darts.datasets import AirPassengersDataset
# Read data:
series = AirPassengersDataset().load()
series = series.astype(np.float32)
# Create training and validation sets:
train, val = series.split_after(pd.Timestamp("19590101"))
# Normalize the time series (note: we avoid fitting the transformer on the validation set)
transformer = Scaler()
train_transformed = transformer.fit_transform(train)
val_transformed = transformer.transform(val)
series_transformed = transformer.transform(series)
torch.cuda.set_device(0)
my_model = RNNModel(
model="RNN",
hidden_dim=20,
dropout=0,
batch_size=16,
n_epochs=300,
optimizer_kwargs={"lr": 1e-3},
model_name="Air_RNN",
log_tensorboard=True,
random_state=42,
training_length=20,
input_chunk_length=14,
force_reset=True,
)
my_model.fit(train_transformed, val_series=val_transformed)
```
**Expected behavior**
I want it works properly.
**System (please complete the following information):**
- Python version: 3.9
- darts version: 0.24.0
- torch version: 1.12.0
**Additional context**
Detail error:
```shell
GPU available: True (cuda), used: True
TPU available: False, using: 0 TPU cores
IPU available: False, using: 0 IPUs
HPU available: False, using: 0 HPUs
---------------------------------------------------------------------------
RuntimeError Traceback (most recent call last)
Cell In[9], line 40
24 torch.cuda.set_device(0)
26 my_model = RNNModel(
27 model="RNN",
28 hidden_dim=20,
(...)
38 force_reset=True,
39 )
---> 40 my_model.fit(train_transformed, val_series=val_transformed)
File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/darts/utils/torch.py:112, in random_method.<locals>.decorator(self, *args, **kwargs)
110 with fork_rng():
111 manual_seed(self._random_instance.randint(0, high=MAX_TORCH_SEED_VALUE))
--> 112 return decorated(self, *args, **kwargs)
File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/darts/models/forecasting/torch_forecasting_model.py:705, in TorchForecastingModel.fit(self, series, past_covariates, future_covariates, val_series, val_past_covariates, val_future_covariates, trainer, verbose, epochs, max_samples_per_ts, num_loader_workers)
699 # call super fit only if user is actually fitting the model
700 super().fit(
701 series=seq2series(series),
702 past_covariates=seq2series(past_covariates),
703 future_covariates=seq2series(future_covariates),
704 )
--> 705 return self.fit_from_dataset(*params)
File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/darts/utils/torch.py:112, in random_method.<locals>.decorator(self, *args, **kwargs)
110 with fork_rng():
111 manual_seed(self._random_instance.randint(0, high=MAX_TORCH_SEED_VALUE))
--> 112 return decorated(self, *args, **kwargs)
File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/darts/models/forecasting/torch_forecasting_model.py:897, in TorchForecastingModel.fit_from_dataset(self, train_dataset, val_dataset, trainer, verbose, epochs, num_loader_workers)
846 @random_method
847 def fit_from_dataset(
848 self,
(...)
854 num_loader_workers: int = 0,
855 ) -> "TorchForecastingModel":
856 """
857 Train the model with a specific :class:`darts.utils.data.TrainingDataset` instance.
858 These datasets implement a PyTorch ``Dataset``, and specify how the target and covariates are sliced
(...)
895 Fitted model.
896 """
--> 897 self._train(
898 *self._setup_for_train(
899 train_dataset=train_dataset,
900 val_dataset=val_dataset,
901 trainer=trainer,
902 verbose=verbose,
903 epochs=epochs,
904 num_loader_workers=num_loader_workers,
905 )
906 )
907 return self
File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/darts/models/forecasting/torch_forecasting_model.py:1041, in TorchForecastingModel._train(self, trainer, model, train_loader, val_loader)
1038 ckpt_path = self.load_ckpt_path
1039 self.load_ckpt_path = None
-> 1041 trainer.fit(
1042 model,
1043 train_dataloaders=train_loader,
1044 val_dataloaders=val_loader,
1045 ckpt_path=ckpt_path,
1046 )
1047 self.model = model
1048 self.trainer = trainer
File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/pytorch_lightning/trainer/trainer.py:531, in Trainer.fit(self, model, train_dataloaders, val_dataloaders, datamodule, ckpt_path)
529 model = _maybe_unwrap_optimized(model)
530 self.strategy._lightning_module = model
--> 531 call._call_and_handle_interrupt(
532 self, self._fit_impl, model, train_dataloaders, val_dataloaders, datamodule, ckpt_path
533 )
File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/pytorch_lightning/trainer/call.py:41, in _call_and_handle_interrupt(trainer, trainer_fn, *args, **kwargs)
39 try:
40 if trainer.strategy.launcher is not None:
---> 41 return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs)
42 return trainer_fn(*args, **kwargs)
44 except _TunerExitException:
File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/pytorch_lightning/strategies/launchers/multiprocessing.py:99, in _MultiProcessingLauncher.launch(self, function, trainer, *args, **kwargs)
97 self._check_torchdistx_support()
98 if self._start_method in ("fork", "forkserver"):
---> 99 _check_bad_cuda_fork()
101 # The default cluster environment in Lightning chooses a random free port number
102 # This needs to be done in the main process here before starting processes to ensure each rank will connect
103 # through the same port
104 assert self._strategy.cluster_environment is not None
File /data/tushihao/miniconda3/envs/jupyter/lib/python3.9/site-packages/lightning_fabric/strategies/launchers/multiprocessing.py:189, in _check_bad_cuda_fork()
187 if _IS_INTERACTIVE:
188 message += " You will have to restart the Python kernel."
--> 189 raise RuntimeError(message)
RuntimeError: Lightning can't create new processes if CUDA is already initialized. Did you manually call `torch.cuda.*` functions, have moved the model to the device, or allocated memory on the GPU any other way? Please remove any such calls, or change the selected strategy. You will have to restart the Python kernel.
```
| closed | 2023-07-17T07:24:45Z | 2023-07-18T15:40:08Z | https://github.com/unit8co/darts/issues/1901 | [
"bug",
"triage"
] | iDestro | 2 |
quasarstream/python-ffmpeg-video-streaming | dash | 43 | Setting custom ffmpeg/ffprobe binary path | I'd like to use python-ffmpeg-video-streaming and set a custom path to the ffmpeg and ffprobe binaries (as opposed to using the value found in the path), as I have multiple custom ffmpeg builds available on my system.
I see the run_command call in _media.py:
` run_command: bool = True, ffmpeg_bin: str = 'ffmpeg', monitor: callable = None, **options):`
And in _command_builder.py:
`def command_builder(ffmpeg_bin: str, media):`
How can I override this value without having to modify the code myself & recompile? | closed | 2020-10-30T15:22:42Z | 2020-11-01T13:59:16Z | https://github.com/quasarstream/python-ffmpeg-video-streaming/issues/43 | [
"question"
] | CraigSample | 1 |
ijl/orjson | numpy | 341 | Problem installing orjson package | When I am installing orjson, I get this error. It was working fine an hour back and I see that something got updated here in the meanwhile. Can anybody help me please?
```
Collecting orjson
Downloading orjson-3.8.6.tar.gz (655 kB)
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ 655.1/655.1 kB 47.3 MB/s eta 0:00:00
Installing build dependencies: started
Installing build dependencies: finished with status 'done'
Getting requirements to build wheel: started
Getting requirements to build wheel: finished with status 'done'
Preparing metadata (pyproject.toml): started
Preparing metadata (pyproject.toml): finished with status 'error'
error: subprocess-exited-with-error
ร Preparing metadata (pyproject.toml) did not run successfully.
โ exit code: 1
โฐโ> [6 lines of output]
Cargo, the Rust package manager, is not installed or is not on PATH.
This package requires Rust and Cargo to compile extensions. Install it through
the system's package manager or via https://rustup.rs/
Checking for Rust toolchain....
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
error: metadata-generation-failed
ร Encountered error while generating package metadata.
โฐโ> See above for output.
``` | closed | 2023-02-09T15:27:54Z | 2023-02-10T09:40:59Z | https://github.com/ijl/orjson/issues/341 | [] | saicharan0112 | 7 |
modelscope/data-juicer | data-visualization | 81 | [Bug]: NameError: name 'fingerprint_warnings' is not defined TypeError: cannot pickle 'OpenCC' object | ### Before Reporting ๆฅๅไนๅ
- [X] I have pulled the latest code of main branch to run again and the bug still existed. ๆๅทฒ็ปๆๅไบไธปๅๆฏไธๆๆฐ็ไปฃ็ ๏ผ้ๆฐ่ฟ่กไนๅ๏ผ้ฎ้ขไปไธ่ฝ่งฃๅณใ
- [X] I have read the [README](https://github.com/alibaba/data-juicer/blob/main/README.md) carefully and no error occurred during the installation process. (Otherwise, we recommend that you can ask a question using the Question template) ๆๅทฒ็ปไป็ป้
่ฏปไบ [README](https://github.com/alibaba/data-juicer/blob/main/README_ZH.md) ไธ็ๆไฝๆๅผ๏ผๅนถไธๅจๅฎ่ฃ
่ฟ็จไธญๆฒกๆ้่ฏฏๅ็ใ๏ผๅฆๅ๏ผๆไปฌๅปบ่ฎฎๆจไฝฟ็จQuestionๆจกๆฟๅๆไปฌ่ฟ่กๆ้ฎ๏ผ
### Search before reporting ๅ
ๆ็ดข๏ผๅๆฅๅ
- [X] I have searched the Data-Juicer [issues](https://github.com/alibaba/data-juicer/issues) and found no similar bugs. ๆๅทฒ็ปๅจ [issueๅ่กจ](https://github.com/alibaba/data-juicer/issues) ไธญๆ็ดขไฝๆฏๆฒกๆๅ็ฐ็ฑปไผผ็bugๆฅๅใ
### OS ็ณป็ป
ubuntu
### Installation Method ๅฎ่ฃ
ๆนๅผ
pip
### Data-Juicer Version Data-Juicer็ๆฌ
v0.1.2
### Python Version Python็ๆฌ
3.8
### Describe the bug ๆ่ฟฐ่ฟไธชbug

### To Reproduce ๅฆไฝๅค็ฐ
```
# configs/demo/process.yaml
# global parameters
project_name: 'demo-process'
dataset_path: './demos/data/demo-dataset.jsonl' # path to your dataset directory or file
np: 4 # number of subprocess to process your dataset
export_path: './outputs/demo-process/demo-processed.jsonl'
# process schedule
# a list of several process operators with their arguments
process:
# - language_id_score_filter:
# lang: 'zh'
- chinese_convert_mapper:
mode: 's2t'
```
```
python tools/process_data.py --config configs/demo/process.yaml
```
ๆฅ้ๅฆไธ้ขๆชๅพ
ไฝฟ็จไปฅไธๅ็ฌ่ๆฌๆต่ฏๆฏๆญฃๅธธ็๏ผ
```
from data_juicer.ops.mapper.chinese_convert_mapper import ChineseConvertMapper
text = {"text": "ๆๅจ้ฆฌ่ทฏ้ๅซฒ็ท่ฎ"}
op = ChineseConvertMapper(mode = 't2s')
aa = op.process(text)
print(aa)
```
### Configs ้
็ฝฎไฟกๆฏ
_No response_
### Logs ๆฅ้ๆฅๅฟ
_No response_
### Screenshots ๆชๅพ
_No response_
### Additional ้ขๅคไฟกๆฏ
_No response_ | closed | 2023-11-16T09:39:32Z | 2023-11-17T08:28:45Z | https://github.com/modelscope/data-juicer/issues/81 | [
"bug"
] | simplew2011 | 3 |
seleniumbase/SeleniumBase | web-scraping | 2,759 | When using ThreadPoolExecutor, a minor problem occurs. | Now, I'm using like below.
``` python
def setup_driver() -> Driver:
driver = Driver(browser="chrome", headless=True, no_sandbox=True)
return driver
```
``` python
# def crawling(driver): driver = setup_driver() driver.get(url) ...
with ThreadPoolExecutor(max_workers=4) as executor:
results[A] = executor.submit(
crawling
)
results[B] = executor.submit(
crawling
)
results[C] = executor.submit(
crawling
)
results[D] = executor.submit(
crawling
)
```
This doesn't work well in parallel. How can I fix it?
It worked fine when using Selenium's Webdriver, but I want to use it(Seleniumbase) because of chromedriver's automatic download feature.
``` python
def setup_driver() -> webdriver.Chrome:
options = Options()
options.add_argument("--headless")
options.add_argument("--no-sandbox")
options.add_argument("--disable-dev-shm-usage")
driver = webdriver.Chrome(service=Service(), options=options)
return driver
```
In above, it worked fine in local, but when running in environments like AWS Lambda, the Chrome WebDriver was not automatically installed properly.
Thank you. | closed | 2024-05-07T06:30:09Z | 2024-05-07T13:18:00Z | https://github.com/seleniumbase/SeleniumBase/issues/2759 | [
"invalid usage"
] | hwk06023 | 1 |
explosion/spaCy | nlp | 13,351 | Incorrect detection of sentence boundaries, if last sentence missing eos symbol for trf model | ## How to reproduce the behaviour
```
In [69]: len(list(spacy.load("en_core_web_trf")("The first sentence. The second sentence. The last one").sents))
Out[69]: 1 <<<<<<<<<<<<<<<<<<<<<< WRONG
In [70]: len(list(spacy.load("en_core_web_trf")("The first sentence. The second sentence. The last one.").sents))
Out[70]: 3
In [71]: len(list(spacy.load("en_core_web_sm")("The first sentence. The second sentence. The last one").sents))
Out[71]: 3
In [72]: len(list(spacy.load("en_core_web_sm")("The first sentence. The second sentence. The last one.").sents))
Out[72]: 3
```
## Your Environment
* Operating System: max os x 10.3
* Python Version Used: 3.11
* spaCy Version Used: 3.7.4
* Environment Information:
```
en_core_web_trf.__version__ >> '3.7.3'
en_core_web_sm.__version__ >> '3.7.1'
``` | closed | 2024-02-25T20:50:50Z | 2024-02-27T12:58:36Z | https://github.com/explosion/spaCy/issues/13351 | [
"feat / parser",
"perf / accuracy"
] | koder-ua | 0 |
widgetti/solara | jupyter | 563 | Functions run very slow when they are called under use_thread | **Problem**
A function runs very slow when it is run under `solara.use_thread`.
**How much slow:**
On my Mac, it is about 20 times. On Hugging Face (which uses Linux), 26 times slower.
**How to reproduce:**
I deployed the [code](https://huggingface.co/spaces/hkayabilisim/test_solara_use_thread/blob/main/test_thread.py) to https://huggingface.co/spaces/hkayabilisim/test_solara_use_thread for you to repeat the case.
For demonstration, I used the prime checking code here: https://github.com/widgetti/solara/blob/master/solara/website/pages/api/use_thread.py
To test the code, use large primes such as 160423, 203789, 364289, 991961, 1203793, 1667321, 3704053.
**Screenshot**
<img width="770" alt="image" src="https://github.com/widgetti/solara/assets/2515171/a93a6131-2081-4dc0-bc3f-4e89c9265b72">
| closed | 2024-03-17T19:55:39Z | 2024-03-27T05:59:38Z | https://github.com/widgetti/solara/issues/563 | [] | hkayabilisim | 3 |
indico/indico | flask | 6,735 | When using the scroll bar with the mouse in select items in a regform, scrolling does not work | **Describe the bug**
If you use a select item in a regform that has enough items to enable scrolling (like country), the user may scroll with the wheel / touchpad, but if the scroll bar is clicked on to scroll "the old way", the options closes instead of allowing to scroll by clicking.
**To Reproduce**
1. create an event and a regform, add the country field
2. try to scroll via clicking
**Expected behavior**
scrolling should be possible via scrollbar
**Additional context**
[sandbox event.](https://sandbox.getindico.io/event/4588/registrations/1570/)
| closed | 2025-02-06T18:36:22Z | 2025-02-07T10:33:40Z | https://github.com/indico/indico/issues/6735 | [
"bug"
] | knusperman | 1 |
joerick/pyinstrument | django | 33 | Maximum recursion depth with python 3.4 | Hi,
Thanks for a great tool! I get a maximum recursion depth when running pyinstrument with python 3.4.3, but it completes fine with 2.7.9. Full error log below.
I have these versions for both python 3.4.3 and 2.7.9
```
pyinstrument (2.0.0)
pyinstrument-cext (0.1.2)
```
```
Traceback (most recent call last):
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/runpy.py", line 170, in _run_module_as_main
"__main__", mod_spec)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/__main__.py", line 138, in <module>
main()
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/__main__.py", line 101, in main
f.write(profiler.output(renderer=renderer, **renderer_kwargs))
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/profiler.py", line 93, in output
return renderer.render(self.starting_frame(root=root))
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/profiler.py", line 82, in starting_frame
return self.first_interesting_frame()
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/profiler.py", line 69, in first_interesting_frame
while len(frame.children) <= 1:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 74, in time
for child in self.children:
File "/Users/kxs624/.pyenv/versions/3.4.3/lib/python3.4/site-packages/pyinstrument/frame.py", line 138, in children
self._children = sorted(unsorted_children, key=methodcaller('time'), reverse=True)
RuntimeError: maximum recursion depth exceeded while calling a Python object
```
| closed | 2018-01-10T04:49:21Z | 2018-02-11T16:24:43Z | https://github.com/joerick/pyinstrument/issues/33 | [] | ksahlin | 3 |
deezer/spleeter | deep-learning | 258 | Tensorflow/stream_executor/cuda/cuda_dnn.cc:329 Could not create cudnn handle | I have been getting the above error when running the train command. Spleeter is running on a Windows 10 laptop core i7 gen 9. With a nvidia geforce gtx 1660 ti. A friend has lent me their gaming laptop to do these experiments.
I have installed CUDA toolset 10.0 (i have also tried with 10.1 and 10.2).
I have copied the cuDNN files as described by the nvidia documentation.
I installed Anaconda3
spleeter was installed using conda (conda install -c conda-forge spleeter-gpu)
The separation works fine, and runs in a few seconds.
This is something to do with the CUDA installation, but I am not sure where to look. Anyone out there who has this working any pointers would be grateful.
| closed | 2020-02-03T22:22:19Z | 2020-02-06T20:30:14Z | https://github.com/deezer/spleeter/issues/258 | [
"question"
] | stickyninja3 | 3 |
Avaiga/taipy | automation | 1,683 | [๐ BUG] Table width issue with persistent pane and d-flex | ### What went wrong? ๐ค
The table takes more than the screen width when having a persistent pane.
I am using *with* to create the content of the pane. If I use another page to put it inside the pane, the persistent pane will not even work.
### Expected Behavior
This shouldn't change the layout of the application where there is no pane. And if there is a pane, the layout should just change so that the pane appears by the side of the app.
### Steps to Reproduce Issue
Create a folder data and put this CSV in it.
[modified_supermarkt_sales_plus.csv](https://github.com/user-attachments/files/16664692/modified_supermarkt_sales_plus.csv)
Add this code and run it:
```python
from taipy.gui import Gui, notify
import pandas as pd
import taipy.gui.builder as tgb
import json
# Load and prepare data
data = pd.read_csv("data/modified_supermarkt_sales_plus.csv")
data["Date"] = pd.to_datetime(data["Date"])
data["Review"] = ["[Review](Review)" for _ in range(len(data))]
data["Total ($)"] = data["Total"]
data["Total (โฌ)"] = data["Total"] * 1.2
displayed_data = data.copy()
# Initialize state variables with default values
show_city_info_pane = True
selected_view = "Simple view"
selected_currency = "USD"
selected_dates = [data["Date"].min().date(), data["Date"].max().date()]
selected_prices = [0, 5000]
selected_city = "All"
selected_product_line = "All"
selected_branch = "All"
rate_info = "Good"
rate_price = "Good"
open_dialog_review = False
selected_row_for_review = None
# Load city information from a JSON file
city_info_dict = {}
# Function to filter the data based on selected criteria
def filter(state):
filtered_data = state.data
if state.selected_city != "All":
filtered_data = filtered_data[filtered_data["City"] == state.selected_city]
if state.selected_product_line != "All":
filtered_data = filtered_data[
filtered_data["Product_line"] == state.selected_product_line
]
if state.selected_branch != "All":
filtered_data = filtered_data[filtered_data["Branch"] == state.selected_branch]
filtered_data = filtered_data[
(filtered_data["Date"].dt.date >= state.selected_dates[0])
& (filtered_data["Total"] >= state.selected_prices[0])
& (filtered_data["Total"] <= state.selected_prices[1])
]
state.displayed_data = filtered_data
state.city_info_partial.update_content(state, build_city_info(state.displayed_data))
# Function to convert the total values based on the selected currency
def convert(state):
if state.selected_currency == "USD":
state.displayed_data["Total"] = state.displayed_data["Total ($)"]
elif state.selected_currency == "EUR":
state.displayed_data["Total"] = state.displayed_data["Total (โฌ)"]
state.refresh("displayed_data")
# Function to handle the review submission
def send_review(state, id, payload):
state.open_dialog_review = False
# Build basic filters section
def build_basic_filters():
tgb.text("### Basic Filters", mode="md")
tgb.selector(
value="{selected_product_line}",
lov=["All"] + data["Product_line"].unique().tolist(),
dropdown=True,
filter=True,
label="Product Line",
on_change=filter,
class_name="fullwidth",
)
tgb.selector(
value="{selected_city}",
lov=["All"] + data["City"].unique().tolist(),
dropdown=True,
filter=True,
label="City",
on_change=filter,
class_name="fullwidth",
)
tgb.selector(
value="{selected_branch}",
lov=["All"] + data["Branch"].unique().tolist(),
dropdown=True,
filter=True,
label="Branch",
on_change=filter,
class_name="fullwidth",
)
# Build conversion section
def build_conversion():
tgb.text("### Conversion", mode="md")
tgb.selector(
value="{selected_currency}",
lov=["USD", "EUR"],
dropdown=True,
label="Currency",
on_change=convert,
class_name="fullwidth",
)
tgb.text("Date Range")
tgb.date_range(
"{selected_dates}", label_start="Start", label_end="End", on_change=filter
)
tgb.text("Price Range")
tgb.slider(
"{selected_prices}",
min=0,
max=5000,
on_change=filter,
continuous=False,
width="100%",
)
# Function to handle the review process
def open_review(state: State, var_name: str, payload: dict):
index = payload["index"]
data = getattr(state, var_name).copy()
state.selected_row_for_review = data.iloc[index].to_frame().T
state.open_dialog_review = True
with tgb.Page() as review_page:
tgb.text("Rate info", mode="md")
tgb.table("{selected_row_for_review}")
tgb.selector(
value="{rate_info}",
lov=["Good", "Bad"],
dropdown=True,
label="Rate info",
class_name="fullwidth", # native in 4.0
)
tgb.text("Rate price", mode="md")
tgb.selector(
value="{rate_price}",
lov=["Good", "Bad"],
dropdown=True,
label="Rate price",
class_name="fullwidth", # native in 4.0
)
# Build city information pane
def build_city_info(displayed_data):
with tgb.Page() as page:
tgb.text("### City Information", mode="md")
for city in displayed_data["City"].unique():
with tgb.expandable(title=city, expanded=False):
tgb.text(
city_info_dict.get(city, "No information available."), mode="md"
)
return page
# Build the main GUI page
with tgb.Page() as page:
with tgb.part(class_name="container d-flex"):
with tgb.part():
tgb.text("Sales Insights", class_name="h1 text-center")
with tgb.layout("1 1 1", gap="20px", columns__mobile="1"):
with tgb.part():
build_basic_filters()
with tgb.part():
build_conversion()
tgb.html("hr")
tgb.toggle(
value="{selected_view}",
lov=["Simple view", "Advanced view", "Raw view"],
)
with tgb.part(render="{selected_view=='Raw view'}"):
tgb.table(
"{data}",
on_action=open_review,
filter=True,
)
with tgb.part(render="{selected_view=='Simple view'}"):
tgb.table(
"{displayed_data}",
columns=["Date", "City", "Product_line", "Total", "Review"],
group_by__City=True,
group_by__Product_line=True,
apply_Total="mean",
filter=True,
on_action=open_review,
)
with tgb.part(render="{selected_view=='Advanced view'}"):
tgb.table(
"{displayed_data}",
columns=[
"City",
"Product_line",
"Total",
"Quantity",
"Tax_5%",
"Total",
"Date",
"Review",
],
filter=True,
on_action=open_review,
)
def open_info_pane(state):
state.show_city_info_pane = True
tgb.button(
"City info",
on_action=open_info_pane,
id="open_pane",
)
with tgb.pane(
open="{show_city_info_pane}",
width="300px",
persistent=True,
anchor="right",
):
tgb.part(partial="{city_info_partial}")
tgb.dialog(
page="review_page",
open="{open_dialog_review}",
on_action=send_review,
labels=["Cancel", "Send"],
width="500px",
title="Review the selected row",
)
# Define pages for the GUI
pages = {"page": page, "review_page": review_page}
# Run the GUI application
if __name__ == "__main__":
gui = Gui(pages=pages)
city_info_partial = gui.add_partial(build_city_info(displayed_data))
gui.run(title="Sales", port=2452)
```
### Version of Taipy
develop - 4.0.0.dev0
### Acceptance Criteria
- [ ] Ensure new code is unit tested, and check code coverage is at least 90%.
- [ ] Create related issue in taipy-doc for documentation and Release Notes.
### Code of Conduct
- [X] I have checked the [existing issues](https://github.com/Avaiga/taipy/issues?q=is%3Aissue+).
- [ ] I am willing to work on this issue (optional) | closed | 2024-08-19T19:51:35Z | 2024-08-23T14:37:33Z | https://github.com/Avaiga/taipy/issues/1683 | [
"๐ฐ GUI",
"๐ฅMalfunction",
"๐จ Priority: Medium"
] | FlorianJacta | 1 |
hpcaitech/ColossalAI | deep-learning | 6,102 | [BUG]: assert grad_chunk.l2_norm is not None | ### Is there an existing issue for this bug?
- [X] I have searched the existing issues
### ๐ Describe the bug
Modify the code to adapt to qwen2vl(transformers.Qwen2VLForConditionalGeneration) and find that the loss can be calculated, but partial chunk : grad_chunk.l2_norm is None.........(LLM is ok)

modified source code to print more information๏ผ

result:

### Environment
_No response_ | closed | 2024-10-25T09:55:40Z | 2024-11-11T13:22:28Z | https://github.com/hpcaitech/ColossalAI/issues/6102 | [
"bug"
] | liangzz1991 | 12 |
quantmind/pulsar | asyncio | 217 | How to communicate between two applications in MutiApp? | i create two application, flask wsgi server and Echo server in MutiApp,
can i send data to Echo server from browser?
How to communicate between two applications in MutiApp?
| closed | 2016-03-16T02:58:08Z | 2016-03-16T22:54:07Z | https://github.com/quantmind/pulsar/issues/217 | [
"question",
"moved"
] | bright-pan | 2 |
iterative/dvc | data-science | 10,431 | dvc repro: pygit2 update breaking changes | # Bug Report
<!--
## Issue name
Issue names must follow the pattern `command: description` where the command is the dvc command that you are trying to run. The description should describe the consequence of the bug.
Example: `repro: doesn't detect input changes`
-->
## Description
<!--
A clear and concise description of what the bug is.
-->
pygit2 renamed GIT_OBJ_COMMIT to GIT_OBJECT_COMMIT (I think)
because we get error GIT_OBJ_COMMIT not found.
### Reproduce
<!--
Step list of how to reproduce the bug
-->
install newest dvc , run dvc repro
<!--
Example:
1. dvc init
2. Copy dataset.zip to the directory
3. dvc add dataset.zip
4. dvc run -d dataset.zip -o model ./train.sh
5. modify dataset.zip
6. dvc repro
-->
### Expected
<!--
A clear and concise description of what you expect to happen.
-->
no crash
### Environment information
<!--
This is required to ensure that we can reproduce the bug.
-->
**Output of `dvc doctor`:**
```console
$ dvc doctor
```
**Additional Information (if any):**
<!--
Please check https://github.com/iterative/dvc/wiki/Debugging-DVC on ways to gather more information regarding the issue.
If applicable, please also provide a `--verbose` output of the command, eg: `dvc add --verbose`.
If the issue is regarding the performance, please attach the profiling information and the benchmark comparisons.
-->
you can fix it by using pip install pygit2==1.14.1
| closed | 2024-05-19T08:50:29Z | 2024-05-22T15:20:48Z | https://github.com/iterative/dvc/issues/10431 | [
"bug",
"p1-important",
"dependencies"
] | SimiPro | 3 |
laurentS/slowapi | fastapi | 156 | slowapi uses limits==2.8.0 which contains usage of deprecated pkg_resources | **Describe the bug**
Our application uses `slowapi==0.1.7` which makes use of `limits==2.8.0`
The problem is `limits==2.8.0` uses `pkg_resources` which is deprecated and when we run tests they fail with:
```python
src/common/limiter.py:1: in <module>
from slowapi.extension import Limiter
.nox/test-3-10-item/lib/python3.10/site-packages/slowapi/__init__.py:1: in <module>
from .extension import Limiter, _rate_limit_exceeded_handler
.nox/test-3-10-item/lib/python3.10/site-packages/slowapi/extension.py:25: in <module>
from limits import RateLimitItem # type: ignore
.nox/test-3-10-item/lib/python3.10/site-packages/limits/__init__.py:5: in <module>
from . import _version, aio, storage, strategies
.nox/test-3-10-item/lib/python3.10/site-packages/limits/aio/__init__.py:1: in <module>
from . import storage, strategies
.nox/test-3-10-item/lib/python3.10/site-packages/limits/aio/storage/__init__.py:6: in <module>
from .base import MovingWindowSupport, Storage
.nox/test-3-10-item/lib/python3.10/site-packages/limits/aio/storage/base.py:5: in <module>
from limits.storage.registry import StorageRegistry
.nox/test-3-10-item/lib/python3.10/site-packages/limits/storage/__init__.py:12: in <module>
from .base import MovingWindowSupport, Storage
.nox/test-3-10-item/lib/python3.10/site-packages/limits/storage/base.py:6: in <module>
from limits.util import LazyDependency
.nox/test-3-10-item/lib/python3.10/site-packages/limits/util.py:11: in <module>
import pkg_resources
.nox/test-3-10-item/lib/python3.10/site-packages/pkg_resources/__init__.py:121: in <module>
warnings.warn("pkg_resources is deprecated as an API", DeprecationWarning)
E DeprecationWarning: pkg_resources is deprecated as an API
```
I tried to manually edit `.nox/test-3-10-item/lib/python3.10/site-packages/limits/util.py` and apply the fix (Remove deprecated use of pkg_resources and switch to importlib_resource) and we don't get the warning anymore.
**To Reproduce**
On Python 3.10 we get this warning when running tests through nox
**Expected behavior**
`limits` fixed the issue in 3.3.0 https://github.com/alisaifee/limits/blob/master/HISTORY.rst#v330
`slowapi` should use a more updated version of `limits`
**Screenshots**
(none)
**Your app (please complete the following information):**
Deps used:
```python
fastapi==0.95.0
limits==2.8.0
slowapi==0.1.7
```
**Python**: Python 3.10.9
**Additional context**
The best would be if `slowapi` could move to `limits>=3.3.0` so this problem would be solved.
In alternative, a way to suppress `E DeprecationWarning: pkg_resources is deprecated as an API` could work as well.
Please note that this error is not happening on all our devs machines so I suspect other people may have some particular setting which is suppressing these warnings. The problem is that we can't find what it is.
Thanks | closed | 2023-04-06T16:14:34Z | 2023-04-07T11:34:54Z | https://github.com/laurentS/slowapi/issues/156 | [] | andreagrandi | 0 |
alteryx/featuretools | scikit-learn | 2,119 | Adding spark to the latest dependency checker lowered the checker's reported pandas version | Making an issue to track this for now. Possibly pyspark 3.3 will allow us to use pandas 1.4 with pyspark | closed | 2022-06-16T22:19:38Z | 2022-07-11T18:58:21Z | https://github.com/alteryx/featuretools/issues/2119 | [] | rwedge | 1 |
yihong0618/running_page | data-visualization | 450 | ่ฏทๆไธไธyihong็ๅทฅไฝๆต | ๆๆณจๆๅฐ๏ผyihong็running_page้จ็ฝฒๅจvercelไธใgithub actionๆฏๅคฉไผ่ท๏ผRUN_TYPEๆฏpassใ็ๆฅๅฟไนๆฒกๆๆๅๆฐๆฎใ
scriptsไธญ็data.dbไนๆฒกๆๆดๆฐใ
ๆไปฅyihongๆฏๅฆไฝๆดๆฐๆฐๆฎ็ใๆไธไธช็งๆไปๅบ๏ผ | closed | 2023-07-15T07:48:10Z | 2023-07-16T02:38:23Z | https://github.com/yihong0618/running_page/issues/450 | [] | agassiyzh | 1 |
davidsandberg/facenet | tensorflow | 1,010 | ็จ่ชๅทฑ็ๆฐๆฎๅ้ข่ฎญ็ปๆจกๅ้ๆฐ่ฎญ็ปtrain_softmax.pyๅบ้ | TypeError: Cannot create initializer for non-floating point type. | open | 2019-04-23T08:00:29Z | 2021-07-26T08:02:33Z | https://github.com/davidsandberg/facenet/issues/1010 | [] | Galaxy-yyyyy | 4 |
DistrictDataLabs/yellowbrick | scikit-learn | 552 | Add FAQ section to docs | **Describe the solution you'd like**
It would be great to add an FAQ section to the docs with answers to commonly asked questions (e.g., how to set the title; change the size; change the colormap).
**Is your feature request related to a problem? Please describe.**
I saw in [this issue](https://github.com/DistrictDataLabs/yellowbrick/issues/551) a suggestion for an FAQ; it sounds like it would be helpful for new users to have an FAQ section in the docs. | closed | 2018-08-09T16:31:18Z | 2018-08-15T02:12:45Z | https://github.com/DistrictDataLabs/yellowbrick/issues/552 | [
"type: documentation"
] | peterespinosa | 2 |
JaidedAI/EasyOCR | machine-learning | 1,136 | 1.7.1, Macbook M1 silicon, default Reader cannot read text, have to disable GPU to fix issue | After upgarded to 1.7.1, previously working code not able to work.
`reader = easyocr.Reader(['ch_sim','en'])`
`result = reader.readtext(file_path)`
result will be empty.
Must specify `gpu=False` to make code work. | open | 2023-09-08T06:02:14Z | 2023-09-08T06:02:36Z | https://github.com/JaidedAI/EasyOCR/issues/1136 | [] | webdkt | 0 |
hyperspy/hyperspy | data-visualization | 3,344 | Dark spot locator algorithm | #### Describe the functionality you would like to see.
A function that automatically detects the positions and radii of dark spots in an image, and the ability to perform manual corrections by the clicking in a figure to add or remove positions.
#### Describe the context
I've developed a function in matplotlib that uses the skimage [Laplacian of Gaussian blob detection algorithm](https://scikit-image.org/docs/stable/api/skimage.feature.html#skimage.feature.blob_log) for locating and counting threading dislocations (dark spots) in cathodoluminescence panchromatic images ) I would then use these results to calculate a threading dislocation density in InGaN samples.
#### Additional information
I've managed to do this with HyperSpy Markers, where the results and marker positions are saved to Signal2D metadata for panchromatic image. However, I'm still working on how to do the manual adjustments with a hyperspy figures.
```
def dark_spot_locator(image, max_sigma=30, min_sigma=12, num_sigma=30, threshold = 0.1,
permanent = False, color = 'r', linewidth = 2, invert_img = True):
import numpy as np
import matplotlib.pyplot as plt
import hyperspy.api as hs
import lumispy as lum
from skimage.util import invert
from skimage.feature import blob_log
# Assigns image intensity values of Signal2D
image_array = image.data
# Inverts color as blob_log detects white blobs on black background
if invert_img == True:
image_array = invert(image_array)
# Laplacian of Gaussian algorithm, outputting array of shape [..., 3]
# containing dark spot offsets and radii
blobs = blob_log(image_array,
max_sigma=max_sigma,
num_sigma=num_sigma,
threshold=threshold)
# Scales pixel pos to real space, and multiplies by reccommended factor
# of sqrt(2) for radii
y = blobs[:, 0] * image.axes_manager[0].scale
x = blobs[:, 1]* image.axes_manager[0].scale
r = blobs[:, 2] * image.axes_manager[0].scale * np.sqrt(2)
# Declares markers object, and takes in offsets and radii found by LoG alogorithm
markers = hs.plot.markers.Circles(
sizes=r,
offsets=np.stack((x, y), axis=1),
edgecolor=color,
linewidth=linewidth,
units = 'width')
# Adds markers to Signal2D
image.add_marker(markers, permanent = permanent)
image = hs.load('Data\M1910-3_pmt_map05.dm4')
dark_spot_locator(image, permanent = True, max_sigma=5,
min_sigma=0.5, num_sigma=5, threshold = 0.01, invert_img = True)
```
Raw panchromatic image, showing dark spots (threading dislocations) and dark lines (misfit dislocations):

LoG algorithm found positions in matplotlib function and figure:

LoG algorithm found positions (green) + manual corrections make by clicking on incorrect positions to remove and adding other positions (red)

This has definitely been personally useful for my mentioned cathodolumescence data but we (@jlaehne) wonder if there are other potential applications for which it would be worth integrating a feature like this into LumiSpy/HyperSpy.
When I load the .dm4 CL image into hyperspy it stores it as a Signal2D. It doesn't look like LumiSpy currently has a signal for storing monochromatic/panchromatic images. We were wondering if a feature like this would be best placed in LumiSpy in which case we would consider how to best implement it with a signal for CL pan/mono images, or if on the otherhand this may be more suited to be upstreamed to HyperSpy.
| closed | 2024-03-28T17:54:44Z | 2024-09-17T15:39:40Z | https://github.com/hyperspy/hyperspy/issues/3344 | [] | aidanc151 | 8 |
miguelgrinberg/python-socketio | asyncio | 868 | "KeyError: 0" while calling emit to a room from a celery + gevent task | **Describe the bug**
Howdy, I am seeing the following error `KeyError: 0` while calling emit to a room from a celery + gevent task. Thanks in advance!
**To Reproduce**
Steps to reproduce the behavior:
1. Setup with Celery with gevent and Kombu
2. Send event to room from celery task with gevent
3. receive error on emit
Python 3.9
Django / Celery app
**Expected behavior**
A clear and concise description of what you expected to happen.
**Logs**
```
trunk-player-worker_1 | [2022-02-06 17:55:07,941: INFO/MainProcess] Task radio.tasks.broadcast_transmission[fc227c55-2ddc-4fae-86a7-8bc83c9faba9] succeeded in 0.028156599029898643s: None
trunk-player_1 | Traceback (most recent call last):
trunk-player_1 | File "src/gevent/greenlet.py", line 906, in gevent._gevent_cgreenlet.Greenlet.run
trunk-player_1 | File "/usr/local/lib/python3.9/site-packages/socketio/pubsub_manager.py", line 168, in _thread
trunk-player_1 | self._handle_emit(data)
trunk-player_1 | File "/usr/local/lib/python3.9/site-packages/socketio/pubsub_manager.py", line 116, in _handle_emit
trunk-player_1 | super(PubSubManager, self).emit(message['event'], message['data'],
trunk-player_1 | File "/usr/local/lib/python3.9/site-packages/socketio/base_manager.py", line 165, in emit
trunk-player_1 | for sid, eio_sid in self.get_participants(namespace, room):
trunk-player_1 | File "/usr/local/lib/python3.9/site-packages/socketio/base_manager.py", line 43, in get_participants
trunk-player_1 | participants = ns[room[0]]._fwdm.copy() if room[0] in ns else {}
trunk-player_1 | KeyError: 0
trunk-player_1 | 2022-02-06T23:55:07Z <Thread at 0x7ff4b49339d0: <bound method PubSubManager._thread of <socketio.kombu_manager.KombuManager object at 0x7ff4b5fbcb20>>> failed with KeyError
```
**Additional context**
The sio.emit is running in a celery task
This also seems to stall the rabbitMQ Queue
This also appears to break the UWSGI socket io as it stops responding but stays connected
code:
```python
@shared_task
def broadcast_transmission(event: str, room: str, data: dict):
try:
mgr = socketio.KombuManager(
os.getenv("CELERY_BROKER_URL", "ampq://user:pass@127.0.0.1/")
)
sio = socketio.Server(
async_mode="gevent", client_manager=mgr, logger=False, engineio_logger=False
)
sio.emit(event, data, room=room)
except Exception as e:
if settings.SEND_TELEMETRY:
capture_exception(e)
raise(e)
``` | closed | 2022-02-07T02:01:00Z | 2022-02-07T02:15:11Z | https://github.com/miguelgrinberg/python-socketio/issues/868 | [] | MaxwellDPS | 1 |
xorbitsai/xorbits | numpy | 56 | using arrow dtype bugfix | closed | 2022-12-08T11:24:38Z | 2022-12-08T11:24:59Z | https://github.com/xorbitsai/xorbits/issues/56 | [] | UranusSeven | 0 |
|
AUTOMATIC1111/stable-diffusion-webui | deep-learning | 15,879 | [Bug]: | ### Checklist
- [ ] The issue exists after disabling all extensions
- [X] The issue exists on a clean installation of webui
- [ ] The issue is caused by an extension, but I believe it is caused by a bug in the webui
- [X] The issue exists in the current version of the webui
- [X] The issue has not been reported before recently
- [ ] The issue has been reported before but has not been fixed yet
### What happened?
Fresh Installation of automatic 1111. I had been stuck on this error for a while. I swapped through a few versions of pythons(from 3.10.6 to 3.10.11 to 3.12.3.) However, I noticed that at the top it still says
"Python 3.10.6 (tags/v3.10.6:9c7b4bd, Aug 1 2022, 21:53:49) [MSC v.1932 64 bit (AMD64)]
Version: v1.9.3" even though I swapped to 3.10.11 and 3.12.3.
The main error is that it cannot import one file into python.
I understand its probably something wrong with my Python or something, could anybody tell me how to fix this?
### Steps to reproduce the problem
1. Download a fresh install
Double Click webui-user.bat
### What should have happened?
WebUi should have started
### What browsers do you use to access the UI ?
_No response_
### Sysinfo
[System.txt](https://github.com/AUTOMATIC1111/stable-diffusion-webui/files/15438075/System.txt)
### Console logs
```Shell
Python 3.10.6 (tags/v3.10.6:9c7b4bd, Aug 1 2022, 21:53:49) [MSC v.1932 64 bit (AMD64)]
Version: v1.9.3
Commit hash: 1c0a0c4c26f78c32095ebc7f8af82f5c04fca8c0
Launching Web UI with arguments:
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
Traceback (most recent call last):
File "C:\Users\Admin\Downloads\sd.webui\webui\launch.py", line 48, in <module>
main()
File "C:\Users\Admin\Downloads\sd.webui\webui\launch.py", line 44, in main
start()
File "C:\Users\Admin\Downloads\sd.webui\webui\modules\launch_utils.py", line 465, in start
import webui
File "C:\Users\Admin\Downloads\sd.webui\webui\webui.py", line 13, in <module>
initialize.imports()
File "C:\Users\Admin\Downloads\sd.webui\webui\modules\initialize.py", line 39, in imports
from modules import processing, gradio_extensons, ui # noqa: F401
File "C:\Users\Admin\Downloads\sd.webui\webui\modules\processing.py", line 18, in <module>
import modules.sd_hijack
File "C:\Users\Admin\Downloads\sd.webui\webui\modules\sd_hijack.py", line 5, in <module>
from modules import devices, sd_hijack_optimizations, shared, script_callbacks, errors, sd_unet, patches
File "C:\Users\Admin\Downloads\sd.webui\webui\modules\sd_hijack_optimizations.py", line 13, in <module>
from modules.hypernetworks import hypernetwork
File "C:\Users\Admin\Downloads\sd.webui\webui\modules\hypernetworks\hypernetwork.py", line 8, in <module>
import modules.textual_inversion.dataset
File "C:\Users\Admin\Downloads\sd.webui\webui\modules\textual_inversion\dataset.py", line 12, in <module>
from modules import devices, shared, images
File "C:\Users\Admin\Downloads\sd.webui\webui\modules\images.py", line 22, in <module>
from modules import sd_samplers, shared, script_callbacks, errors
File "C:\Users\Admin\Downloads\sd.webui\webui\modules\sd_samplers.py", line 5, in <module>
from modules import sd_samplers_kdiffusion, sd_samplers_timesteps, sd_samplers_lcm, shared, sd_samplers_common, sd_schedulers
File "C:\Users\Admin\Downloads\sd.webui\webui\modules\sd_samplers_kdiffusion.py", line 3, in <module>
import k_diffusion.sampling
File "C:\Users\Admin\Downloads\sd.webui\webui\repositories\k-diffusion\k_diffusion\__init__.py", line 1, in <module>
from . import augmentation, config, evaluation, external, gns, layers, models, sampling, utils
File "C:\Users\Admin\Downloads\sd.webui\webui\repositories\k-diffusion\k_diffusion\evaluation.py", line 6, in <module>
import clip
File "C:\Users\Admin\Downloads\sd.webui\system\python\lib\site-packages\clip\__init__.py", line 1, in <module>
from .clip import *
File "C:\Users\Admin\Downloads\sd.webui\system\python\lib\site-packages\clip\clip.py", line 6, in <module>
from pkg_resources import packaging
ImportError: cannot import name 'packaging' from 'pkg_resources' (C:\Users\Admin\Downloads\sd.webui\system\python\lib\site-packages\pkg_resources\__init__.py)
Press any key to continue . . .
```
### Additional information
_No response_ | open | 2024-05-24T17:27:42Z | 2024-05-28T18:46:11Z | https://github.com/AUTOMATIC1111/stable-diffusion-webui/issues/15879 | [
"bug-report"
] | CrashedPlatz | 2 |
rasbt/watermark | jupyter | 8 | %install_ext is deprecated... | IPython deprecated the `%install_ext` magic (https://github.com/ipython/ipython/issues/8634). The replacement story is using a package to distribute the extension and then use `%load_ext package.extension` to load the extension.
Ah, and shamless plug: if you don't want to work with your own package, there is also now https://github.com/ipython-contrib/IPython-extensions watermark would be a great addition there!
| closed | 2015-10-09T11:01:48Z | 2016-05-21T18:35:32Z | https://github.com/rasbt/watermark/issues/8 | [] | jankatins | 1 |
TencentARC/GFPGAN | pytorch | 466 | T | https://github.com/kubernetes-client/python/issues/991 | open | 2023-11-10T09:46:22Z | 2023-11-10T09:46:22Z | https://github.com/TencentARC/GFPGAN/issues/466 | [] | AbdullahAlForman | 0 |
tflearn/tflearn | data-science | 703 | tflearn DNN examples segmentation fault | I tried out the titanic example, but running gives a segmentation fault. I reduced the number of lines and it will segfault with the following:
import tflearn
net = tflearn.input_data(shape=[None, 6])
model = tflearn.DNN(net)
It does come back and I can print the "model" variable so it seems to occur on exit of this simple script.
On the "titanic" example, it will segfault on:
model.fit(data, labels, n_epoch=10, batch_size=16, show_metric=True)
I traced this to:
self.summ_writer = writer_summary(
self.tensorboard_dir + run_id, self.session.graph)
in helpers/trainers.py. "wrtier_summary" prints as follows:
<class 'tensorflow.python.summary.writer.writer.FileWriter'>
I have also tried finetuning.py and it segfaults as well.
I am running:
OS: Fedora 19 (old, but everything else works)
Python v2.7.5
tensorflow: v1.0.1
numpy: 1.12.1
scipy: 0.19.0
I am running this in a virtualenv environment and everything was installed with pip in this environment. Is this a known issue with this environment?
| open | 2017-04-08T02:09:23Z | 2017-04-18T02:24:32Z | https://github.com/tflearn/tflearn/issues/703 | [] | remkonoteboom | 3 |
Evil0ctal/Douyin_TikTok_Download_API | fastapi | 347 | [Feature request] Can I get the "paid partnership" information of the tiktok video? | **Is your feature request related to a problem? Please describe.**
I found the API can't return the "paid partnership" information of the video. For example, https://www.tiktok.com/@bobbyparrish/video/7249131513927781675, it shows it is a "paid partnership" video, but i can't find this information returned by API. Can you help check this, thanks!

| closed | 2024-04-06T17:23:57Z | 2024-04-15T01:08:05Z | https://github.com/Evil0ctal/Douyin_TikTok_Download_API/issues/347 | [
"enhancement"
] | YepJin | 3 |
pydantic/pydantic | pydantic | 11,211 | A strange new bug with HttpUrl in v2.10.4 | ### Initial Checks
- [X] I confirm that I'm using Pydantic V2
### Description
I get `'{"a":"a=\'a\'"}'` which is not a correct dump. Everything works fine in v2.9.
### Example Code
```Python
from pydantic import HttpUrl, BaseModel
class A(BaseModel):
a: str
class B(BaseModel):
a: HttpUrl | A
b = B(a=A(a="a"))
b.model_dump_json()
```
### Python, Pydantic & OS Version
```Text
pydantic version: 2.10.4
pydantic-core version: 2.27.2
pydantic-core build: profile=release pgo=false
install path: D:\Conda\envs\personal\Lib\site-packages\pydantic
python version: 3.10.14 | packaged by Anaconda, Inc. | (main, Mar 21 2024, 16:20:14) [MSC v.1916 64 bit (AMD64)]
platform: Windows-10-10.0.19045-SP0
related packages: fastapi-0.115.0 pydantic-settings-2.5.2 typing_extensions-4.12.2
commit: unknown
```
| closed | 2025-01-03T08:38:50Z | 2025-01-08T15:56:30Z | https://github.com/pydantic/pydantic/issues/11211 | [
"bug V2"
] | atodniAr | 2 |
mwaskom/seaborn | matplotlib | 3,046 | Support "layer labels" with legend | It should be possible to represent a a _layer_ in the legend, in addition to a _mapping_. An example usecase would be
```python
(
so.Plot(data, "x", "y")
.add(so.Line(color="r"), so.PolyFit(1), label="1st degree")
.add(so.Line(color="b"), so.Polyfit(2), label="2nd degree")
.add(so.Dots())
)
```
We'd like this plot to have a legend with two entries: a red line with the label "1st degree" and a blue line with the label "2nd degree".
Complications!
- How does this interact with mapped properties? Should this only be allowed if there's no mapping for that layer? Or should we get a legend with the mapped properties set to a "neutral" value (similar to when you have, e.g. two separate legends for color and linewidth)?
- How should this legend appear by default? Perhaps as a separate legend panel from the variable legends? Should it always be first? If we want it to have a title, where does that happen?
- How would we refer to this legend in TBD legend customization interface?
- Implementation-wise, should this expand the functionality of `Plot._update_legend_contents`, or should we separate out methods for "labeled layer" legends and mapping legends?
- Is there anything else we might want to (or users might expect us to) use the concept of a "layer label" for? | closed | 2022-09-26T01:15:22Z | 2023-08-31T13:01:22Z | https://github.com/mwaskom/seaborn/issues/3046 | [
"objects-plot",
"feature"
] | mwaskom | 3 |
robotframework/robotframework | automation | 4,669 | Importing custom library not working if module name differ class name | Hello, the documentation says:
> Python classes are always inside a module. If the name of a class implementing a library is the same as the name of the module, Robot Framework allows dropping the class name when importing the library. [...] **If the module name and class name are different**, libraries must be taken into use using both module and class names, such as **mymodule.MyLibrary** or parent.submodule.MyLib.
Simple reproduction code:
**CustomLibrary.py**
```python
from robot.api.deco import library, keyword
@library
class Lib:
@keyword
def add(self, a, b):
return a+b
def some_method(a, b):
return a+b
```
**test.robot**
```
*** Settings ***
Library CustomLibrary.Lib
*** Test Cases ***
Example Test
${result} add ${1} ${2}
Log to Console result = ${result}
```
Both files are on the same level (for simplify demonstration purpose)
I'm getting `No keyword with name 'add' found.` error.
However changing `Library CustomLibrary.Lib` into `Library CustomLibrary.py` and `add` into `Some Method` works perfectly. I found [some old examples](https://www.youtube.com/watch?v=Y8pD5-SP_0c) that shows that my way should work.
Python: 3.11 64bit
Windows: 10 64bit
robotframework: 6.0.2 | closed | 2023-03-02T13:29:20Z | 2023-03-06T08:29:25Z | https://github.com/robotframework/robotframework/issues/4669 | [] | Redysz | 10 |
jupyter/nbgrader | jupyter | 938 | Accessing scores per question from assignments | First, a question: Is there a way to export grades, or interact with the gradebook so as to get the grades per assignment, per question?
As far as I can tell, there is not (although perhaps possible with a custom Exporter?).
Unless I've missed something that would currently make this possible, I think this might be a useful feature:
- It would allow for better auditing of assignments, seeing where students are having trouble, tracking problem topics, and individual outlier questions (that suggests a problem with the Q).
- It would also be useful to deal grading quirks: it would be easier to recalculate total assignment scores including or excluding certain questions, if issues arise (I have this issue right now in which a bad test case can sometimes fail, and I can't find a clear way to exclude this question from the overall grades).
Perhaps there is something in the underlying grading that makes this type of logging / exporting difficult, but if anyone knows a suggested way to do this, I would like to try it. | closed | 2018-03-01T01:54:15Z | 2022-12-05T06:58:28Z | https://github.com/jupyter/nbgrader/issues/938 | [
"duplicate"
] | TomDonoghue | 9 |
davidsandberg/facenet | tensorflow | 895 | Poor finetuning accuracy | I excluded the Logits layer and finetuned the model with 40 identities. The resulting model has got poor accuracy. It detects even unknown person as 100% known person. why so? | closed | 2018-10-10T14:18:11Z | 2019-04-15T13:51:59Z | https://github.com/davidsandberg/facenet/issues/895 | [] | Zumbalamambo | 0 |
ARM-DOE/pyart | data-visualization | 682 | Add nearest neighbor and most common in ROI in map_gates_to_grid | It would be useful to have a nearest neighbor and most common value within the radius of influence in the gridding routines. I have some idea how to do this in the python implementation (and have tested it, but it is sloooooow), but in the fast c-code code written by @jjhelmus - ummm, yeah.
Use case would be gridding hydrometeor identification from polar data (you don't do HID on gridded data - shame! shame!), or for tags related to use of particular rainfall algorithms applied at the polar level, etc. | closed | 2017-08-28T18:54:29Z | 2018-05-22T13:26:51Z | https://github.com/ARM-DOE/pyart/issues/682 | [] | swnesbitt | 2 |
modin-project/modin | data-science | 6,828 | BUG: The csv file cannot be read if there are square brackets in the csv file path or full path. | ### Modin version checks
- [X] I have checked that this issue has not already been reported.
- [X] I have confirmed this bug exists on the latest released version of Modin.
- [ ] I have confirmed this bug exists on the main branch of Modin. (In order to do this you can follow [this guide](https://modin.readthedocs.io/en/stable/getting_started/installation.html#installing-from-the-github-master-branch).)
### Reproducible Example
```python
df = pd.read_csv("/home/user/[DE]/[PROJECT]/total_data_231127.csv",
names=['seq', 'A', 'B'],
usecols=['A','B'],
sep="\t",
dtype=str,
na_values=['\\N'],
)
```
### Issue Description
I thought it was a Korean (UTF-8) problem at first, so I tested it several times to find the error.
(Because, I am using the Korean version of Ubuntu 22.04 and Korean is included in the file path.)
After removing Korean, I tried to read_csv using the file full path, but the same error occurred.
Below is the test screen using square brackets.
(Using "MODIN_ENGINE" = "dask")


I tried these tests just in case, but it was no use.


Currently, I'm using it very well after removing square brackets.
However I think it would be better to leave such a report, so I'm writing in a bug report.
Thank you.
### Expected Behavior
FileNotFoundError: [Errno 2] No such file or directory: '/home/user/[DE]/[PROJECT]/total_data_231127.csv'
### Error Logs
<details>
```python-traceback
Replace this line with the error backtrace (if applicable).
```
</details>
### Installed Versions
<details>
- Ubuntu 22.04.2 LTS (Korean)
- conda 22.9.0
- jupyter notebook 6.5.3
- pip list print
...
dask 2023.5.0
modin 0.23.1.post0
modin-spreadsheet 0.1.2
...
</details>
| open | 2023-12-15T01:49:32Z | 2024-06-28T22:35:39Z | https://github.com/modin-project/modin/issues/6828 | [
"bug ๐ฆ",
"External",
"P3"
] | JacobKwon | 4 |
darrenburns/posting | automation | 50 | Support for navigating by direction and custom keybindings | Hello!
First off, I love the look and feel of the interface. I'm using the galaxy theme and it's just so beautiful it almost makes me cry.
Just one thing about the navigation, I'm personally very fond of having ctrl+hjkl for moving between splits/panes etc. This interface would be complete for me if there was a possibility to set directional navigation.
I looked into the codebase, to start with just for setting some keybinds since I couldn't find a way to set it in config, but I also wanted to see how the directional nav could be done, and I'm a little puzzled (disclaimer, I'm not very experienced with Python or Textual). It seems to be quite hard with the structure of the screen -> widgets, to get which widget exists in a certain direction, instead of next/prev.
I wanted to just do a hacky solution with `get_widget_by_id` (and finding the ids by `self.walk_children(Widget)`), but it turns out that the `get_widget_by_id` only goes downward in the tree.
Do you have any pointers regarding how such an implementation might be done? I could probably spend some time trying to implement it, not guaranteeing that it will be a great implementation though. Perhaps just point me to relevant documentation, and maybe where you'd see such an addition belonging. And if you don't feel like having such an option, then I'd appreciate if you could help me find a hacky solution that could work so that I can have a local fork and use that. Like just having a fixed layout and focusing by ID:s based on where I am located.
Another thing, I really think it would be nice to have keymaps as an option in the config file. It'd have to be something like:
```yaml
keymaps:
main_screen:
send_request: ctrl+enter
url_bar:
focus_next: down
```
But I see that it would require quite a bit of work. Same goes there though, I could try implementing that. But a few pointers would probably help, if you've thought about it perhaps.
Thanks for creating a fantastic program! | closed | 2024-07-19T21:16:46Z | 2024-08-02T22:58:43Z | https://github.com/darrenburns/posting/issues/50 | [] | albinjon | 0 |
PokeAPI/pokeapi | api | 307 | Wrong evolution chain for Wurmple? | [JSON-Response](https://pokeapi.co/api/v2/evolution-chain/135/)
To me it looks like the chain shows:
wurmple -> silcoon -> ( beautifly | cascoon - > dustox )
I hope this visualization makes it clear. | closed | 2017-10-08T00:03:11Z | 2017-10-08T16:16:30Z | https://github.com/PokeAPI/pokeapi/issues/307 | [
"question"
] | HansWurst90 | 2 |
davidsandberg/facenet | tensorflow | 964 | VGG19 Model | Can some one share VGG19 architecture, modified to run with facenet code? | open | 2019-01-29T09:44:52Z | 2019-03-20T12:21:00Z | https://github.com/davidsandberg/facenet/issues/964 | [] | Shahnawazgrewal | 1 |
cvat-ai/cvat | pytorch | 9,219 | Bug Report: Missing Label Numbers in CVAT Prevents Setting Default Labels Beyond 10 | ### Actions before raising this issue
- [x] I searched the existing issues and did not find anything similar.
- [x] I read/searched [the docs](https://docs.cvat.ai/docs/)
### Steps to Reproduce
_No response_
### Expected Behavior
_No response_
### Possible Solution
_No response_
### Context
In CVAT, the Labels section does not display assigned numbers for labels. Because of this, I cannot set a default label using hotkeys (Ctrl + [number]) when I have more than 10 labels. Since only the first 10 labels are accessible via keyboard shortcuts, selecting labels beyond that requires manual selection, which slows down the annotation process.
docs:

my env:

### Environment
```Markdown
``` | open | 2025-03-17T08:39:50Z | 2025-03-17T08:39:50Z | https://github.com/cvat-ai/cvat/issues/9219 | [
"bug"
] | ntl-y | 0 |
Buuntu/fastapi-react | fastapi | 44 | Add posts to react-admin and db migrations as example | Can serve as an example of how to handle relations | open | 2020-05-26T17:40:23Z | 2020-05-26T17:40:23Z | https://github.com/Buuntu/fastapi-react/issues/44 | [
"enhancement"
] | Buuntu | 0 |
automagica/automagica | automation | 30 | Could not install in vscode | I got this error while instalilng the pip
Could not find a version that satisfies the requirement sty==1.0.0b2 (from Automagica==0.3.1) (from versions: )
No matching distribution found for sty==1.0.0b2 (from Automagica==0.3.1)
Can you help me? | closed | 2018-09-04T19:28:27Z | 2018-11-20T07:56:21Z | https://github.com/automagica/automagica/issues/30 | [] | Cemoo | 4 |
polakowo/vectorbt | data-visualization | 693 | Doc Bug report | https://github.com/polakowo/vectorbt/blob/master/examples/MACDVolume.ipynb
<img width="1407" alt="image" src="https://github.com/polakowo/vectorbt/assets/22385710/369bc694-034b-4df6-a8f3-7d0ceb7e9034">
It is disappear
| open | 2024-03-05T11:45:59Z | 2024-03-16T10:52:21Z | https://github.com/polakowo/vectorbt/issues/693 | [] | GithubJaimeLee | 1 |
aidlearning/AidLearning-FrameWork | jupyter | 188 | other tablet archecture @aidlearning . please is there the posibility of having aidlearning apk that can be installed on other phone or tablets with archecture other than aarch64. I have a 32bit tablet and it does not install on it. thanks | closed | 2021-09-26T10:03:31Z | 2021-10-01T16:45:54Z | https://github.com/aidlearning/AidLearning-FrameWork/issues/188 | [] | ichit | 1 |
|
Asabeneh/30-Days-Of-Python | pandas | 157 | V | V | closed | 2021-04-28T17:10:08Z | 2021-07-05T21:58:51Z | https://github.com/Asabeneh/30-Days-Of-Python/issues/157 | [] | eknatx | 0 |
manrajgrover/halo | jupyter | 46 | Disable spinners when not outputting to a TTY | I am using halo for the spinners and it works absolutely fine while running the management command from the command line, however I do not want the spinners while running my test suite. This is cluttering my tests and I want to disable it. I don't see an option to disable it. The only other option is to write the output to a file.
Is it possible to disable the spinners when not outputting to a TTY ?
## Description
- Disable spinners when not outputting to a TTY
### System settings
- Operating System: Ubuntu 16.04.3 LTS
- Python version: 3.6.3
- Halo version: 0.0.10
| closed | 2018-01-24T05:10:31Z | 2018-02-20T19:24:14Z | https://github.com/manrajgrover/halo/issues/46 | [
"bug"
] | vidyarani-dg | 9 |
xuebinqin/U-2-Net | computer-vision | 119 | What does u2net_train.py expect in train_data/DUTS/DUTS-TR/DUTS-TR/gt_aug/? | Obviously I'm a complete newbie to ML and all this. I put my jpgs to
train_data/DUTS/DUTS-TR/DUTS-TR/im_aug/
and they're recognized, but I'm supposed to put some label png files to
train_data/DUTS/DUTS-TR/DUTS-TR/gt_aug/
as this error message says:
FileNotFoundError: No such file: '/home/user/Desktop/U-2-Net/train_data/DUTS/DUTS-TR/DUTS-TR/gt_aug/0500.png'
What does it expect and how do I create these label files? | open | 2020-12-07T14:21:54Z | 2021-01-02T19:10:10Z | https://github.com/xuebinqin/U-2-Net/issues/119 | [] | coveritytest | 4 |
OpenInterpreter/open-interpreter | python | 659 | Cryptic error when now API key set in a system | ### Describe the bug
Interpreter fails if there's no API key set in system variable, even though I self-host the model and don't need any key.
```
Open Interpreter will require approval before running code.
Use interpreter -y to bypass this.
Press CTRL-C to exit.
> hello
Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new
LiteLLM.Info: If you need to debug this error, use `litellm.set_verbose=True'.
Traceback (most recent call last):
File "E:\Python\3.10\lib\site-packages\interpreter\core\respond.py", line 44, in respond
for chunk in interpreter._llm(messages_for_llm):
File "E:\Python\3.10\lib\site-packages\interpreter\llm\convert_to_coding_llm.py", line 19, in coding_llm
for chunk in text_llm(messages):
File "E:\Python\3.10\lib\site-packages\interpreter\llm\setup_text_llm.py", line 117, in base_llm
return litellm.completion(**params)
File "E:\Python\3.10\lib\site-packages\litellm\utils.py", line 798, in wrapper
raise e
File "E:\Python\3.10\lib\site-packages\litellm\utils.py", line 757, in wrapper
result = original_function(*args, **kwargs)
File "E:\Python\3.10\lib\site-packages\litellm\timeout.py", line 53, in wrapper
result = future.result(timeout=local_timeout_duration)
File "E:\Python\3.10\lib\concurrent\futures\_base.py", line 458, in result
return self.__get_result()
File "E:\Python\3.10\lib\concurrent\futures\_base.py", line 403, in __get_result
raise self._exception
File "E:\Python\3.10\lib\site-packages\litellm\timeout.py", line 42, in async_func
return func(*args, **kwargs)
File "E:\Python\3.10\lib\site-packages\litellm\main.py", line 1195, in completion
raise exception_type(
File "E:\Python\3.10\lib\site-packages\litellm\utils.py", line 3173, in exception_type
raise e
File "E:\Python\3.10\lib\site-packages\litellm\utils.py", line 2530, in exception_type
raise original_exception
File "E:\Python\3.10\lib\site-packages\litellm\main.py", line 447, in completion
raise e
File "E:\Python\3.10\lib\site-packages\litellm\main.py", line 429, in completion
response = openai.ChatCompletion.create(
File "E:\Python\3.10\lib\site-packages\openai\api_resources\chat_completion.py", line 25, in create
return super().create(*args, **kwargs)
File "E:\Python\3.10\lib\site-packages\openai\api_resources\abstract\engine_api_resource.py", line 151, in create
) = cls.__prepare_create_request(
File "E:\Python\3.10\lib\site-packages\openai\api_resources\abstract\engine_api_resource.py", line 108, in __prepare_create_request
requestor = api_requestor.APIRequestor(
File "E:\Python\3.10\lib\site-packages\openai\api_requestor.py", line 139, in __init__
self.api_key = key or util.default_api_key()
File "E:\Python\3.10\lib\site-packages\openai\util.py", line 186, in default_api_key
raise openai.error.AuthenticationError(
openai.error.AuthenticationError: No API key provided. You can set your API key in code using 'openai.api_key = <API-KEY>', or you can set the environment variable OPENAI_API_KEY=<API-KEY>). If your API key is stored in a file, you can point the openai module at it with 'openai.api_key_path = <PATH>'. You can generate API keys in the OpenAI web interface. See https://platform.openai.com/account/api-keys for details.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "E:\Python\3.10\lib\runpy.py", line 196, in _run_module_as_main
return _run_code(code, main_globals, None,
File "E:\Python\3.10\lib\runpy.py", line 86, in _run_code
exec(code, run_globals)
File "E:\Python\3.10\Scripts\interpreter.exe\__main__.py", line 7, in <module>
File "E:\Python\3.10\lib\site-packages\interpreter\core\core.py", line 26, in cli
cli(self)
File "E:\Python\3.10\lib\site-packages\interpreter\cli\cli.py", line 254, in cli
interpreter.chat()
File "E:\Python\3.10\lib\site-packages\interpreter\core\core.py", line 88, in chat
for _ in self._streaming_chat(message=message, display=display):
File "E:\Python\3.10\lib\site-packages\interpreter\core\core.py", line 109, in _streaming_chat
yield from terminal_interface(self, message)
File "E:\Python\3.10\lib\site-packages\interpreter\terminal_interface\terminal_interface.py", line 62, in terminal_interface
for chunk in interpreter.chat(message, display=False, stream=True):
File "E:\Python\3.10\lib\site-packages\interpreter\core\core.py", line 117, in _streaming_chat
yield from self._respond()
File "E:\Python\3.10\lib\site-packages\interpreter\core\core.py", line 143, in _respond
yield from respond(self)
File "E:\Python\3.10\lib\site-packages\interpreter\core\respond.py", line 91, in respond
raise Exception(f"{output}\n\nThere might be an issue with your API key(s).\n\nTo reset your API key (we'll use OPENAI_API_KEY for this example, but you may need to reset your ANTHROPIC_API_KEY, HUGGINGFACE_API_KEY, etc):\n Mac/Linux: 'export OPENAI_API_KEY=your-key-here',\n Windows: 'setx OPENAI_API_KEY your-key-here' then restart terminal.\n\n")
Exception: Traceback (most recent call last):
File "E:\Python\3.10\lib\site-packages\interpreter\core\respond.py", line 44, in respond
for chunk in interpreter._llm(messages_for_llm):
File "E:\Python\3.10\lib\site-packages\interpreter\llm\convert_to_coding_llm.py", line 19, in coding_llm
for chunk in text_llm(messages):
File "E:\Python\3.10\lib\site-packages\interpreter\llm\setup_text_llm.py", line 117, in base_llm
return litellm.completion(**params)
File "E:\Python\3.10\lib\site-packages\litellm\utils.py", line 798, in wrapper
raise e
File "E:\Python\3.10\lib\site-packages\litellm\utils.py", line 757, in wrapper
result = original_function(*args, **kwargs)
File "E:\Python\3.10\lib\site-packages\litellm\timeout.py", line 53, in wrapper
result = future.result(timeout=local_timeout_duration)
File "E:\Python\3.10\lib\concurrent\futures\_base.py", line 458, in result
return self.__get_result()
File "E:\Python\3.10\lib\concurrent\futures\_base.py", line 403, in __get_result
raise self._exception
File "E:\Python\3.10\lib\site-packages\litellm\timeout.py", line 42, in async_func
return func(*args, **kwargs)
File "E:\Python\3.10\lib\site-packages\litellm\main.py", line 1195, in completion
raise exception_type(
File "E:\Python\3.10\lib\site-packages\litellm\utils.py", line 3173, in exception_type
raise e
File "E:\Python\3.10\lib\site-packages\litellm\utils.py", line 2530, in exception_type
raise original_exception
File "E:\Python\3.10\lib\site-packages\litellm\main.py", line 447, in completion
raise e
File "E:\Python\3.10\lib\site-packages\litellm\main.py", line 429, in completion
response = openai.ChatCompletion.create(
File "E:\Python\3.10\lib\site-packages\openai\api_resources\chat_completion.py", line 25, in create
return super().create(*args, **kwargs)
File "E:\Python\3.10\lib\site-packages\openai\api_resources\abstract\engine_api_resource.py", line 151, in create
) = cls.__prepare_create_request(
File "E:\Python\3.10\lib\site-packages\openai\api_resources\abstract\engine_api_resource.py", line 108, in __prepare_create_request
requestor = api_requestor.APIRequestor(
File "E:\Python\3.10\lib\site-packages\openai\api_requestor.py", line 139, in __init__
self.api_key = key or util.default_api_key()
File "E:\Python\3.10\lib\site-packages\openai\util.py", line 186, in default_api_key
raise openai.error.AuthenticationError(
openai.error.AuthenticationError: No API key provided. You can set your API key in code using 'openai.api_key = <API-KEY>', or you can set the environment variable OPENAI_API_KEY=<API-KEY>). If your API key is stored in a file, you can point the openai module at it with 'openai.api_key_path = <PATH>'. You can generate API keys in the OpenAI web interface. See https://platform.openai.com/account/api-keys for details.
There might be an issue with your API key(s).
To reset your API key (we'll use OPENAI_API_KEY for this example, but you may need to reset your ANTHROPIC_API_KEY, HUGGINGFACE_API_KEY, etc):
Mac/Linux: 'export OPENAI_API_KEY=your-key-here',
Windo
```
### Reproduce
#### This works
`interpreter -ab "http://localhost:1234/v1" -m openai/TheBloke/dolphin-2.1-mistral-7B-GGUF --context_window 4000 --max_tokens 2000` **--api_key NONE**
#### This doesn't
`interpreter -ab "http://localhost:1234/v1" -m openai/TheBloke/dolphin-2.1-mistral-7B-GGUF --context_window 4000 --max_tokens 2000`
### Expected behavior
Open Interpreter should start with default / empty key and connect to the local server, as the key is not required.
### Screenshots
_No response_
### Open Interpreter version
0.1.10
### Python version
3.10.10
### Operating System name and version
Windows 11
### Additional context
LLM is self-hosted in LM Studio. | closed | 2023-10-19T10:13:11Z | 2024-02-12T17:40:25Z | https://github.com/OpenInterpreter/open-interpreter/issues/659 | [
"Bug",
"External"
] | DjWarmonger | 6 |
MaxHalford/prince | scikit-learn | 112 | ca.transform() for columns: compute principal coordinates for the columns | Hi,
How to compute principal coordinates for the columns, where columns are new ones. Indeed, the function ca.transform() is only for rows, not for columns.
Thanks,
Clรฉlia
| closed | 2021-05-18T15:36:49Z | 2021-05-18T15:50:19Z | https://github.com/MaxHalford/prince/issues/112 | [] | clelialopez | 0 |
deepset-ai/haystack | machine-learning | 8,239 | Update tests in core integration: Amazon bedrock | closed | 2024-08-15T14:26:07Z | 2024-08-19T13:33:49Z | https://github.com/deepset-ai/haystack/issues/8239 | [
"P1"
] | shadeMe | 0 |
|
koxudaxi/datamodel-code-generator | fastapi | 1,941 | CLI options `use-standard-collections` and `use-union-operator` are not being applied for input file type `graphql` | **Describe the bug**
CLI options `use-standard-collections` and `use-union-operator` are not being applied for input file type `graphql`
**To Reproduce**
Example schema:
```graphql
type DummyType {
id: ID!
label: String!
value: Int
}
type DummyType2 {
list_types: [DummyType]
}
```
Used commandline:
```
$ datamodel-codegen --input test.graphql --input-file-type=graphql --output-model-type pydantic_v2.BaseModel --use-standard-collections --use-union-operator --output test_models.py
```
**Expected behavior**
`Optional` should be replaced by union operator `|`, e.g. `Optional[str]` should be replaced by `str | None`. `List` should be replaced by standard type `list`
**Version:**
- OS: MAC OS Sonoma 14.4.1
- Python version: 3.11.9
- datamodel-code-generator version: 0.25.6
**Additional context**
Add any other context about the problem here.
| closed | 2024-04-30T09:01:15Z | 2024-07-04T15:21:09Z | https://github.com/koxudaxi/datamodel-code-generator/issues/1941 | [] | paultop6 | 2 |
kornia/kornia | computer-vision | 2,597 | Add an alias `average_enpoint_error = aepe` (or viceversa) | i'd also add an alias `average_enpoint_error = aepe` (or viceversa) to have more explicit name functions
_Originally posted by @edgarriba in https://github.com/kornia/kornia/pull/2500#discussion_r1335526211_
| closed | 2023-10-03T08:49:33Z | 2023-10-10T22:36:21Z | https://github.com/kornia/kornia/issues/2597 | [
"enhancement :rocket:"
] | edgarriba | 2 |
dsdanielpark/Bard-API | api | 91 | Suggest new chat feature | * A new chat option for each new chat a person wants to do.
* The chats should be saved so that users can review them at any time.
* The chat threads could be organized by date, topic, or user.
* The chat threads could also be tagged with keywords so that users can easily search for them.
| closed | 2023-07-04T15:30:07Z | 2024-02-06T05:54:24Z | https://github.com/dsdanielpark/Bard-API/issues/91 | [] | itsmeshashank13 | 4 |
psf/black | python | 4,396 | INTERNAL ERROR while formatting (log included) | <!--
Please make sure that the bug is not already fixed either in newer versions or the
current development version. To confirm this, you have three options:
1. Update Black's version if a newer release exists: `pip install -U black`
2. Use the online formatter at <https://black.vercel.app/?version=main>, which will use
the latest main branch. Note that the online formatter currently runs on
an older version of Python and may not support newer syntax, such as the
extended f-string syntax added in Python 3.12.
3. Or run _Black_ on your machine:
- create a new virtualenv (make sure it's the same Python version);
- clone this repository;
- run `pip install -e .[d]`;
- run `pip install -r test_requirements.txt`
- make sure it's sane by running `python -m pytest`; and
- run `black` like you did last time.
-->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
I was formatting my code when an error broke. The console said that I needed to report it, so here I am
**To Reproduce**
<!--
Minimal steps to reproduce the behavior with source code and Black's configuration.
-->
[blk_zse1vd0s.log](https://github.com/user-attachments/files/16110113/blk_zse1vd0s.log)
I have no idea how to reproduce it, but I'll attach a log file.
The resulting error is:
> INTERNAL ERROR: Black produced code that is not equivalent to the source. Please report a bug on https://github.com/psf/black/issues. This diff might be helpful: C:\Users\XXX\AppData\Local\Temp\blk_zse1vd0s.log
**Expected behavior**
<!-- A clear and concise description of what you expected to happen. -->
Not this...
**Environment**
<!-- Please complete the following information: -->
- Black's version: <!-- e.g. [main] --> 24.4.2
- OS and Python version: <!-- e.g. [Linux/Python 3.7.4rc1] --> Win 10, Python 3.10
**Additional context**
<!-- Add any other context about the problem here. -->
Good luck and thank you :)
| closed | 2024-07-05T13:15:02Z | 2024-07-05T23:36:12Z | https://github.com/psf/black/issues/4396 | [
"T: bug",
"F: fmtskip"
] | teverus | 2 |
unytics/bigfunctions | data-visualization | 33 | [new]: `xml2json(xml)` | ### Check the idea has not already been suggested
- [X] I could not find my idea in [existing issues](https://github.com/unytics/bigfunctions/issues?q=is%3Aissue+is%3Aopen+label%3Anew-bigfunction)
### Edit the title above with self-explanatory function name and argument names
- [X] The function name and the argument names I entered in the title above seems self explanatory to me.
### BigFunction Description as it would appear in the documentation
convert xml to json string using a js library
Inspired from https://github.com/salrashid123/bq-udf-xml
### Examples of (arguments, expected output) as they would appear in the documentation
`<a><b>foo</b></a>` --> `{"a": {"b": "foo"}}` | closed | 2022-12-07T22:36:44Z | 2023-04-08T15:01:56Z | https://github.com/unytics/bigfunctions/issues/33 | [
"good first issue",
"new-bigfunction"
] | unytics | 0 |
Johnserf-Seed/TikTokDownload | api | 249 | [Feature] ๅธๆๅ ๅ
ฅ้ๅฝๅไธ่ฝฝ็ๆไปถๅ่ฝ | **ๆ่ฟฐๆจๆณ่ฆ็ๅ่ฝ**
TikTokDownload ๅฝไปค่ก่ฟ่กๆถๅ้ขๅ ไธชๅฏ้้้กน๏ผ่ชๅฎไนไธ่ฝฝๆไปถ็ๆไปถๅ
**ๆ่ฟฐไฝ ๆณ่ฆๆฟๆข็ๅ่ฝๆ้่ฆไฟฎๆน็่งฃๅณๆนๆก**
| closed | 2022-11-14T10:30:48Z | 2022-11-27T11:45:34Z | https://github.com/Johnserf-Seed/TikTokDownload/issues/249 | [
"้ๆฑๅปบ่ฎฎ(enhancement)"
] | SageJustus | 0 |
piskvorky/gensim | nlp | 3,367 | Phrases.analyse_sentence() performs a greedy search for phrases | <!--
**IMPORTANT**:
- Use the [Gensim mailing list](https://groups.google.com/forum/#!forum/gensim) to ask general or usage questions. Github issues are only for bug reports.
- Check [Recipes&FAQ](https://github.com/RaRe-Technologies/gensim/wiki/Recipes-&-FAQ) first for common answers.
Github bug reports that do not include relevant information and context will be closed without an answer. Thanks!
-->
#### Problem description
When computing phrases, it is desirable that `Phrases.analyze_sentence()` implement a certain look-ahead and return phrases with higher scores. Let's say we have bigrams `('a_b', 0.25)` and `('b_c', 0.45)` and a `s = ['a', 'b', 'c']`. At present, the return value of `phrases.analyze_sentence(s)` is going to be:
```
[('a_b', 0.25),
('c'), None]
```
although, since the second bigram has a higher score, it makes sense for the return value to be:
```
[('a', None),
('b_c'), 0.45]
```
Is there value in implementing this, optimised version of `analyze_sentence()`?
_[Update]_ PS: Would it be okay to (eventually) raise a PR for this?
#### Versions
```
Linux-4.19.0-21-cloud-amd64-x86_64-with-debian-10.12
Python 3.7.12 | packaged by conda-forge | (default, Oct 26 2021, 06:08:53)
[GCC 9.4.0]
Bits 64
NumPy 1.19.5
SciPy 1.7.3
gensim 4.2.0
FAST_VERSION 0
```
| open | 2022-07-21T01:36:17Z | 2022-07-25T18:04:27Z | https://github.com/piskvorky/gensim/issues/3367 | [] | chaturv3di | 1 |
ultralytics/ultralytics | machine-learning | 19,519 | model.train() returns None when training on multiple GPUs. | ### Search before asking
- [x] I have searched the Ultralytics YOLO [issues](https://github.com/ultralytics/ultralytics/issues) and found no similar bug report.
### Ultralytics YOLO Component
Train
### Bug
When training the YOLO model (both v11 and v12) on multiple GPUs (n_gpu > 1), the results returned from `model.train()` is always `None`.
Hereโs an example code block I used to test this issue:
`results = model.train(
data = "./dataset.yaml",
epochs=10,
fraction=0.2,
batch=64,
imgsz=128,
conf=0.3,
iou=0.2,
device = [0,1],
)`
Here is a screenshot showing the type of `results` output:

### Environment
Ultralytics 8.3.82 ๐ Python-3.11.10 torch-2.5.1+cu124 CUDA:0 (Tesla T4, 14931MiB)
Setup complete โ
(4 CPUs, 14.6 GB RAM, 34.2/94.3 GB disk)
OS Linux-6.1.123+-x86_64-with-glibc2.35
Environment Linux
Python 3.11.10
Install pip
Path /opt/conda/lib/python3.11/site-packages/ultralytics
RAM 14.64 GB
Disk 34.2/94.3 GB
CPU Intel Xeon 2.30GHz
CPU count 4
GPU Tesla T4, 14931MiB
GPU count 2
CUDA 12.4
numpy โ
1.26.4<=2.1.1,>=1.23.0
matplotlib โ
3.9.2>=3.3.0
opencv-python โ
4.11.0.86>=4.6.0
pillow โ
10.4.0>=7.1.2
pyyaml โ
6.0.2>=5.3.1
requests โ
2.32.3>=2.23.0
scipy โ
1.14.1>=1.4.1
torch โ
2.5.1+cu124>=1.8.0
torch โ
2.5.1+cu124!=2.4.0,>=1.8.0; sys_platform == "win32"
torchvision โ
0.20.1+cu124>=0.9.0
tqdm โ
4.67.1>=4.64.0
psutil โ
6.1.0
py-cpuinfo โ
9.0.0
pandas โ
2.1.4>=1.1.4
seaborn โ
0.13.2>=0.11.0
ultralytics-thop โ
2.0.14>=2.0.0
{'OS': 'Linux-6.1.123+-x86_64-with-glibc2.35',
'Environment': 'Linux',
'Python': '3.11.10',
'Install': 'pip',
'Path': '/opt/conda/lib/python3.11/site-packages/ultralytics',
'RAM': '14.64 GB',
'Disk': '34.2/94.3 GB',
'CPU': 'Intel Xeon 2.30GHz',
'CPU count': 4,
'GPU': 'Tesla T4, 14931MiB',
'GPU count': 2,
'CUDA': '12.4',
'Package Info': {'numpy': 'โ
1.26.4<=2.1.1,>=1.23.0',
'matplotlib': 'โ
3.9.2>=3.3.0',
'opencv-python': 'โ
4.11.0.86>=4.6.0',
'pillow': 'โ
10.4.0>=7.1.2',
'pyyaml': 'โ
6.0.2>=5.3.1',
'requests': 'โ
2.32.3>=2.23.0',
'scipy': 'โ
1.14.1>=1.4.1',
'torch': 'โ
2.5.1+cu124!=2.4.0,>=1.8.0; sys_platform == "win32"',
'torchvision': 'โ
0.20.1+cu124>=0.9.0',
'tqdm': 'โ
4.67.1>=4.64.0',
'psutil': 'โ
6.1.0',
'py-cpuinfo': 'โ
9.0.0',
'pandas': 'โ
2.1.4>=1.1.4',
'seaborn': 'โ
0.13.2>=0.11.0',
'ultralytics-thop': 'โ
2.0.14>=2.0.0'}}
### Minimal Reproducible Example
`results = model.train(
data = "./dataset.yaml",
epochs=10,
fraction=0.2,
batch=64,
imgsz=128,
conf=0.3,
iou=0.2,
device = [0,1],
)`
`print(type(results))`
`print(results.save_dir)`
### Additional
_No response_
### Are you willing to submit a PR?
- [ ] Yes I'd like to help by submitting a PR! | closed | 2025-03-04T20:59:43Z | 2025-03-18T00:13:22Z | https://github.com/ultralytics/ultralytics/issues/19519 | [
"bug",
"devops"
] | xiaoyu-wen-1118 | 4 |
pydantic/FastUI | fastapi | 318 | How to use a custom Oauth provider (e.g. with demo app) | Would it be possible to add an example custom oauth provider?
Iโm having an issue getting it to work just right with fastui. | open | 2024-05-22T14:18:33Z | 2024-06-27T19:11:51Z | https://github.com/pydantic/FastUI/issues/318 | [] | jimkring | 2 |
benlubas/molten-nvim | jupyter | 105 | [Help] Questions regarding demo & overall usage experience | Hello, I tried molten first time today. I have never had experience with notebooks in nvim before. First impressions are very good and promising.
However, I can't understand how can I do some basic things, which were demonstrated in the video attached to the readme.
- Currently my cell outputs are only shown when the cursor hovers chunk of code being executed. I don't want it to disappear after I move cursor away. How can I configure it this way?
- There is an options to execute line of code, and operator like `<MoltenEvaluateOperator>j20`, which is usable, but clearly not the smoothest workflow for cell execution. Is there an option to execute entire cell? I noticed that in demo it seems to be working this way.
| closed | 2023-12-27T08:55:17Z | 2023-12-27T16:44:11Z | https://github.com/benlubas/molten-nvim/issues/105 | [] | V3RGANz | 5 |
benbusby/whoogle-search | flask | 280 | [QUESTION] How to update herokuapp to new whoogle version via web? | See the title so i am asking how to do it via herokuweb without delete the app
| closed | 2021-04-11T05:57:37Z | 2021-04-12T13:58:14Z | https://github.com/benbusby/whoogle-search/issues/280 | [
"question"
] | itsnotsaved | 1 |
scikit-image/scikit-image | computer-vision | 7,423 | Rolling ball algorithm scales badly with radius | ### Description:
I was playing with rolling ball background subtraction, and naively tried to invoke `skimage.restoration.rolling_ball(img, radius=50)` on a single-plane 582x200 RGB image. I noticed it eating CPU like crazy, so I did some further testing. Here is what I found:
radius | time (s) | s/r
-- | -- | --
5 | 0.627 | 0.1254
10 | 3.83 | 0.383
15 | 10 | 0.6666666667
20 | 20.7 | 1.035
25 | 39.4 | 1.576
30 | 67 | 2.233333333

Conversely, the following Groovy script in Fiji takes 17 ms to complete on my system with radius=50 on the same image:
```groovy
#@ ImagePlus imp
#@ Double(value=50) radius
import ij.plugin.filter.BackgroundSubtracter
start = System.currentTimeMillis()
BackgroundSubtracter bs = new BackgroundSubtracter()
bs.rollingBallBackground(imp.getProcessor(), radius, false, false, false, false, false)
end = System.currentTimeMillis()
println(end - start)
imp.updateAndDraw()
```
Is this a known limitation of the algorithm as currently implemented in skimage?
### Way to reproduce:
```python
img = np.random.random([200, 582, 3])
from skimage.restoration import rolling_ball
rolling_ball(img, radius=50)
```
### Version information:
```Shell
3.11.6 | packaged by conda-forge | (main, Oct 3 2023, 10:40:35) [GCC 12.3.0]
Linux-6.5.0-28-generic-x86_64-with-glibc2.38
scikit-image version: 0.22.0
numpy version: 1.26.2
```
I also experienced the same issue using skimage within Pyodide (not sure which versions, sorry).
Given that the default radius for the `rolling_ball` routine as currently written is 100, I feel like I must be doing something wrong? | open | 2024-05-15T19:03:07Z | 2024-11-14T02:39:16Z | https://github.com/scikit-image/scikit-image/issues/7423 | [
":sleeping: Dormant",
":bug: Bug"
] | ctrueden | 11 |
streamlit/streamlit | deep-learning | 10,281 | Full page reload when opening dialog | ### Checklist
- [x] I have searched the [existing issues](https://github.com/streamlit/streamlit/issues) for similar issues.
- [x] I added a very descriptive title to this issue.
- [x] I have provided sufficient information below to help reproduce this issue.
### Summary
Hello, I am currently on the nightly version. I want to show a dialog on my page. The weird thing is that my full page is reloading when I try to show a dialog. It takes a lot of time before the dialog shows up. How can I fix that? It is just a dialog with a single button. No large datasets involved.
### Reproducible Code Example
```Python
```
### Steps To Reproduce
_No response_
### Expected Behavior
_No response_
### Current Behavior
_No response_
### Is this a regression?
- [ ] Yes, this used to work in a previous version.
### Debug info
- Streamlit version:
- Python version:
- Operating System:
- Browser:
### Additional Information
_No response_ | closed | 2025-01-29T12:44:13Z | 2025-01-30T18:28:13Z | https://github.com/streamlit/streamlit/issues/10281 | [
"type:bug",
"status:awaiting-user-response"
] | Martijn3161 | 6 |
jupyterhub/repo2docker | jupyter | 449 | Docs page has lost its sidebar | https://repo2docker.readthedocs.io/en/latest/ now doesn't show a menu/sidebar anymore and seems to miss the title below the logo.
<img width="843" alt="screen shot 2018-10-19 at 18 01 12" src="https://user-images.githubusercontent.com/1448859/47229855-0b85b780-d3c9-11e8-8ff8-ece8a1b3a872.png">
cc @choldgraf the sphinx templates level 11 wizard ๐งโโ๏ธ
| closed | 2018-10-19T16:02:35Z | 2018-10-20T06:41:17Z | https://github.com/jupyterhub/repo2docker/issues/449 | [] | betatim | 1 |
jina-ai/serve | fastapi | 5,966 | Endless Waiting executor0 but the service works normally. | the code to run from #5959
```
DEBUG executor0/rep-0@39476 Setting signal handlers [07/14/23 08:58:52]
DEBUG executor0/rep-0@39476 Signal handlers already set
DEBUG executor0-replica-set@39272 Waiting for ReplicaSet to start successfully [07/14/23 08:58:52]
โ Waiting ... โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ 0/2 -:--:--DEBUG gateway/rep-0@39477 Setting signal handlers [07/14/23 08:58:52]
DEBUG gateway-replica-set@39272 Waiting for ReplicaSet to start successfully [07/14/23 08:58:52]
โ Waiting ... โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ 0/2 -:--:--DEBUG gateway/rep-0@39477 Signal handlers already set
DEBUG gateway/rep-0@39477 adding connection for deployment executor0/heads/0 to grpc://0.0.0.0:62917 [07/14/23 08:58:52]
DEBUG gateway/rep-0@39477 create_connection connection for executor0 to grpc://0.0.0.0:62917
DEBUG gateway/rep-0@39477 create_connection connection for executor0 to grpc://0.0.0.0:62917
DEBUG gateway/rep-0@39477 connection for deployment executor0/heads/0 to grpc://0.0.0.0:62917 added
DEBUG gateway/rep-0@39477 Setting up GRPC server
DEBUG gateway/rep-0@39477 Get all endpoints from TopologyGraph
DEBUG gateway/rep-0@39477 Running GatewayRuntime warmup
DEBUG gateway/rep-0@39477 Getting Endpoints data from executor0
DEBUG gateway/rep-0@39477 starting warmup task for deployment executor0
DEBUG gateway/rep-0@39477 gRPC call to executor0 for EndpointDiscovery errored, with error <AioRpcError of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused"
debug_error_string = "UNKNOWN:Failed to pick subchannel {created_time:"2023-07-14T08:58:52.240257381+08:00", children:[UNKNOWN:failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection
refused {grpc_status:14, created_time:"2023-07-14T08:58:52.240253323+08:00"}]}"
> and for the 1th time.
DEBUG gateway/rep-0@39477 resetting connection for executor0 to 0.0.0.0:62917
DEBUG gateway/rep-0@39477 create_connection connection for executor0 to 0.0.0.0:62917
โ น Waiting executor0 gateway... โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ 0/2 0:00:01DEBUG gateway/rep-0@39477 gRPC call to executor0 for EndpointDiscovery errored, with error <AioRpcError of RPC that terminated with: [07/14/23 08:58:53]
status = StatusCode.UNAVAILABLE
details = "failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused"
debug_error_string = "UNKNOWN:Failed to pick subchannel {created_time:"2023-07-14T08:58:53.241945446+08:00", children:[UNKNOWN:failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection
refused {grpc_status:14, created_time:"2023-07-14T08:58:53.241944256+08:00"}]}"
> and for the 2th time.
DEBUG gateway/rep-0@39477 resetting connection for executor0 to 0.0.0.0:62917
DEBUG gateway/rep-0@39477 create_connection connection for executor0 to 0.0.0.0:62917
โ ฆ Waiting executor0 gateway... โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ 0/2 0:00:02DEBUG gateway/rep-0@39477 gRPC call to executor0 for EndpointDiscovery errored, with error <AioRpcError of RPC that terminated with: [07/14/23 08:58:55]
status = StatusCode.UNAVAILABLE
details = "failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused"
debug_error_string = "UNKNOWN:Failed to pick subchannel {created_time:"2023-07-14T08:58:55.150302259+08:00", children:[UNKNOWN:failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection
refused {created_time:"2023-07-14T08:58:55.150299709+08:00", grpc_status:14}]}"
> and for the 3th time.
DEBUG gateway/rep-0@39477 resetting connection for executor0 to 0.0.0.0:62917
DEBUG gateway/rep-0@39477 create_connection connection for executor0 to 0.0.0.0:62917
โ ง Waiting executor0 gateway... โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ 0/2 0:00:05DEBUG gateway/rep-0@39477 gRPC call to executor0 for EndpointDiscovery errored, with error <AioRpcError of RPC that terminated with: [07/14/23 08:58:57]
status = StatusCode.UNAVAILABLE
details = "failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused"
debug_error_string = "UNKNOWN:Failed to pick subchannel {created_time:"2023-07-14T08:58:57.65715233+08:00", children:[UNKNOWN:failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused
{created_time:"2023-07-14T08:58:57.657149449+08:00", grpc_status:14}]}"
> and for the 4th time.
DEBUG gateway/rep-0@39477 gRPC call for executor0 failed, retries exhausted
DEBUG gateway/rep-0@39477 resetting connection for executor0 to 0.0.0.0:62917
DEBUG gateway/rep-0@39477 create_connection connection for executor0 to 0.0.0.0:62917
WARNIโฆ gateway/rep-0@39477 Getting endpoints failed: failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused. Waiting for another trial
โ Waiting executor0 gateway... โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ 0/2 0:00:06DEBUG gateway/rep-0@39477 Getting Endpoints data from executor0 [07/14/23 08:58:58]
โ Waiting executor0 gateway... โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ 0/2 0:00:09DEBUG gateway/rep-0@39477 gRPC call to executor0 for EndpointDiscovery errored, with error <AioRpcError of RPC that terminated with: [07/14/23 08:59:01]
status = StatusCode.UNAVAILABLE
details = "failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused"
debug_error_string = "UNKNOWN:Failed to pick subchannel {created_time:"2023-07-14T08:59:01.975496864+08:00", children:[UNKNOWN:failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection
refused {grpc_status:14, created_time:"2023-07-14T08:59:01.975494113+08:00"}]}"
> and for the 1th time.
DEBUG gateway/rep-0@39477 resetting connection for executor0 to 0.0.0.0:62917
DEBUG gateway/rep-0@39477 create_connection connection for executor0 to 0.0.0.0:62917
Loading checkpoint shards: 100%|โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ| 7/7 [00:11<00:00, 1.67s/it]
โ ฆ Waiting executor0 gateway... โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ 0/2 0:00:17DEBUG gateway/rep-0@39477 gRPC call to executor0 for EndpointDiscovery errored, with error <AioRpcError of RPC that terminated with: [07/14/23 08:59:09]
status = StatusCode.UNAVAILABLE
details = "failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection refused"
debug_error_string = "UNKNOWN:Failed to pick subchannel {created_time:"2023-07-14T08:59:09.637254313+08:00", children:[UNKNOWN:failed to connect to all addresses; last error: UNKNOWN: Failed to connect to remote host: Connection
refused {grpc_status:14, created_time:"2023-07-14T08:59:09.637251379+08:00"}]}"
> and for the 2th time.
โ Waiting executor0 gateway... โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ 0/2 0:00:17DEBUG gateway/rep-0@39477 resetting connection for executor0 to 0.0.0.0:62917
DEBUG gateway/rep-0@39477 create_connection connection for executor0 to 0.0.0.0:62917
โ Waiting executor0 gateway... โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ 0/2 0:00:19DEBUG executor0/rep-0@39476 <__main__.GLMInference object at 0x7f129e307f40> is successfully loaded! [07/14/23 08:59:11]
DEBUG executor0/rep-0@39476 Setting up GRPC server
INFO executor0/rep-0@39476 start server bound to 0.0.0.0:62917
DEBUG executor0/rep-0@39476 server bound to 0.0.0.0:62917 started
DEBUG executor0/rep-0@39476 GRPC server setup successful
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC [07/14/23 08:59:11]
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:12]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:12.556621633+08:00", grpc_status:4}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:13]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:13.666544143+08:00", grpc_status:4}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:14]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:14.774546914+08:00", grpc_status:4}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:15]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {grpc_status:4, created_time:"2023-07-14T08:59:15.882533002+08:00"}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:16]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {grpc_status:4, created_time:"2023-07-14T08:59:16.990531376+08:00"}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC [07/14/23 08:59:17]
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:18]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:18.098532957+08:00", grpc_status:4}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:19]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {grpc_status:4, created_time:"2023-07-14T08:59:19.206557554+08:00"}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:20]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:20.313535613+08:00", grpc_status:4}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC
โ ผ Waiting executor0 gateway... โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ 0/2 0:00:29DEBUG executor0/rep-0@39476 got an endpoint discovery request [07/14/23 08:59:21]
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:21]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:21.421539286+08:00", grpc_status:4}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
โ ผ Waiting executor0 gateway... โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ 0/2 0:00:29DEBUG gateway/rep-0@39477 Got all endpoints from TopologyGraph {'/chat', '/chat_manas', '_jina_dry_run_'} [07/14/23 08:59:21]
INFO gateway/rep-0@39477 start server bound to 0.0.0.0:19002
DEBUG gateway/rep-0@39477 server bound to 0.0.0.0:19002 started
DEBUG gateway/rep-0@39477 GRPC server setup successful
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC
DEBUG gateway/rep-0@39272 ready and listening [07/14/23 08:59:21]
DEBUG gateway-replica-set@39272 ReplicaSet started successfully [07/14/23 08:59:21]
DEBUG gateway@39272 Deployment started successfully [07/14/23 08:59:21]
โ ฆ Waiting executor0 gateway... โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ 0/2 0:00:29DEBUG executor0/rep-0@39476 recv _status request
DEBUG gateway/rep-0@39477 completed warmup task in 29.353113174438477s.
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:22]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {grpc_status:4, created_time:"2023-07-14T08:59:22.529546379+08:00"}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:23]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:23.636531418+08:00", grpc_status:4}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:25]
status = StatusCode.UNAVAILABLE
details = "failed to connect to all addresses; last error: UNKNOWN: HTTP proxy returned response code 407"
debug_error_string = "UNKNOWN:Failed to pick subchannel {created_time:"2023-07-14T08:59:25.646676539+08:00", children:[UNKNOWN:failed to connect to all addresses; last error: UNKNOWN: HTTP proxy returned response code 407
{created_time:"2023-07-14T08:59:25.646673591+08:00", grpc_status:14}]}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:26]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {grpc_status:4, created_time:"2023-07-14T08:59:26.754537615+08:00"}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:27]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {grpc_status:4, created_time:"2023-07-14T08:59:27.863538749+08:00"}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:28]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {grpc_status:4, created_time:"2023-07-14T08:59:28.971538289+08:00"}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC [07/14/23 08:59:29]
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:30]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:30.078544313+08:00", grpc_status:4}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:31]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:31.185541927+08:00", grpc_status:4}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
DEBUG executor0/rep-0@39272 Checking readiness to 0.0.0.0:62917 with protocol GRPC
DEBUG executor0/rep-0@39272 Exception: <AioRpcError of RPC that terminated with: [07/14/23 08:59:32]
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "UNKNOWN:Deadline Exceeded {created_time:"2023-07-14T08:59:32.29353679+08:00", grpc_status:4}"
>
DEBUG executor0/rep-0@39272 Server on 0.0.0.0:62917 with protocol GRPC is not yet ready
``` | closed | 2023-07-14T01:11:22Z | 2023-07-19T06:51:13Z | https://github.com/jina-ai/serve/issues/5966 | [] | wqh17101 | 41 |
holoviz/panel | plotly | 7,015 | pyodide-worker raising error with pyodide 0.26.1 | I was trying to upgrade the pyodide and wasm code + documentation to latest version. I found that `--pyodide-worker` will not work with latest version of pyodide 0.26.1 and panel 1.5.0b2
## Reproduce
- `pip install panel==1.5.0-b.2`
- Follow the instructions in the reference example https://panel.holoviz.org/how_to/wasm/convert.html
- after the conversion manually update 0.25.0 to 0.26.1 in the script.js file.
See the error in the browser console

```bash
PythonError: Traceback (most recent call last):
File "/lib/python312.zip/_pyodide/_base.py", line 574, in eval_code_async
await CodeRunner(
File "/lib/python312.zip/_pyodide/_base.py", line 394, in run_async
coroutine = eval(self.code, globals, locals)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/lib/python312.zip/contextlib.py", line 158, in __exit__
self.gen.throw(value)
File "/lib/python3.12/site-packages/panel/util/parameters.py", line 51, in edit_readonly
yield
File "<exec>", line 8, in <module>
File "/lib/python3.12/site-packages/param/parameterized.py", line 2319, in update
restore = dict(self_._update(arg, **kwargs))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/lib/python3.12/site-packages/param/parameterized.py", line 2352, in _update
self_._batch_call_watchers()
File "/lib/python312.zip/contextlib.py", line 158, in __exit__
self.gen.throw(value)
File "/lib/python3.12/site-packages/param/parameterized.py", line 284, in _batch_call_watchers
yield
File "/lib/python3.12/site-packages/param/parameterized.py", line 2546, in _batch_call_watchers
self_._execute_watcher(watcher, events)
File "/lib/python3.12/site-packages/param/parameterized.py", line 2506, in _execute_watcher
watcher.fn(*args, **kwargs)
File "/lib/python3.12/site-packages/panel/reactive.py", line 412, in _param_change
self._apply_update(named_events, properties, model, ref)
File "/lib/python312.zip/contextlib.py", line 158, in __exit__
self.gen.throw(value)
File "/lib/python3.12/site-packages/panel/io/document.py", line 375, in unlocked
yield
File "/lib/python3.12/site-packages/panel/reactive.py", line 316, in _apply_update
self._update_model(events, msg, root, model, doc, comm)
File "/lib/python3.12/site-packages/panel/reactive.py", line 344, in _update_model
model.update(**msg)
File "/lib/python3.12/site-packages/bokeh/core/has_props.py", line 483, in update
setattr(self, k, v)
File "/lib/python3.12/site-packages/bokeh/core/has_props.py", line 336, in __setattr__
return super().__setattr__(name, value)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/lib/python3.12/site-packages/bokeh/core/property/descriptors.py", line 332, in __set__
self._set(obj, old, value, setter=setter)
File "/lib/python3.12/site-packages/bokeh/core/property/descriptors.py", line 620, in _set
self._trigger(obj, old, value, hint=hint, setter=setter)
File "/lib/python3.12/site-packages/bokeh/core/property/descriptors.py", line 698, in _trigger
obj.trigger(self.name, old, value, hint, setter)
File "/lib/python3.12/site-packages/bokeh/model/model.py", line 571, in trigger
super().trigger(descriptor.name, old, new, hint=hint, setter=setter)
File "/lib/python3.12/site-packages/bokeh/util/callback_manager.py", line 186, in trigger
self.document.callbacks.notify_change(cast(Model, self), attr, old, new, hint, setter, invoke)
File "/lib/python3.12/site-packages/bokeh/document/callbacks.py", line 251, in notify_change
self.trigger_on_change(event)
File "/lib/python3.12/site-packages/bokeh/document/callbacks.py", line 423, in trigger_on_change
invoke_with_curdoc(doc, invoke_callbacks)
File "/lib/python312.zip/contextlib.py", line 158, in __exit__
self.gen.throw(value)
File "/lib/python3.12/site-packages/bokeh/io/doc.py", line 82, in patch_curdoc
yield
File "/lib/python3.12/site-packages/bokeh/document/callbacks.py", line 453, in invoke_with_curdoc
return f()
^^^
File "/lib/python3.12/site-packages/bokeh/document/callbacks.py", line 422, in invoke_callbacks
cb(event)
File "/lib/python3.12/site-packages/panel/io/pyodide.py", line 346, in pysync
dispatch_fn(json_patch, pyodide.ffi.to_js(buffer_map), msg_id)
File "https://mnr-jupyterhub.de-prod.dk/mt-uk/user/masma/vscode/proxy/8000/pyodide/script.js", line 4, in sendPatch
pyodide.ffi.JsException: DataCloneError: Failed to execute 'postMessage' on 'DedicatedWorkerGlobalScope': [object Map] could not be cloned.
at new_error (pyodide.asm.js:10:9965)
at pyodide.asm.wasm:0x16dbfd
at pyodide.asm.wasm:0x177431
at _PyEM_TrampolineCall_JS (pyodide.asm.js:10:125894)
at pyodide.asm.wasm:0x1c2eaf
at pyodide.asm.wasm:0x2c7c0f
at pyodide.asm.wasm:0x20a884
at pyodide.asm.wasm:0x1c359c
at pyodide.asm.wasm:0x1c38ab
at pyodide.asm.wasm:0x1c3929
at pyodide.asm.wasm:0x29e95d
at pyodide.asm.wasm:0x2a4f54
at pyodide.asm.wasm:0x1c3a69
at pyodide.asm.wasm:0x1c36d2
at pyodide.asm.wasm:0x176a95
at callPyObjectKwargs (pyodide.asm.js:10:64068)
at Module.callPyObjectMaybePromising (pyodide.asm.js:10:65316)
at wrapper (pyodide.asm.js:10:27006)
at Cn.e.port1.onmessage (pyodide.asm.js:10:102154)
``` | closed | 2024-07-26T06:02:39Z | 2024-08-02T14:43:57Z | https://github.com/holoviz/panel/issues/7015 | [] | MarcSkovMadsen | 15 |
PaddlePaddle/models | nlp | 5,254 | PaddleCVๆจกๅๅบไธ็3Dๆฃๆต้็M3D-RPNๆจกๅๆๆกฃๆ้ฎ้ข | ่ฟไธช็ฎๅฝไธ็ๆๆกฃ
ๅฟซ้ๅผๅง็ฎๅฝไธ็cd M3D-RPNๅ
่ฟ่กln -s /path/to/kitti dataset/kittiๆ็คบ๏ผ
ln:failed to create symbolic link "dataset/kitti":ๆฒกๆ่ฟไธช็ฎๅฝ
ๅฆๅค๏ผๆฌๆจกๅๆฏๅจ1.8ไธๅผๅ็๏ผไปไนๆถๅ่ฝๆไพไธช2.0็็ๆฌๅข
| open | 2021-02-01T03:26:28Z | 2024-02-26T05:09:17Z | https://github.com/PaddlePaddle/models/issues/5254 | [] | Sqhttwl | 2 |
K3D-tools/K3D-jupyter | jupyter | 59 | better check/warning on data inconsistency | `import k3d
import numpy as np
plot = k3d.plot()
plot += k3d.points(np.array([[0,0,0], [0,0,2.5]]), colors = [0xff0000])
plot.display()
`
Produces an empty plot. The user mistake is that one color is given for 2 points. It should checked and error displayed. | closed | 2017-06-13T12:26:27Z | 2017-10-22T06:30:35Z | https://github.com/K3D-tools/K3D-jupyter/issues/59 | [] | marcinofulus | 3 |
ultralytics/yolov5 | deep-learning | 12,504 | fine tune problem | > @joryoker you can quickly fine-tune your trained custom dataset model by using transfer learning. Transfer learning allows you to build upon your existing model by leveraging the knowledge it has gained from the initial training.
>
> To perform fine-tuning with your 10 new images, follow these steps:
>
> 1. Load your previously trained model and weights.
> 2. Freeze the layers of the model except for the last few layers.
> 3. Train the model using your 10 new images for a few epochs, allowing it to adapt to the new data.
> 4. Optionally, you can gradually unfreeze more layers and continue training if needed.
>
> By fine-tuning with a smaller dataset, you can save considerable training time while still improving the model's performance on the new images. Remember to use a learning rate appropriate for fine-tuning.
>
> If you need more detailed guidance, the [YOLOv5 official tutorials](https://docs.ultralytics.com/yolov5/tutorials/) provide step-by-step instructions on various topics including transfer learning.
>
> Let me know if you have any further questions or need additional assistance!
I'm a little confused. My initial data set has 30,000 pieces of data, and then I added dozens of new data for fine tune (based on the model trained on the previous initial data set). The new data is similarity as the initial data set. , and the final accuracy of the training is very high, but for new data, the detection results are not as good as the retrained model, and there are always a few error detections. Moreover, when the model is trained, the accuracy has high fluctuations, which is not as smooth as retraining. Is it because there is too little data and the impact on loss is relatively small? If that's the case, I still need to retrain. The base model already use "strip_optimizer(weights)" make requires_grad = False,and use lower learning rate(0.0001) to fine tune.

_Originally posted by @BossCrab-jyj in https://github.com/ultralytics/yolov5/issues/12069#issuecomment-1855075732_
| closed | 2023-12-14T05:38:34Z | 2023-12-18T01:28:06Z | https://github.com/ultralytics/yolov5/issues/12504 | [] | BossCrab-jyj | 0 |
cupy/cupy | numpy | 8,718 | `import cupyx.scipy.signal` emits FutureWarning | ### Description
The warning should not be emitted.
### To Reproduce
```py
>>> import cupyx.scipy.signal
/home/maehashi/Development/cupy/cupyx/jit/_interface.py:173: FutureWarning: cupyx.jit.rawkernel is experimental. The interface can change in the future.
cupy._util.experimental('cupyx.jit.rawkernel')
```
### Installation
Built from GitHub source
### Environment
```
# Paste the output here
```
### Additional Information
_No response_ | open | 2024-11-02T04:03:47Z | 2024-11-05T04:44:35Z | https://github.com/cupy/cupy/issues/8718 | [
"cat:bug",
"prio:high"
] | kmaehashi | 0 |
ultralytics/ultralytics | computer-vision | 19,832 | Is there an error in the padding of the validation set? Is it allowed not to pad to imgsz? | ### Search before asking
- [x] I have searched the Ultralytics YOLO [issues](https://github.com/ultralytics/ultralytics/issues) and [discussions](https://github.com/orgs/ultralytics/discussions) and found no similar questions.
### Question
I found that the training set was resized directly, but the validation set was scaled to keep the aspect ratio. The training set is easy to understand, but the validation set is not padded to the imgsz size after being scaled proportionally, but padded to 24 in length and 32 in width. Is this a bug or intentional? I did not find the code for the validation set to do this step.I did it very simply, and accidentally discovered that the anchor was not 80*80. I further discovered that the image size of the validation set was not transformed according to imgsz, for example, imgsz (320, 320), the image size (720, 1280), the actual size was (224, 352), it was not filled to 320, but 32 more, and then 224 did not make up 320.
I don't know if it's intentional, like allowing not to pad to the imgsz size, but it's important to me where the code is doing the extra padding and why.

### Additional
_No response_ | open | 2025-03-23T10:34:15Z | 2025-03-23T15:57:33Z | https://github.com/ultralytics/ultralytics/issues/19832 | [
"question",
"detect"
] | Hua2544 | 10 |
tqdm/tqdm | pandas | 1,231 | Using generator with process_map crashes, despite passing `total` parameter to tqdm_class | When using process_map on a generator, it crashes if `len` can't be called. For example:
```python
โช cd /var/www/cl/docker/nginx/ && sudo docker-compose exec cl-python /opt/courtlistener/manage.py dump_anon_docket_html --offset 74850 --processes 4 ; bush "Done making html"
0%| | 0/2703724 [0Traceback (most recent call last):
File "/opt/courtlistener/manage.py", line 15, in <module>
main()
File "/opt/courtlistener/manage.py", line 11, in main
execute_from_command_line(sys.argv)
File "/usr/local/lib/python3.8/site-packages/django/core/management/__init__.py", line 401, in execute_from_command_line
utility.execute()
File "/usr/local/lib/python3.8/site-packages/django/core/management/__init__.py", line 395, in execute
self.fetch_command(subcommand).run_from_argv(self.argv)
File "/usr/local/lib/python3.8/site-packages/django/core/management/base.py", line 330, in run_from_argv
self.execute(*args, **cmd_options)
File "/usr/local/lib/python3.8/site-packages/django/core/management/base.py", line 371, in execute
output = self.handle(*args, **options)
File "/opt/courtlistener/cl/corpus_importer/management/commands/dump_anon_docket_html.py", line 68, in handle
make_html(options)
File "/opt/courtlistener/cl/corpus_importer/management/commands/dump_anon_docket_html.py", line 40, in make_html
process_map(
File "/usr/local/lib/python3.8/site-packages/tqdm/contrib/concurrent.py", line 130, in process_map
return _executor_map(ProcessPoolExecutor, fn, *iterables, **tqdm_kwargs)
File "/usr/local/lib/python3.8/site-packages/tqdm/contrib/concurrent.py", line 61, in _executor_map
kwargs["total"] = len(iterables[0])
TypeError: object of type 'generator' has no len()
Sentry is attempting to send 1 pending error messages
Waiting up to 2 seconds
Press Ctrl-C to quit
```
That's a lot, here's the important part:
```python
kwargs["total"] = len(iterables[0])
TypeError: object of type 'generator' has no len()
```
What's funny about that is that I *passed* the total parameter to my `tqdm_class` parameter:
```python
progress_bar = tqdm(
total=total,
)
process_map(
_write_item_to_disk,
objects,
max_workers=options["processes"],
tqdm_class=progress_bar,
total=total,
)
```
It makes sense when you look in the code though. There, you find that the `process_map` calls `_executor_map` with whatever tqdm_kwargs it gets:
```python
return _executor_map(ProcessPoolExecutor, fn, *iterables, **tqdm_kwargs)
```
And `_executor_map` does:
```python
if "total" not in kwargs:
kwargs["total"] = len(iterables[0])
```
Which crashes.
The fix is probably either:
1. If there's a `tqdm_class` parameter, `_executor_map` should use the `total` parameter it provides, instead of using the kwarg provided to `process_map`; or
2. The documentation should be updated to note that the `total` parameter is a supported argument in the `process_map` function.
I think number 2 is easier (no code change), but number 1 makes more sense because it provides a simpler API for how to set the `total` parameter. You could do a combination of the two also:
3. Update the docs to note that total can be provided to the `process_map`, but prefer the `total` parameter that is passed to the `tqdm_class` class, if there is one.
Version info:
4.59.0 3.8.8 (default, Mar 12 2021, 19:44:18)
[GCC 8.3.0] linux | closed | 2021-08-19T21:48:24Z | 2021-08-22T21:32:50Z | https://github.com/tqdm/tqdm/issues/1231 | [
"p0-bug-critical โข",
"to-merge โฐ",
"c1-quick ๐"
] | mlissner | 1 |
huggingface/datasets | pytorch | 7,066 | One subset per file in repo ? | Right now we consider all the files of a dataset to be the same data, e.g.
```
single_subset_dataset/
โโโ train0.jsonl
โโโ train1.jsonl
โโโ train2.jsonl
```
but in cases like this, each file is actually a different subset of the dataset and should be loaded separately
```
many_subsets_dataset/
โโโ animals.jsonl
โโโ trees.jsonl
โโโ metadata.jsonl
```
It would be nice to detect those subsets automatically using a simple heuristic. For example we can group files together if their paths names are the same except some digits ? | open | 2024-07-23T12:43:59Z | 2024-07-23T12:43:59Z | https://github.com/huggingface/datasets/issues/7066 | [] | lhoestq | 0 |
google/seq2seq | tensorflow | 155 | InvalidArgumentError while training on GPU | Hi
After 1000 steps, when evaluation starts, I get the following error when training on GPU. I did the exact same training with the same data (200.000 training sentences) and parameters (example_config_nmt_small) on CPU without this error. Could anyone point me in the right direction as to what I did wrong exactly?
> File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/seq2seq/seq2seq/contrib/seq2seq/decoder.py", line 287, in dynamic_decode
swap_memory=swap_memory)
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/ops/control_flow_ops.py", line 2605, in while_loop
result = context.BuildLoop(cond, body, loop_vars, shape_invariants)
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/ops/control_flow_ops.py", line 2438, in BuildLoop
pred, body, original_loop_vars, loop_vars, shape_invariants)
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/ops/control_flow_ops.py", line 2388, in _BuildLoop
body_result = body(*packed_vars_for_body)
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/seq2seq/seq2seq/contrib/seq2seq/decoder.py", line 240, in body
decoder_finished) = decoder.step(time, inputs, state)
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/seq2seq/seq2seq/decoders/attention_decoder.py", line 187, in step
time=time_, outputs=outputs, state=cell_state, sample_ids=sample_ids)
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/seq2seq/seq2seq/contrib/seq2seq/helper.py", line 134, in next_inputs
time=time, outputs=outputs, state=state, sample_ids=sample_ids)
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/seq2seq/seq2seq/decoders/attention_decoder.py", line 155, in att_next_inputs
name=name)
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/seq2seq/seq2seq/contrib/seq2seq/helper.py", line 204, in next_inputs
lambda: nest.map_structure(read_from_ta, self._input_tas))
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/ops/control_flow_ops.py", line 1745, in cond
_, res_f = context_f.BuildCondBranch(fn2)
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/ops/control_flow_ops.py", line 1639, in BuildCondBranch
r = fn()
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/seq2seq/seq2seq/contrib/seq2seq/helper.py", line 204, in <lambda>
lambda: nest.map_structure(read_from_ta, self._input_tas))
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/util/nest.py", line 302, in map_structure
structure[0], [func(*x) for x in entries])
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/util/nest.py", line 302, in <listcomp>
structure[0], [func(*x) for x in entries])
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/seq2seq/seq2seq/contrib/seq2seq/helper.py", line 201, in read_from_ta
return inp.read(next_time)
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/ops/tensor_array_ops.py", line 250, in read
name=name)
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/ops/gen_data_flow_ops.py", line 2421, in _tensor_array_read_v3
name=name)
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/framework/op_def_library.py", line 763, in apply_op
op_def=op_def)
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/framework/ops.py", line 2327, in create_op
original_op=self._default_original_op, op_def=op_def)
File "/home/pricie/milan/VirtualEnvironment/GPU/seq2seq/lib/python3.5/site-packages/tensorflow/python/framework/ops.py", line 1226, in __init__
self._traceback = _extract_stack()
> InvalidArgumentError (see above for traceback): Tried to read from index 49 but array size is: 49
> [[Node: model/att_seq2seq/decode/attention_decoder_1/decoder/while/CustomHelperNextInputs/TrainingHelperNextInputs/cond/TensorArrayReadV3 = TensorArrayReadV3[_class=["loc:@model/att_seq2seq/decode/TrainingHelper/TensorArray"], dtype=DT_FLOAT, _device="/job:localhost/replica:0/task:0/cpu:0"](model/att_seq2seq/decode/attention_decoder_1/decoder/while/CustomHelperNextInputs/TrainingHelperNextInputs/cond/TensorArrayReadV3/Switch, model/att_seq2seq/decode/attention_decoder_1/decoder/while/CustomHelperNextInputs/TrainingHelperNextInputs/cond/TensorArrayReadV3/Switch_1/_437, model/att_seq2seq/decode/attention_decoder_1/decoder/while/CustomHelperNextInputs/TrainingHelperNextInputs/cond/TensorArrayReadV3/Switch_2)]]
> [[Node: model/att_seq2seq/decode/attention_decoder_1/decoder/while/CustomHelperNextInputs/TrainingHelperNextInputs/cond/TensorArrayReadV3/_423 = _Recv[client_terminated=false, recv_device="/job:localhost/replica:0/task:0/gpu:0", send_device="/job:localhost/replica:0/task:0/cpu:0", send_device_incarnation=1, tensor_name="edge_1011_model/att_seq2seq/decode/attention_decoder_1/decoder/while/CustomHelperNextInputs/TrainingHelperNextInputs/cond/TensorArrayReadV3", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/gpu:0"](^_cloopmodel/att_seq2seq/decode/attention_decoder_1/decoder/while/NextIteration_2/_250)]]
Thanks in advance. | closed | 2017-04-12T11:13:59Z | 2017-04-12T15:09:59Z | https://github.com/google/seq2seq/issues/155 | [] | ghost | 1 |
Lightning-AI/pytorch-lightning | deep-learning | 20,451 | Error: Invalid value for '--accelerator': 'auto' is not one of 'cpu', 'gpu', 'cuda', 'mps', 'tpu'. | ### Bug description
when run this codes
```
fabric run ./path/to/train.py \
--devices=auto \
--accelerator=auto
```
get problem below
```
Error: Invalid value for '--accelerator': 'auto' is not one of 'cpu', 'gpu', 'cuda', 'mps', 'tpu'.
```
### What version are you seeing the problem on?
v2.4
### How to reproduce the bug
_No response_
### Error messages and logs
```
# Error messages and logs here please
```
### Environment
<details>
<summary>Current environment</summary>
```
#- PyTorch Lightning Version (e.g., 2.4.0):
#- PyTorch Version (e.g., 2.4):
#- Python version (e.g., 3.12):
#- OS (e.g., Linux):
#- CUDA/cuDNN version:
#- GPU models and configuration:
#- How you installed Lightning(`conda`, `pip`, source):
```
</details>
### More info
_No response_
cc @borda @justusschock @awaelchli | open | 2024-11-26T12:55:57Z | 2024-12-07T06:33:03Z | https://github.com/Lightning-AI/pytorch-lightning/issues/20451 | [
"feature",
"help wanted",
"fabric",
"ver: 2.4.x"
] | wkd88 | 2 |
profusion/sgqlc | graphql | 108 | sessions support for requests | Can you publish the latest changes to the pypi please? for local development I can use direct github link to the latest commit as dependency I'd like to refer to it from my published package (setup.py doesn't seem to support direct links) | closed | 2020-07-10T15:47:09Z | 2020-07-10T16:45:52Z | https://github.com/profusion/sgqlc/issues/108 | [] | vavalomi | 2 |
slackapi/bolt-python | fastapi | 742 | Testing slack bot events to our chatbot in Python | Issue: We are currently unable to receive the payloads from the Slackbot. We reached out to Slack Support via Feedback and Aubrey from the support desk executed a test and they notified us about below:
`Your configured Event Subscriptions Request URL endpoint is not successfully acknowledging the Slack event callbacks meaning that even after receiving the event callback, your endpoint is not sending back the expected HTTP 200 status code, but rather it's sending back HTTP 403 which indicates that the callback requests were "Forbidden" so it seems that your endpoint is rejecting our event callbacks.`
Goal: Our goal is for the slack events to be registered properly in the chatbot, currently we are not receiving it so we are also exploring other ways to be able to test this and identify what is wrong.
#### The `slack_bolt` version
slack==0.0.2
slack-bolt==1.6.0
slack-sdk==3.18.4
slackblocks==0.2.5
slackclient==2.7.2
slackeventsapi==3.0.1
#### Python runtime version
Python 3.7.11
#### OS info
#1 SMP Sat Aug 27 02:44:20 EDT 2022
#### Problem: The chatbot does not seem to receive the payload from slack.
Current situation: Once we attempt to trigger an event, the chatbot remains the same saying "โกBolt App is Running"
#### What we have tried:
โข Reinstalling the application to ensure that all my changes were saved and were reflected properly in my Slack Environment
โข Testing the communication by curling through to the URL (it responded with challenge parameter).
โข Testing the communication by entering some text via DM / channel communication and opening the homepage.
Since we wanted to test the below events, we tried executing appropriate actions for these:
- app_home_opened
- message.channels
- message.groups
- message.im
##### For the items that the team asked us for earlier, it is included below:
โข Slack API ID: A043X1Y2MK6
โข URL of Slack Workspace: https://docomo-data-strategy.slack.com/
โข What event type are you expecting to get? We subscribed to below events. We wanted to try app_home_opened, message.im, message.channels, etc.
## Requirements
Please read the [Contributing guidelines](https://github.com/slackapi/bolt-python/blob/main/.github/contributing.md) and [Code of Conduct](https://slackhq.github.io/code-of-conduct) before creating this issue or pull request. By submitting, you are agreeing to those rules.
| closed | 2022-10-19T06:36:18Z | 2022-10-20T05:55:44Z | https://github.com/slackapi/bolt-python/issues/742 | [
"question",
"need info"
] | jdimapilis02 | 2 |
roboflow/supervision | computer-vision | 953 | TRACKING AND CENTER ID IN LINE COUNTER SOLUTION | ### Search before asking
- [X] I have searched the Supervision [issues](https://github.com/roboflow/supervision/issues) and found no similar feature requests.
### Question
Hello Supervision, I'm working with your line crossing counting solution I created this .py script, it works in common scenarios and with the camera in the ideal position (which is great), but I'm working in a scenario where the camera is positioned on the ceiling and pointing towards the floor... it can detect it, but the tracking does not remain constant because it is apparently looking for a center ID, which I think is hard to happen as the image is not three dimensional.... how could it adapt the tracking or this center ID for this type of scenario... The model that Iยดm using here itยดs alredy a custom model

### Additional
Iยดm using this code to run the inference/counting:
import os
import numpy as np
import supervision as sv
from ultralytics import YOLO
# Define o caminho do vรญdeo fonte e do vรญdeo alvo
#HOME = os.getcwd()
SOURCE_VIDEO_PATH = os.path.join( "Identificacao-vendedor.mp4")
TARGET_VIDEO_PATH = SOURCE_VIDEO_PATH.replace(".mp4", "_output.mp4")
# Inicializaรงรฃo do modelo YOLO
MODEL = "best (5).pt"
model = YOLO(MODEL)
model.fuse()
# Define as classes de interesse
selected_classes = [0,1] # Exemplo: ID 0 para 'pessoa'
# Configuraรงรฃo da zona de linha e anotadores
LINE_START = sv.Point(300, 600)
LINE_END = sv.Point(292, 50)
line_zone = sv.LineZone(start=LINE_START, end=LINE_END)
line_zone_annotator = sv.LineZoneAnnotator(thickness=4, text_thickness=4, text_scale=1)
# Configuraรงรฃo do rastreador ByteTrack
byte_tracker = sv.ByteTrack(track_thresh=0.05, track_buffer=30, match_thresh=0.8, frame_rate=24)
# Anotadores adicionais
box_annotator = sv.BoxAnnotator(thickness=3, text_thickness=0, text_scale=0.5)
trace_annotator = sv.TraceAnnotator(thickness=4, trace_length=50)
# Funรงรฃo de callback para processamento de cada frame
def callback(frame: np.ndarray, index: int) -> np.ndarray:
results = model(frame, verbose=True, device='0', conf=0.015, iou=0.02, imgsz=1280)[0]
detections = sv.Detections.from_ultralytics(results)
detections = detections[np.isin(detections.class_id, selected_classes)]
detections = byte_tracker.update_with_detections(detections)
line_zone.trigger(detections)
labels = [f"#{tracker_id} {model.model.names[class_id]} {confidence:0.2f}" for _, _, confidence, class_id, tracker_id, _ in detections]
annotated_frame = trace_annotator.annotate(scene=frame.copy(), detections=detections)
annotated_frame = box_annotator.annotate(scene=annotated_frame, detections=detections, labels=labels)
annotated_frame = line_zone_annotator.annotate(annotated_frame, line_counter=line_zone)
return annotated_frame
# Funรงรฃo para processar o vรญdeo
def process_video(source_path: str, target_path: str, callback, debug: bool = False) -> None:
source_video_info = sv.VideoInfo.from_video_path(video_path=source_path)
with sv.VideoSink(target_path=target_path, video_info=source_video_info) as sink:
for index, frame in enumerate(sv.get_video_frames_generator(source_path=source_path)):
result_frame = callback(frame, index)
sink.write_frame(frame=result_frame)
if debug and index == 0: # Exemplo de condiรงรฃo para depuraรงรฃo
break
# Chamada da funรงรฃo para processar o vรญdeo
process_video(source_path=SOURCE_VIDEO_PATH, target_path=TARGET_VIDEO_PATH, callback=callback, debug=False)
| closed | 2024-02-28T12:03:05Z | 2024-02-28T12:12:23Z | https://github.com/roboflow/supervision/issues/953 | [
"question"
] | Rasantis | 1 |
OFA-Sys/Chinese-CLIP | nlp | 240 | ไฝฟ็จ่ชๅทฑๆๅปบ็ๆฐๆฎ้๏ผๅพฎ่ฐlossไธไธ้ | ๆจๅฅฝ๏ผ่ฏทๆไธไธ๏ผๆๆณ็จclipๆฅๅพฎ่ฐไธไธชๅ็ฑปไปปๅกใ
ๅฎ้ชไฝฟ็จไบcifar-10ๆๅปบไบไธไธชๆฐๆฎ้ใtextไธบ็ฑปๅซ็ไธญๆๅ็งฐ๏ผไธๅ
ฑๆ5w็่ฎญ็ป้ๅ1w็้ช่ฏ้ใ
jsonไธญ็ๆ ทไพๅฆไธ
{"text_id":2, "text":"้ธ","image_ids":[0]}
{"text_id":1, "text":"ๆฑฝ่ฝฆ","image_ids":[4001]}
ๆฐๆฎ่ฝฌๆข่ฟ็จๅ่ฎญ็ป่ฟ็จๆฒกๆๆฅ้ใ
่ฎญ็ป่ฟ็จไธญ๏ผlossๅชๆ็ฌฌไธไธชepochไธ้๏ผๅ็ปญไธ็ด้่กใไธๅจ่ฎญ็ป้ๅ้ช่ฏ้ไธๅ็กฎ็้ฝๅพไฝ๏ผไฝไบzero-shot็ๆๆใ
่ฏท้ฎไธ่ฟๅฏ่ฝๆฏไปไน้ฎ้ข๏ผ | closed | 2023-12-27T06:35:20Z | 2024-09-20T02:45:16Z | https://github.com/OFA-Sys/Chinese-CLIP/issues/240 | [] | LiqunW | 2 |
csurfer/pyheat | matplotlib | 6 | use pyheat for cgi | I watch the code,find pyheat can't be uesd for cgi | closed | 2017-02-13T03:13:29Z | 2017-02-21T02:37:58Z | https://github.com/csurfer/pyheat/issues/6 | [] | wheat-attack | 2 |
Neoteroi/BlackSheep | asyncio | 272 | Router blueprint | ##### _Note: consider using [Discussions](https://github.com/Neoteroi/BlackSheep/discussions) to open a conversation about new featuresโฆ_
**๐ Feature Request**
i don't know, though have tried to check, but i don't know if there is anything like blueprint like of flask or Router like of Fastapi, will really make sense, rather than creating new app instances and mounting it can be really costly.
Thanks for blacksheep | closed | 2022-08-02T17:30:40Z | 2023-07-15T16:39:41Z | https://github.com/Neoteroi/BlackSheep/issues/272 | [
"document"
] | Kokoserver | 4 |
healthchecks/healthchecks | django | 834 | Export and import support | I have a self hosted healthchecks instance and I had to change my VPS last month. I exported all the database and tried to import back but it's not being easy. Maybe if there is a option to export and import in some known format like json or yaml.
Is there something like that in the roadmap for this project? If not, maybe I can give a shot. | open | 2023-05-30T12:23:29Z | 2024-07-29T07:46:45Z | https://github.com/healthchecks/healthchecks/issues/834 | [
"feature"
] | pfsmorigo | 1 |
plotly/plotly.py | plotly | 4,986 | make Kaleido 1.0 (new architecture) the default for static image generation | open | 2025-01-28T15:42:59Z | 2025-02-13T21:14:59Z | https://github.com/plotly/plotly.py/issues/4986 | [
"feature",
"P1"
] | gvwilson | 0 |
|
Hironsan/BossSensor | computer-vision | 32 | Problems about './store/model.h5' | When I read โboss_train.pyโ,I found a constant named "FILE_PATH".I can't find its value "./store/model.h5" in any folders.Is it a forein file? | open | 2020-01-07T10:55:39Z | 2020-01-07T11:08:43Z | https://github.com/Hironsan/BossSensor/issues/32 | [] | SCYtelils | 1 |
junyanz/pytorch-CycleGAN-and-pix2pix | deep-learning | 1,319 | Is it possible to change shape of origin image using cyclegan? | Thanks for your contribution. I am trying to convert the 3D model to anime style, it works well so far. In fake B image, the model outlines the lines on the edges to make it more anime. Also color have change a little.
Besides lines and color, I wanna a little bit shape changes. Is it possible? | closed | 2021-09-23T04:03:56Z | 2021-12-02T21:07:28Z | https://github.com/junyanz/pytorch-CycleGAN-and-pix2pix/issues/1319 | [] | codexq123 | 1 |
coqui-ai/TTS | pytorch | 3,989 | [Bug] Install on clean Ubuntu 22.04 fails in SudachiPy build.... | ### Describe the bug
I'm trying to install TTS on an ARM server with a clean version of Ubuntu 22.04 after "apt install python3-pip".
Here's the bug in the log section.
TIA
### To Reproduce
Start with a clean version of 22.04 on an arm server.
Run "apt install python3-pip"
Run "pip install TTS"
### Expected behavior
_No response_
### Logs
```shell
pip install TTS
Collecting TTS
Using cached TTS-0.22.0.tar.gz (1.7 MB)
Installing build dependencies ... done
Getting requirements to build wheel ... done
Preparing metadata (pyproject.toml) ... done
Collecting unidecode>=1.3.2
Using cached Unidecode-1.3.8-py3-none-any.whl (235 kB)
Collecting umap-learn>=0.5.1
Using cached umap_learn-0.5.6-py3-none-any.whl (85 kB)
Collecting coqpit>=0.0.16
Using cached coqpit-0.0.17-py3-none-any.whl (13 kB)
Collecting torch>=2.1
Using cached torch-2.4.1-cp310-cp310-manylinux2014_aarch64.whl (89.7 MB)
Collecting bangla
Using cached bangla-0.0.2-py2.py3-none-any.whl (6.2 kB)
Collecting spacy[ja]>=3
Using cached spacy-3.7.6.tar.gz (1.3 MB)
Installing build dependencies ... done
Getting requirements to build wheel ... done
Installing backend dependencies ... done
Preparing metadata (pyproject.toml) ... done
Collecting bnnumerizer
Using cached bnnumerizer-0.0.2.tar.gz (4.7 kB)
Preparing metadata (setup.py) ... done
Collecting einops>=0.6.0
Using cached einops-0.8.0-py3-none-any.whl (43 kB)
Collecting inflect>=5.6.0
Using cached inflect-7.3.1-py3-none-any.whl (34 kB)
Collecting trainer>=0.0.32
Using cached trainer-0.0.36-py3-none-any.whl (51 kB)
Collecting flask>=2.0.1
Using cached flask-3.0.3-py3-none-any.whl (101 kB)
Collecting scipy>=1.11.2
Using cached scipy-1.14.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (35.6 MB)
Collecting packaging>=23.1
Using cached packaging-24.1-py3-none-any.whl (53 kB)
Collecting fsspec>=2023.6.0
Using cached fsspec-2024.9.0-py3-none-any.whl (179 kB)
Collecting nltk
Using cached nltk-3.9.1-py3-none-any.whl (1.5 MB)
Collecting gruut[de,es,fr]==2.2.3
Using cached gruut-2.2.3.tar.gz (73 kB)
Preparing metadata (setup.py) ... done
Collecting anyascii>=0.3.0
Using cached anyascii-0.3.2-py3-none-any.whl (289 kB)
Collecting tqdm>=4.64.1
Using cached tqdm-4.66.5-py3-none-any.whl (78 kB)
Collecting cython>=0.29.30
Using cached Cython-3.0.11-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (3.5 MB)
Collecting transformers>=4.33.0
Using cached transformers-4.44.2-py3-none-any.whl (9.5 MB)
Collecting numba>=0.57.0
Using cached numba-0.60.0-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl (3.4 MB)
Collecting g2pkk>=0.1.1
Using cached g2pkk-0.1.2-py3-none-any.whl (25 kB)
Collecting encodec>=0.1.1
Using cached encodec-0.1.1.tar.gz (3.7 MB)
Preparing metadata (setup.py) ... done
Collecting pyyaml>=6.0
Using cached PyYAML-6.0.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (718 kB)
Collecting scikit-learn>=1.3.0
Using cached scikit_learn-1.5.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (12.5 MB)
Collecting aiohttp>=3.8.1
Using cached aiohttp-3.10.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (1.2 MB)
Collecting torchaudio
Using cached torchaudio-2.4.1-cp310-cp310-manylinux2014_aarch64.whl (1.7 MB)
Collecting jamo
Using cached jamo-0.4.1-py3-none-any.whl (9.5 kB)
Collecting jieba
Using cached jieba-0.42.1.tar.gz (19.2 MB)
Preparing metadata (setup.py) ... done
Collecting pandas<2.0,>=1.4
Using cached pandas-1.5.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (11.4 MB)
Collecting bnunicodenormalizer
Using cached bnunicodenormalizer-0.1.7-py3-none-any.whl (23 kB)
Collecting pypinyin
Using cached pypinyin-0.52.0-py2.py3-none-any.whl (833 kB)
Collecting numpy==1.22.0
Using cached numpy-1.22.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (13.4 MB)
Collecting soundfile>=0.12.0
Using cached soundfile-0.12.1-py2.py3-none-any.whl (24 kB)
Collecting pysbd>=0.3.4
Using cached pysbd-0.3.4-py3-none-any.whl (71 kB)
Collecting hangul-romanize
Using cached hangul_romanize-0.1.0-py3-none-any.whl (4.6 kB)
Collecting librosa>=0.10.0
Using cached librosa-0.10.2.post1-py3-none-any.whl (260 kB)
Collecting matplotlib>=3.7.0
Using cached matplotlib-3.9.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (8.2 MB)
Collecting num2words
Using cached num2words-0.5.13-py3-none-any.whl (143 kB)
Requirement already satisfied: Babel<3.0.0,>=2.8.0 in /usr/lib/python3/dist-packages (from gruut[de,es,fr]==2.2.3->TTS) (2.8.0)
Collecting dateparser~=1.1.0
Using cached dateparser-1.1.8-py2.py3-none-any.whl (293 kB)
Collecting gruut-ipa<1.0,>=0.12.0
Using cached gruut-ipa-0.13.0.tar.gz (101 kB)
Preparing metadata (setup.py) ... done
Collecting gruut_lang_en~=2.0.0
Using cached gruut_lang_en-2.0.1.tar.gz (15.3 MB)
Preparing metadata (setup.py) ... done
Collecting jsonlines~=1.2.0
Using cached jsonlines-1.2.0-py2.py3-none-any.whl (7.6 kB)
Collecting networkx<3.0.0,>=2.5.0
Using cached networkx-2.8.8-py3-none-any.whl (2.0 MB)
Collecting python-crfsuite~=0.9.7
Using cached python-crfsuite-0.9.10.tar.gz (478 kB)
Preparing metadata (setup.py) ... done
Collecting gruut_lang_fr~=2.0.0
Using cached gruut_lang_fr-2.0.2.tar.gz (10.9 MB)
Preparing metadata (setup.py) ... done
Collecting gruut_lang_de~=2.0.0
Using cached gruut_lang_de-2.0.1.tar.gz (18.1 MB)
Preparing metadata (setup.py) ... done
Collecting gruut_lang_es~=2.0.0
Using cached gruut_lang_es-2.0.1.tar.gz (31.4 MB)
Preparing metadata (setup.py) ... done
Collecting async-timeout<5.0,>=4.0
Using cached async_timeout-4.0.3-py3-none-any.whl (5.7 kB)
Collecting multidict<7.0,>=4.5
Using cached multidict-6.0.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (126 kB)
Collecting frozenlist>=1.1.1
Using cached frozenlist-1.4.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (238 kB)
Collecting aiohappyeyeballs>=2.3.0
Using cached aiohappyeyeballs-2.4.0-py3-none-any.whl (12 kB)
Requirement already satisfied: attrs>=17.3.0 in /usr/lib/python3/dist-packages (from aiohttp>=3.8.1->TTS) (21.2.0)
Collecting yarl<2.0,>=1.0
Using cached yarl-1.9.11-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (462 kB)
Collecting aiosignal>=1.1.2
Using cached aiosignal-1.3.1-py3-none-any.whl (7.6 kB)
Collecting Jinja2>=3.1.2
Using cached jinja2-3.1.4-py3-none-any.whl (133 kB)
Collecting blinker>=1.6.2
Using cached blinker-1.8.2-py3-none-any.whl (9.5 kB)
Collecting click>=8.1.3
Using cached click-8.1.7-py3-none-any.whl (97 kB)
Collecting itsdangerous>=2.1.2
Using cached itsdangerous-2.2.0-py3-none-any.whl (16 kB)
Collecting Werkzeug>=3.0.0
Using cached werkzeug-3.0.4-py3-none-any.whl (227 kB)
Requirement already satisfied: more-itertools>=8.5.0 in /usr/lib/python3/dist-packages (from inflect>=5.6.0->TTS) (8.10.0)
Collecting typeguard>=4.0.1
Using cached typeguard-4.3.0-py3-none-any.whl (35 kB)
Collecting pooch>=1.1
Using cached pooch-1.8.2-py3-none-any.whl (64 kB)
Collecting typing-extensions>=4.1.1
Using cached typing_extensions-4.12.2-py3-none-any.whl (37 kB)
Collecting soxr>=0.3.2
Using cached soxr-0.5.0.post1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (221 kB)
Collecting audioread>=2.1.9
Using cached audioread-3.0.1-py3-none-any.whl (23 kB)
Collecting lazy-loader>=0.1
Using cached lazy_loader-0.4-py3-none-any.whl (12 kB)
Collecting decorator>=4.3.0
Using cached decorator-5.1.1-py3-none-any.whl (9.1 kB)
Collecting librosa>=0.10.0
Using cached librosa-0.10.2-py3-none-any.whl (260 kB)
Using cached librosa-0.10.1-py3-none-any.whl (253 kB)
Using cached librosa-0.10.0.post2-py3-none-any.whl (253 kB)
Collecting pooch<1.7,>=1.0
Using cached pooch-1.6.0-py3-none-any.whl (56 kB)
Collecting librosa>=0.10.0
Using cached librosa-0.10.0.post1-py3-none-any.whl (252 kB)
Using cached librosa-0.10.0-py3-none-any.whl (252 kB)
Collecting joblib>=0.14
Using cached joblib-1.4.2-py3-none-any.whl (301 kB)
Collecting msgpack>=1.0
Using cached msgpack-1.0.8-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (376 kB)
Collecting matplotlib>=3.7.0
Using cached matplotlib-3.9.1.post1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (8.2 MB)
Using cached matplotlib-3.9.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (8.2 MB)
Using cached matplotlib-3.8.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (11.4 MB)
Collecting kiwisolver>=1.3.1
Using cached kiwisolver-1.4.7-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (1.4 MB)
Collecting fonttools>=4.22.0
Using cached fonttools-4.53.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (4.6 MB)
Collecting python-dateutil>=2.7
Using cached python_dateutil-2.9.0.post0-py2.py3-none-any.whl (229 kB)
Collecting pillow>=8
Using cached pillow-10.4.0-cp310-cp310-manylinux_2_28_aarch64.whl (4.4 MB)
Collecting cycler>=0.10
Using cached cycler-0.12.1-py3-none-any.whl (8.3 kB)
Collecting contourpy>=1.0.1
Using cached contourpy-1.3.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (308 kB)
Requirement already satisfied: pyparsing>=2.3.1 in /usr/lib/python3/dist-packages (from matplotlib>=3.7.0->TTS) (2.4.7)
Collecting docopt>=0.6.2
Using cached docopt-0.6.2.tar.gz (25 kB)
Preparing metadata (setup.py) ... done
Collecting llvmlite<0.44,>=0.43.0dev0
Using cached llvmlite-0.43.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (42.9 MB)
Requirement already satisfied: pytz>=2020.1 in /usr/lib/python3/dist-packages (from pandas<2.0,>=1.4->TTS) (2022.1)
Collecting threadpoolctl>=3.1.0
Using cached threadpoolctl-3.5.0-py3-none-any.whl (18 kB)
Collecting scipy>=1.11.2
Using cached scipy-1.14.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (35.6 MB)
Using cached scipy-1.13.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (33.7 MB)
Using cached scipy-1.13.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (33.7 MB)
Using cached scipy-1.12.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (34.8 MB)
Using cached scipy-1.11.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (32.9 MB)
Collecting cffi>=1.0
Using cached cffi-1.17.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (448 kB)
Collecting cymem<2.1.0,>=2.0.2
Using cached cymem-2.0.8-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (44 kB)
Collecting typer<1.0.0,>=0.3.0
Using cached typer-0.12.5-py3-none-any.whl (47 kB)
Requirement already satisfied: setuptools in /usr/lib/python3/dist-packages (from spacy[ja]>=3->TTS) (59.6.0)
Collecting spacy-loggers<2.0.0,>=1.0.0
Using cached spacy_loggers-1.0.5-py3-none-any.whl (22 kB)
Collecting wasabi<1.2.0,>=0.9.1
Using cached wasabi-1.1.3-py3-none-any.whl (27 kB)
Requirement already satisfied: requests<3.0.0,>=2.13.0 in /usr/lib/python3/dist-packages (from spacy[ja]>=3->TTS) (2.25.1)
Collecting murmurhash<1.1.0,>=0.28.0
Using cached murmurhash-1.0.10-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (29 kB)
Collecting pydantic!=1.8,!=1.8.1,<3.0.0,>=1.7.4
Using cached pydantic-2.9.0-py3-none-any.whl (434 kB)
Collecting catalogue<2.1.0,>=2.0.6
Using cached catalogue-2.0.10-py3-none-any.whl (17 kB)
Collecting spacy-legacy<3.1.0,>=3.0.11
Using cached spacy_legacy-3.0.12-py2.py3-none-any.whl (29 kB)
Collecting langcodes<4.0.0,>=3.2.0
Using cached langcodes-3.4.0-py3-none-any.whl (182 kB)
Collecting srsly<3.0.0,>=2.4.3
Using cached srsly-2.4.8-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (488 kB)
Collecting thinc<8.3.0,>=8.2.2
Using cached thinc-8.2.5-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (868 kB)
Collecting preshed<3.1.0,>=3.0.2
Using cached preshed-3.0.9-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl (150 kB)
Collecting weasel<0.5.0,>=0.1.0
Using cached weasel-0.4.1-py3-none-any.whl (50 kB)
Collecting sudachipy!=0.6.1,>=0.5.2
Using cached SudachiPy-0.6.8.tar.gz (161 kB)
Installing build dependencies ... done
Getting requirements to build wheel ... error
error: subprocess-exited-with-error
ร Getting requirements to build wheel did not run successfully.
โ exit code: 1
โฐโ> [36 lines of output]
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/pip/_vendor/pep517/in_process/_in_process.py", line 363, in <module>
main()
File "/usr/lib/python3/dist-packages/pip/_vendor/pep517/in_process/_in_process.py", line 345, in main
json_out['return_val'] = hook(**hook_input['kwargs'])
File "/usr/lib/python3/dist-packages/pip/_vendor/pep517/in_process/_in_process.py", line 130, in get_requires_for_build_wheel
return hook(config_settings)
File "/usr/lib/python3/dist-packages/setuptools/build_meta.py", line 162, in get_requires_for_build_wheel
return self._get_build_requires(
File "/usr/lib/python3/dist-packages/setuptools/build_meta.py", line 143, in _get_build_requires
self.run_setup()
File "/usr/lib/python3/dist-packages/setuptools/build_meta.py", line 267, in run_setup
super(_BuildMetaLegacyBackend,
File "/usr/lib/python3/dist-packages/setuptools/build_meta.py", line 158, in run_setup
exec(compile(code, __file__, 'exec'), locals())
File "setup.py", line 18, in <module>
setup(
File "/usr/lib/python3/dist-packages/setuptools/__init__.py", line 153, in setup
return distutils.core.setup(**attrs)
File "/usr/lib/python3/dist-packages/setuptools/_distutils/core.py", line 109, in setup
_setup_distribution = dist = klass(attrs)
File "/usr/lib/python3/dist-packages/setuptools/dist.py", line 459, in __init__
_Distribution.__init__(
File "/usr/lib/python3/dist-packages/setuptools/_distutils/dist.py", line 293, in __init__
self.finalize_options()
File "/usr/lib/python3/dist-packages/setuptools/dist.py", line 836, in finalize_options
for ep in sorted(loaded, key=by_order):
File "/usr/lib/python3/dist-packages/setuptools/dist.py", line 835, in <lambda>
loaded = map(lambda e: e.load(), filtered)
File "/usr/lib/python3/dist-packages/pkg_resources/__init__.py", line 2464, in load
self.require(*args, **kwargs)
File "/usr/lib/python3/dist-packages/pkg_resources/__init__.py", line 2487, in require
items = working_set.resolve(reqs, env, installer, extras=self.extras)
File "/usr/lib/python3/dist-packages/pkg_resources/__init__.py", line 782, in resolve
raise VersionConflict(dist, req).with_context(dependent_req)
pkg_resources.VersionConflict: (setuptools 59.6.0 (/usr/lib/python3/dist-packages), Requirement.parse('setuptools>=62.4'))
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
error: subprocess-exited-with-error
ร Getting requirements to build wheel did not run successfully.
โ exit code: 1
โฐโ> See above for output.
note: This error originates from a subprocess, and is likely not a problem with pip.
```
### Environment
```shell
Not finished.
```
### Additional context
_No response_ | closed | 2024-09-06T12:47:00Z | 2024-12-28T11:58:05Z | https://github.com/coqui-ai/TTS/issues/3989 | [
"bug",
"wontfix"
] | pwayner | 8 |
miguelgrinberg/microblog | flask | 314 | Part XVI: Full-Text Search / Constant Redirects | I am enjoying your tutorial, thank you. I just wanted to reference this Issue #219
I am using your files from the Git Repo, and I ran into the same issue #219. I found this via Google Fu.
[https://stackoverflow.com/questions/61237524/validating-get-params-with-wtforms-in-flask](https://stackoverflow.com/questions/61237524/validating-get-params-with-wtforms-in-flask)
Lead me to update:
app/main/forms.py
YOURS:
```
if 'csrf_enabled' not in kwargs:
kwargs['csrf_enabled'] = False
```
I added:
```
if 'meta' not in kwargs:
kwargs['meta'] = {'csrf': False}
```
By doing so allowed "g.search_form.validate" (app/main/routes.py) to return True when I put data into the 'q' field. Until I had that, I would always be redirected.
```
192.168.1.188 - - [22/Jan/2022 16:57:15] "GET /search?q=crap HTTP/1.1" 302 -
192.168.1.188 - - [22/Jan/2022 16:57:15] "GET /explore HTTP/1.1" 200 -
```
| closed | 2022-01-22T22:08:02Z | 2022-01-23T00:16:18Z | https://github.com/miguelgrinberg/microblog/issues/314 | [
"bug"
] | nuonce | 2 |
recommenders-team/recommenders | machine-learning | 1,918 | [FEATURE] Find a way to run the tests for external contributors after approval from core devs | ### Description
<!--- Describe your expected feature in detail -->
Currently, external contributors can do PRs, but after a core dev approves and executes to run the tests, there is an error:
```
Run azure/login@v1
Error: Az CLI Login failed. Please check the credentials and make sure az is installed on the runner. For more information refer https://aka.ms/create-secrets-for-GitHub-workflows
```
See example #1916
### Expected behavior with the suggested feature
<!--- For example: -->
<!--- *Adding algorithm xxx will help people understand more about xxx use case scenarios. -->
### Other Comments
Related to https://github.com/microsoft/recommenders/pull/1840 | closed | 2023-04-11T10:47:26Z | 2023-04-11T10:49:12Z | https://github.com/recommenders-team/recommenders/issues/1918 | [
"enhancement"
] | miguelgfierro | 1 |
keras-team/keras | data-science | 20,686 | Error after upgrade from 3.6 to 3.7 | I have a model, that use combined input.
Using version prior to 3.7 it works very fine, when using this to predict

Dropping the tf.function and using a normal predict also works fine
Output is typical something like this

Upgrading to keras 3.7, using the same model result in this output

Do you see the difference?
shape[0] for x and b is now None
But the real problem is the prediction, that is no longer correct
Just to be sure, I print the shape of x and b before clalling the tf.function
(1238, 42)
(1238, 15)
It looks like there is a problem finding the batch size for combined input, so for now I will have to stay at 3.6 | open | 2024-12-24T16:30:40Z | 2024-12-31T13:56:13Z | https://github.com/keras-team/keras/issues/20686 | [
"type:Bug"
] | ThorvaldAagaard | 4 |
pallets-eco/flask-sqlalchemy | flask | 350 | http://flask-sqlalchemy.pocoo.org/2.1/ down? | http://flask-sqlalchemy.pocoo.org/2.1/
down?
| closed | 2015-11-19T06:08:33Z | 2020-12-05T21:31:20Z | https://github.com/pallets-eco/flask-sqlalchemy/issues/350 | [] | rlam3 | 3 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.