repo_name
stringlengths 9
75
| topic
stringclasses 30
values | issue_number
int64 1
203k
| title
stringlengths 1
976
| body
stringlengths 0
254k
| state
stringclasses 2
values | created_at
stringlengths 20
20
| updated_at
stringlengths 20
20
| url
stringlengths 38
105
| labels
sequencelengths 0
9
| user_login
stringlengths 1
39
| comments_count
int64 0
452
|
---|---|---|---|---|---|---|---|---|---|---|---|
keras-team/keras | machine-learning | 20,378 | keras.io examples using model.stateless_call aren't setting training=True and aren't managing non_trainable state | both the examples
* https://keras.io/guides/writing_a_custom_training_loop_in_jax/
* https://keras.io/guides/distributed_training_with_jax/
have a training loop that uses `stateless_call`, but neither sets `training=True` ( which means they are using the `training=False` behaviour )
i guess it's not strictly relevant to the first example, which has no interesting non_trainables, but the second example has batch norm and dropout.
additionally i don't see where this non_trainable state would be managed?
* re: batchnorm; each replica might have the same params, but is getting a different batch, so would collect different batch norm stats after the stateless_call. these aren't being aggregated?
* re: dropout; the RNGKeys for dropout are in the non_trainable, but never change. shouldn't these be split after each forward pass?
| closed | 2024-10-18T12:28:37Z | 2024-10-18T21:14:00Z | https://github.com/keras-team/keras/issues/20378 | [] | matpalm | 2 |
AutoViML/AutoViz | scikit-learn | 90 | Cannot even import autoviz? | Not able to even import autoviz after installation. Thought to try it, but doesn't work.
Please see the screenshot below:



| closed | 2023-08-01T08:31:48Z | 2023-08-08T14:57:39Z | https://github.com/AutoViML/AutoViz/issues/90 | [] | mkumar73 | 2 |
MolSSI/cookiecutter-cms | pytest | 143 | Where to write the dependency? | Sorry this is a very simple question, I have selected the
```
Select dependency_source:
1 - conda-forge
```
I wonder where should I add dependencies for the installation?
Normally, I would add the dependencies to `install_requires` in `setup.py`, but that would mean that the dependency is installed via pip.
I wonder where should I put the dependencies? | closed | 2021-10-28T12:38:08Z | 2021-10-28T16:41:01Z | https://github.com/MolSSI/cookiecutter-cms/issues/143 | [] | xiki-tempula | 4 |
huggingface/datasets | nlp | 6,537 | Adding support for netCDF (*.nc) files | ### Feature request
netCDF (*.nc) is a file format for storing multidimensional scientific data, which is used by packages like `xarray` (labelled multi-dimensional arrays in Python). It would be nice to have native support for netCDF in `datasets`.
### Motivation
When uploading *.nc files onto Huggingface Hub through the `datasets` API, I would like to be able to preview the dataset without converting it to another format.
### Your contribution
I can submit a PR, provided I have the time. | open | 2023-12-27T09:27:29Z | 2023-12-27T20:46:53Z | https://github.com/huggingface/datasets/issues/6537 | [
"enhancement"
] | shermansiu | 3 |
ultralytics/ultralytics | pytorch | 19,243 | Run failed when using nms in batch infer | ### Search before asking
- [x] I have searched the Ultralytics YOLO [issues](https://github.com/ultralytics/ultralytics/issues) and found no similar bug report.
### Ultralytics YOLO Component
Export
### Bug
I want to convert the model to ONNX format and integrate NMS into the model. However, after conversion, when I try to perform batch inference with dynamic shapes, I get a Split error. When I change the batch size during conversion, the error numbers also change accordingly. It only works correctly when the preset batch size matches the inference batch size. Is this because some tensor values were converted to constants somewhere? I hope to get an answer.
err message:
```
[E:onnxruntime:, sequential_executor.cc:516 ExecuteKernel] Non-zero status code returned while running Split node. Name:'/Split_2' Status Message: Cannot split using values in 'split' attribute. Axis=0 Input shape={3,8400} NumOutputs=1 Num entries in 'split' (must equal number of outputs) was 1 Sum of sizes in 'split' (must equal size of selected axis) was 1
```
the split node:
<img width="740" alt="Image" src="https://github.com/user-attachments/assets/a99745fd-0b79-4d2d-b6ed-efc33ddeb666" />
### Environment
ultralytics 8.3.74
torch 2.4.0
onnx 1.17.0
onnxruntime-gpu 1.19.0
opencv-python 4.10.0.84
python 3.12.4
### Minimal Reproducible Example
export code:
```
from ultralytics import YOLO
import cv2
import numpy as np
model = YOLO('yolov11n-face.pt') # load a pretrained YOLOv8n detection model
image = cv2.imread('../datasets/face_test/image.jpg')
results = model((image, image2, image3, image4)) # predict on an image
model.export(format="onnx", nms=True, dynamic=True, simplify=True)
```
test code:
```
image = cv2.imread('../datasets/face_test/image1.jpg')
image2 = cv2.imread('../datasets/face_test/image2.jpg')
results = model((image, image2)) # predict on an image
output_path = 'yolov11n-face.onnx'
image_list = [image, image2]
w = torch.Tensor([i.shape[0] for i in image_list]).to(model.device)
h = torch.Tensor([i.shape[1] for i in image_list]).to(model.device)
in_tensor = preprocess(image_list) # (2, 640, 640, 3)
onnx_model = onnx.load(output_path)
onnx.checker.check_model(onnx_model)
ort_session = onnxruntime.InferenceSession(output_path)
ort_inputs = {
ort_session.get_inputs()[0].name: to_numpy(in_tensor.cpu())
}
# print(ort_inputs)
for _ in range(1):
ort_outs = ort_session.run(None, ort_inputs)
print('onnx out:')
print(ort_outs)
```
### Additional
_No response_
### Are you willing to submit a PR?
- [ ] Yes I'd like to help by submitting a PR! | closed | 2025-02-14T08:43:59Z | 2025-02-18T13:31:04Z | https://github.com/ultralytics/ultralytics/issues/19243 | [
"bug",
"fixed",
"exports"
] | sajomanaka | 11 |
timkpaine/lantern | plotly | 141 | add replay for lantern live | closed | 2018-01-24T15:13:58Z | 2018-09-13T18:48:46Z | https://github.com/timkpaine/lantern/issues/141 | [
"feature"
] | timkpaine | 1 |
|
syrupy-project/syrupy | pytest | 764 | Colored 1.5 broke syrupy | **Describe the bug**
I run `poetry update` this morning and it updated `colored` to 1.5. Which is a legit update according to pyrupy `pyproject.toml`. From the change log of `colored` I can see they renamed some functions like `bg` to `back` and `fg` to `fore`. These changes broke syrupy.
<!-- A clear and concise description of what the bug is. -->
**To reproduce**
Update to colored 1.5 and try using syrupy in a CI pipeline on github/gitlab
<!--
Steps to reproduce the behavior:
1. Setup '...'
2. Run command '....'
3. See error
-->
**Expected behavior**
<!-- A clear and concise description of what you expected to happen. -->
It should not break or it should not allow to update to 1.5
**Screenshots**
<!-- If applicable, add screenshots to help explain your problem. -->
**Environment (please complete the following information):**
- OS: I am not sure whatever my CI is using, I think Alpine
- Syrupy Version: 3.0.6
- Python Version: 3.11
**Additional context**
<!-- Add any other context about the problem here. -->
| closed | 2023-06-19T07:46:07Z | 2023-06-20T11:44:49Z | https://github.com/syrupy-project/syrupy/issues/764 | [] | dfioravanti | 4 |
serengil/deepface | deep-learning | 945 | Get Landmarks | Dear all,
maybe I am too stupid to find it. How can I get the Landmarks when I am using extract_faces?
kr
Dirk | closed | 2024-01-08T10:36:16Z | 2024-01-08T10:59:55Z | https://github.com/serengil/deepface/issues/945 | [
"question"
] | DirkWuerdemann | 1 |
jacobgil/pytorch-grad-cam | computer-vision | 545 | `SemanticSegmentationTarget` throws tensor device casting error | ## Bug
The following code
```python
targets = [
SemanticSegmentationTarget(2, np.ones((128, 128), dtype=np.float32))
]
with GradCAM(model=m, target_layers=l, reshape_transform=reshape_transform) as cam:
mask = cam(input_tensor=timg, targets=targets)
```
throws a RuntimeError
```
RuntimeError: Expected all tensors to be on the same device, but found at least two devices, mps:0 and cpu!
```
## Full Stack Trace
```
{
"name": "RuntimeError",
"message": "Expected all tensors to be on the same device, but found at least two devices, mps:0 and cpu!",
"stack": "---------------------------------------------------------------------------
RuntimeError Traceback (most recent call last)
Cell In[5], line 30
24 targets = [
25 SemanticSegmentationTarget(2, np.ones((128, 128), dtype=np.float32))
26 # SemanticSegmentationTarget(10, np.ones((128, 128), dtype=np.float32))
27 ]
29 with GradCAM(model=m, target_layers=l, reshape_transform=reshape_transform) as cam:
---> 30 mask = cam(
31 input_tensor=timg,
32 targets=targets)
34 pt.imshow(show_on_image(img, mask[0]))
File ~/.pyenv/versions/3.12.3/lib/python3.12/site-packages/pytorch_grad_cam/base_cam.py:186, in BaseCAM.__call__(self, input_tensor, targets, aug_smooth, eigen_smooth)
183 if aug_smooth is True:
184 return self.forward_augmentation_smoothing(input_tensor, targets, eigen_smooth)
--> 186 return self.forward(input_tensor, targets, eigen_smooth)
File ~/.pyenv/versions/3.12.3/lib/python3.12/site-packages/pytorch_grad_cam/base_cam.py:98, in BaseCAM.forward(self, input_tensor, targets, eigen_smooth)
96 if self.uses_gradients:
97 self.model.zero_grad()
---> 98 loss = sum([target(output) for target, output in zip(targets, outputs)])
99 loss.backward(retain_graph=True)
101 # In most of the saliency attribution papers, the saliency is
102 # computed with a single target layer.
103 # Commonly it is the last convolutional layer.
(...)
108 # use all conv layers for example, all Batchnorm layers,
109 # or something else.
File ~/.pyenv/versions/3.12.3/lib/python3.12/site-packages/pytorch_grad_cam/utils/model_targets.py:68, in SemanticSegmentationTarget.__call__(self, model_output)
67 def __call__(self, model_output):
---> 68 return (model_output[self.category, :, :] * self.mask).sum()
69 return (model_output[self.category, :, :] * self.mask.to(model_output.device)).sum()
RuntimeError: Expected all tensors to be on the same device, but found at least two devices, mps:0 and cpu!"
}
```
## Version
grad-cam==1.5.4
on **Python 3.12.3**
| closed | 2024-12-12T09:11:55Z | 2024-12-12T09:54:00Z | https://github.com/jacobgil/pytorch-grad-cam/issues/545 | [] | shreyass-ranganatha | 0 |
deepset-ai/haystack | pytorch | 8,289 | Remove deprecated `SentenceWindowRetrieval` component | It was left as a stubbed component in 2.4 to aid backward compatibility with existing pipelines. It should be removed in 2.5. | closed | 2024-08-26T12:21:59Z | 2024-08-28T08:04:53Z | https://github.com/deepset-ai/haystack/issues/8289 | [
"P2",
"2.x"
] | shadeMe | 0 |
hankcs/HanLP | nlp | 693 | 人名识别导致分词错误 | <!--
注意事项和版本号必填,否则不回复。若希望尽快得到回复,请按模板认真填写,谢谢合作。
-->
## 注意事项
请确认下列注意事项:
* 我已仔细阅读下列文档,都没有找到答案:
- [首页文档](https://github.com/hankcs/HanLP)
- [wiki](https://github.com/hankcs/HanLP/wiki)
- [常见问题](https://github.com/hankcs/HanLP/wiki/FAQ)
* 我已经通过[Google](https://www.google.com/#newwindow=1&q=HanLP)和[issue区检索功能](https://github.com/hankcs/HanLP/issues)搜索了我的问题,也没有找到答案。
* 我明白开源社区是出于兴趣爱好聚集起来的自由社区,不承担任何责任或义务。我会礼貌发言,向每一个帮助我的人表示感谢。
* [x] 我在此括号内输入x打钩,代表上述事项确认完毕。
## 版本号
<!-- 发行版请注明jar文件名去掉拓展名的部分;GitHub仓库版请注明master还是portable分支 -->
当前最新版本号是:1.5.2
我使用的版本是:1.3.4
<!--以上属于必填项,以下可自由发挥-->
## 我的问题
句子 :
```
商改后要买另外的保险保费怎么算
```
分词结果:
```
[商改后/nr, 要买/nz, 另外/c, 的/ude1, 保险/n, 保费/n, 怎么/ryv, 算/v]
```
其中 商改后 被识别为一个人名
## 复现问题
<!-- 你是如何操作导致产生问题的?比如修改了代码?修改了词典或模型?-->
### 步骤
1. 首先……
2. 然后……
3. 接着……
### 触发代码
```
public void testIssue1234() throws Exception
{
System.out.println(StandardTokenizer.segment("商改后要买另外的保险保费怎么算"));
}
```
### 期望输出
<!-- 你希望输出什么样的正确结果?-->
```
[商改/v,后/, 要买/nz, 另外/c, 的/ude1, 保险/n, 保费/n, 怎么/ryv, 算/v]
```
### 实际输出
```
[商改后/nr, 要买/nz, 另外/c, 的/ude1, 保险/n, 保费/n, 怎么/ryv, 算/v]
```
## 其他信息
<!-- 任何可能有用的信息,包括截图、日志、配置文件、相关issue等等。-->
| closed | 2017-11-27T03:27:56Z | 2017-12-02T04:22:36Z | https://github.com/hankcs/HanLP/issues/693 | [
"improvement"
] | iwaller | 3 |
scrapy/scrapy | python | 6,024 | Problem with Twisted AttributeError: 'EPollReactor' object has no attribute '_handleSignals' | ### Description
Since the release of Twisted 23.8.0 recently, I've had problems running Scrapy. I had to fix to the previous version, Twisted==22.10.0
### Steps to Reproduce
1. Install Scrapy in a new environment
2. Run .start() from an CrawlerProcess object
**Expected behavior:** Scrapy to crawl
**Actual behavior:**

**Reproduces how often:** Always in the new version
### Versions
Scrapy>=2.9.0
### Additional context
I've seen this fix on [stackoverflow](https://stackoverflow.com/questions/76995567/error-when-crawl-data-epollreactor-object-has-no-attribute-handlesignals)
| closed | 2023-08-29T00:04:21Z | 2023-10-17T13:08:25Z | https://github.com/scrapy/scrapy/issues/6024 | [
"bug"
] | RobertoHigor | 8 |
ckan/ckan | api | 7,639 | Solution: Integration with Tableau, Power BI and Data Studio | **Problem research and evidence**
1. Integration with top data visualization tools was suggested during the discussion: https://github.com/ckan/ckan/issues/7499 @jqnatividad concluded that it's an optimal way to enable top data visualization experience.
2. During the research there were 3 respondents that mentioned integrations with BI Tools:
- Customizable dashboard - linking Tableau, Power BI, Google Data Studio
- Make it easier to send data to Tableau, Power BI
- Integration with Tableau
⚡️ If you have a datastore you can integrate a BI tool.
**Solution hypothesis**
Solution should provide a view of Tableau, Power BI and Data Studio on CKAN pages. It'll enable publishers enrich datasets published with visualizations and stories based on them.
Solution should enable data sending to data visualization tool with from CKAN interface.
**Solution description**
TBD
**Questions to consider**
Is this change going to break current installations?
Can we provide a backwards compatibility?
How easy is gonna be for current implementations to migrate to this new release?
Do current versions of CKAN have the adequate resources/support to migrate to this new version?
Are we going to change the database schema?
Are we going to change the API?
Are we going to deprecate Interfaces? | open | 2023-06-10T15:48:27Z | 2023-08-01T06:44:58Z | https://github.com/ckan/ckan/issues/7639 | [
"To Discuss"
] | thegostev | 3 |
microsoft/nni | data-science | 5,702 | How to use loss instead of accuray in HPO | **Describe the issue**: I just started using nni and followed the /hpo_quickstart_pytorch/ example to inplement.
But, instead of accuracy, I need to thave the main monitor and optimize for the loss, and the lower loss, the better, obviously.
I used nni.report_intermediate_result(valLossSingle) to report the loss. I think I need to let the tuner know that this is actually loss, rather than accuracy. How to do that?
**Environment**:
- NNI version: 2.7
- Training service (local|remote|pai|aml|etc): local
- Client OS: ubuntu
- Server OS (for remote mode only):
- Python version: 3.10
- PyTorch/TensorFlow version: 2
- Is conda/virtualenv/venv used?: conda
- Is running in Docker?: no
**Configuration**:
- Experiment config (remember to remove secrets!):
- Search space:
**Log message**:
- nnimanager.log:
- dispatcher.log:
- nnictl stdout and stderr:
<!--
Where can you find the log files:
LOG: https://github.com/microsoft/nni/blob/master/docs/en_US/Tutorial/HowToDebug.md#experiment-root-director
STDOUT/STDERR: https://nni.readthedocs.io/en/stable/reference/nnictl.html#nnictl-log-stdout
-->
**How to reproduce it?**: | open | 2023-10-27T06:48:29Z | 2023-10-27T06:48:29Z | https://github.com/microsoft/nni/issues/5702 | [] | 0i0i0i | 0 |
tflearn/tflearn | data-science | 1,071 | tflearn Installation issues. | Hello,
I've been trying to install tflearn module for a while now and haven't been able to do so. I have checked most documentations and other problems but none seem to be usefull.
I am using anaconda extension. I have tensorflow installed, it's current installed version is 1.8 and as far as I understood it should be higher or equal to 1.0.
I can't use anaconda navigator for some reason but the prompt is working just fine and that's where I've been trying.
`import tflearn
curses is not supported on this machine (please install/reinstall curses for an optimal experience)
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "C:\Users\AppData\Local\Continuum\anaconda3\lib\site-packages\tflearn-0.3.2-py3.6.egg\tflearn\__init__.py", line 62, in <module>
File "C:\Users\AppData\Local\Continuum\anaconda3\lib\site-packages\tflearn-0.3.2-py3.6.egg\tflearn\datasets\__init__.py", line 1, in <module>
# -*- coding: utf-8 -*-
File "C:\Users\AppData\Local\Continuum\anaconda3\lib\site-packages\tflearn-0.3.2-py3.6.egg\tflearn\datasets\cifar10.py", line 16, in <module>
File "C:\Users\AppData\Local\Continuum\anaconda3\lib\site-packages\tflearn-0.3.2-py3.6.egg\tflearn\data_utils.py", line 7, in <module>
File "C:\Users\AppData\Local\Continuum\anaconda3\lib\site-packages\PIL\Image.py", line 60, in <module>
from . import _imaging as core
ImportError: DLL load failed: The specified module could not be found.`
That's the error that comes up when I try to import it from the prompt. It seems like it is installed but not importable
`~\AppData\Local\Continuum\anaconda3\lib\site-packages\PIL\Image.py in <module>()
58 # Also note that Image.core is not a publicly documented interface,
59 # and should be considered private and subject to change.
---> 60 from . import _imaging as core
61 if PILLOW_VERSION != getattr(core, 'PILLOW_VERSION', None):
62 raise ImportError("The _imaging extension was built for another "
ImportError: DLL load failed: The specified module could not be found.`
And that's a jupyter's notebook import error.
Please guide me!
| closed | 2018-07-03T13:03:30Z | 2018-07-04T11:29:42Z | https://github.com/tflearn/tflearn/issues/1071 | [] | Nadedic | 1 |
fastapiutils/fastapi-utils | fastapi | 340 | [BUG] Note to the installation guide | Hi! your instalation [guide](https://fastapi-utils.davidmontague.xyz/#installation) differs your guide in [github](https://github.com/dmontagu/fastapi-utils?tab=readme-ov-file#installation)
<img width="1089" alt="image" src="https://github.com/user-attachments/assets/226b09f3-44f2-46ba-876a-0c15d64788b7">
<img width="1025" alt="image" src="https://github.com/user-attachments/assets/8b0a1071-20ea-40d3-afe5-19ba6d97ce15">
| closed | 2024-10-20T09:42:59Z | 2024-11-11T14:18:23Z | https://github.com/fastapiutils/fastapi-utils/issues/340 | [
"bug"
] | ArtyomIvlev | 1 |
apache/airflow | data-science | 47,452 | Rename DAG (from models/dag) to SchedulerDAG -- or remove it! | ### Body
It is easily confused with the DAG in task sdk.
Ideally we can remove. At least rename it.
comment from ash
> That should be renamed as SchedulerDAG really -- (or possibly it can be removed now) -- it's a) a compat import while that part isn't complete, and b) where scheduler specific DAG functions go
### Committer
- [x] I acknowledge that I am a maintainer/committer of the Apache Airflow project. | open | 2025-03-06T15:01:54Z | 2025-03-06T15:04:11Z | https://github.com/apache/airflow/issues/47452 | [
"area:Scheduler",
"kind:meta"
] | dstandish | 0 |
pydata/pandas-datareader | pandas | 9 | TST: make sure tests work on python 2.6 | closed | 2015-01-22T05:34:58Z | 2015-01-22T05:53:37Z | https://github.com/pydata/pandas-datareader/issues/9 | [] | davidastephens | 0 |
|
deezer/spleeter | tensorflow | 677 | [Discussion] can we convert the pretrained model to saved model format? | can we directly download the model and convert to saved model format, then load from Cpp API? | closed | 2021-11-09T12:51:14Z | 2021-11-11T11:43:38Z | https://github.com/deezer/spleeter/issues/677 | [
"question"
] | imiskolee | 0 |
autokey/autokey | automation | 36 | Help menu/main page on github have dead links | Google Code closed down. As such the FAQ/Online Help/Scripting Help links lead to dead sites. They probably should be changed to Archive.org links or hosted elsewhere preferably. The Paypal link is defunct as well.
| closed | 2016-10-20T09:38:54Z | 2016-11-08T05:12:09Z | https://github.com/autokey/autokey/issues/36 | [] | pensive-fuzz | 1 |
JaidedAI/EasyOCR | pytorch | 923 | AttributeError: '_MultiProcessingDataLoaderIter' object has no attribute 'next' | Hi.
I am currently trying to train my own model.
I obtained the dataset as required and modified the config file. However, I get this error when trying to train.
I have already tried to decrease workers in config file.
I also tried to modify the dataset.py file, line 101 from
image, text = data_loader_iter.next()
to
image, text = next(data_loader_iter)
However, the error persist.
Thanks | closed | 2023-01-03T17:19:37Z | 2023-01-14T09:02:52Z | https://github.com/JaidedAI/EasyOCR/issues/923 | [] | proclaim5584 | 3 |
newpanjing/simpleui | django | 206 | models中图片字段展示 | 模型中我存储的图片路径为“/blog/static/images/blog/timg_PrPLvaa.jpg”,这是我访问静态资源路径,但是后台显示该字段时是用a标签包裹,自动拼接的路径不对,“http://192.168.3.28:5058/admin/blog/article/1/change/blog/static/images/blog/timg_PrPLvaa.jpg”,前边拼接的可以设计成可以让用户自定义修改,要不然点击访问不了该图片。再不然也可以不自动进行拼接,让用户保存完整链接。
**留下你的联系方式,以便与你取得联系**
QQ:13716.7399
| closed | 2019-12-23T08:53:06Z | 2019-12-26T09:23:05Z | https://github.com/newpanjing/simpleui/issues/206 | [
"enhancement"
] | User-Name-Chao | 0 |
vaexio/vaex | data-science | 1,968 | [BUG-REPORT] to_pandas_df giving error Urgent!!! | @JovanVeljanoski @maartenbreddels Hey There Can you please help me in this, its little Urgent
I'm trying to do the conversion of vaex dataframe to pandas dataframe but its giving me error shared down below. At first I've taken two dataframes and joined them on more than column with merging solution and after that I've tried converting it to pandas data frame.
Both the dataframes have these columns and datatypes:
one data frame is named as df and another is named as product
first image is of df dataframe:

second image is for product dataframe

These two tables are joined using below code:

After this we get dataframe Named megedStuff which gives us the joins. Also many time I saw vaex is not good at handling empty table cells or nan values or missing values.
When I try to convert this mergedStuff to pandas df it shows below error and also I'm attaching data with this. Maybe after converting it to csvs it will work fine but I'm not sure.
[product.csv](https://github.com/vaexio/vaex/files/8231496/product.csv)
[df.csv](https://github.com/vaexio/vaex/files/8231497/df.csv)
ArrowIndexErrorTraceback (most recent call last)
<ipython-input-11-968fb59ca07f> in <module>
----> 1 mergedStuff1 = mergedStuff.to_pandas_df()
/opt/conda/lib/python3.7/site-packages/vaex/dataframe.py in to_pandas_df(self, column_names, selection, strings, virtual, index_name, parallel, chunk_size, array_type)
3242 return iterator()
3243 else:
-> 3244 return create_pdf(self.to_dict(column_names=column_names, selection=selection, parallel=parallel, array_type=array_type))
3245
3246 @docsubst
/opt/conda/lib/python3.7/site-packages/vaex/dataframe.py in to_dict(self, column_names, selection, strings, virtual, parallel, chunk_size, array_type)
3157 yield i1, i2, dict(list(zip(column_names, [array_types.convert(chunk, array_type) for chunk in chunks])))
3158 return iterator()
-> 3159 return dict(list(zip(column_names, [array_types.convert(chunk, array_type) for chunk in self.evaluate(column_names, selection=selection, parallel=parallel)])))
3160
3161 @_hidden
/opt/conda/lib/python3.7/site-packages/vaex/dataframe.py in evaluate(self, expression, i1, i2, out, selection, filtered, array_type, parallel, chunk_size, progress)
2996 return self.evaluate_iterator(expression, s1=i1, s2=i2, out=out, selection=selection, filtered=filtered, array_type=array_type, parallel=parallel, chunk_size=chunk_size, progress=progress)
2997 else:
-> 2998 return self._evaluate_implementation(expression, i1=i1, i2=i2, out=out, selection=selection, filtered=filtered, array_type=array_type, parallel=parallel, chunk_size=chunk_size, progress=progress)
2999
3000 @docsubst
/opt/conda/lib/python3.7/site-packages/vaex/dataframe.py in _evaluate_implementation(self, expression, i1, i2, out, selection, filtered, array_type, parallel, chunk_size, raw, progress)
6351 arrays[expression][i1:i2] = blocks[i]
6352 if expression_to_evaluate:
-> 6353 df.map_reduce(assign, lambda *_: None, expression_to_evaluate, progress=progress, ignore_filter=False, selection=selection, pre_filter=use_filter, info=True, to_numpy=False, name="evaluate")
6354 def finalize_result(expression):
6355 expression_obj = expression
/opt/conda/lib/python3.7/site-packages/vaex/dataframe.py in map_reduce(self, map, reduce, arguments, progress, delay, info, to_numpy, ignore_filter, pre_filter, name, selection)
427 progressbar.add_task(task, f'map reduce: {name}')
428 task = self.executor.schedule(task)
--> 429 return self._delay(delay, task)
430
431 def apply(self, f, arguments=None, vectorize=False, multiprocessing=True):
/opt/conda/lib/python3.7/site-packages/vaex/dataframe.py in _delay(self, delay, task, progressbar)
1686 return task
1687 else:
-> 1688 self.execute()
1689 return task.get()
1690
/opt/conda/lib/python3.7/site-packages/vaex/dataframe.py in execute(self)
410 print(repr(task))
411 if self.executor.tasks:
--> 412 self.executor.execute()
413
414 async def execute_async(self):
/opt/conda/lib/python3.7/site-packages/vaex/execution.py in execute(self)
306
307 def execute(self):
--> 308 for _ in self.execute_generator():
309 pass # just eat all elements
310
/opt/conda/lib/python3.7/site-packages/vaex/execution.py in execute_generator(self, use_async)
433 dataset.row_count,
434 progress=progress,
--> 435 cancel=lambda: self._cancel(run), unpack=True, run=run, use_async=use_async)
436 duration_wallclock = time.time() - t0
437 logger.debug("executing took %r seconds", duration_wallclock)
/opt/conda/lib/python3.7/site-packages/vaex/multithreading.py in map(self, callable, iterator, count, on_error, progress, cancel, unpack, use_async, **kwargs_extra)
104 iterator = (loop.run_in_executor(self, lambda value=value: wrapped(value)) for value in cancellable_iter())
105 else:
--> 106 iterator = super(ThreadPoolIndex, self).map(wrapped, cancellable_iter())
107 total = 0
108 iterator = iter(buffer(iterator, self._max_workers + 3))
/opt/conda/lib/python3.7/concurrent/futures/_base.py in map(self, fn, timeout, chunksize, *iterables)
573 end_time = timeout + time.monotonic()
574
--> 575 fs = [self.submit(fn, *args) for args in zip(*iterables)]
576
577 # Yield must be hidden in closure so that the futures are submitted
/opt/conda/lib/python3.7/concurrent/futures/_base.py in <listcomp>(.0)
573 end_time = timeout + time.monotonic()
574
--> 575 fs = [self.submit(fn, *args) for args in zip(*iterables)]
576
577 # Yield must be hidden in closure so that the futures are submitted
/opt/conda/lib/python3.7/site-packages/vaex/multithreading.py in cancellable_iter()
90 chunk_iterator = iterator
91 def cancellable_iter():
---> 92 for value in chunk_iterator:
93 yield value
94 if cancelled:
/opt/conda/lib/python3.7/site-packages/vaex/dataset.py in chunk_iterator(self, columns, chunk_size, reverse)
647 raise KeyError(f'Oops, you tried to get column {name}, but you renamed it to {rename}')
648 columns = [self.reverse.get(name, name) for name in columns]
--> 649 for i1, i2, chunks in self.original.chunk_iterator(columns, chunk_size, reverse=reverse):
650 yield i1, i2, {self.renaming.get(name, name): ar for name, ar in chunks.items()}
651
/opt/conda/lib/python3.7/site-packages/vaex/dataset.py in chunk_iterator(self, columns, chunk_size, reverse)
1165 if column in self._dropped_names:
1166 raise KeyError(f'Oops, you tried to get column {column} while it is actually dropped')
-> 1167 yield from self.original.chunk_iterator(columns, chunk_size=chunk_size, reverse=reverse)
1168
1169 def hashed(self):
/opt/conda/lib/python3.7/site-packages/vaex/join.py in chunk_iterator(self, *args, **kwargs)
62
63 def chunk_iterator(self, *args, **kwargs):
---> 64 yield from self.original.chunk_iterator(*args, **kwargs)
65
66 def hashed(self):
/opt/conda/lib/python3.7/site-packages/vaex/dataset.py in chunk_iterator(self, columns, chunk_size, reverse)
1235 for (i1, i2, ichunks), (j1, j2, jchunks) in zip(
1236 self.left.chunk_iterator(columns_left, chunk_size, reverse=reverse),
-> 1237 self.right.chunk_iterator(columns_right, chunk_size, reverse=reverse)):
1238 # TODO: if one of the datasets does not respect the chunk_size (e.g. parquet)
1239 # this might fail
/opt/conda/lib/python3.7/site-packages/vaex/dataset.py in chunk_iterator(self, columns, chunk_size, reverse)
908 # TODO: we may be able to do this slightly more efficient by first
909 # materializing the columns
--> 910 yield from self._default_chunk_iterator(self._columns, columns, chunk_size, reverse=reverse)
911
912 def slice(self, start, end):
/opt/conda/lib/python3.7/site-packages/vaex/dataset.py in _default_chunk_iterator(self, array_map, columns, chunk_size, reverse)
506 def _default_chunk_iterator(self, array_map, columns, chunk_size, reverse=False):
507 for i1, i2, reader in self._default_lazy_chunk_iterator(array_map, columns, chunk_size, reverse):
--> 508 yield i1, i2, reader()
509
510 @abstractmethod
/opt/conda/lib/python3.7/site-packages/vaex/dataset.py in reader(i1, i2)
497 i2 = min((i + 1) * chunk_size, self.row_count)
498 def reader(i1=i1, i2=i2):
--> 499 chunks = {k: array_map[k][i1:i2] for k in columns}
500 length = i2 - i1
501 for name, chunk in chunks.items():
/opt/conda/lib/python3.7/site-packages/vaex/dataset.py in <dictcomp>(.0)
497 i2 = min((i + 1) * chunk_size, self.row_count)
498 def reader(i1=i1, i2=i2):
--> 499 chunks = {k: array_map[k][i1:i2] for k in columns}
500 length = i2 - i1
501 for name, chunk in chunks.items():
/opt/conda/lib/python3.7/site-packages/vaex/column.py in __getitem__(self, slice)
375 take_indices[mask] = 0
376 if isinstance(ar_unfiltered, supported_arrow_array_types):
--> 377 ar = ar_unfiltered.take(vaex.array_types.to_arrow(take_indices))
378 else:
379 ar = ar_unfiltered[take_indices]
/opt/conda/lib/python3.7/site-packages/pyarrow/array.pxi in pyarrow.lib.Array.take()
/opt/conda/lib/python3.7/site-packages/pyarrow/compute.py in take(data, indices, boundscheck, memory_pool)
452 """
453 options = TakeOptions(boundscheck=boundscheck)
--> 454 return call_function('take', [data, indices], options, memory_pool)
455
456
/opt/conda/lib/python3.7/site-packages/pyarrow/_compute.pyx in pyarrow._compute.call_function()
/opt/conda/lib/python3.7/site-packages/pyarrow/_compute.pyx in pyarrow._compute.Function.call()
/opt/conda/lib/python3.7/site-packages/pyarrow/error.pxi in pyarrow.lib.pyarrow_internal_check_status()
/opt/conda/lib/python3.7/site-packages/pyarrow/error.pxi in pyarrow.lib.check_status()
ArrowIndexError: Index 21245 out of bounds | closed | 2022-03-11T11:31:22Z | 2022-05-27T13:26:09Z | https://github.com/vaexio/vaex/issues/1968 | [
"difficult to reproduce"
] | ashsharma96 | 3 |
OFA-Sys/Chinese-CLIP | nlp | 331 | 能否通过文本向量化中加入指令优化图片召回率 | 目前能否能通过在文本向量化中加入类似“为这个句子生成向量以用于检索图片:”,的指令来优化图片的召回率 | open | 2024-07-22T06:41:48Z | 2024-07-22T06:41:48Z | https://github.com/OFA-Sys/Chinese-CLIP/issues/331 | [] | Amphetaminewei | 0 |
mitmproxy/pdoc | api | 246 | `IndentationError` when using a multi-line decorator for a class method | #### Problem Description
I encounter an `IndentationError` when trying to generate the documentation of a class which contains a multi-line decorator on a class method.
I believe that this is a limitation of the [`_dedent` method](https://github.com/mitmproxy/pdoc/blob/main/pdoc/doc_ast.py#L210) but I do not have a good proposal for a solution yet. If we can discuss and figure out a good approach on how to resolve this I am happy to help with fixing the issue :+1:
What would need to happen is that the actual function definition also gets dedented rather than simply the line following the decorator. See the example below.
#### Steps to reproduce the behavior:
I was able to break down the issue to the following example:
```python
import pytest
class TestDummy:
"""Test Dummy."""
@pytest.mark.parametrize(
"arg,string",
[
["hello", "world"],
],
)
def test_dummy(self, arg, string):
"""Dummy test method."""
assert arg != string
```
<details>
<summary>
Using `pdoc -o tmp example.py` generates the following error:
</summary>
```python-traceback
Traceback (most recent call last):
File "/home/max/Git/cobib/.direnv/python-3.9.2/bin/pdoc", line 33, in <module>
sys.exit(load_entry_point('pdoc', 'console_scripts', 'pdoc')())
File "/home/max/Git/pdoc/pdoc/__main__.py", line 153, in cli
pdoc.pdoc(
File "/home/max/Git/pdoc/pdoc/__init__.py", line 386, in pdoc
write(doc.Module(m))
File "/home/max/Git/pdoc/pdoc/__init__.py", line 355, in write
outfile.write_bytes(r(mod).encode())
File "/home/max/Git/pdoc/pdoc/__init__.py", line 367, in r
return render.html_module(module=mod, all_modules=all_modules)
File "/home/max/Git/pdoc/pdoc/render.py", line 70, in html_module
return env.select_template(
File "/home/max/Git/cobib/.direnv/python-3.9.2/lib/python3.9/site-packages/jinja2/environment.py", line 1090, in render
self.environment.handle_exception()
File "/home/max/Git/cobib/.direnv/python-3.9.2/lib/python3.9/site-packages/jinja2/environment.py", line 832, in handle_exception
reraise(*rewrite_traceback_stack(source=source))
File "/home/max/Git/cobib/.direnv/python-3.9.2/lib/python3.9/site-packages/jinja2/_compat.py", line 28, in reraise
raise value.with_traceback(tb)
File "/home/max/Git/pdoc/pdoc/templates/default/module.html.jinja2", line 650, in top-level template code
{%- if loop.nextitem %}.{% endif -%}
File "/home/max/Git/pdoc/pdoc/templates/frame.html.jinja2", line 18, in top-level template code
<body>{% block body %}{% endblock %}</body>
File "/home/max/Git/pdoc/pdoc/templates/default/module.html.jinja2", line 722, in block "body"
{% block module_contents %}
File "/home/max/Git/pdoc/pdoc/templates/default/module.html.jinja2", line 729, in block "module_contents"
{{ member(m) }}
File "/home/max/Git/cobib/.direnv/python-3.9.2/lib/python3.9/site-packages/jinja2/runtime.py", line 679, in _invoke
rv = self._func(*arguments)
File "/home/max/Git/pdoc/pdoc/templates/default/module.html.jinja2", line 557, in template
{{ function(doc) }}
File "/home/max/Git/cobib/.direnv/python-3.9.2/lib/python3.9/site-packages/jinja2/runtime.py", line 679, in _invoke
rv = self._func(*arguments)
File "/home/max/Git/pdoc/pdoc/templates/default/module.html.jinja2", line 523, in template
{{ decorators(fn) }}
File "/home/max/Git/cobib/.direnv/python-3.9.2/lib/python3.9/site-packages/jinja2/runtime.py", line 679, in _invoke
rv = self._func(*arguments)
File "/home/max/Git/pdoc/pdoc/templates/default/module.html.jinja2", line 514, in template
{% for d in doc.decorators if not d.startswith("@_") %}
File "/home/max/Git/cobib/.direnv/python-3.9.2/lib/python3.9/site-packages/jinja2/environment.py", line 471, in getattr
return getattr(obj, attribute)
File "/usr/lib/python3.9/functools.py", line 969, in __get__
val = self.func(instance)
File "/home/max/Git/pdoc/pdoc/doc.py", line 767, in decorators
for t in doc_ast.parse(obj).decorator_list:
File "/home/max/Git/pdoc/pdoc/doc_ast.py", line 60, in parse
return _parse_function(src)
File "/home/max/Git/pdoc/pdoc/doc_ast.py", line 196, in _parse_function
tree = ast.parse(_dedent(source))
File "/usr/lib/python3.9/ast.py", line 50, in parse
return compile(source, filename, mode, flags,
File "<unknown>", line 7
def test_dummy(self, arg, string):
IndentationError: unexpected indent
```
</details>
For additional details, this is what `_dedent` generates at the moment:
```python
@pytest.mark.parametrize(
"arg,string",
[
["hello", "world"],
],
)
def test_dummy(self, arg, string):
"""Dummy test method."""
assert arg != string
```
The dedented string is not causing any problems but the remaining indent of the `def test_dummy` line eventually causes the error to arise.
#### System Information
A fresh checkout of `main`:
```
% pdoc --version
pdoc: 6.4.2 (+15, commit 8d133d2)
Python: 3.9.2
Platform: Linux-5.11.11-arch1-1-x86_64-with-glibc2.33
```
| closed | 2021-04-10T15:20:00Z | 2021-04-11T06:49:26Z | https://github.com/mitmproxy/pdoc/issues/246 | [
"bug"
] | mrossinek | 0 |
Lightning-AI/pytorch-lightning | data-science | 20,253 | Cannot turn off sampler injection at inference time. | ### Bug description
I want to use a custom distributed batch sampler at inference time.
The sampler looks like this:
```python
class DistributedInferenceBatchSampler(DistributedSampler):
def __init__(self, dataset: Dataset,
batch_size: int = 1,
num_replicas: Optional[int] = None,
rank: Optional[int] = None,
shuffle: bool = False,
seed: int = 0,
drop_last: bool = False,
) -> None:
super().__init__(dataset, num_replicas=num_replicas, rank=rank,
shuffle=shuffle, seed=seed, drop_last=drop_last)
# do stuff
# sort data indices by datapoint length, batch up
# subsample batches for current rank
self.batches = # nested list [[b1_1, b1_2, ...], [b2_1, b2_2, ...], ...]
def __iter__(self) -> Iterator[T_co]:
return iter(self.batches)
def __len__(self) -> int:
return len(self.batches)
```
I use this dataloader inside my data module:
```python
class DataModule(LightningDataModule):
# ...
def predict_dataloader(self, rank=None, num_replicas=None):
# define Dataset 'data'
bsampler = DistributedInferenceBatchSampler(dataset=data,
batch_size=4,
num_replicas=num_replicas,
rank=rank)
data_loader = DataLoader(data,
batch_sampler=bsampler)
return data_loader
```
I'm running inference like so:
```python
trainer = Trainer( strategy='auto',
use_distributed_sampler=False, # using custom distributed batchsampler
accelerator=gpu,
deterministic=False,
enable_progress_bar=True,
enable_model_summary=True,
devices=devices
)
trainer.predict(model=self._model, datamodule=self._datamodule)
```
However, Lightning tries to replace my batch sampler despite the `use_distributed_sample=False` flag because it always does so in predict mode, and fails because the sampler doesn't have the same signature as a Pytorch `BatchSampler`.
I've tried wrapping my custom `DistributedInferenceBatchSampler` like so:
```python
class BatchSamplerWrapper(BatchSampler):
def __init__(self, sampler, batch_size=1, drop_last=False):
self.sampler = sampler
self.batch_size = batch_size # ignored
self.drop_last = drop_last # ignored
def __iter__(self):
for batch in self.sampler:
yield batch
def __len__(self):
return len(self.sampler)
class DataModule(LightningDataModule):
# ...
def predict_dataloader(self, rank=None, num_replicas=None):
# define Dataset 'data'
bsampler = DistributedInferenceBatchSampler(dataset=data,
batch_size=4,
num_replicas=num_replicas,
rank=rank)
wrapper = BatchSamplerWrapper(bsampler, batch_size=4, drop_last=False)
data_loader = DataLoader(data,
batch_sampler=wrapper)
return data_loader
```
However, Lightning replaces my `bsampler` inside the wrapper with a `torch.utils.data.sampler.SequentialSampler` which leads to `BatchSamplerWrapper.__iter__()` not having the intended behaviour. It returns an `int` rather than a list of `int`s, leading to:
```
Set the environment variable HYDRA_FULL_ERROR=1 for a complete stack trace.
Error executing job with overrides: []
Traceback (most recent call last):
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/codebase/IgFlow/multiflow/experiments/inference.py", line 18, in sample
run.sample()
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/codebase/IgFlow/multiflow/experiments/model_run.py", line 125, in sample
trainer.predict(model=self._model, datamodule=self._datamodule)
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/pytorch_lightning/trainer/trainer.py", line 864, in predict
return call._call_and_handle_interrupt(
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/pytorch_lightning/trainer/call.py", line 43, in _call_and_handle_interrupt
return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs)
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/pytorch_lightning/strategies/launchers/subprocess_script.py", line 102, in launch
return function(*args, **kwargs)
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/pytorch_lightning/trainer/trainer.py", line 903, in _predict_impl
results = self._run(model, ckpt_path=ckpt_path)
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/pytorch_lightning/trainer/trainer.py", line 989, in _run
results = self._run_stage()
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/pytorch_lightning/trainer/trainer.py", line 1030, in _run_stage
return self.predict_loop.run()
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/pytorch_lightning/loops/utilities.py", line 182, in _decorator
return loop_run(self, *args, **kwargs)
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/pytorch_lightning/loops/prediction_loop.py", line 119, in run
batch, batch_idx, dataloader_idx = next(data_fetcher)
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/pytorch_lightning/loops/fetchers.py", line 127, in __next__
batch = super().__next__()
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/pytorch_lightning/loops/fetchers.py", line 56, in __next__
batch = next(self.iterator)
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/pytorch_lightning/utilities/combined_loader.py", line 326, in __next__
out = next(self._iterator)
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/pytorch_lightning/utilities/combined_loader.py", line 132, in __next__
out = next(self.iterators[0])
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/torch/utils/data/dataloader.py", line 633, in __next__
data = self._next_data()
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/torch/utils/data/dataloader.py", line 1345, in _next_data
return self._process_data(data)
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/torch/utils/data/dataloader.py", line 1371, in _process_data
data.reraise()
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/torch/_utils.py", line 644, in reraise
raise exception
TypeError: Caught TypeError in DataLoader worker process 0.
Original Traceback (most recent call last):
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/torch/utils/data/_utils/worker.py", line 308, in _worker_loop
data = fetcher.fetch(index)
File "/data/nagagpu03/not-backed-up/nvme00/vavourakis/miniforge3/envs/mflow/lib/python3.10/site-packages/torch/utils/data/_utils/fetch.py", line 51, in fetch
data = [self.dataset[idx] for idx in possibly_batched_index]
TypeError: 'int' object is not iterable
```
I just want to turn off this sampler-replacing behaviour. I have a similar setup during training (rather than inference) and that works fine (no wrappers required, either).
### What version are you seeing the problem on?
v2.1
### How to reproduce the bug
_No response_
### Error messages and logs
```
# Error messages and logs here please
```
### Environment
_No response_
### More info
_No response_ | open | 2024-09-06T11:29:47Z | 2024-09-06T11:30:00Z | https://github.com/Lightning-AI/pytorch-lightning/issues/20253 | [
"bug",
"needs triage",
"ver: 2.1.x"
] | ovavourakis | 0 |
seleniumbase/SeleniumBase | web-scraping | 3,570 | Intermittent proxy-with-auth not being set before page load | ## Intermittent proxy-with-auth not being set before page load
Since the Manifest V3 extension conversion, there is a short delay before the auth is set, which may lead to an auth prompt appearing if navigating to a page immediately after the web browser is opened when using proxy-with-auth. This delay is a fraction of a second. It can be avoided by added a tiny delay after the web browser is opened (before navigating to a URL). | closed | 2025-02-26T17:23:29Z | 2025-02-26T22:43:31Z | https://github.com/seleniumbase/SeleniumBase/issues/3570 | [
"bug"
] | mdmintz | 1 |
pytest-dev/pytest-qt | pytest | 303 | Use PyPI token in GitHub actions | I just noticed this:
https://github.com/pytest-dev/pytest-qt/blob/e8080068d3ed215051231d5d5fd43fd1268e4eaa/.github/workflows/main.yml#L83-L86
It looks like https://github.com/pypa/pypi-support/issues/98 was taken care of. | closed | 2020-05-08T15:22:15Z | 2020-08-12T20:35:16Z | https://github.com/pytest-dev/pytest-qt/issues/303 | [] | The-Compiler | 1 |
home-assistant/core | asyncio | 140,838 | SmartThing Washer start/stop | ### The problem
Can't start or stop the washing machine with the switch component included in home assistant integration.
The command should use the samsungce.washerOperatingState capability instead of the switch. With the command start and pause or cancel.
### What version of Home Assistant Core has the issue?
core-2025.3.3
### What was the last working version of Home Assistant Core?
_No response_
### What type of installation are you running?
Home Assistant OS
### Integration causing the issue
SmartThings
### Link to integration documentation on our website
https://www.home-assistant.io/integrations/smartthings
### Diagnostics information
_No response_
### Example YAML snippet
```yaml
```
### Anything in the logs that might be useful for us?
```txt
```
### Additional information
_No response_ | closed | 2025-03-17T23:17:58Z | 2025-03-19T15:20:21Z | https://github.com/home-assistant/core/issues/140838 | [
"integration: smartthings"
] | kimzrn | 2 |
python-gino/gino | sqlalchemy | 12 | Support additional table arguments | Simply follow the SQLAlchemy ORM way, support `__table_args__`:
http://docs.sqlalchemy.org/en/latest/orm/extensions/declarative/table_config.html | closed | 2017-07-22T02:00:21Z | 2017-08-04T06:07:24Z | https://github.com/python-gino/gino/issues/12 | [
"help wanted",
"task"
] | fantix | 0 |
man-group/arctic | pandas | 189 | ChunkStore - Delete does not update metadata properly | does not update rows or chunk count when a chunk_range is specified
| closed | 2016-08-04T19:15:44Z | 2016-08-05T13:57:50Z | https://github.com/man-group/arctic/issues/189 | [] | bmoscon | 1 |
allenai/allennlp | pytorch | 5,308 | TextFieldTensors are not supported as input to a model Head | <!--
Please fill this template entirely and do not erase any of it.
We reserve the right to close without a response bug reports which are incomplete.
If you have a question rather than a bug, please ask on [Stack Overflow](https://stackoverflow.com/questions/tagged/allennlp) rather than posting an issue here.
-->
## Checklist
<!-- To check an item on the list replace [ ] with [x]. -->
- [x] I have verified that the issue exists against the `main` branch of AllenNLP.
- [x] I have read the relevant section in the [contribution guide](https://github.com/allenai/allennlp/blob/main/CONTRIBUTING.md#bug-fixes-and-new-features) on reporting bugs.
- [x] I have checked the [issues list](https://github.com/allenai/allennlp/issues) for similar or identical bug reports.
- [x] I have checked the [pull requests list](https://github.com/allenai/allennlp/pulls) for existing proposed fixes.
- [x] I have checked the [CHANGELOG](https://github.com/allenai/allennlp/blob/main/CHANGELOG.md) and the [commit log](https://github.com/allenai/allennlp/commits/main) to find out if the bug was already fixed in the main branch.
- [x] I have included in the "Description" section below a traceback from any exceptions related to this bug.
- [x] I have included in the "Related issues or possible duplicates" section beloew all related issues and possible duplicate issues (If there are none, check this box anyway).
- [x] I have included in the "Environment" section below the name of the operating system and Python version that I was using when I discovered this bug.
- [x] I have included in the "Environment" section below the output of `pip freeze`.
- [ ] I have included in the "Steps to reproduce" section below a minimally reproducible example.
## Description
<!-- Please provide a clear and concise description of what the bug is here. -->
I'm trying to pass a `TextFieldTensor` to a `Head` using the `multitask` Model.
For context, the `ClassifierHead` doesn't support `TextFieldTensors` for sequence decoding tasks. I created a simple head which inherits from the `AutoRegressiveSeqDecoder`, and puts the encoder text and mask into a dictionary before passing to the forward of the decoder.
```python
@Head.register("seq_decoder_tokens")
class SeqDecoderTokensHead(AutoRegressiveSeqDecoder):
@overrides
def forward(
self,
encoded_text: torch.Tensor,
encoded_text_mask: torch.Tensor,
target_tokens: TextFieldTensors = None,
) -> dict[str, torch.Tensor]:
encoder_out = {
"encoder_outputs": encoded_text,
"source_mask": encoded_text_mask,
}
return super().forward(encoder_out, target_tokens)
```
The error, as in the traceback below, is with the `make_inputs_for_task` function.
https://github.com/allenai/allennlp/blob/7d4a67263d7a210aca22d4f2b03e8568d3c34a48/allennlp/models/multitask.py#L114-L119
From the type annotation, I'm assuming that it might not be supported by the model so I might be barking up the wrong tree? I assumed that I could just pass the variable to the Head and it would work.
I'm not sure this is the best solution but I've made a fork with one possible solution, if it helps?
https://github.com/amitkparekh/allennlp/commit/f1e0630d83ed0d459f1efa167e4b2bad76f80bb2
<details>
<summary><b>Python traceback:</b></summary>
<p>
<!-- Paste the traceback from any exception (if there was one) in between the next two lines below -->
```
File "/Users/amit/Develop/pokerface/.venv/lib/python3.9/site-packages/allennlp/models/multitask.py", line 125, in <listcomp>
return [whole_batch_input[i] for i in task_indices[task]]
File "/Users/amit/Develop/pokerface/.venv/lib/python3.9/site-packages/allennlp/models/multitask.py", line 125, in make_inputs_for_task
return [whole_batch_input[i] for i in task_indices[task]]
File "/Users/amit/Develop/pokerface/.venv/lib/python3.9/site-packages/allennlp/models/multitask.py", line 138, in <dictcomp>
head_arguments = {key: make_inputs_for_task(head_name, value) for key, value in head_arguments.items()}
File "/Users/amit/Develop/pokerface/.venv/lib/python3.9/site-packages/allennlp/models/multitask.py", line 138, in forward
head_arguments = {key: make_inputs_for_task(head_name, value) for key, value in head_arguments.items()}
File "/Users/amit/Develop/pokerface/.venv/lib/python3.9/site-packages/torch/nn/modules/module.py", line 889, in _call_impl
result = self.forward(*input, **kwargs)
File "/Users/amit/Develop/pokerface/.venv/lib/python3.9/site-packages/allennlp/training/gradient_descent_trainer.py", line 351, in batch_outputs
output_dict = self._pytorch_model(**batch)
File "/Users/amit/Develop/pokerface/.venv/lib/python3.9/site-packages/allennlp/training/gradient_descent_trainer.py", line 458, in _train_epoch
batch_outputs = self.batch_outputs(batch, for_training=True)
File "/Users/amit/Develop/pokerface/.venv/lib/python3.9/site-packages/allennlp/training/gradient_descent_trainer.py", line 727, in _try_train
train_metrics = self._train_epoch(epoch)
File "/Users/amit/Develop/pokerface/.venv/lib/python3.9/site-packages/allennlp/training/gradient_descent_trainer.py", line 706, in train
metrics, epoch = self._try_train()
File "/Users/amit/Develop/pokerface/.venv/lib/python3.9/site-packages/allennlp/commands/train.py", line 543, in run
return self.trainer.train()
File "/Users/amit/Develop/pokerface/.venv/lib/python3.9/site-packages/allennlp/commands/train.py", line 470, in _train_worker
metrics = train_loop.run()
File "/Users/amit/Develop/pokerface/.venv/lib/python3.9/site-packages/allennlp/commands/train.py", line 240, in train_model
model = _train_worker(
File "/Users/amit/Develop/pokerface/.venv/lib/python3.9/site-packages/allennlp/commands/train.py", line 171, in train_model_from_file
return train_model(
File "/Users/amit/Develop/pokerface/.venv/lib/python3.9/site-packages/allennlp/commands/train.py", line 111, in train_model_from_args
train_model_from_file(
File "/Users/amit/Develop/pokerface/.venv/lib/python3.9/site-packages/allennlp/commands/__init__.py", line 121, in main
args.func(args)
File "/Users/amit/Develop/pokerface/scraps/debug_allennlp.py", line 35, in <module>
main()
File "/usr/local/Cellar/python@3.9/3.9.5/Frameworks/Python.framework/Versions/3.9/lib/python3.9/runpy.py", line 87, in _run_code
exec(code, run_globals)
File "/usr/local/Cellar/python@3.9/3.9.5/Frameworks/Python.framework/Versions/3.9/lib/python3.9/runpy.py", line 97, in _run_module_code
_run_code(code, mod_globals, init_globals,
File "/usr/local/Cellar/python@3.9/3.9.5/Frameworks/Python.framework/Versions/3.9/lib/python3.9/runpy.py", line 268, in run_path
return _run_module_code(code, init_globals, run_name,
File "/usr/local/Cellar/python@3.9/3.9.5/Frameworks/Python.framework/Versions/3.9/lib/python3.9/runpy.py", line 87, in _run_code
exec(code, run_globals)
File "/usr/local/Cellar/python@3.9/3.9.5/Frameworks/Python.framework/Versions/3.9/lib/python3.9/runpy.py", line 197, in _run_module_as_main (Current frame)
return _run_code(code, main_globals, None,
```
</p>
</details>
## Related issues or possible duplicates
- None
## Environment
<!-- Provide the name of operating system below (e.g. OS X, Linux) -->
OS: OS X
<!-- Provide the Python version you were using (e.g. 3.7.1) -->
Python version: 3.9.5
<details>
<summary><b>Output of <code>pip freeze</code>:</b></summary>
<p>
<!-- Paste the output of `pip freeze` in between the next two lines below -->
```
allennlp @ file:///Users/amit/Library/Caches/pypoetry/artifacts/b1/ea/84/1d6e28cb885e4d55ac8939f78e408f7deb86ea845630a64f10d7700a62/allennlp-2.5.0-py3-none-any.whl
allennlp-models @ file:///Users/amit/Library/Caches/pypoetry/artifacts/77/31/1f/2ef3c0ed6db0dc41de744411f56b478ab042f091dda6cd079b72b12c53/allennlp_models-2.5.0-py3-none-any.whl
appdirs @ file:///Users/amit/Library/Caches/pypoetry/artifacts/47/cf/4f/4ef02fb715aa36daeebad18cc5570126159c659c41c7b5ec46a7387d9b/appdirs-1.4.4-py2.py3-none-any.whl
appnope @ file:///Users/amit/Library/Caches/pypoetry/artifacts/fd/ae/b8/2382438588ec752ed602c2dcab2a0678b30ff15f2d9a30267ff97ecd64/appnope-0.1.2-py2.py3-none-any.whl
argon2-cffi @ file:///Users/amit/Library/Caches/pypoetry/artifacts/3c/87/0e/fc0a0440e3e84e11c88d9d2d049f9aff2fdc4e7493dd19af67381b1d05/argon2_cffi-20.1.0-cp37-abi3-macosx_10_6_intel.whl
astor @ file:///Users/amit/Library/Caches/pypoetry/artifacts/1f/03/36/982d1222edac5e8cb8ac6e0464249747fa800d4fb04728a99153ecfe4d/astor-0.8.1-py2.py3-none-any.whl
async-generator @ file:///Users/amit/Library/Caches/pypoetry/artifacts/fe/72/db/709736555b02c2d1ae90038b7b05138b15e24edde3aa7556fc2507a90f/async_generator-1.10-py3-none-any.whl
attrs @ file:///Users/amit/Library/Caches/pypoetry/artifacts/6f/a9/ee/569c37f69a8c365ee41d2340aeac0214ee8c0086b8d8db43a21545204b/attrs-21.2.0-py2.py3-none-any.whl
backcall @ file:///Users/amit/Library/Caches/pypoetry/artifacts/43/8e/e8/4e598704edf6fb4a53d552ea511c04e9958dcf850897760e5387878b99/backcall-0.2.0-py2.py3-none-any.whl
backports.csv @ file:///Users/amit/Library/Caches/pypoetry/artifacts/30/84/1a/81a42cff31ce7f0b7a86ab54e2cbcb610d96fa8b735d63bdb7251e91cb/backports.csv-1.0.7-py2.py3-none-any.whl
bandit @ file:///Users/amit/Library/Caches/pypoetry/artifacts/6a/05/4f/98680ab175e4b595c2d1b775974c208b6b20c05448a52944374c2db4b0/bandit-1.7.0-py3-none-any.whl
beartype @ file:///Users/amit/Library/Caches/pypoetry/artifacts/c3/fe/bd/c04bccf2fa951904264c6dc16cbdff35bc2aec170d65b9afb879841dfa/beartype-0.7.1-py3-none-any.whl
beautifulsoup4 @ file:///Users/amit/Library/Caches/pypoetry/artifacts/eb/47/47/287c1b8a386f9437d562f9221ae959756bc5bbfcd541c38c17968dfe8a/beautifulsoup4-4.9.3-py3-none-any.whl
black @ file:///Users/amit/Library/Caches/pypoetry/artifacts/6a/38/11/b77f947a81ed86d08787f98d076035bedb21abd3d3c57129221264c3ea/black-21.6b0-py3-none-any.whl
bleach @ file:///Users/amit/Library/Caches/pypoetry/artifacts/15/bd/f0/eaed67c8e6d37dda902603474339528f29bde3f7ecc2bb4b8874e0da87/bleach-3.3.0-py2.py3-none-any.whl
blis @ file:///Users/amit/Library/Caches/pypoetry/artifacts/96/b3/59/60398be6c97784bb3ba6ef5d76e86ce6bc530e27ca83de6d106b5dadd3/blis-0.7.4-cp39-cp39-macosx_10_9_x86_64.whl
boto3 @ file:///Users/amit/Library/Caches/pypoetry/artifacts/5e/59/0d/9ce4b54813b37237c51d31bb0d3c0853123334f6309a2088f81ec49c03/boto3-1.17.109-py2.py3-none-any.whl
botocore @ file:///Users/amit/Library/Caches/pypoetry/artifacts/34/8e/85/14e136ebe48249de4155082aa1b2aa772ec3f7d415dd1c60067ed3a2ad/botocore-1.20.109-py2.py3-none-any.whl
cachetools @ file:///Users/amit/Library/Caches/pypoetry/artifacts/04/ca/d7/8af05dc8ccae1212d4151afd99960369c2415b26bc13ed3bbb288c4f5a/cachetools-4.2.2-py3-none-any.whl
catalogue @ file:///Users/amit/Library/Caches/pypoetry/artifacts/af/cf/ab/153c9326701d6adc746e129e3e05991a55bd43b2f19e3c52d7ccc9a7b4/catalogue-1.0.0-py2.py3-none-any.whl
certifi @ file:///Users/amit/Library/Caches/pypoetry/artifacts/cd/2c/dc/e5bfda594e18f3f1e9af9f11e13581014d821425f325f3220b3ed2c337/certifi-2021.5.30-py2.py3-none-any.whl
cffi @ file:///Users/amit/Library/Caches/pypoetry/artifacts/6e/aa/04/2c3c9401654c8f5580dc8965817a99e8ad464a0987e17149061aadfcbf/cffi-1.14.6-cp39-cp39-macosx_10_9_x86_64.whl
cfgv @ file:///Users/amit/Library/Caches/pypoetry/artifacts/6b/52/b7/27617ac43f25c9962779813593809288745c414fd878b968cc3d91ca6c/cfgv-3.3.0-py2.py3-none-any.whl
chardet @ file:///Users/amit/Library/Caches/pypoetry/artifacts/11/63/f9/797eda27963177a6b75a340f62aa194d462ea69e6b0dbb77a651fa2b62/chardet-4.0.0-py2.py3-none-any.whl
checklist @ file:///Users/amit/Library/Caches/pypoetry/artifacts/04/84/f3/1324eec13577715f52121b3073ab37792c08483aae7faa7d22b7dd5e1d/checklist-0.0.11.tar.gz
cheroot @ file:///Users/amit/Library/Caches/pypoetry/artifacts/3c/cf/87/c9bb0e3b0d4c43affeb8f9714d5791b61c97750bc3a2ee35d276d425b5/cheroot-8.5.2-py2.py3-none-any.whl
CherryPy @ file:///Users/amit/Library/Caches/pypoetry/artifacts/dd/d5/0c/5289a45f52e9aa001f7b8c2b9c792377e79ad572bc759f1a692d160818/CherryPy-18.6.1-py2.py3-none-any.whl
click @ file:///Users/amit/Library/Caches/pypoetry/artifacts/ae/32/83/e159324c1bd58177322f4e45f598d500fe22544bff20f53f55cf749da8/click-8.0.1-py3-none-any.whl
colorama @ file:///Users/amit/Library/Caches/pypoetry/artifacts/9e/b3/11/7d87ac44fdb2d557301f1f4086a37c080d1482a98751abe7cdbabbad26/colorama-0.4.4-py2.py3-none-any.whl
commonmark @ file:///Users/amit/Library/Caches/pypoetry/artifacts/11/56/f6/d054064b623fab5c7e4420f60d931f49fea2dacdebe1dc991201010c84/commonmark-0.9.1-py2.py3-none-any.whl
configparser @ file:///Users/amit/Library/Caches/pypoetry/artifacts/4e/17/fd/30c9e84dc4b951f3587a4d5eb2894e20105120991793ff3e9f3a60d787/configparser-5.0.2-py3-none-any.whl
conllu @ file:///Users/amit/Library/Caches/pypoetry/artifacts/a9/0b/c7/419ecfd4c8064217550121b01eb4c617ebb82d54c06c775a2d683abbe4/conllu-4.4-py2.py3-none-any.whl
cryptography @ file:///Users/amit/Library/Caches/pypoetry/artifacts/c7/6d/36/40b404429242377c0424cb9957a04887a4be7399a62c1505197845c09f/cryptography-3.4.7-cp36-abi3-macosx_10_10_x86_64.whl
cymem @ file:///Users/amit/Library/Caches/pypoetry/artifacts/b8/f4/20/752f31c5ea9976672f58722949ff87b94832958712b6e3ba60e831421e/cymem-2.0.5-cp39-cp39-macosx_10_9_x86_64.whl
darglint @ file:///Users/amit/Library/Caches/pypoetry/artifacts/bb/64/25/6fa28703f05a524eef407d4425af6290f242eb131574704386e677624b/darglint-1.8.0-py3-none-any.whl
decorator @ file:///Users/amit/Library/Caches/pypoetry/artifacts/99/d8/37/35167f3a4175b089109325a5ee11846ac4de416442972573b87351396d/decorator-5.0.9-py3-none-any.whl
defusedxml @ file:///Users/amit/Library/Caches/pypoetry/artifacts/2b/69/07/7b13f7eaf3a4d7af737dcebe24d3d17b1c2a2f457fbddf746f5642bc43/defusedxml-0.7.1-py2.py3-none-any.whl
dill @ file:///Users/amit/Library/Caches/pypoetry/artifacts/26/56/9e/73963d2285e6c700801f185e8c1d28f1f971c09aaa411cec9b799a5fca/dill-0.3.4-py2.py3-none-any.whl
distlib @ file:///Users/amit/Library/Caches/pypoetry/artifacts/6f/35/97/6a255392aaa7200818a8cd0f4b014ef2e0a086bab49dd568780f367ef5/distlib-0.3.2-py2.py3-none-any.whl
docker-pycreds @ file:///Users/amit/Library/Caches/pypoetry/artifacts/9b/0b/be/891931da9caf5e55102337a635d3a7eeeb92c93b4bd39c24d0810f1f25/docker_pycreds-0.4.0-py2.py3-none-any.whl
docutils @ file:///Users/amit/Library/Caches/pypoetry/artifacts/43/28/92/79000933ad30371dc938d9b368a9000e20ac0bb467a716c19ef1fbd3c7/docutils-0.17.1-py2.py3-none-any.whl
entrypoints @ file:///Users/amit/Library/Caches/pypoetry/artifacts/63/c1/af/bbfdd91bcb544e62ac8f1567ef23c243cb188d1a9cb933532999c9bbb0/entrypoints-0.3-py2.py3-none-any.whl
eradicate @ file:///Users/amit/Library/Caches/pypoetry/artifacts/12/ce/ac/197035fe6d51568abb7ea160f5ad416d2164a2010005e8356b8229e550/eradicate-2.0.0.tar.gz
feedparser @ file:///Users/amit/Library/Caches/pypoetry/artifacts/d1/81/87/0f3c1c0b02176b2bf05af85261d8ce7522e4d241e5d9f7b3f0ec4f2a10/feedparser-6.0.8-py3-none-any.whl
filelock @ file:///Users/amit/Library/Caches/pypoetry/artifacts/7e/c4/97/2cfbeab3cc292d0b4290cb7cab0b969b3002dc24f6dd5944cbe340e684/filelock-3.0.12-py3-none-any.whl
flake8 @ file:///Users/amit/Library/Caches/pypoetry/artifacts/3f/57/d5/11722093c13092cc3bfc3dd7c88aef6f8e4d5ac97cfe5fd054d5aba412/flake8-3.9.2-py2.py3-none-any.whl
flake8-bandit @ file:///Users/amit/Library/Caches/pypoetry/artifacts/7e/e4/46/e15782d941f9cde39b64ca5b636180f47573f2b2c9315be56b55152f17/flake8_bandit-2.1.2.tar.gz
flake8-broken-line @ file:///Users/amit/Library/Caches/pypoetry/artifacts/51/ea/87/37348b281b73d7df44fc46b09c0430e2984e991df11998e2e9bb459fce/flake8_broken_line-0.3.0-py3-none-any.whl
flake8-bugbear @ file:///Users/amit/Library/Caches/pypoetry/artifacts/15/95/75/8c7a4504d7eda1a394c3d14349b88577ff2d98d59941944c95bd8672ba/flake8_bugbear-21.4.3-py36.py37.py38-none-any.whl
flake8-commas @ file:///Users/amit/Library/Caches/pypoetry/artifacts/c3/47/1d/7f7fac0c58b2bd2bf7361bcba0bceba1c81c365cab5e1de352fa7fac68/flake8_commas-2.0.0-py2.py3-none-any.whl
flake8-comprehensions @ file:///Users/amit/Library/Caches/pypoetry/artifacts/4e/10/91/04adae987aa18ee9463db75694dccee7bf7d5518118462f18252bd3e0a/flake8_comprehensions-3.5.0-py3-none-any.whl
flake8-debugger @ file:///Users/amit/Library/Caches/pypoetry/artifacts/66/04/47/7bef98a8d237eb17cbfbcb803343be1c79e2c0674ceba163717b6c8e1b/flake8_debugger-4.0.0-py3-none-any.whl
flake8-docstrings @ file:///Users/amit/Library/Caches/pypoetry/artifacts/e0/85/e9/6b482a11d48cf26e1170d9f5bf0b044a5a6c9b816ffe70945e90fc3e56/flake8_docstrings-1.6.0-py2.py3-none-any.whl
flake8-eradicate @ file:///Users/amit/Library/Caches/pypoetry/artifacts/fc/3d/a0/58427b14b0a6d33587f3d6896e695615272c37c3ff2c89d6155b5155f6/flake8_eradicate-1.1.0-py3-none-any.whl
flake8-isort @ file:///Users/amit/Library/Caches/pypoetry/artifacts/11/37/1c/68fb64c6704b9c2468f711b83090590abc2c8295eeafbac9a167f32e0a/flake8_isort-4.0.0-py2.py3-none-any.whl
flake8-polyfill @ file:///Users/amit/Library/Caches/pypoetry/artifacts/28/17/cc/952c11cd5ffb2608137557f928dc4f9365b4dbe1e2a6015eeea78583ac/flake8_polyfill-1.0.2-py2.py3-none-any.whl
flake8-quotes @ file:///Users/amit/Library/Caches/pypoetry/artifacts/75/61/73/b33ec4139bc79d01b0748fb1ae5889fbd6bd544c6f35521fb4dd981b1a/flake8-quotes-3.2.0.tar.gz
flake8-rst-docstrings @ file:///Users/amit/Library/Caches/pypoetry/artifacts/69/c4/19/62afcee08756b2cc746fb4d585d02dd819556e0b0d30a6cc7acb6a101a/flake8_rst_docstrings-0.2.3-py3-none-any.whl
flake8-string-format @ file:///Users/amit/Library/Caches/pypoetry/artifacts/24/89/bb/7ce8e216f8c7289aa8a2ad4c44f30f87af6c7cdaf5d510110d566d66ec/flake8_string_format-0.3.0-py2.py3-none-any.whl
ftfy @ file:///Users/amit/Library/Caches/pypoetry/artifacts/da/3c/cd/9230817d48f70d575b300d027e9d3845ffe7dec691cfeca5959d022536/ftfy-6.0.3.tar.gz
future @ file:///Users/amit/Library/Caches/pypoetry/artifacts/f8/58/55/86be1f567b212fdd98854d12815964a49db8fb1bcff725018e5f95c61d/future-0.18.2.tar.gz
gitdb @ file:///Users/amit/Library/Caches/pypoetry/artifacts/82/af/0d/fc5992ac7ef8a227e6b9705aa6de550211814dd0318b857530d3306d02/gitdb-4.0.7-py3-none-any.whl
GitPython @ file:///Users/amit/Library/Caches/pypoetry/artifacts/89/d2/fc/aacfc97469f68f6b2da4db532a2e04ba3ba94e09728e3ea6d4444e0dd2/GitPython-3.1.18-py3-none-any.whl
google-api-core @ file:///Users/amit/Library/Caches/pypoetry/artifacts/5e/fa/f1/f104f6d061efd1ddb611c0cb4fcc9252c5e15459515b8991136cdf8886/google_api_core-1.31.0-py2.py3-none-any.whl
google-auth @ file:///Users/amit/Library/Caches/pypoetry/artifacts/c7/96/64/403c53d4b77d3b92517777391e7598970fd56386f0bc6098f99801e59f/google_auth-1.32.1-py2.py3-none-any.whl
google-cloud-core @ file:///Users/amit/Library/Caches/pypoetry/artifacts/ed/24/7b/fd351e28e6811f5ef5718d536c84e43df631d72487e2870d11a840b8b2/google_cloud_core-1.7.1-py2.py3-none-any.whl
google-cloud-storage @ file:///Users/amit/Library/Caches/pypoetry/artifacts/52/25/b2/b8d3db4a638b27dd17a77802f8eae114ac39325db23cd0f7c58781e35a/google_cloud_storage-1.38.0-py2.py3-none-any.whl
google-crc32c @ file:///Users/amit/Library/Caches/pypoetry/artifacts/ac/6b/f2/c976910125e4a02d8b127cb4138fc9ac6167c7c971a76bc1bde216ee7b/google_crc32c-1.1.2-cp39-cp39-macosx_10_14_x86_64.whl
google-resumable-media @ file:///Users/amit/Library/Caches/pypoetry/artifacts/7d/de/88/a84ae5cef0a9895612e0c4db686aab010ff824edd1aeceb3906c3cd7e0/google_resumable_media-1.3.1-py2.py3-none-any.whl
googleapis-common-protos @ file:///Users/amit/Library/Caches/pypoetry/artifacts/8d/40/1f/21e71977c3d547b27842caaafc2e420c9a6dd40a745cce4b61673e3be3/googleapis_common_protos-1.53.0-py2.py3-none-any.whl
h5py @ file:///Users/amit/Library/Caches/pypoetry/artifacts/21/20/f5/670da7f96cdc48c98aad052bf28c0efc57bf865ff1f9b2c50ae8d6b2a3/h5py-3.3.0-cp39-cp39-macosx_10_9_x86_64.whl
huggingface-hub @ file:///Users/amit/Library/Caches/pypoetry/artifacts/40/86/15/ea367547cd99a3a52f226c2b2b7fd5d28b0b7c0e1323eee2909b46cc31/huggingface_hub-0.0.13-py3-none-any.whl
hypothesis @ file:///Users/amit/Library/Caches/pypoetry/artifacts/62/44/a7/47fd46593add79f575e8418b9aac6d209749e120cdd07fbf309533c548/hypothesis-6.14.1-py3-none-any.whl
identify @ file:///Users/amit/Library/Caches/pypoetry/artifacts/88/59/31/ca587f87a94f22f5d7f086dd53b95373139c023f75974e9293131680b7/identify-2.2.11-py2.py3-none-any.whl
idna @ file:///Users/amit/Library/Caches/pypoetry/artifacts/ef/7f/a9/19cc0b8760bdf6f696290c06532496f8bb29fbdaad044f852fed00ec82/idna-2.10-py2.py3-none-any.whl
iniconfig @ file:///Users/amit/Library/Caches/pypoetry/artifacts/fa/b0/c6/10cfac68c9e6de9d2a1678366ca89fd9292b362c1760dbe758e41691cb/iniconfig-1.1.1-py2.py3-none-any.whl
ipykernel @ file:///Users/amit/Library/Caches/pypoetry/artifacts/30/98/d5/b01a5306b6404f2c2862d298a4f8649f5c2579c0f5973f59838fe3fc2b/ipykernel-5.5.5-py3-none-any.whl
ipython @ file:///Users/amit/Library/Caches/pypoetry/artifacts/1e/56/a1/122ee0fd1f99ee5a3e81bfe0366288a488583ed0be289fb244e1663376/ipython-7.25.0-py3-none-any.whl
ipython-genutils @ file:///Users/amit/Library/Caches/pypoetry/artifacts/b4/31/01/6f96480580d1674cab0b5e26dc9fca7bbdf7a2fd5811a7807a92436268/ipython_genutils-0.2.0-py2.py3-none-any.whl
ipywidgets @ file:///Users/amit/Library/Caches/pypoetry/artifacts/75/67/72/e14b677150e119dacbd4bf8559095cb47df20ea3c6ecc37bac97964b4a/ipywidgets-7.6.3-py2.py3-none-any.whl
iso-639 @ file:///Users/amit/Library/Caches/pypoetry/artifacts/f0/a6/d6/17ede193e09cf4ac45d787cbef2e1c78ff7dff5a58775728212204bbd0/iso-639-0.4.5.tar.gz
isort @ file:///Users/amit/Library/Caches/pypoetry/artifacts/e8/65/0b/2aee1c8017c733cd6111eb34137491a1456d443396b6d283fba8e0d4a3/isort-5.9.2-py3-none-any.whl
jaraco.classes @ file:///Users/amit/Library/Caches/pypoetry/artifacts/d0/3c/d2/f157bfb781b294c3d68a29e898ab39327bc2397eea1b42cf8afdfda14b/jaraco.classes-3.2.1-py3-none-any.whl
jaraco.collections @ file:///Users/amit/Library/Caches/pypoetry/artifacts/7d/46/b7/579da18cb7f7d7a7d7cd19be8b2d5f5cf3449b1a019c7ec307333b2346/jaraco.collections-3.3.0-py3-none-any.whl
jaraco.functools @ file:///Users/amit/Library/Caches/pypoetry/artifacts/4e/9d/9f/96376949fa50b18fc96d90f93d24360b467689922c512daa6beee4d08b/jaraco.functools-3.3.0-py3-none-any.whl
jaraco.text @ file:///Users/amit/Library/Caches/pypoetry/artifacts/d8/d0/b6/f9d37687fabea73a57ffc167a243a09255bc57413476984bab40bd0984/jaraco.text-3.5.0-py3-none-any.whl
jedi @ file:///Users/amit/Library/Caches/pypoetry/artifacts/2a/5b/d6/62e1f4e7b392c3e7f8258bbe3159dff695814a46e65547cd547ca0fedb/jedi-0.18.0-py2.py3-none-any.whl
Jinja2 @ file:///Users/amit/Library/Caches/pypoetry/artifacts/21/2e/46/0a76ea6f6a15e594c9828a85a781f1cee8ed5a1b77e361305645f9e1f4/Jinja2-3.0.1-py3-none-any.whl
jmespath @ file:///Users/amit/Library/Caches/pypoetry/artifacts/2c/f0/52/b0ba93d941bd49c8719dee7ca27d2096bf96e17948667388c3ee2ac8f8/jmespath-0.10.0-py2.py3-none-any.whl
joblib @ file:///Users/amit/Library/Caches/pypoetry/artifacts/28/41/26/7ed6532cff9d56b8d2878f93bc289c075f338c12aa7d630862aae39d45/joblib-1.0.1-py3-none-any.whl
jsonnet @ file:///Users/amit/Library/Caches/pypoetry/artifacts/60/4e/2d/acde747a02049d38e6dbda9dc3fbf64f03bf2e14c8e9ad04f07edcc66b/jsonnet-0.17.0.tar.gz
jsonschema @ file:///Users/amit/Library/Caches/pypoetry/artifacts/db/1d/66/ad84fa70cc987bd4aad68be808562321cdab3cb03f4d5d7714a0e0571c/jsonschema-3.2.0-py2.py3-none-any.whl
jupyter @ file:///Users/amit/Library/Caches/pypoetry/artifacts/bb/e8/12/09df1332820a1126a780ab09cec78d2f50457f79bcd0cb2fbb07b19ef4/jupyter-1.0.0-py2.py3-none-any.whl
jupyter-client @ file:///Users/amit/Library/Caches/pypoetry/artifacts/8c/14/4d/593d81015262d7306a1f89c900a11dfb7d2f18ee37f8ef17ba7ab983bf/jupyter_client-6.2.0-py3-none-any.whl
jupyter-console @ file:///Users/amit/Library/Caches/pypoetry/artifacts/4b/2f/b8/119f975fb811a5e911beacc7a1bb8f8e1154254fb33204e753131e7aca/jupyter_console-6.4.0-py3-none-any.whl
jupyter-core @ file:///Users/amit/Library/Caches/pypoetry/artifacts/91/8f/22/1377f102bb4478eb073c741714348b7cbb6518d221da9c232cfc5242b3/jupyter_core-4.7.1-py3-none-any.whl
jupyterlab-pygments @ file:///Users/amit/Library/Caches/pypoetry/artifacts/de/ef/fc/5883436de4b7865f082f7cba0e0e0ff5fbf229fe55d6e7d5431a6080f4/jupyterlab_pygments-0.1.2-py2.py3-none-any.whl
jupyterlab-widgets @ file:///Users/amit/Library/Caches/pypoetry/artifacts/26/96/3a/11068104cd5f33e8f5437f04e75fb2220aa927c47ae5f01a7477acc169/jupyterlab_widgets-1.0.0-py3-none-any.whl
lmdb @ file:///Users/amit/Library/Caches/pypoetry/artifacts/fd/60/fa/5d8bc278b7132dd8c6a50b32b0b4076f31080cee889a3e59167a93af64/lmdb-1.2.1-cp39-cp39-macosx_10_14_x86_64.whl
lxml @ file:///Users/amit/Library/Caches/pypoetry/artifacts/9e/44/6a/570737853888f173f84e160c5772c792bfd10ea0385a76c138c94b23fc/lxml-4.6.3-cp39-cp39-macosx_10_9_x86_64.whl
MarkupSafe @ file:///Users/amit/Library/Caches/pypoetry/artifacts/20/e4/29/5b1a93d4ee8437f01551437cffbb57ba6744c59796443ca99051473f75/MarkupSafe-2.0.1-cp39-cp39-macosx_10_9_x86_64.whl
matplotlib-inline @ file:///Users/amit/Library/Caches/pypoetry/artifacts/d8/4c/e1/8b0adf0d076721db928e74aaaf5cf1d73cec52ea2a27ebabe3e9be8957/matplotlib_inline-0.1.2-py3-none-any.whl
mccabe @ file:///Users/amit/Library/Caches/pypoetry/artifacts/96/5e/5f/21ae5296697ca7f94de4da6e21d4936d74029c352a35202e4c339a4253/mccabe-0.6.1-py2.py3-none-any.whl
mistune @ file:///Users/amit/Library/Caches/pypoetry/artifacts/33/31/4c/2d69dc65d06d1c8f8b00b8e995e24bae97fce2e1f8ec5d8d2d98e852da/mistune-0.8.4-py2.py3-none-any.whl
more-itertools @ file:///Users/amit/Library/Caches/pypoetry/artifacts/47/d4/7d/526affa62eb0c76eec19004cbbf18cc4e55f51c665e92b88bb2ed25752/more_itertools-8.8.0-py3-none-any.whl
munch @ file:///Users/amit/Library/Caches/pypoetry/artifacts/c3/f9/98/c46b861b1fe10f4d4fecd0ed8752a968be33d2c7e698b70589015aa0b2/munch-2.5.0-py2.py3-none-any.whl
murmurhash @ file:///Users/amit/Library/Caches/pypoetry/artifacts/03/a3/c8/88bb7a1608c5b641c88c47b5b0c9b5377ddbc103ca4574acf4af3f00b4/murmurhash-1.0.5-cp39-cp39-macosx_10_9_x86_64.whl
mypy @ file:///Users/amit/Library/Caches/pypoetry/artifacts/ab/61/06/b62a10ea75c111856b223b2921035d273fd3d746110fde9ff4bb20fc0c/mypy-0.901-cp39-cp39-macosx_10_9_x86_64.whl
mypy-extensions @ file:///Users/amit/Library/Caches/pypoetry/artifacts/92/45/bf/1807ce854ff668d92602207a37bfa9316def2a3f257bd03c4c5be4bc9b/mypy_extensions-0.4.3-py2.py3-none-any.whl
nbclient @ file:///Users/amit/Library/Caches/pypoetry/artifacts/cd/06/72/d0468da165d742240bf6e43191ef3a942576956bf30426df2cba0ac668/nbclient-0.5.3-py3-none-any.whl
nbconvert @ file:///Users/amit/Library/Caches/pypoetry/artifacts/cd/57/5f/0aebcf8edff99a9ea65795c24b549346cdb70b238cf2c200de0bf86f0f/nbconvert-6.1.0-py3-none-any.whl
nbformat @ file:///Users/amit/Library/Caches/pypoetry/artifacts/36/a7/e7/e1a0c1c54f6151e23afd51bc71e3f6e0b24a96dd1e693b92dd9a4e4ab3/nbformat-5.1.3-py3-none-any.whl
nest-asyncio @ file:///Users/amit/Library/Caches/pypoetry/artifacts/51/1a/ee/4f904dc67e6f59f0c8b75ef59d0111e4fb57dd4a884ad19d289ab31a77/nest_asyncio-1.5.1-py3-none-any.whl
nltk @ file:///Users/amit/Library/Caches/pypoetry/artifacts/18/6d/ee/ffa73af7527056102cff0b73ffa10fc5a7ffa898f9214d546e6ec70b57/nltk-3.6.2-py3-none-any.whl
nodeenv @ file:///Users/amit/Library/Caches/pypoetry/artifacts/54/f0/a1/53d1e469f8e160bad013c23411375fd63d4ea70cda0fc649fcb244ca7e/nodeenv-1.6.0-py2.py3-none-any.whl
notebook @ file:///Users/amit/Library/Caches/pypoetry/artifacts/8a/53/13/86803f6ca277cd2b45ce841ae8843ab6eb51d45b1d3a8fe96eb2382a07/notebook-6.4.0-py3-none-any.whl
numpy @ file:///Users/amit/Library/Caches/pypoetry/artifacts/d6/2b/b1/16975744b34a9b3ec7bd901b286178933b525ca493a93edd7441d3b807/numpy-1.21.0-cp39-cp39-macosx_10_9_x86_64.whl
orjson @ file:///Users/amit/Library/Caches/pypoetry/artifacts/c8/c9/04/a1cc49302292dbbe34795ed7d5a881c50cf02f4f813f336f7493949870/orjson-3.6.0-cp39-cp39-macosx_10_9_x86_64.macosx_11_0_arm64.macosx_10_9_universal2.whl
overrides @ file:///Users/amit/Library/Caches/pypoetry/artifacts/24/45/16/62e842b5cdff34f5106ee676232cbcc7d7a1333e4900d111bca737b13a/overrides-3.1.0.tar.gz
packaging @ file:///Users/amit/Library/Caches/pypoetry/artifacts/f9/4f/09/c91a145b26102e014fd6e33bd8c7b87306c8e1d4a771158f34dd13210e/packaging-21.0-py3-none-any.whl
pandocfilters @ file:///Users/amit/Library/Caches/pypoetry/artifacts/2b/8b/2b/6cd1e4385f3f7f98a25f05764a4ea3f2f20d1db00612ef79e25bb90fe9/pandocfilters-1.4.3.tar.gz
parso @ file:///Users/amit/Library/Caches/pypoetry/artifacts/36/cd/ab/a8c3a5df337bc6f34a10f3f385417b62cdfebe2873ac2fec38206af0db/parso-0.8.2-py2.py3-none-any.whl
pastel @ file:///Users/amit/Library/Caches/pypoetry/artifacts/da/84/f3/3e4d8b15eabeba62960ed9d3ccc1e30b7ae5f1b93e6c28d291c67eaf93/pastel-0.2.1-py2.py3-none-any.whl
pathspec @ file:///Users/amit/Library/Caches/pypoetry/artifacts/40/8b/b2/80a6945971d8475bcc04d09afec4845855dd74f68da6a4c18bbf8f7784/pathspec-0.8.1-py2.py3-none-any.whl
pathtools @ file:///Users/amit/Library/Caches/pypoetry/artifacts/ce/ff/c7/31da76336d55d51d979a50868616c867c7b2ea6f2d2084b8c744726ae7/pathtools-0.1.2.tar.gz
patternfork-nosql @ file:///Users/amit/Library/Caches/pypoetry/artifacts/fa/53/d7/a04c2b1cd20312460da3f82ca634ac259fc581089956ed73763c0757cc/patternfork_nosql-3.6.tar.gz
pbr @ file:///Users/amit/Library/Caches/pypoetry/artifacts/c7/6a/db/482c805b950fd5a0ece81e7a9a063cb1aa99169ca73fba511759c9db30/pbr-5.6.0-py2.py3-none-any.whl
pdfminer.six @ file:///Users/amit/Library/Caches/pypoetry/artifacts/21/c6/47/9aded02cbdd666be599af42cbfc0c6cf4ad847f177acf882ee11ff2f19/pdfminer.six-20201018-py3-none-any.whl
pep8-naming @ file:///Users/amit/Library/Caches/pypoetry/artifacts/a4/93/c7/a3b9b8b4aef682b4caa67015d897aff3d064860a460124ad8a23b6f45f/pep8_naming-0.11.1-py2.py3-none-any.whl
pexpect @ file:///Users/amit/Library/Caches/pypoetry/artifacts/5c/c2/43/b54fe59cab7e831df35401c8e6840162bf4a2ae5862604e7bc22db3000/pexpect-4.8.0-py2.py3-none-any.whl
pickleshare @ file:///Users/amit/Library/Caches/pypoetry/artifacts/b5/48/a1/d2b823337003d531d87cf0d503ef28bb579703a74d14ad24a88863d616/pickleshare-0.7.5-py2.py3-none-any.whl
Pillow @ file:///Users/amit/Library/Caches/pypoetry/artifacts/d7/25/9f/10bf1b0f46db0f11f87fcf319db8e4a8d0acd2326569efb025a0b8b7aa/Pillow-8.3.1-cp39-cp39-macosx_10_10_x86_64.whl
plac @ file:///Users/amit/Library/Caches/pypoetry/artifacts/a5/cb/88/6e55cfacecccbab7d6a03f0d05004c5a39c03a39265423788255714111/plac-1.1.3-py2.py3-none-any.whl
pluggy @ file:///Users/amit/Library/Caches/pypoetry/artifacts/29/58/fc/ed8b7451d3ef91a6465024f5656141da996e7aafd4d41a1659629a75e7/pluggy-0.13.1-py2.py3-none-any.whl
poethepoet @ file:///Users/amit/Library/Caches/pypoetry/artifacts/36/42/75/cc7c77f4b9ac69fd8470c7524490be4f0c722f6f7b98d14da11fc2774e/poethepoet-0.10.0-py3-none-any.whl
pokerface==0.1.0
portend @ file:///Users/amit/Library/Caches/pypoetry/artifacts/9e/0a/b1/aab32a1b4dbdac42ada198c9c4378651e9a6fba9698662d1e1838a7100/portend-2.7.1-py3-none-any.whl
pre-commit @ file:///Users/amit/Library/Caches/pypoetry/artifacts/7b/f6/8d/c5425b5811fbbbe85e82cc7980de882a7b239984e2338412a964d62d8f/pre_commit-2.13.0-py2.py3-none-any.whl
preshed @ file:///Users/amit/Library/Caches/pypoetry/artifacts/f9/3c/c4/47684f5aebfe9d8a986343a71cbbdef4bcbc7494eb8d08857621978f33/preshed-3.0.5-cp39-cp39-macosx_10_9_x86_64.whl
prometheus-client @ file:///Users/amit/Library/Caches/pypoetry/artifacts/ce/fa/c7/255d5bc58b6e1a7cc1e9d4eb3c5a993bea6af1aa016b5008147ab8beb7/prometheus_client-0.11.0-py2.py3-none-any.whl
promise @ file:///Users/amit/Library/Caches/pypoetry/artifacts/d6/c6/43/95f1e737b1dd79d3a5ac6cfb264a889716bab4cd9d28a9bc8c69591d53/promise-2.3.tar.gz
prompt-toolkit @ file:///Users/amit/Library/Caches/pypoetry/artifacts/c0/cc/91/2b2506bf1a53afc81d06682da5dce00b5f700a51ee2a8452dc8b98af15/prompt_toolkit-3.0.19-py3-none-any.whl
protobuf @ file:///Users/amit/Library/Caches/pypoetry/artifacts/88/20/9e/47da88de518b7b11435b39354a7b9a334acfae01916f4aee875b908765/protobuf-3.17.3-cp39-cp39-macosx_10_9_x86_64.whl
psutil @ file:///Users/amit/Library/Caches/pypoetry/artifacts/c5/45/be/f328230784273b5b602263092a0788e3055d7fa032dc1dcb0b1583bcb9/psutil-5.8.0-cp39-cp39-macosx_10_9_x86_64.whl
ptyprocess @ file:///Users/amit/Library/Caches/pypoetry/artifacts/2a/29/5d/0cdc5ec916431d60f03d2f725c54edbfa9fe53700b75fdfee209a3291e/ptyprocess-0.7.0-py2.py3-none-any.whl
py @ file:///Users/amit/Library/Caches/pypoetry/artifacts/60/79/0b/c48bd9c2a989aa8b1eb7a67cd02b053c10734f2e4e5665f7995f09999c/py-1.10.0-py2.py3-none-any.whl
py-rouge @ file:///Users/amit/Library/Caches/pypoetry/artifacts/31/31/4f/cc7585fdf5aec32c5b688726f52c2238f959caba4f6a65950f1a932745/py_rouge-1.1-py3-none-any.whl
py-spy @ file:///Users/amit/Library/Caches/pypoetry/artifacts/23/40/09/d09dd2f0516f0d0e73fe6a8dcfc9aaa5c13ffc2ebe7361b237391b0625/py_spy-0.3.7-py2.py3-none-macosx_10_9_x86_64.whl
pyasn1 @ file:///Users/amit/Library/Caches/pypoetry/artifacts/7b/3a/54/42ce43b579bda01b9d79022fb733811594441e7a32e9f9a5a98f672bdc/pyasn1-0.4.8-py2.py3-none-any.whl
pyasn1-modules @ file:///Users/amit/Library/Caches/pypoetry/artifacts/dd/b8/4f/b56433e0354274a31074995e01b8671751e9f0ed0001f5254e5b03a54f/pyasn1_modules-0.2.8-py2.py3-none-any.whl
pycodestyle @ file:///Users/amit/Library/Caches/pypoetry/artifacts/4c/30/97/026c283ef67eb248e5b7e6fad1f8ffb99dae50c11fd93eb939fd7c1f46/pycodestyle-2.7.0-py2.py3-none-any.whl
pycparser @ file:///Users/amit/Library/Caches/pypoetry/artifacts/37/8e/5a/0ea4f84bc7f11e0e3468110efa2c7783241ea7eaa63a92a751de06f78f/pycparser-2.20-py2.py3-none-any.whl
pydantic @ file:///Users/amit/Library/Caches/pypoetry/artifacts/96/6b/c5/3defa89e523826db389eb2be9b0b808835134ab64c68e8193da2ac7d51/pydantic-1.8.2-cp39-cp39-macosx_10_9_x86_64.whl
pydocstyle @ file:///Users/amit/Library/Caches/pypoetry/artifacts/75/e7/e5/1acad15a51efd39cf39259c7888c205fd787a92efea28f7afc5a9e315c/pydocstyle-6.1.1-py3-none-any.whl
pyflakes @ file:///Users/amit/Library/Caches/pypoetry/artifacts/eb/c4/2c/47fcc1b3f387b1f7033e85b3ac6ee7772338461a8de8ac3977c6a7dcc1/pyflakes-2.3.1-py2.py3-none-any.whl
Pygments @ file:///Users/amit/Library/Caches/pypoetry/artifacts/d1/62/9b/10957d050758a3da079375961ec00a6c83b71a90eefd0199361d9d54de/Pygments-2.9.0-py3-none-any.whl
Pyment @ file:///Users/amit/Library/Caches/pypoetry/artifacts/e0/cc/65/a915b66c524613d67c0dfcd3b9e99a2916908d106a16a5909520b29036/Pyment-0.3.3-py2.py3-none-any.whl
pyparsing @ file:///Users/amit/Library/Caches/pypoetry/artifacts/92/0f/cf/effdcd5d76a6186df0969f85b3b030284ff8058936d5016540b5258ea3/pyparsing-2.4.7-py2.py3-none-any.whl
pyrsistent @ file:///Users/amit/Library/Caches/pypoetry/artifacts/a3/50/5c/55747fd13209ee597a3e45d1ad70eef133559516e470c05a47c29365b0/pyrsistent-0.18.0-cp39-cp39-macosx_10_9_x86_64.whl
pytest @ file:///Users/amit/Library/Caches/pypoetry/artifacts/17/a3/46/eb89acf91c8962553e409da649186ff5e3d2c1c93195f0643e7dfd1b57/pytest-6.2.4-py3-none-any.whl
python-dateutil @ file:///Users/amit/Library/Caches/pypoetry/artifacts/93/67/cf/49f56d9e954addcfc50e5ffc9faee013c2eb00c6d77d56c6a22cb33b54/python_dateutil-2.8.1-py2.py3-none-any.whl
python-docx @ file:///Users/amit/Library/Caches/pypoetry/artifacts/7f/3f/b0/ca05b61dd6a8beb8bc8317700154416271ddda4db5425c92e9d780cba7/python-docx-0.8.11.tar.gz
python-dotenv @ file:///Users/amit/Library/Caches/pypoetry/artifacts/a6/61/71/a8300bb6be27750f8810f5d2a0c070e220ebc1a416f0837d4bbc283391/python_dotenv-0.17.1-py2.py3-none-any.whl
pytz @ file:///Users/amit/Library/Caches/pypoetry/artifacts/b0/a7/8d/54de3ab4d1ff29abbbca1e9ccbaefdc2a1b290138311b84f73bee16de1/pytz-2021.1-py2.py3-none-any.whl
PyYAML @ file:///Users/amit/Library/Caches/pypoetry/artifacts/b6/55/22/537845ea953a4d8d5006f11bdd1b03824425d7f809d5a7ae8efbbeab95/PyYAML-5.4.1-cp39-cp39-macosx_10_9_x86_64.whl
pyzmq @ file:///Users/amit/Library/Caches/pypoetry/artifacts/0f/06/79/957f3cfec70cf4f952ea59dd7a62281c8687323814b532588d3690e2ec/pyzmq-22.1.0-cp39-cp39-macosx_10_15_universal2.whl
qtconsole @ file:///Users/amit/Library/Caches/pypoetry/artifacts/88/04/9d/bfea17c2892fc9d54fe273876b3c28dc50389a75af0a43c4e5db123bbd/qtconsole-5.1.1-py3-none-any.whl
QtPy @ file:///Users/amit/Library/Caches/pypoetry/artifacts/7b/28/5f/a53ed26195df09abacb0aa3383076185f0496f7dc9f7b496180c7316a1/QtPy-1.9.0-py2.py3-none-any.whl
regex @ file:///Users/amit/Library/Caches/pypoetry/artifacts/0a/f5/b9/ac8ceed381bfa6deb8be166808f27353d5565085d6b762baac40befdff/regex-2021.7.6-cp39-cp39-macosx_10_9_x86_64.whl
requests @ file:///Users/amit/Library/Caches/pypoetry/artifacts/22/0a/9d/0df883fbffbb406d0cddbb35e881e4ac6bfb8f0dee8733056b6a054bf7/requests-2.25.1-py2.py3-none-any.whl
restructuredtext-lint @ file:///Users/amit/Library/Caches/pypoetry/artifacts/b6/09/80/91d176f17ba9a28291203e41600b294aa26214e185082bcb0cc3543588/restructuredtext_lint-1.3.2.tar.gz
rich @ file:///Users/amit/Library/Caches/pypoetry/artifacts/d8/09/d1/c1e297bde3276fe5289c3fc9cf6804eb531e28372b527f854b1427b012/rich-10.5.0-py3-none-any.whl
rsa @ file:///Users/amit/Library/Caches/pypoetry/artifacts/27/21/1f/fea99b1c1766c11c2c47dd961d7773ebab5c6acbf730200bd2e021b836/rsa-4.7.2-py3-none-any.whl
s3transfer @ file:///Users/amit/Library/Caches/pypoetry/artifacts/8f/96/42/4ec7dc1795d747cb348df8b6aad3b471251863f3eab457ce33668cf8a1/s3transfer-0.4.2-py2.py3-none-any.whl
sacremoses @ file:///Users/amit/Library/Caches/pypoetry/artifacts/96/d8/48/cdb78bf884395d731e5af8316b4de4517cd3b3b2b7bd28ede180216c83/sacremoses-0.0.45-py3-none-any.whl
scikit-learn @ file:///Users/amit/Library/Caches/pypoetry/artifacts/04/83/07/6b7befd90140a96d92542597559789a8e6f99ebcd86366336d38472ea6/scikit_learn-0.24.2-cp39-cp39-macosx_10_13_x86_64.whl
scipy @ file:///Users/amit/Library/Caches/pypoetry/artifacts/d1/03/cb/ba3c405831dab5ed08abcc721f597e4374beb9ac88f9b0a426d101b02b/scipy-1.6.1-cp39-cp39-macosx_10_9_x86_64.whl
Send2Trash @ file:///Users/amit/Library/Caches/pypoetry/artifacts/44/b5/57/aae8fa9005db0161e198bc50572d47f162988a2e2f3434bd4d8e6f78a0/Send2Trash-1.7.1-py3-none-any.whl
sentencepiece @ file:///Users/amit/Library/Caches/pypoetry/artifacts/0c/cd/fe/92b552fb5511ad2ab9906a6b85474f6a7b6312629fe9524f11bff2933d/sentencepiece-0.1.96-cp39-cp39-macosx_10_6_x86_64.whl
sentry-sdk @ file:///Users/amit/Library/Caches/pypoetry/artifacts/29/89/c1/911ea3ca1b49d945fd159b841827b63a00f57c439d6c0c26cb7aca5f7c/sentry_sdk-1.3.0-py2.py3-none-any.whl
sgmllib3k @ file:///Users/amit/Library/Caches/pypoetry/artifacts/48/41/c1/47c574e94f31057312eab350c2a7e7b75d1105eb5b673a14efe485c128/sgmllib3k-1.0.0.tar.gz
shortuuid @ file:///Users/amit/Library/Caches/pypoetry/artifacts/80/85/8d/5bdb9fbab8b4fc7bd9599a4982cac0ae2498f4c863d13869d4e1e7b722/shortuuid-1.0.1-py3-none-any.whl
six @ file:///Users/amit/Library/Caches/pypoetry/artifacts/08/9f/47/c16ae03035fc69eaf100ea39657a49baaeef714e25a52575710c34cd48/six-1.16.0-py2.py3-none-any.whl
smmap @ file:///Users/amit/Library/Caches/pypoetry/artifacts/fb/95/d9/27c304575d15e0faf1b64e46ec12611c0a683b4d9d6aa459850d5a77df/smmap-4.0.0-py2.py3-none-any.whl
snowballstemmer @ file:///Users/amit/Library/Caches/pypoetry/artifacts/c7/56/66/7613028d4906686fd240574f9e4ec773d99d60753a515f163d21b44935/snowballstemmer-2.1.0-py2.py3-none-any.whl
sortedcontainers @ file:///Users/amit/Library/Caches/pypoetry/artifacts/b9/80/e1/4bdfa349488797fd308ecbe48f4fad57a3245777fb47c8741730583262/sortedcontainers-2.4.0-py2.py3-none-any.whl
soupsieve @ file:///Users/amit/Library/Caches/pypoetry/artifacts/20/16/55/4a9893b172bb2a7815f46f6a947ff3506dd241ea679377ffdc0b2c811e/soupsieve-2.2.1-py3-none-any.whl
spacy @ file:///Users/amit/Library/Caches/pypoetry/artifacts/7d/eb/24/0b844c7393780507707bd2faddc484095f21df0f6d0bc69be88d6d7137/spacy-2.3.7-cp39-cp39-macosx_10_9_x86_64.whl
srsly @ file:///Users/amit/Library/Caches/pypoetry/artifacts/4e/7f/ef/750f5466b050be2d76de8aaf076c3ffcdff5cd0a5af0cf11ef6ebea3c6/srsly-1.0.5-cp39-cp39-macosx_10_9_x86_64.whl
stevedore @ file:///Users/amit/Library/Caches/pypoetry/artifacts/c2/31/a7/c2802b19c1cba9407f4254c97dacd72884d3b27c63bbbb3ada4edbf3a8/stevedore-3.3.0-py3-none-any.whl
subprocess32 @ file:///Users/amit/Library/Caches/pypoetry/artifacts/b9/91/2e/cc8d3ccbf05fa27ee73859de9d02ef1a7eba84ed701970db1063a1848d/subprocess32-3.5.4.tar.gz
tempora @ file:///Users/amit/Library/Caches/pypoetry/artifacts/b9/f8/26/4304b7c3157148ded2a9f388c0c3e6b1c0016147ff5995928d8db2ce8b/tempora-4.1.1-py3-none-any.whl
tensorboardX @ file:///Users/amit/Library/Caches/pypoetry/artifacts/f6/18/d3/c562b4cfa8f42a7172c29cfea8ded406d87b3e79e6bcc0032dc3eace99/tensorboardX-2.4-py2.py3-none-any.whl
termcolor @ file:///Users/amit/Library/Caches/pypoetry/artifacts/a2/5d/c7/e4ccb3b3bb8d3e3aff995fb6ffb12cfc78bbc8affa283907ee5eb5a5a5/termcolor-1.1.0.tar.gz
terminado @ file:///Users/amit/Library/Caches/pypoetry/artifacts/ac/6f/f6/464b3b1e95eff0b0b464b918cbcf0b6a2f81b4fda008b8570450a25a7a/terminado-0.10.1-py3-none-any.whl
testfixtures @ file:///Users/amit/Library/Caches/pypoetry/artifacts/87/49/bd/0bf2640dd53740f9dc21bb1eecb8093631a71ea4703c8ef1a3a7bdd42d/testfixtures-6.17.1-py2.py3-none-any.whl
testpath @ file:///Users/amit/Library/Caches/pypoetry/artifacts/1e/2d/08/76691a9e7e429930fb378dd96f760de96f2686841c47da2b35a04c5aad/testpath-0.5.0-py3-none-any.whl
thinc @ file:///Users/amit/Library/Caches/pypoetry/artifacts/71/1d/6d/f57a2fc5d6752b008f1583c84234c8909921813d7f863b65fac52afdc4/thinc-7.4.5-cp39-cp39-macosx_10_9_x86_64.whl
threadpoolctl @ file:///Users/amit/Library/Caches/pypoetry/artifacts/7b/b5/49/550b4953bb841e92404d74b7c7671139fd8bbdcec26c2e89c1843fcb76/threadpoolctl-2.2.0-py3-none-any.whl
tokenizers @ file:///Users/amit/Library/Caches/pypoetry/artifacts/e3/67/87/626827d63f69f9c1357c553076ccf0ee9721c9764ef434fa3654531f3f/tokenizers-0.10.3-cp39-cp39-macosx_10_11_x86_64.whl
toml @ file:///Users/amit/Library/Caches/pypoetry/artifacts/6b/6a/c9/53b19f7870a77d855e8b05ecdc98193944e5d246dafe11bbcad850ecba/toml-0.10.2-py2.py3-none-any.whl
tomlkit @ file:///Users/amit/Library/Caches/pypoetry/artifacts/fd/06/32/b79e75623225a9b5af79899482b9c2933c2fa2c6fb0eff80fcec10ae48/tomlkit-0.7.2-py2.py3-none-any.whl
torch @ file:///Users/amit/Library/Caches/pypoetry/artifacts/15/c9/5c/5d6856af4e4b18034b2e50b2cc08ba37035276b823f21dfa97a2e454b5/torch-1.8.1-cp39-none-macosx_10_9_x86_64.whl
torchvision @ file:///Users/amit/Library/Caches/pypoetry/artifacts/b2/82/34/8438fea281caa4f92140f0bdd1ff59ee8bc2e3e0b35d68bfc3d0d14111/torchvision-0.9.1-cp39-cp39-macosx_10_9_x86_64.whl
tornado @ file:///Users/amit/Library/Caches/pypoetry/artifacts/8a/d4/49/50c44b642217b81c4d63991cb2560ebb78029f8090bd4a23c1c11a4ac0/tornado-6.1-cp39-cp39-macosx_10_9_x86_64.whl
tqdm @ file:///Users/amit/Library/Caches/pypoetry/artifacts/1a/46/38/2897fecc5f3ff99d118d7dd77d749365e163113a712337b2b28837bedd/tqdm-4.61.2-py2.py3-none-any.whl
traitlets @ file:///Users/amit/Library/Caches/pypoetry/artifacts/f3/6b/36/998ab52c38eb1c4820cdef1e66043ddebb64e04535f88dbfd04486ce03/traitlets-5.0.5-py3-none-any.whl
transformers @ file:///Users/amit/Library/Caches/pypoetry/artifacts/ae/3e/26/9a99e73cfd7d44463aad270e850034478e6258f45900175d268a0cd976/transformers-4.5.1-py3-none-any.whl
types-orjson @ file:///Users/amit/Library/Caches/pypoetry/artifacts/e3/ac/f5/782a3257b6f17101317a753b694ce3dffc227c7eee054f8633cbb18700/types_orjson-0.1.1-py2.py3-none-any.whl
typing-extensions @ file:///Users/amit/Library/Caches/pypoetry/artifacts/3d/38/26/2c9b521373bbaf207e658ec81f51aa2a8af7454bfe4d7c15743a6533d5/typing_extensions-3.10.0.0-py3-none-any.whl
urllib3 @ file:///Users/amit/Library/Caches/pypoetry/artifacts/d8/4b/3f/9e8027e7f15b2f99244ad505328c3cf87912ad87446c1c8e89efacf731/urllib3-1.25.11-py2.py3-none-any.whl
virtualenv @ file:///Users/amit/Library/Caches/pypoetry/artifacts/13/ab/72/1f00c98674e32cc5fcae271032c164917dfd53be7c009e119f1cf47f8d/virtualenv-20.4.7-py2.py3-none-any.whl
wandb @ file:///Users/amit/Library/Caches/pypoetry/artifacts/49/f4/c8/324b20beeceb351e72c821219d999d460442c4b4ff903122f29979ab5e/wandb-0.10.33-py2.py3-none-any.whl
wasabi @ file:///Users/amit/Library/Caches/pypoetry/artifacts/d5/9d/af/58d834e926bfc5371fb8208596bdec3d5824083600c5681f98ce0790d7/wasabi-0.8.2-py3-none-any.whl
wcwidth @ file:///Users/amit/Library/Caches/pypoetry/artifacts/7d/f4/60/0737157bb9711fec72c70dff523aa54491eef317e0d586cf5388ff0908/wcwidth-0.2.5-py2.py3-none-any.whl
webencodings @ file:///Users/amit/Library/Caches/pypoetry/artifacts/ed/d4/da/61384706cfac042ba3bd148746d66e50695463993be117c7c8dadeef7a/webencodings-0.5.1-py2.py3-none-any.whl
wemake-python-styleguide @ file:///Users/amit/Library/Caches/pypoetry/artifacts/b2/30/6d/aff16dd6cc6e7169bd34e8f2e8feff71a3cc593d6ebb617acc9cf7e927/wemake_python_styleguide-0.15.3-py3-none-any.whl
widgetsnbextension @ file:///Users/amit/Library/Caches/pypoetry/artifacts/eb/b0/c5/e9e106309ddf8d2cbebcd3c9f2c2be8c7c7346f58d8ff7ace4196371d8/widgetsnbextension-3.5.1-py2.py3-none-any.whl
word2number @ file:///Users/amit/Library/Caches/pypoetry/artifacts/91/7b/91/fd4e6b1580eb2a2f0bb8b725ba137628acb0adb21522a3ff9d69e6f5e1/word2number-1.1.zip
zc.lockfile @ file:///Users/amit/Library/Caches/pypoetry/artifacts/b5/a8/c8/e94e98335e585be92e35e5d07dd8a75e5c2e7774c8bd24410160f9cfe0/zc.lockfile-2.0-py2.py3-none-any.whl
```
</p>
</details>
## Steps to reproduce
I wasn't sure how to do this step but I will if I'm not making any sense and you'd like me to.
<details>
<summary><b>Example source:</b></summary>
<p>
<!-- Add a fully runnable example in between the next two lines below that will reproduce the bug -->
```
```
</p>
</details>
| closed | 2021-07-12T15:44:45Z | 2021-07-26T18:05:23Z | https://github.com/allenai/allennlp/issues/5308 | [
"bug"
] | amitkparekh | 3 |
AUTOMATIC1111/stable-diffusion-webui | deep-learning | 15,322 | [Bug]: Checkpoint / LoRA gallery slows down the browser | ### Checklist
- [X] The issue exists after disabling all extensions
- [X] The issue exists on a clean installation of webui
- [ ] The issue is caused by an extension, but I believe it is caused by a bug in the webui
- [X] The issue exists in the current version of the webui
- [X] The issue has not been reported before recently
- [ ] The issue has been reported before but has not been fixed yet
### What happened?
Since the last release, the checkpoint / LoRA galleries slow down Firefox to the point that Firefox shows a "This site slows down Firefox [stop]" banner at top of the page, if you have many checkpoints/LoRA.
### Steps to reproduce the problem
1. Have many checkpoints and/or LoRA
2. Click the Checkpoints or LoRa tab
### What should have happened?
Before the last release, it had no noticable delay. Now it takes quite some time until the tab switches, the webui gets unresponsive and Firefox warns that the site is slowing down the browser.
If wonder if it previously had some kind of lazy-loading in place that was removed or similar things. I do not even need to scroll in the last, changing the tab from Generation to Lora alone is enough to slow down the browser before I even see the first rectangle. Most Lora do not have preview images but only the standard grey placeholder.
### What browsers do you use to access the UI ?
Mozilla Firefox
### Sysinfo
(can add later if needed as I would need to remove sensitive information)
### Console logs
```Shell
(I don't see anything relevant)
```
### Additional information
_No response_ | open | 2024-03-19T10:01:20Z | 2024-03-19T10:01:20Z | https://github.com/AUTOMATIC1111/stable-diffusion-webui/issues/15322 | [
"bug-report"
] | allo- | 0 |
oegedijk/explainerdashboard | dash | 101 | It is able to support RNN models like LSTM? (3D array INPUT) | It is possible to adapt this set of dashboards for RNN-based models, such as LSTM networks.
The main problem I am encountering is in the input dataset which, being pre-processed, does not have the traditional linear structure, but is produced by the application of window-based techniques, for example.
The INPUT it's a 3D array

| closed | 2021-03-15T17:20:02Z | 2021-05-04T07:54:09Z | https://github.com/oegedijk/explainerdashboard/issues/101 | [] | Alex891 | 5 |
junyanz/pytorch-CycleGAN-and-pix2pix | pytorch | 1,207 | Question: Why does a Pixel Discriminator work? | As far as I understand the Pixel Discriminator is equivalent to a PatchGAN of size 1x1, meaning the decoder tries to decide if each pixel is real or fake. But without taking into account surrounding pixels, what information does it use to assign a label and how is it also able to recover high frequency contents of the images?
I have been working on a project and the only stable training I could get was using the pixel discriminator. With the patch gan the discriminator started off too good compared to the generator and the network couldn't learn. Hence my curiosity. | open | 2020-12-03T16:40:41Z | 2020-12-17T03:40:59Z | https://github.com/junyanz/pytorch-CycleGAN-and-pix2pix/issues/1207 | [] | dpelacani | 3 |
OFA-Sys/Chinese-CLIP | nlp | 15 | CN-CLIPViT-B/16 模型在Flickr30K-CN数据集上的zero-shot结果复现不出来 | 你好,我使用你们开源的CN-CLIPViT-B/16 模型在MUGE上复现的结果和你们汇报的结果差不多

但是在Flickr30K-CN数据集上的结果差很远,不知道问题出在哪里,麻烦帮忙看下,谢谢。

| closed | 2022-11-18T11:30:07Z | 2022-11-28T06:30:05Z | https://github.com/OFA-Sys/Chinese-CLIP/issues/15 | [
"solved in expectation",
"已解决待验证"
] | ddNTP | 17 |
slackapi/bolt-python | fastapi | 1,032 | Rename `logger` to `logger_` in `slack_bolt.kwargs_injection.args.Args` | It is common practice to define a variable called `logger` in a script. However, the `slack_bolt.kwargs_injection.args.Args` function used as a decorator also defines `logger`, so if you want to call your normal logger in your script you have to give it a different name. It can also be unexpected behaviour, as the author may think they are calling the `logger` from their module, but actually they are calling a different logger instance altogether.
I understand this would be a breaking change, but I think it would improve usability.
### Category (place an `x` in each of the `[ ]`)
* [ ] **slack_bolt.App** and/or its core components
* [ ] **slack_bolt.async_app.AsyncApp** and/or its core components
* [ ] Adapters in **slack_bolt.adapter**
* [x] Others
## Requirements
Please read the [Contributing guidelines](https://github.com/slackapi/bolt-python/blob/main/.github/contributing.md) and [Code of Conduct](https://slackhq.github.io/code-of-conduct) before creating this issue or pull request. By submitting, you are agreeing to those rules.
| closed | 2024-02-22T10:50:55Z | 2024-02-28T00:26:15Z | https://github.com/slackapi/bolt-python/issues/1032 | [
"enhancement",
"good first issue"
] | ChrisHills463 | 7 |
stanfordnlp/stanza | nlp | 1,340 | Dependency parsing: is there some way to discourage multiple `nsubj` dependents? | A very weird English tree produced by Stanza 1.6.0 in the [demo](http://stanza.run/):
> My cousin my extremely rude colleague admired last year chewed the chicken enthusiastically.
<img width="1171" alt="image" src="https://github.com/stanfordnlp/stanza/assets/985263/a6468f67-9d09-4ad5-8f3e-4cbc92d08e6b">
In UD, no word is allowed to have multiple (plain) `nsubj` dependents. But "admired" has two.
Is there a recommended alternate training or decoding method in Stanza that could avoid this sort of problem? | open | 2024-01-30T04:00:00Z | 2025-02-22T18:20:13Z | https://github.com/stanfordnlp/stanza/issues/1340 | [
"question"
] | nschneid | 8 |
Yorko/mlcourse.ai | data-science | 711 | No demo assignment in topic-3 | [jupyter_english/topic03_decision_trees_kNN](https://github.com/Yorko/mlcourse.ai/tree/main/jupyter_english/topic03_decision_trees_kNN) and [jupyter_french/topic03_decision_trees_kNN](https://github.com/Yorko/mlcourse.ai/tree/main/jupyter_french/topic03_decision_trees_kNN) has no .ipynb files for the demo assignment and solution to it.
Can you add them to the repo? | closed | 2022-09-05T15:33:33Z | 2022-09-07T14:26:24Z | https://github.com/Yorko/mlcourse.ai/issues/711 | [] | aulasau | 3 |
cleanlab/cleanlab | data-science | 401 | Add tutorial notebook to docs.cleanlab.ai | Make tutorial notebook appear on docs.cleanlab.ai (and verify it looks good on the website) by adding it to:
1) https://github.com/cleanlab/cleanlab/blob/master/docs/source/tutorials/index.rst
2) https://github.com/cleanlab/cleanlab/blob/master/docs/source/index.rst | closed | 2022-09-06T00:52:03Z | 2022-09-09T01:27:29Z | https://github.com/cleanlab/cleanlab/issues/401 | [] | elisno | 0 |
gee-community/geemap | streamlit | 1,952 | No layer style UI for ee.ImageCollection | ### Environment Information
Colab
Mon Mar 25 21:27:30 2024 UTC
OS Linux CPU(s) 2 Machine x86_64
Architecture 64bit RAM 12.7 GiB Environment IPython
Python 3.10.12 (main, Nov 20 2023, 15:14:05) [GCC 11.4.0]
geemap 0.32.0 ee 0.1.394 ipyleaflet 0.18.2
folium 0.14.0 jupyterlab Module not found notebook 6.5.5
ipyevents 2.0.2 geopandas 0.13.2
### Description
The layer visualization editing tool is not available when the input to `Map.add_layer` is an `ee.ImageCollection`. If an image collection is passed, I believe `.mosaic()` is called so that image tiles can be requested for viewing. Since we're visualizing an `ee.Image` object, we should be able to set and apply visualization parameters (consistent with Code Editor behavior).
### What I Did
1. Add an `ee.ImageCollection` object to a `Map` with `add_layer`
2. Display the map and click layer gear icon to set visualization properties (no tool "vis params are uneditable")
3. Mosaic the `ee.ImageCollection` first and then add to a map
4. Display the new map and click layer gear icon to set visualization properties (tool appears and WAI).
```py
# [NEW CELL] Construct an ee.ImageCollection.
col = ee.ImageCollection([ee.Image(1).byte(), ee.Image(10).byte()])
# [NEW CELL] Display the ee.ImageCollection.
m0 = geemap.Map()
m0.add_layer(col)
m0
# [NEW CELL] Display the .mosaic() of the ee.ImageCollection.
m1 = geemap.Map()
m1.add_layer(col.mosaic())
m1
```

| closed | 2024-03-25T21:48:17Z | 2024-03-26T02:42:30Z | https://github.com/gee-community/geemap/issues/1952 | [
"bug"
] | jdbcode | 0 |
pytest-dev/pytest-django | pytest | 499 | Making django_db_blocker un-opt-outable is zealous and lame | Use of `django_db_blocker` to cripple use of the database outside of specially tagged tests and fixtures interferes with Django model `get_or_create` calls at import time, which is how some systems ensure the presence of fundamental data across environments.
Making this behavior not configurable is disruptive and annoying. It results in the need to patch the patcher, a la
```
# settings/test.py
from pytest_django import plugin
plugin._blocking_manager = mock.MagicMock()
```
Please consider a means to disable this behavior through settings. | closed | 2017-07-28T22:42:23Z | 2018-03-09T09:46:13Z | https://github.com/pytest-dev/pytest-django/issues/499 | [] | jamesob | 6 |
mwaskom/seaborn | data-science | 3,392 | Swarmplot and markers | I am exploring the `swarmplot` functionality provided by seaborn. I tried to follow [stackoverflow suggestions](https://stackoverflow.com/questions/52878845/swarmplot-with-hue-affecting-marker-beyond-color) on changing the default marker for groups, but without success.
Looking at the documentation of the functionality provided by `swarmplot`, it says "additional parameters are passed to the `scatter` functions giving as an example:
```python
import seaborn as sns
tips = sns.load_dataset("tips")
sns.swarmplot(
data=tips, x="total_bill", y="day",
marker="x", linewidth=1,
)
```
However, combining this with e.g. `hue`, does not give the expected result (and I found a hint in #1595 that it is also not possible:
```python
sns.swarmplot(
data=tips, x="total_bill", y="day",
marker="x", linewidth=1,
hue='sex'
)
```

I would have expected that the crosses are now in blue and orange, but this is not the case. Happy to have a look at the code or update the documentation, but a hint where to start would be great. | closed | 2023-06-19T13:29:03Z | 2023-06-19T15:53:40Z | https://github.com/mwaskom/seaborn/issues/3392 | [] | enryH | 4 |
Buuntu/fastapi-react | fastapi | 100 | Form validation errors on Sign Up and Login pages | I found a bug where the frontend errors out when you press the submit button on empty fields for the login and sign up pages.
Here is a gif of the error for the sign up page. The same error occurred with login page as well.

| closed | 2020-08-07T04:52:52Z | 2020-08-13T02:48:20Z | https://github.com/Buuntu/fastapi-react/issues/100 | [] | adamnieto | 2 |
noirbizarre/flask-restplus | flask | 95 | Namespaces should be lazy registerable | As @frol suggested, for big APIs splitted on multiple files, Namespace should be the only class to import.
As Blueprints for Flask, it should provide all necessary methods to not require the API instance in the file and to be lazy registerable.
The target is being able to write:
``` python
# feature.py
from flask_restplus import Namespace, Resource
ns = Namespace('feature', 'Some namespace')
model = ns.model('Model', {})
@ns.route('/')
class MyFeature(Resource):
@ns.marshal_with(model)
def get(self):
pass
```
``` python
# api.py
from flask_restplus import Api
from feature import ns
api = Api()
api.register_namespace(ns)
```
| closed | 2015-11-10T08:54:30Z | 2016-01-17T19:22:48Z | https://github.com/noirbizarre/flask-restplus/issues/95 | [
"enhancement"
] | noirbizarre | 8 |
pytorch/pytorch | python | 149,223 | Inconsistent results in using torch.jit.script API from API documentation. | ### 🐛 Describe the bug
Expects an AttributeError is raised since the ignored_method is skipped in compiling as per API documentation.
```python
def test_script_module_with_ignored_method(self):
class IgnoredMethodModule(nn.Module):
def forward(self, x):
return x * 2
@torch.jit.ignore
def ignored_method(self, x):
return x * 3
module = IgnoredMethodModule()
scripted_module = torch.jit.script(module)
input_tensor = torch.tensor(5)
self.assertEqual(scripted_module(input_tensor), module(input_tensor))
# Ensure ignored method is not part of the scripted module
with self.assertRaises(AttributeError):
scripted_module.ignored_method(input_tensor)
```
```
======================================================================
FAIL: test_script_module_with_ignored_method (__main__.TestTorchJitScript)
----------------------------------------------------------------------
Traceback (most recent call last):
File "/home/user/projects/api_guided_testgen/out/bug_detect_gpt4o/exec/basic_rag_apidoc/torch/torch.jit.script.py", line 70, in test_script_module_with_ignored_method
scripted_module.ignored_method(input_tensor)
AssertionError: AttributeError not raised
```
### Versions
Collecting environment information...
PyTorch version: 2.5.0
Is debug build: False
CUDA used to build PyTorch: None
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.5 LTS (x86_64)
GCC version: (Ubuntu 9.5.0-1ubuntu1~22.04) 9.5.0
Clang version: 14.0.0-1ubuntu1.1
CMake version: version 3.22.1
Libc version: glibc-2.35
Python version: 3.9.21 (main, Dec 11 2024, 16:24:11) [GCC 11.2.0] (64-bit runtime)
Python platform: Linux-6.8.0-52-generic-x86_64-with-glibc2.35
Is CUDA available: False
CUDA runtime version: No CUDA
CUDA_MODULE_LOADING set to: N/A
GPU models and configuration: No CUDA
Nvidia driver version: No CUDA
cuDNN version: No CUDA
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
CPU:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Address sizes: 39 bits physical, 48 bits virtual
Byte Order: Little Endian
CPU(s): 4
On-line CPU(s) list: 0-3
Vendor ID: GenuineIntel
Model name: Intel(R) Xeon(R) E-2224G CPU @ 3.50GHz
CPU family: 6
Model: 158
Thread(s) per core: 1
Core(s) per socket: 4
Socket(s): 1
Stepping: 10
CPU max MHz: 4700.0000
CPU min MHz: 800.0000
BogoMIPS: 6999.82
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb pti ssbd ibrs ibpb stibp tpr_shadow flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid mpx rdseed adx smap clflushopt intel_pt xsaveopt xsavec xgetbv1 xsaves dtherm ida arat pln pts hwp hwp_notify hwp_act_window hwp_epp vnmi md_clear flush_l1d arch_capabilities
Virtualization: VT-x
L1d cache: 128 KiB (4 instances)
L1i cache: 128 KiB (4 instances)
L2 cache: 1 MiB (4 instances)
L3 cache: 8 MiB (1 instance)
NUMA node(s): 1
NUMA node0 CPU(s): 0-3
Vulnerability Gather data sampling: Mitigation; Microcode
Vulnerability Itlb multihit: KVM: Mitigation: VMX disabled
Vulnerability L1tf: Mitigation; PTE Inversion; VMX conditional cache flushes, SMT disabled
Vulnerability Mds: Mitigation; Clear CPU buffers; SMT disabled
Vulnerability Meltdown: Mitigation; PTI
Vulnerability Mmio stale data: Mitigation; Clear CPU buffers; SMT disabled
Vulnerability Reg file data sampling: Not affected
Vulnerability Retbleed: Mitigation; IBRS
Vulnerability Spec rstack overflow: Not affected
Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl
Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization
Vulnerability Spectre v2: Mitigation; IBRS; IBPB conditional; STIBP disabled; RSB filling; PBRSB-eIBRS Not affected; BHI Not affected
Vulnerability Srbds: Mitigation; Microcode
Vulnerability Tsx async abort: Mitigation; TSX disabled
Versions of relevant libraries:
[pip3] numpy==2.0.1
[pip3] torch==2.5.0
[pip3] torchaudio==2.5.0
[pip3] torchvision==0.20.0
[conda] blas 1.0 mkl
[conda] cpuonly 2.0 0 pytorch
[conda] ffmpeg 4.3 hf484d3e_0 pytorch
[conda] libjpeg-turbo 2.0.0 h9bf148f_0 pytorch
[conda] mkl 2023.1.0 h213fc3f_46344
[conda] mkl-service 2.4.0 py39h5eee18b_2
[conda] mkl_fft 1.3.11 py39h5eee18b_0
[conda] mkl_random 1.2.8 py39h1128e8f_0
[conda] numpy 2.0.1 py39h5f9d8c6_1
[conda] numpy-base 2.0.1 py39hb5e798b_1
[conda] pytorch 2.5.0 py3.9_cpu_0 pytorch
[conda] pytorch-mutex 1.0 cpu pytorch
[conda] torchaudio 2.5.0 py39_cpu pytorch
[conda] torchvision 0.20.0 py39_cpu pytorch
cc @EikanWang @jgong5 @wenzhe-nrv @sanchitintel | open | 2025-03-14T20:54:32Z | 2025-03-14T23:20:43Z | https://github.com/pytorch/pytorch/issues/149223 | [
"oncall: jit"
] | sjh0849 | 0 |
mwaskom/seaborn | data-visualization | 3,769 | Smaller mark width for overlapping data with `so.Hist(common_bins=False)` | When doing a `so.Hist(common_bins=False)`, if the bins for each group overlap, the width calculated for each mark is smaller that it should be.
Here's a minimal working example, where I have a dataset `A`, and its x-shifted version `B = A + shift`. In each row, I'm plotting a different shift, and when they start overlapping, the bar width is smaller than the bin width.

```python
import matplotlib.pyplot as plt
import numpy as np
import pandas as pd
import seaborn.objects as so
def plot(ax, shift: float):
data = np.random.default_rng(0).normal(size=50)
df = pd.DataFrame({"A": data, "B": data + shift}).melt()
return (
so.Plot(df, x="value", color="variable")
.add(so.Bars(), so.Hist(common_bins=False))
.on(ax)
.plot()
)
shifts = [4.5, 4, 3.9, 3.8]
fig, axes = plt.subplots(len(shifts), sharex=True, gridspec_kw={"hspace": 0})
for ax, shift in zip(axes, shifts):
plot(ax, shift)
ax.set(ylabel=f"{shift = }")
```
I could trace it to this width calculation:
https://github.com/mwaskom/seaborn/blob/b4e5f8d261d6d5524a00b7dd35e00a40e4855872/seaborn/_core/plot.py#L1453
which ends up running the following line for all groups as one:
https://github.com/mwaskom/seaborn/blob/b4e5f8d261d6d5524a00b7dd35e00a40e4855872/seaborn/_core/scales.py#L467
If the bin edges are `[0, 1, 2]` and `[0.5, 1.5, 2.5]` for each group, it calculates the bin width from `[0, 0.5, 1, 1.5, ...]` and finds a width of 0.5 instead of a width of 1.
Maybe this is not a bug but something by design when there is overlap between marks?
In case it is a bug, I could contribute a fix, but would probably need some direction as to where to fix it.
Thanks! | open | 2024-10-19T21:33:56Z | 2024-10-24T12:11:39Z | https://github.com/mwaskom/seaborn/issues/3769 | [] | maurosilber | 1 |
albumentations-team/albumentations | machine-learning | 1,633 | [Tech debt] Improve interface for CoarseDropout | * `min_holes`, `max_holes` => `num_holes_range`
* `min_height`, `max_height` => `hole_height_range`
* `min_width`, `max_width` => `hole_width_range`
=>
We can update transform to use new signature, keep old as working, but mark as deprecated.
----
PR could be similar to https://github.com/albumentations-team/albumentations/pull/1615 | closed | 2024-04-05T19:29:31Z | 2024-04-19T22:01:33Z | https://github.com/albumentations-team/albumentations/issues/1633 | [
"good first issue",
"Tech debt"
] | ternaus | 1 |
databricks/koalas | pandas | 2,228 | gotImport Error | ImportError Traceback (most recent call last)
Cell In[8], line 4
2 import numpy as np
3 from collections.abc import Iterable
----> 4 import databricks.koalas as ks
5 import matplotlib.pyplot as plt
7 from pyspark import SparkContext, SparkConf
File ~\AppData\Local\Packages\PythonSoftwareFoundation.Python.3.10_qbz5n2kfra8p0\LocalCache\local-packages\Python310\site-packages\databricks\koalas\__init__.py:74
66 raise RuntimeError(
67 "Please explicitly unset 'ARROW_PRE_0_15_IPC_FORMAT' environment variable in both "
68 "driver and executor sides. It is required to set this environment variable only "
69 "when you use pyarrow>=0.15 and pyspark<3.0."
70 )
73 from databricks.koalas.frame import DataFrame
---> 74 from databricks.koalas.indexes import Index, MultiIndex
75 from databricks.koalas.series import Series
76 from databricks.koalas.typedef import pandas_wraps
File ~\AppData\Local\Packages\PythonSoftwareFoundation.Python.3.10_qbz5n2kfra8p0\LocalCache\local-packages\Python310\site-packages\databricks\koalas\indexes.py:51
49 from databricks.koalas.frame import DataFrame
50 from databricks.koalas.missing.indexes import _MissingPandasLikeIndex, _MissingPandasLikeMultiIndex
---> 51 from databricks.koalas.series import Series, _col
52 from databricks.koalas.utils import (
53 compare_allow_null,
54 compare_disallow_null,
(...)
61 validate_bool_kwarg,
62 )
63 from databricks.koalas.internal import _InternalFrame, NATURAL_ORDER_COLUMN_NAME
File ~\AppData\Local\Packages\PythonSoftwareFoundation.Python.3.10_qbz5n2kfra8p0\LocalCache\local-packages\Python310\site-packages\databricks\koalas\series.py:22
20 import re
21 import inspect
---> 22 from collections import Iterable, OrderedDict
23 from functools import partial, wraps, reduce
24 from typing import Any, Generic, List, Optional, Tuple, TypeVar, Union
ImportError: cannot import name 'Iterable' from 'collections' (C:\Program Files\WindowsApps\PythonSoftwareFoundation.Python.3.10_3.10.2800.0_x64__qbz5n2kfra8p0\lib\collections\__init__.py) | open | 2023-03-22T08:55:33Z | 2024-07-05T08:52:28Z | https://github.com/databricks/koalas/issues/2228 | [] | ParthRMehta | 1 |
tensorpack/tensorpack | tensorflow | 923 | Monitor tensors every batch | Default epoch monitors for training are quite useful.
Could you give an example of how to monitor one tensor's value **every batch** ?(and write it to tensorboard file.)
Thanks a lot.
| closed | 2018-10-05T20:19:07Z | 2018-10-25T20:32:44Z | https://github.com/tensorpack/tensorpack/issues/923 | [
"duplicate",
"usage"
] | XinDongol | 4 |
deepset-ai/haystack | nlp | 8,932 | Remove folder from core integrations https://github.com/deepset-ai/haystack-core-integrations/tree/main/nodes | closed | 2025-02-25T10:56:39Z | 2025-02-25T15:26:30Z | https://github.com/deepset-ai/haystack/issues/8932 | [
"P2"
] | julian-risch | 0 |
|
huggingface/datasets | nlp | 6,655 | Cannot load the dataset go_emotions | ### Describe the bug
When I run the following code I get an exception;
`go_emotions = load_dataset("go_emotions")`
> AttributeError Traceback (most recent call last)
Cell In[6], [line 1](vscode-notebook-cell:?execution_count=6&line=1)
----> [1](vscode-notebook-cell:?execution_count=6&line=1) go_emotions = load_dataset("go_emotions")
[2](vscode-notebook-cell:?execution_count=6&line=2) data = go_emotions.data
File [c:\Users\hijik\anaconda3\Lib\site-packages\datasets\load.py:2523](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2523), in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, verification_mode, ignore_verifications, keep_in_memory, save_infos, revision, token, use_auth_token, task, streaming, num_proc, storage_options, trust_remote_code, **config_kwargs)
[2518](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2518) verification_mode = VerificationMode(
[2519](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2519) (verification_mode or VerificationMode.BASIC_CHECKS) if not save_infos else VerificationMode.ALL_CHECKS
[2520](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2520) )
[2522](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2522) # Create a dataset builder
-> [2523](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2523) builder_instance = load_dataset_builder(
[2524](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2524) path=path,
[2525](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2525) name=name,
[2526](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2526) data_dir=data_dir,
[2527](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2527) data_files=data_files,
[2528](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2528) cache_dir=cache_dir,
[2529](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2529) features=features,
[2530](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2530) download_config=download_config,
[2531](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2531) download_mode=download_mode,
[2532](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2532) revision=revision,
[2533](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2533) token=token,
[2534](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2534) storage_options=storage_options,
[2535](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2535) trust_remote_code=trust_remote_code,
[2536](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/load.py:2536) _require_default_config_name=name is None,
...
---> [63](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/utils/_dill.py:63) if issubclass(obj_type, transformers.PreTrainedTokenizerBase):
[64](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/utils/_dill.py:64) pklregister(obj_type)(_save_transformersPreTrainedTokenizerBase)
[66](file:///C:/Users/hijik/anaconda3/Lib/site-packages/datasets/utils/_dill.py:66) # Unwrap `torch.compile`-ed functions
AttributeError: module 'transformers' has no attribute 'PreTrainedTokenizerBase'
Output is truncated. View as a [scrollable element](command:cellOutput.enableScrolling?10bc0728-6947-456e-9a3e-f056872b04c6) or open in a [text editor](command:workbench.action.openLargeOutput?10bc0728-6947-456e-9a3e-f056872b04c6). Adjust cell output [settings](command:workbench.action.openSettings?%5B%22%40tag%3AnotebookOutputLayout%22%5D)...
### Steps to reproduce the bug
```
from datasets import load_dataset
go_emotions = load_dataset("go_emotions")
```
### Expected behavior
Should simply load the variable with the data from the file
### Environment info
Copy-and-paste the text below in your GitHub issue.
- `datasets` version: 2.16.1
- Platform: Windows-10-10.0.22631-SP0
- Python version: 3.11.4
- `huggingface_hub` version: 0.20.3
- PyArrow version: 11.0.0
- Pandas version: 1.5.3
- `fsspec` version: 2023.10.0 | open | 2024-02-09T12:15:39Z | 2024-02-12T09:35:55Z | https://github.com/huggingface/datasets/issues/6655 | [] | arame | 4 |
sqlalchemy/alembic | sqlalchemy | 387 | Foreign key relationships are not created when use_alter is used. | **Migrated issue, originally created by tomkcook ([@tomkcook](https://github.com/tomkcook))**
The basic problem is that when a column is marked as ForeignKey and the use_alter option is set to True, the foreign key constraint is not created in the database. The database is postgresql 9.5.
```
$ pip freeze
alembic==0.8.7
click==6.6
Flask==0.11.1
Flask-Alembic==2.0.1
Flask-Login==0.3.2
Flask-SQLAlchemy==2.1
GeoAlchemy2==0.3.0
gunicorn==19.6.0
itsdangerous==0.24
Jinja2==2.8
Mako==1.0.4
MarkupSafe==0.23
pkg-resources==0.0.0
psycopg2==2.6.2
python-editor==1.0.1
SQLAlchemy==1.0.15
Werkzeug==0.11.11
```
My test application is attached. Briefly, I have these models:
```
from test import db
from sqlalchemy.dialects.postgres import UUID
from sqlalchemy.sql import func
class User(db.Model):
id = db.Column(UUID, primary_key = True, default=func.uuid_generate_v1())
name = db.Column(db.String)
class Post(db.Model):
id = db.Column(UUID, primary_key = True, default=func.uuid_generate_v1())
name = db.Column(db.String)
owner_id = db.Column(UUID, db.ForeignKey('user.id', use_alter=True))
owner = db.relationship("User", backref=db.backref('posts', lazy='dynamic'))
if __name__ == '__main__':
me = User(name = 'me')
db.session.add(me)
db.session.add(Post(name = 'My Post', owner = me))
db.session.commit()
```
When I generate a revision I get this:
```
""".
Revision ID: f4f67ed72eac
Revises:
Create Date: 2016-09-12 13:26:02.697400
"""
# revision identifiers, used by Alembic.
revision = 'f4f67ed72eac'
down_revision = None
branch_labels = ('default',)
depends_on = None
from alembic import op
import sqlalchemy as sa
from sqlalchemy.dialects import postgresql
def upgrade():
### commands auto generated by Alembic - please adjust! ###
op.create_table('post',
sa.Column('id', postgresql.UUID(), nullable=False),
sa.Column('name', sa.String(), nullable=True),
sa.Column('owner_id', postgresql.UUID(), nullable=True),
sa.ForeignKeyConstraint(['owner_id'], ['test.user.id'], use_alter=True),
sa.PrimaryKeyConstraint('id'),
schema='test'
)
op.create_table('user',
sa.Column('id', postgresql.UUID(), nullable=False),
sa.Column('name', sa.String(), nullable=True),
sa.PrimaryKeyConstraint('id'),
schema='test'
)
### end Alembic commands ###
def downgrade():
### commands auto generated by Alembic - please adjust! ###
op.drop_table('user', schema='test')
op.drop_table('post', schema='test')
### end Alembic commands ###
```
As you can see, there is a ForeignKeyConstraint created. But when I run alembic upgrade to that version, the table in the database doesn't have any foreign key constraints:
```
test=> \d user
Table "test.user"
Column | Type | Modifiers
--------+-------------------+-----------
id | uuid | not null
name | character varying |
Indexes:
"user_pkey" PRIMARY KEY, btree (id)
test=> \d post
Table "test.post"
Column | Type | Modifiers
----------+-------------------+-----------
id | uuid | not null
name | character varying |
owner_id | uuid |
Indexes:
"post_pkey" PRIMARY KEY, btree (id)
```
If the `use_alter` flag is omitted, then foreign key constraints are created correctly.
----------------------------------------
Attachments: [test.tar.bz2](../wiki/imported_issue_attachments/387/test.tar.bz2)
| closed | 2016-09-12T13:42:35Z | 2022-07-07T14:33:56Z | https://github.com/sqlalchemy/alembic/issues/387 | [
"bug",
"duplicate",
"recursive FK issue"
] | sqlalchemy-bot | 3 |
ultrafunkamsterdam/undetected-chromedriver | automation | 1,026 | Disable javascript not working | Hey there, I tried to disable the javascript running on the undetected_chromedriver but without luck.
This command does not work:
`self.chrome_options.add_argument('--disable-javascript')`
This alternative is not supported by the undetected_chromedriver
` prefs = {'profile.default_content_setting_values': {
'images': 2,
'javascript': 2,
'geolocation': 2,
'popups': 2,
'cookies': 2}}
self.chrome_options.add_experimental_option("prefs", prefs)
`
Not sure if there's another way to do it, but both methods above work on normal selenium chrome.
Thank you. | closed | 2023-02-03T17:25:35Z | 2023-02-03T17:51:52Z | https://github.com/ultrafunkamsterdam/undetected-chromedriver/issues/1026 | [] | ZhangPeng4242 | 0 |
JoeanAmier/TikTokDownloader | api | 152 | 【建议】main 启动的时候添加一下启动参数 | python mian.py 5 直接web页面启动 | open | 2024-02-06T12:25:15Z | 2024-02-06T14:26:46Z | https://github.com/JoeanAmier/TikTokDownloader/issues/152 | [] | scccy | 2 |
Asabeneh/30-Days-Of-Python | flask | 306 | No license file | The `license.md` file helps with how your files can be edited and even forked also, guides how pull request can be done on your main branch.
It should be added.
I can work on it. | open | 2022-10-01T23:49:00Z | 2024-04-09T02:08:23Z | https://github.com/Asabeneh/30-Days-Of-Python/issues/306 | [] | chemben17 | 2 |
allenai/allennlp | nlp | 4,974 | Publish allennlp-server to pypi | It's annoying that it's separate. It doesn't get proper test coverage this way. We should either support this feature properly, or not at all. | closed | 2021-02-12T01:03:36Z | 2021-04-06T18:29:48Z | https://github.com/allenai/allennlp/issues/4974 | [
"Feature request"
] | dirkgr | 14 |
scikit-image/scikit-image | computer-vision | 7,607 | Missing scikit-image nightly wheels on MacOS | We installed the nightlies in Dask CI (usually from here: https://pypi.anaconda.org/scientific-python-nightly-wheels/simple), but we are getting errors that the wheels can't be found there anymore (and seem to be gone), is this intended and if yes is there another source? | closed | 2024-11-13T15:40:32Z | 2024-11-15T19:03:19Z | https://github.com/scikit-image/scikit-image/issues/7607 | [
":warning: Priority",
":robot: type: Infrastructure"
] | phofl | 6 |
huggingface/datasets | computer-vision | 7,193 | Support of num_workers (multiprocessing) in map for IterableDataset | ### Feature request
Currently, IterableDataset doesn't support setting num_worker in .map(), which results in slow processing here. Could we add support for it? As .map() can be run in the batch fashion (e.g., batch_size is default to 1000 in datasets), it seems to be doable for IterableDataset as the regular Dataset.
### Motivation
Improving data processing efficiency
### Your contribution
Testing | open | 2024-10-02T18:34:04Z | 2024-10-03T09:54:15Z | https://github.com/huggingface/datasets/issues/7193 | [
"enhancement"
] | getao | 1 |
schemathesis/schemathesis | pytest | 2,409 | [BUG] Response validation fails when using authentication and custom session | ### Checklist
- [x] I checked the [FAQ section](https://schemathesis.readthedocs.io/en/stable/faq.html#frequently-asked-questions) of the documentation
- [x] I looked for similar issues in the [issue tracker](https://github.com/schemathesis/schemathesis/issues)
- [x] I am using the latest version of Schemathesis
### Describe the bug
Validating responses fails when an endpoint specifies a `security` parameter and a custom test client is passed as the `session` argument to `case.call_and_validate()`, such as in [this example from the documentation](https://schemathesis.readthedocs.io/en/stable/auth.html#custom-test-client-in-python-tests).
### To Reproduce
🚨 **Mandatory** 🚨: Steps to reproduce the behavior:
Running this test example with `pytest` should reproduce the issue:
```python
from typing import Annotated
from fastapi import FastAPI, Security
from fastapi.security import APIKeyHeader
from starlette_testclient import TestClient
import schemathesis
app = FastAPI()
@app.get("/", responses={200: {"model": {}}, 403: {"model": {}}})
def root(api_key: Annotated[str, Security(APIKeyHeader(name="x-api-key"))]):
return {"message": "Hello world"}
schemathesis.experimental.OPEN_API_3_1.enable()
schema = schemathesis.from_asgi("/openapi.json", app)
@schema.parametrize()
def test_api(case):
client = TestClient(app)
case.call_and_validate(session=client)
```
This gives me the following error:
```
FAILED tests/test_schema.py::test_api[GET /] - requests.exceptions.ConnectionError: HTTPConnectionPool(host='localhost', port=80): Max retries exceeded with url: / (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7c2e41ffa6...
```
This is the OpenAPI schema:
```json
{
"openapi": "3.1.0",
"info": {
"title": "FastAPI",
"version": "0.1.0"
},
"paths": {
"/": {
"get": {
"summary": "Root",
"operationId": "root__get",
"responses": {
"200": {
"description": "Successful Response",
"content": {
"application/json": {
"schema": {}
}
}
},
"403": {
"description": "Forbidden"
}
},
"security": [
{
"APIKeyHeader": []
}
]
}
}
},
"components": {
"securitySchemes": {
"APIKeyHeader": {
"type": "apiKey",
"in": "header",
"name": "x-api-key"
}
}
}
}
```
### Expected behavior
The test given above should pass
### Environment
```
- OS: Linux
- Python version: 3.12
- Schemathesis version: 3.34.1
- Spec version: 3.1.0
```
### Additional context
I believe the issue is in this check: https://github.com/schemathesis/schemathesis/blob/master/src/schemathesis/specs/openapi/checks.py#L351 The check is creating a `requests.Session` which will lead to an actual HTTP call. It should be using the specified session instead.
| closed | 2024-08-20T08:02:45Z | 2024-08-20T19:39:17Z | https://github.com/schemathesis/schemathesis/issues/2409 | [
"Priority: High",
"Type: Bug"
] | flacerdk | 2 |
errbotio/errbot | automation | 946 | Command & response translations/Internationalization | ### I am...
* [ ] Reporting a bug
* [X] Suggesting a new feature
* [ ] Requesting help with running my bot
* [ ] Requesting help writing plugins
* [ ] Here about something else
### I am running...
* Errbot version: 4.3.5
* OS version: Fedora 25
* Python version: 3.5.2
* Using a virtual environment: yes
### Suggested Feature...
Could we make some kind of decorator that allow us to setup a command alias or translation?
In multi language environment can be useful if we can simplify add some kind of option to translate the command names.
Additionally could be useful the option of make the output of a command translatable based on the User or MUC preference.
Or maybe a initially simple approach that let us run multiple instances of the same bot in different languages. I think that this is not-so-hard(tm) to be done for the responses part, but still requires some work in the command name and params translations. | closed | 2017-01-19T23:09:54Z | 2019-01-05T17:07:54Z | https://github.com/errbotio/errbot/issues/946 | [] | qlixed | 1 |
deezer/spleeter | deep-learning | 600 | Cannot find reference 'get_default_audio_adapter' in 'adapter.py' | <!-- Please respect the title [Discussion] tag. -->
from spleeter.audio.adapter import get_default_audio_adapter but my IDE reported that the reference cannot be found | closed | 2021-03-30T01:03:18Z | 2021-04-02T13:42:46Z | https://github.com/deezer/spleeter/issues/600 | [
"question"
] | cong-x-p | 1 |
nvbn/thefuck | python | 1,318 | _ | _ | closed | 2022-07-05T08:42:36Z | 2022-07-05T14:56:57Z | https://github.com/nvbn/thefuck/issues/1318 | [] | 3Rafal | 0 |
christabor/flask_jsondash | plotly | 9 | Need delete widget option | Doesn't exist.
| closed | 2016-05-03T20:28:19Z | 2016-05-03T21:45:29Z | https://github.com/christabor/flask_jsondash/issues/9 | [] | christabor | 0 |
inducer/pudb | pytest | 48 | Screen flashes at each step | This has always been an issue for me. Each time I step through a line of code, the screen flashes. If the line takes a long time to complete, all I can see is my original prompt (bash or IPython or whatever). This is annoying when stepping through code, because the flashing is disorienting, and it's also annoying when executing a long line of code, because I can no longer see what is executing until it is done. This also happens a lot if there are print statements, as the statement is printed at the "original prompt".
Is this easily fixable, or is it an Urwid issue? I haven't looked at the code yet, but maybe just the screen redrawing commands need to be reorganized a little bit. Can you also reproduce this?
| open | 2012-11-12T00:48:42Z | 2024-08-12T21:49:10Z | https://github.com/inducer/pudb/issues/48 | [
"Bug"
] | asmeurer | 43 |
Evil0ctal/Douyin_TikTok_Download_API | fastapi | 564 | [BUG] 获取直播流失败,恳请老师解答一下 | <img width="783" alt="Image" src="https://github.com/user-attachments/assets/b7d4a6c3-2486-49c0-ac65-cfee3c1da4c6" /> | open | 2025-02-26T05:33:16Z | 2025-02-26T05:33:16Z | https://github.com/Evil0ctal/Douyin_TikTok_Download_API/issues/564 | [
"BUG"
] | Dilemmacpy | 0 |
CorentinJ/Real-Time-Voice-Cloning | deep-learning | 1,196 | Hybrid modeling | I have one question. Is it possible to use the Tortoise tts model together with that to enhance its results? | open | 2023-04-17T15:35:44Z | 2023-04-17T15:35:44Z | https://github.com/CorentinJ/Real-Time-Voice-Cloning/issues/1196 | [] | Tortoise17 | 0 |
Asabeneh/30-Days-Of-Python | matplotlib | 413 | loops | i cant do it:
Find the ten most spoken languages from the data
Find the 10 most populated countries in the world | closed | 2023-07-03T17:24:50Z | 2023-07-03T17:38:34Z | https://github.com/Asabeneh/30-Days-Of-Python/issues/413 | [] | Bidmajnoon | 0 |
davidsandberg/facenet | tensorflow | 1,150 | How to combine two models to a single pb file? | Dear,
My goal is adding more dataset to the pre-trained model to increase the accuracy.
1. I freezed the graph to get the 1st model named **pre-trained-model.pb**
2. Then I trained another model based on my own images followed [this topic](https://github.com/davidsandberg/facenet/wiki/Classifier-training-of-inception-resnet-v1) named **new-model.pb**
How can I merge these 2 model into one **final-model.pb**? or can I feed the 1st model to the second model to produce the final single model (pb)? | open | 2020-04-10T04:37:07Z | 2022-05-20T16:07:53Z | https://github.com/davidsandberg/facenet/issues/1150 | [] | tongvantruong | 1 |
Urinx/WeixinBot | api | 146 | 运行weixin.py报错诶 | ImportError: No module named requests_toolbelt.multipart.encoder | open | 2017-01-18T08:21:48Z | 2017-04-05T08:56:20Z | https://github.com/Urinx/WeixinBot/issues/146 | [] | zy00000 | 2 |
aidlearning/AidLearning-FrameWork | jupyter | 142 | Unfavorable language? How to change the language fully to english !? | Dear Devs, it is a pleasure to use such incredible app, but there are some things which needs to be resolved, the gui is quite good, but many settings are written in Chinese language which is unfavorable to many, pls kindly try to change the whole language system into English which will be appreciated!
Looking forward for your actions.
Thanks
| closed | 2020-12-12T07:00:25Z | 2023-02-20T04:14:26Z | https://github.com/aidlearning/AidLearning-FrameWork/issues/142 | [] | Ryan-Blade | 3 |
Nemo2011/bilibili-api | api | 207 | 【求助】有关于判断done_geetest() | **Python 版本:** 3.10.7
**模块版本:** 15.0.0
**运行环境:** Windows
---
我是一个初学者,正在尝试写一个我自己的bilibili相关的想法来锻炼自己的能力
我参考本库的文档设计了一个窗口,其他一切正常,但我不知道如何在主函数中判断done_geetest()来关闭极验的窗口随后获取验证码.
我一开始想到的就是使用while true来循环判断并根据if来break,但很显然这会导致线程阻塞并使窗口崩溃
随后我想到使用async进行异步处理,但相关的教程过于繁琐导致我找不到实际需要的部分
我也进行了一些尝试,但实际上还是阻塞了
如果有大佬愿意帮助解决不胜感激
| closed | 2023-02-19T17:23:14Z | 2023-02-20T11:02:34Z | https://github.com/Nemo2011/bilibili-api/issues/207 | [
"question"
] | yhzcake | 3 |
iterative/dvc | machine-learning | 10,653 | Dvc commit operation is too slow | # Bug Report
## Description
I use git and dvc to manage my training datasets, which consists of thousands of jsonl files.
After I modify several jsonl files, I use `dvc status && dvc commit`. `dvc status` operation is completed quickly (I know dvc will only hash files once until it gets modified. Here only several jsonl files are modified so `dvc status` operation cost little). However, `dvc commit` operation cost a lot of time.
While `dvc commit` is executing, I see lots of "Checking out xxx/xxx/xxx.jsonl" shows in the terminal, and I believe those jsonl files are not modified. Why dvc need to check out files that are not modified?
### Expected
Assume two files `a.jsonl` and `b.jsonl` are modified, I think `dvc commit` should equal to `dvc add a.jsonl b.jsonl`. However, it seems that `dvc commit` will check out all files tracked by dvc.
I expect `dvc commit` operation skip files which are not modified, so it can be completed quickly.
### Environment information
**Output of `dvc doctor`:**
```console
$ dvc doctor
DVC version: 3.55.2 (pip)
-------------------------
Platform: Python 3.9.19 on Linux-5.14.0-284.25.1.el9_2.x86_64-x86_64-with-glibc2.31
Subprojects:
dvc_data = 3.16.5
dvc_objects = 5.1.0
dvc_render = 1.0.2
dvc_task = 0.4.0
scmrepo = 3.3.7
Supports:
http (aiohttp = 3.10.5, aiohttp-retry = 2.8.3),
https (aiohttp = 3.10.5, aiohttp-retry = 2.8.3),
s3 (s3fs = 2024.6.1, boto3 = 1.35.7)
Config:
Global: /mnt/afs/jiangtan/.config/dvc
System: /etc/xdg/dvc
Cache types: hardlink, symlink
Cache directory: fuse.quarkfs_client on quarkfs_client
Caches: local
Remotes: s3
Workspace directory: fuse.quarkfs_client on quarkfs_client
Repo: dvc, git
Repo.site_cache_dir: /path/to/repo/.dvc/site_cache_dir/repo/eddf3641719990517f0cfc808ea33376
```
| open | 2024-12-17T08:15:02Z | 2024-12-22T18:52:26Z | https://github.com/iterative/dvc/issues/10653 | [
"performance",
"triage"
] | jiangtann | 10 |
Significant-Gravitas/AutoGPT | python | 8,807 | blocks and graphs don't work | ### ⚠️ Search for existing issues first ⚠️
- [X] I have searched the existing issues, and there is no existing issue for my problem
### Which Operating System are you using?
Windows
### Which version of AutoGPT are you using?
Latest Release
### What LLM Provider do you use?
Azure
### Which area covers your issue best?
Installation and setup
### What commit or version are you using?
13da8af170602005b7a51ae527c388758825ed15
### Describe your issue.
After going to the frontend page of AutoGPT, my blocks and graphs do not work
### Upload Activity Log Content
(venv) PS G:\Stable-diffusion\AutoGPT\autogpt_platform\frontend> npm run dev
> frontend@0.3.3 dev
> next dev
▲ Next.js 14.2.13
- Local: http://localhost:5151
- Environments: .env.local
- Experiments (use with caution):
· instrumentationHook
✓ Starting...
○ Compiling /instrumentation ...
✓ Compiled /instrumentation in 3.2s (1256 modules)
✓ Ready in 7.9s
✓ Compiled /src/middleware in 353ms (562 modules)
○ Compiling / ...
✓ Compiled / in 14.8s (7445 modules)
Supabase auth error: AuthSessionMissingError: Auth session missing!
at eval (webpack-internal:///(rsc)/./node_modules/@supabase/auth-js/dist/module/GoTrueClient.js:885:59)
at SupabaseAuthClient._useSession (webpack-internal:///(rsc)/./node_modules/@supabase/auth-js/dist/module/GoTrueClient.js:787:26)
at async SupabaseAuthClient._getUser (webpack-internal:///(rsc)/./node_modules/@supabase/auth-js/dist/module/GoTrueClient.js:877:20)
at async eval (webpack-internal:///(rsc)/./node_modules/@supabase/auth-js/dist/module/GoTrueClient.js:864:20)
at async eval (webpack-internal:///(rsc)/./node_modules/@supabase/auth-js/dist/module/GoTrueClient.js:732:28) {
__isAuthError: true,
status: 400,
code: undefined
}
GET / 200 in 15947ms
⚠ The "images.domains" configuration is deprecated. Please use "images.remotePatterns" configuration instead.
○ Compiling /login ...
✓ Compiled /login in 4.9s (8595 modules)
Supabase auth error: AuthSessionMissingError: Auth session missing!
at eval (webpack-internal:///(rsc)/./node_modules/@supabase/auth-js/dist/module/GoTrueClient.js:885:59)
at SupabaseAuthClient._useSession (webpack-internal:///(rsc)/./node_modules/@supabase/auth-js/dist/module/GoTrueClient.js:787:26)
at process.processTicksAndRejections (node:internal/process/task_queues:95:5)
at async SupabaseAuthClient._getUser (webpack-internal:///(rsc)/./node_modules/@supabase/auth-js/dist/module/GoTrueClient.js:877:20)
at async eval (webpack-internal:///(rsc)/./node_modules/@supabase/auth-js/dist/module/GoTrueClient.js:864:20)
at async eval (webpack-internal:///(rsc)/./node_modules/@supabase/auth-js/dist/module/GoTrueClient.js:732:28) {
__isAuthError: true,
status: 400,
code: undefined
}
GET /login?_rsc=v3pub 200 in 191ms
○ Compiling /signup ...
✓ Compiled /signup in 1627ms (8591 modules)
GET / 200 in 63ms
POST /signup 303 in 838ms
○ Compiling /build ...
✓ Compiled /build in 3.4s (9538 modules)
### Upload Error Log Content
_No response_ | closed | 2024-11-27T12:20:54Z | 2024-12-05T20:07:56Z | https://github.com/Significant-Gravitas/AutoGPT/issues/8807 | [] | amesemyta1 | 3 |
STVIR/pysot | computer-vision | 280 | 如何确定训练RPN的次数? | 请问如何确定训练RPN的次数?麻烦详细说明一下
谢谢 | closed | 2020-01-14T08:53:46Z | 2020-01-19T06:26:42Z | https://github.com/STVIR/pysot/issues/280 | [] | ghost | 1 |
huggingface/pytorch-image-models | pytorch | 1,300 | nexterov in LARS | https://github.com/rwightman/pytorch-image-models/blob/75144395739162a153dae5628320b85b5895634b/timm/optim/lars.py#L70-L73
Why set the `nesterov` as False when calling `__setstate__`? | closed | 2022-06-13T06:50:48Z | 2022-06-14T04:18:17Z | https://github.com/huggingface/pytorch-image-models/issues/1300 | [] | weigq | 3 |
3b1b/manim | python | 1,656 | GLib-GIO-WARNING!!!(Acer Predator) | Describe the error
Run for a few seconds and then Interrupt
Code and Error
Code:
manimgl example_scenes.py GraphExample
Error:
GLib-GIO-WARNING **: 17:58:47.474: Unexpectedly, UWP app `AcerIncorporated.PredatorSenseV30_3.0.3162.0_x64__48frkmn4z8aw4' (AUMId`AcerIncorporated.PredatorSenseV30_48frkmn4z8aw4!CentenialConvert') supports 1 extensions but has no verbs
Environment
OS System: win10
manim version: master
python version:python3.8
@3b1b @bhbr @TonyCrane @eulertour @Sridhar3b1b Waiting for your help. Please Guide me | open | 2021-10-19T12:33:39Z | 2021-10-20T15:38:45Z | https://github.com/3b1b/manim/issues/1656 | [] | vaibhawkhemka | 3 |
tfranzel/drf-spectacular | rest-api | 474 | SimpleJWT integration doesn't take into account the settings AUTH_HEADER_NAME | Hi, just figuring this out today because I had to use a custom header for simplejwt auth
**Describe the bug**
SimpleJwt can be configured by using a dict in the settings.py of a django project.
`SIMPLE_JWT = {
'AUTH_HEADER_NAME' : "HTTP_X_TOKEN" # translate to X-token as header.
}`
But the current implementation doesn't take this settings into account :
`class SimpleJWTScheme(OpenApiAuthenticationExtension):
target_class = 'rest_framework_simplejwt.authentication.JWTAuthentication'
name = 'jwtAuth'
def get_security_definition(self, auto_schema):
from rest_framework_simplejwt.settings import api_settings
if len(api_settings.AUTH_HEADER_TYPES) > 1:
warn(
f'OpenAPI3 can only have one "bearerFormat". JWT Settings specify '
f'{api_settings.AUTH_HEADER_TYPES}. Using the first one.'
)
return {
'type': 'http',
'scheme': 'bearer',
'bearerFormat': api_settings.AUTH_HEADER_TYPES[0],
}`
Where the return should become something I guess like
`{'type':'apiKey', 'in': 'header', 'name': api_settings.SIMPLE_JWT['AUTH_HEADER_NAME']}`
**To Reproduce**
Change simplejwt header setting.
**Expected behavior**
Authentication should scheme should follow simplejwt settings.
I may have time to make a PR if needed.
| closed | 2021-08-02T12:13:57Z | 2021-08-25T13:15:35Z | https://github.com/tfranzel/drf-spectacular/issues/474 | [
"bug",
"fix confirmation pending"
] | Smixi | 14 |
yuka-friends/Windrecorder | streamlit | 153 | bug: 录制时关机或关闭Windrecoder会导致录像文件损坏 | Shutting down or closing Windrecoder during recording might cause the video file to be damaged. | **运行环境**:Win10 Home 22H2, AMD 5800H, 录制编码器 `AMD_h265`, 录制比特率 `150kbps`
**bug描述**:录制时关机或关闭Windrecoder会导致录像文件损坏,而且OCR模块会正常更改文件名为 `xxx-OCRED.mp4`,potplayer无法打开视频文件。

个人推测是因为使用mp4封装文件导致文件损坏,可以尝试使用mkv录制,再在索引文件前封装为mp4(OBS推荐做法)

Edit:相同码率下mkv视频好像比mp4糊一些 | open | 2024-04-24T07:13:06Z | 2024-07-11T09:44:20Z | https://github.com/yuka-friends/Windrecorder/issues/153 | [
"bug",
"P0"
] | RTLiang | 1 |
man-group/arctic | pandas | 542 | Publish arctic releases via conda-forge | Conda is increasingly used in the data science community. Conda-forge provides build infrastructure for many project. Aim to get arctic built automatically for conda:
https://conda-forge.org/#about | closed | 2018-04-20T08:38:06Z | 2019-11-18T11:05:50Z | https://github.com/man-group/arctic/issues/542 | [
"hackathon",
"medium",
"build-and-release"
] | jamesblackburn | 2 |
neuml/txtai | nlp | 365 | Update transcription notebook | Update transcription notebook to incorporate latest models available in transformers added with #362 | closed | 2022-10-17T16:26:03Z | 2022-10-17T16:53:22Z | https://github.com/neuml/txtai/issues/365 | [] | davidmezzetti | 0 |
ultralytics/yolov5 | machine-learning | 13,330 | about save-txt in yolov5-seg | ### Search before asking
- [X] I have searched the YOLOv5 [issues](https://github.com/ultralytics/yolov5/issues) and [discussions](https://github.com/ultralytics/yolov5/discussions) and found no similar questions.
### Question
when activating "save-txt" in yolov5-seg.py, a txt with the coordinates of the predicted region is saved, but i found that the coordinates seem not to be in sequence, that is to say when i use fillpoly in opencv, the coordinates seem unable to form a polygon like the one of prediction. is there a way to make the coordinates in sequence?
我发现启用save-txt后保存的包含预测分割区域的txt里的坐标似乎不是按顺序的(指坐标的保存顺序不是围着分割区域的)?用opencv的fillpoly填充出来的也跟预测的区域不一样。有办法把坐标变成按顺序的吗?

### Additional
_No response_ | open | 2024-09-23T09:40:49Z | 2024-10-27T13:30:36Z | https://github.com/ultralytics/yolov5/issues/13330 | [
"question"
] | Powerfulidot | 5 |
ultralytics/ultralytics | python | 19,091 | Export to NCNN format no longer works with Ultralytics 8.3.71 and torch 2.6.0+cpu on Raspberry Pi 4 | ### Search before asking
- [x] I have searched the Ultralytics YOLO [issues](https://github.com/ultralytics/ultralytics/issues) and found no similar bug report.
### Ultralytics YOLO Component
Export
### Bug
The latest version of `ultralytics 8.3.71` in conjunction with `torch-2.6.0+cpu` causes the `yolo export model=yolo11n.pt format=ncnn` to longer work. It results in the following output, and the model does not get converted to NCNN format. If you downgrade the libraries to `ultralytics 8.3.70` and `torch-2.5.1+cpu`, it does work.
Can you please resolve this issue? Is there any way I can help?
```
yolo export model=my_model.pt format=ncnn
Ultralytics 8.3.71 Python-3.11.2 torch-2.6.0+cpu CPU (Cortex-A72)
YOLO11n summary (fused): 238 layers, 2,583,127 parameters, 0 gradients, 6.3 GFLOPs
PyTorch: starting from 'my_model.pt' with input shape (1, 3, 480, 480) BCHW and output shape(s) (1, 9, 4725) (5.2 MB)
TorchScript: starting export with torch 2.6.0+cpu...
Illegal instruction
```

Thanks to the Ultralytics team for all your great work on this library!
### Environment
(venv) evan@raspberrypi:~/yolo/yolo3 $ yolo checks
Ultralytics 8.3.71 🚀 Python-3.11.2 torch-2.6.0+cpu CPU (Cortex-A72)
Setup complete ✅ (4 CPUs, 7.6 GB RAM, 12.0/27.8 GB disk)
OS Linux-6.6.62+rpt-rpi-v8-aarch64-with-glibc2.36
Environment Linux
Python 3.11.2
Install pip
RAM 7.63 GB
Disk 12.0/27.8 GB
CPU Cortex-A72
CPU count 4
GPU None
GPU count None
CUDA None
numpy ✅ 2.1.1<=2.1.1,>=1.23.0
matplotlib ✅ 3.10.0>=3.3.0
opencv-python ✅ 4.11.0.86>=4.6.0
pillow ✅ 11.1.0>=7.1.2
pyyaml ✅ 6.0.2>=5.3.1
requests ✅ 2.32.3>=2.23.0
scipy ✅ 1.15.1>=1.4.1
torch ✅ 2.6.0>=1.8.0
torch ✅ 2.6.0!=2.4.0,>=1.8.0; sys_platform == "win32"
torchvision ✅ 0.21.0>=0.9.0
tqdm ✅ 4.67.1>=4.64.0
psutil ✅ 6.1.1
py-cpuinfo ✅ 9.0.0
pandas ✅ 2.2.3>=1.1.4
seaborn ✅ 0.13.2>=0.11.0
ultralytics-thop ✅ 2.0.14>=2.0.0
### Minimal Reproducible Example
**Steps to reproduce**
This is all run on a Raspberry Pi 4b 8GB using 64-bit Raspberry Pi OS version 12 (Bookworm). The OS is freshly installed on an SD cardusing Raspberry Pi Imager.
Open a terminal and run the following commands:
1. `sudo apt update && sudo apt upgrade -y`
2. `mkdir yolo && cd yolo`
3. `python3 -m venv venv`
4. `source venv/bin/activate`
5. `pip install ultralytics ncnn`
6. `yolo export model=yolo11n.pt format=ncnn`
The error will occur upon running that last command. The error no longer occurs if you downgrade using `pip install ultralytics==8.3.70 torch==2.5.1`.
### Additional
_No response_
### Are you willing to submit a PR?
- [ ] Yes I'd like to help by submitting a PR! | open | 2025-02-06T02:24:31Z | 2025-03-11T18:36:30Z | https://github.com/ultralytics/ultralytics/issues/19091 | [
"bug",
"embedded",
"exports"
] | EdjeElectronics | 10 |
JaidedAI/EasyOCR | pytorch | 567 | Using custom model with different input size? (rgb: True) | I've successfully trained my own model, thank you so much for the guidance there, but now when I am trying to use the model I made with `rgb: true` (making it have an input of 3 channels) I get errors for size mismatches:
```
RuntimeError: Given groups=1, weight of size [32, 3, 3, 3], expected
input[1, 1, 64, 256] to have 3 channels, but got 1 channels instead
```
This may be due to the downloaded model, or something? I played with `download_enabled` and removing the downloaded model, that gives this error:
```
FileNotFoundError: Missing ./ocr_model/craft_mlt_25k.pth and downloads disabled
```
How do I use the reader **only using** my network? or am I trying to do something that doesn't make any sense? If my training is making great predictions, I'm confused on why I need another model?
```py
reader = easyocr.Reader(
['en'],
# here's where my custom_model.pth file is
model_storage_directory="./ocr_model",
# here's where my custom_model.py and custom_model.yaml live
user_network_directory='./ocr_network',
recog_network='custom_model'
)
reader.readtext('test.jpg')
```
I believe I set `custom_model.yaml` up properly:
```yaml
network_params:
input_channel: 3
output_channel: 256
hidden_size: 256
imgH: 64
lang_list:
- 'en'
character_list: 0123456789!"#$%&'()*+,-./:;<=>?@[\]^_`{|}~ ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz
```
## EDIT
Doing the exact same training but with `rgb: False` and everything works, woo! am I missing out on much without rgb? | closed | 2021-10-13T00:31:36Z | 2022-12-22T02:26:54Z | https://github.com/JaidedAI/EasyOCR/issues/567 | [] | ckcollab | 4 |
ets-labs/python-dependency-injector | asyncio | 94 | Enhancement of catalogs inheritance | Example below should work properly:
``` python
"""Example of catalogs inheritance."""
import dependency_injector as di
class CatalogA(di.AbstractCatalog):
"""Example catalog A."""
p1 = di.Provider()
class CatalogB(CatalogA):
"""Example catalog B."""
p2 = di.Provider()
class CatalogC(CatalogB):
"""Example catalog C."""
p3 = di.Provider()
# `di.AbstractCatalog.providers` attribute is a dict of all available providers,
# including current catalog providers and providers that are inherited
# from parent catalogs:
assert CatalogA.providers == dict(p1=CatalogA.p1)
assert CatalogB.providers == dict(p1=CatalogA.p1, p2=CatalogB.p2)
assert CatalogC.providers == dict(p1=CatalogA.p1, p2=CatalogB.p2, p3=CatalogC.p3)
# `di.AbstractCatalog.inherited_providers` attribute is a dict of all providers that
# are inherited from parent catalogs
assert CatalogA.inherited_providers == dict()
assert CatalogB.inherited_providers == dict(p1=CatalogA.p1)
assert CatalogC.inherited_providers == dict(p1=CatalogA.p1, p2=CatalogB.p2)
# `di.AbstractCatalog.cls_providers` attribute is a dict of current catalog providers:
assert CatalogA.cls_providers == dict(p1=CatalogA.p1)
assert CatalogB.cls_providers == dict(p2=CatalogB.p2)
assert CatalogB.cls_providers == dict(p3=CatalogC.p3)
```
| closed | 2015-10-07T07:46:45Z | 2015-10-07T16:44:57Z | https://github.com/ets-labs/python-dependency-injector/issues/94 | [
"enhancement"
] | rmk135 | 0 |
ansible/awx | django | 15,631 | AWX Office Hours - 11/12/24 | # AWX Office Hours
## Proposed agenda based on topics
- https://github.com/ansible/awx/pull/15627
## What
After a successful Contributor Summit in October 2023, one of the bits of feedback we got was to host a regular time for the Automation Controller (AWX) Team to be available for your folks in the AWX Community, so we are happy to announce a new regular video meeting.
This kind of feedback loop is vital to the success of AWX and the AWX team wants to make it as easy as possible for you - our community - to get involved.
## Where & When
Our next meeting will be held on Tuesday, November 12th, 2024 at [1500 UTC](https://dateful.com/time-zone-converter?t=15:00&tz=UTC)
* [Google Meet](https://meet.google.com/vyk-dfow-cfi)
* Via Phone PIN: 842522378 [Guide](https://support.google.com/meet/answer/9518557)
This meeting is held once a month, on the second Tuesday of the month, at [1500 UTC](https://dateful.com/time-zone-converter?t=15:00&tz=UTC)
## How
Add one topic per comment in this GitHub issue
If you don't have a GitHub account, jump on [#awx:ansible.com](https://matrix.to/#/#awx:ansible.com) on Matrix and we can add the topic for you
## Talk with us
As well as the fortnightly video meeting you can join the Community (inc development team) on Matrix Chat.
* Matrix: [#awx:ansible.com](https://matrix.to/#/#awx:ansible.com) (recomended)
* libera.chat IRC: `#ansible-awx` (If you are already setup on IRC)
The Matrix & IRC channels are bridged, you'll just have a better experience on Matrix
## Links
[AWX YouTube Chanel](https://www.youtube.com/@ansible-awx)
[Previous Meeting](#15319)
[Meeting recording]()
Next Meeting
See you soon! | closed | 2024-11-12T15:24:17Z | 2024-12-10T15:46:14Z | https://github.com/ansible/awx/issues/15631 | [
"needs_triage",
"community"
] | thedoubl3j | 0 |
python-gino/gino | asyncio | 750 | [Suggestion] Docs Translation | gino would be useful project (IMO)
I have a great interest in gino project.
If I has willing to translate Korean, Can You accept my Pull request?
P.S. Happy New Year | closed | 2021-01-02T13:34:09Z | 2021-01-10T21:00:05Z | https://github.com/python-gino/gino/issues/750 | [
"feature request"
] | KimSoungRyoul | 1 |
adbar/trafilatura | web-scraping | 584 | Removing related links at end of article/sidebar on news websites? | Over here in the Media Cloud project we're seeing poor performance on the content extraction task for a variety of pages that include links to other "related" stories at the end of article content. Our use case is trying to extract only article content as text. Do you have advice on tweaks to make to improve that performance? This might be the opposite of #518, because we do _not_ want related links as part of content.
Here's sample code with real examples parsed in a way that looks very similar to our usage. The function returns true if the supplied text is included in the extracted content (the erroneous results, in our use case). Each of these incorrectly includes text that is part of a "related links" type callout that appears _after_ article content. Any advice appreciated.
```python
import trafilatura
import requests
MEDIA_CLOUD_USER_AGENT = 'Mozilla/5.0 (compatible; mediacloud academic archive; mediacloud.org)'
def is_text_in_webpage_content(txt:str, url:str) -> bool:
req = requests.get(url, headers={'User-Agent': MEDIA_CLOUD_USER_AGENT},timeout=30)
parsed = trafilatura.bare_extraction(req.text, only_with_metadata=False, url=url,
include_images=False, include_comments=False)
content_text = parsed['text']
return txt in content_text
print(is_text_in_webpage_content(
'Thai Official', # item on bottom of page in "Latest News" section
'https://www.ibtimes.co.uk/falling-inflation-shifts-focus-when-ecb-could-cut-rates-1722106'))
print(is_text_in_webpage_content(
'HIV from Terrence Higgins to Today', # <li> under the "listen on sounds" banner after article
'https://www.bbc.co.uk/sport/football/67640638'))
print(is_text_in_webpage_content(
'Madhuri Dixit', # title of an item in the featured movie below the main content area
'https://timesofindia.indiatimes.com/videos/lifestyle/fashion/10-indian-saris-every-woman-should-have-in-her-wardrobe/videoshow/105809845.cms'))
print(is_text_in_webpage_content(
'Immigration, Ukraine', # title of an item in the "most popular" sidebar content
'https://www.bfmtv.com/cote-d-azur/nice-25-personnes-expulsees-lors-d-operations-anti-squat-menees-dans-le-quartier-des-liserons_AN-202312150639.html'))
``` | open | 2024-05-03T17:19:12Z | 2024-07-17T17:37:18Z | https://github.com/adbar/trafilatura/issues/584 | [
"bug"
] | rahulbot | 3 |
plotly/dash-core-components | dash | 949 | animation_options in Graph has incorrect default value ("ease" instead of "easing") | The defaultProps for `animation_options` are `{ frame: { redraw: false, }, transition: { duration: 750, ease: 'cubic-in-out', }, }`, but `Plotly.animate` takes an "easing" argument, not "ease".
I don't see any warning in the console for this, so I don't think these arguments are being validated; I can put whatever I like in `animation_options` and never get warnings/errors.
As an aside, it would be helpful if the `dash_core_components` [docs](https://dash.plotly.com/dash-core-components/graph) mentioned that `frame.duration` has to be set at least as long as `transition.duration`, or at least linked to [https://plotly.com/javascript/animations/](https://plotly.com/javascript/animations/), as it's not immediately clear that you can't just arbitrarily set `transition.duration` to higher values. In fact the default `frame.duration` is 500, so the 750 default value here is misleading (maybe just setting `frame.duration` to 750 in the default here would at least highlight to users that it needs to be set). | open | 2021-04-08T15:23:40Z | 2021-04-08T15:23:40Z | https://github.com/plotly/dash-core-components/issues/949 | [] | slishak | 0 |
dmlc/gluon-cv | computer-vision | 1,014 | When training yolov3 with coco, how to merge dog and cat into animal categories? | When training yolov3 with coco, how to merge dog and cat into animal categories? Thx. | closed | 2019-10-30T03:11:20Z | 2021-06-07T07:04:25Z | https://github.com/dmlc/gluon-cv/issues/1014 | [
"Stale"
] | YeShangyuan | 5 |
django-import-export/django-import-export | django | 1,441 | How to add error message when `UNIQUE constraint failed` | when import twice get a long `Traceback (most recent call last):`
```py
class Employee(models.Model):
id_number = models.CharField(max_length=180, null=True, blank=True, unique=True)
```
I want to show the error like admin default error message, like `already exist Employee with this _id_number ` | closed | 2022-05-27T01:15:54Z | 2023-04-12T13:02:24Z | https://github.com/django-import-export/django-import-export/issues/1441 | [
"question"
] | wgf4242 | 4 |
xuebinqin/U-2-Net | computer-vision | 225 | Can .pth change to pytorch .pt file | Hi, may I know what kind of models you train in torch? I would like to change the .pth to pytorch .pt file, since .pth is in conflict with python directory, and I cannot open in android mobile.
Thanks. | open | 2021-07-03T03:51:52Z | 2021-07-03T10:22:04Z | https://github.com/xuebinqin/U-2-Net/issues/225 | [] | Josonlchui | 1 |
sherlock-project/sherlock | python | 1,638 | Issues while runinng | here's the output
pi@raspberrypi:~/sherlock $ python3 sherlock
Traceback (most recent call last):
File "/usr/lib/python3.9/runpy.py", line 197, in _run_module_as_main
return _run_code(code, main_globals, None,
File "/usr/lib/python3.9/runpy.py", line 87, in _run_code
exec(code, run_globals)
File "/home/pi/sherlock/sherlock/__main__.py", line 21, in <module>
import sherlock
File "/home/pi/sherlock/sherlock/sherlock.py", line 12, in <module>
import pandas as pd
File "/home/pi/.local/lib/python3.9/site-packages/pandas/__init__.py", line 16, in <module>
raise ImportError(
ImportError: Unable to import required dependencies:
numpy:
IMPORTANT: PLEASE READ THIS FOR ADVICE ON HOW TO SOLVE THIS ISSUE!
Importing the numpy C-extensions failed. This error can happen for
many reasons, often due to issues with your setup or how NumPy was
installed.
We have compiled some common reasons and troubleshooting tips at:
https://numpy.org/devdocs/user/troubleshooting-importerror.html
Please note and check the following:
* The Python version is: Python3.9 from "/usr/bin/python3"
* The NumPy version is: "1.23.5"
and make sure that they are the versions you expect.
Please carefully study the documentation linked above for further help.
Original error was: libcblas.so.3: cannot open shared object file: No such file
or directory
please lmk how to solve it | closed | 2022-12-14T22:25:41Z | 2023-02-04T17:21:21Z | https://github.com/sherlock-project/sherlock/issues/1638 | [] | Micychalek | 1 |
davidsandberg/facenet | tensorflow | 304 | Why slim.flatten(net) before slim.dropout(net)? | when I read the model code, I found there is a layer slim.flatten(net) before net = slim.dropout(net, dropout_keep_prob, is_training=is_training, scope='Dropout'), it's different with paper.
sorry, I'm new to this area, thank you very much | closed | 2017-06-02T07:35:44Z | 2017-06-06T01:46:59Z | https://github.com/davidsandberg/facenet/issues/304 | [] | chencjiajy | 1 |
pytest-dev/pytest-html | pytest | 295 | [Request]: Can we please add the support to extend the pytest-html to pytest-bdd framework? | This plugin is absolutely working for pytest-bdd framework too. I see the passed/failed/skipped etc. However, I am wondering if we can extend support of this plugin for pytest-bdd framework? As we can see the report with the Feature as well for Feature/Gherkin steps. ?? Any suggestions/thoughts are highly appreciated. | open | 2020-04-24T22:21:43Z | 2020-10-22T22:41:20Z | https://github.com/pytest-dev/pytest-html/issues/295 | [
"enhancement",
"feature"
] | CuriousQA | 6 |
NVlabs/neuralangelo | computer-vision | 163 | The second-order analytical gradients are not all zero as described in the article. | I printed the value of the value of the second derivative, and found that the second derivative is not zero. I understand that theoretically the second derivative of trilinear interpolation should be 0, but why are the results of the code implementation inconsistent?
`gradient = torch.autograd.grad(sdf.sum(), x, create_graph=True)[0]`
`hessian = torch.autograd.grad(gradient.sum(), x, create_graph=True)[0]` | open | 2023-12-01T07:31:46Z | 2023-12-01T07:32:02Z | https://github.com/NVlabs/neuralangelo/issues/163 | [] | ckrsls | 0 |
zihangdai/xlnet | nlp | 24 | Python package | Would you be opposed to turning this into a python package? Something like this: https://github.com/monkeyhippies/xlnet/commit/f0471a242ed5dad5c4be7602b17dd0a96ad6b671
Or is it meant to just be a repo of helpful scripts? | closed | 2019-06-22T07:59:54Z | 2019-06-22T08:20:03Z | https://github.com/zihangdai/xlnet/issues/24 | [] | monkeyhippies | 0 |
keras-team/autokeras | tensorflow | 1,641 | TimeSeriesForecaster: Predict function returns ValueError | Hi,
I've been using AutoKeras for Time Series forecasting but when the model has been trained and I apply the test data, it raises: "ValueError: The prediction data requires the original training data to make predictions on subsequent data points"
but I can't put both the test data and the training data as separate arguments? Do I need to append the test data to the training data for the model to predict or is there something I'm missing?
Setup Details
- OS type and version: Windows 10
- Python: 3.9
- autokeras: 1.0.16
- keras-tuner: 1.0.4
- scikit-learn: 1.0
- numpy: 1.19.5
- pandas: 1.3.4
- tensorflow: 2.5.0 (GPU)
| open | 2021-10-23T17:37:02Z | 2023-03-01T20:25:07Z | https://github.com/keras-team/autokeras/issues/1641 | [] | sparkey63 | 5 |
activeloopai/deeplake | data-science | 2,685 | [BUG] Can NOT run deeplake python library | ### Severity
None
### Current Behavior
Can not run deeplake python library, I already tried to have fresh deployment on different version of python, langchain and deeplake.
### Steps to Reproduce
- Use conda to create and activate new virutal environment
- install all python library,
> pip`` install langchain==0.0.208 deeplake openai tiktoken
> or pip`` install langchain deeplake openai tiktoken
> or pip`` install langchain==0.0.208 deeplake==3.8.2 openai tiktoken
- try import deeplake
all failed with following code
`Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/Users/jamesou/Documents/Projects/langchain/Learning/deeplake.py", line 28, in <module>
db = DeepLake.from_documents(docs, dataset_path=dataset_path, embedding=OpenAIEmbeddings())
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/jamesou/miniforge3/envs/deep/lib/python3.11/site-packages/langchain/vectorstores/base.py", line 317, in from_documents
return cls.from_texts(texts, embedding, metadatas=metadatas, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/Users/jamesou/miniforge3/envs/deep/lib/python3.11/site-packages/langchain/vectorstores/deeplake.py", line 736, in from_texts
deeplake_dataset = cls(
^^^^
File "/Users/jamesou/miniforge3/envs/deep/lib/python3.11/site-packages/langchain/vectorstores/deeplake.py", line 123, in __init__
raise ValueError(
ValueError: Could not import deeplake python package. Please install it with `pip install deeplake`.`
### Expected/Desired Behavior
it should be able to run successfully
### Python Version
3.11.1 or 3.11.6
### OS
OSX 13.6
### IDE
VS Code
### Packages
_No response_
### Additional Context
_No response_
### Possible Solution
_No response_
### Are you willing to submit a PR?
- [ ] I'm willing to submit a PR (Thank you!) | closed | 2023-11-05T09:15:51Z | 2023-11-29T23:49:01Z | https://github.com/activeloopai/deeplake/issues/2685 | [
"bug"
] | jamesoujj | 3 |
PokemonGoF/PokemonGo-Bot | automation | 5,931 | Looking for a quick answer? Not sure if you have found something new ? Check our Discord | [https://discord.gg/n3g5puF](https://discord.gg/n3g5puF)
:)
| open | 2017-02-22T20:13:45Z | 2018-06-11T16:01:03Z | https://github.com/PokemonGoF/PokemonGo-Bot/issues/5931 | [
"Enhancement",
"Approved"
] | pogarek | 2 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.