repo_name
stringlengths
9
75
topic
stringclasses
30 values
issue_number
int64
1
203k
title
stringlengths
1
976
body
stringlengths
0
254k
state
stringclasses
2 values
created_at
stringlengths
20
20
updated_at
stringlengths
20
20
url
stringlengths
38
105
labels
sequencelengths
0
9
user_login
stringlengths
1
39
comments_count
int64
0
452
deeppavlov/DeepPavlov
tensorflow
1,373
Models: Migrate NER to PyTorch
closed
2021-01-12T11:21:58Z
2021-11-30T10:07:52Z
https://github.com/deeppavlov/DeepPavlov/issues/1373
[]
danielkornev
0
albumentations-team/albumentations
machine-learning
2,090
[New transform] Add RandomContrast
Add RandomContrast that has the same API as Kornia's RandomContrast https://kornia.readthedocs.io/en/latest/augmentation.module.html#kornia.augmentation.RandomContrast
closed
2024-11-08T15:43:17Z
2024-11-16T22:57:06Z
https://github.com/albumentations-team/albumentations/issues/2090
[ "enhancement" ]
ternaus
1
praw-dev/praw
api
1,825
requests.exceptions.HTTPError not handled by praw
### Describe the Bug When I was using `submit_image`, praw encountered a 500 internal server error. I have a handler for `praw.exceptions.RedditAPIException`, but that did not get caught because a `requests.exceptions.HTTPError` was raised instead of a `praw.exceptions.RedditAPIException`. I assume that the intent is to have praw except exceptions from "requests" and raise them as praw exceptions (to avoid clients depending on knowledge of internal design). ``` $ python3 -m pip freeze | grep -e requests -e praw praw==7.5.0 prawcore==2.3.0 requests==2.26.0 requests-file==1.4.3 requests-oauthlib==1.3.0 requests-toolbelt==0.9.1 requests-unixsocket==0.2.0 ``` ### Desired Result praw should have raised a `praw.exceptions.RedditAPIException` (or one of the classes that inherit from that) ### Relevant Logs ```shell File "/home/rolf/weather/redditSend.py", line 160, in postImage submission = sub.submit_image( File "/home/rolf/.local/lib/python3.8/site-packages/praw/models/reddit/subreddit.py", line 1188, in submit_image "title": title, File "/home/rolf/.local/lib/python3.8/site-packages/praw/models/reddit/subreddit.py", line 701, in _upload_media response.raise_for_status() File "/home/rolf/.local/lib/python3.8/site-packages/requests/models.py", line 953, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: https://reddit-uploaded-media.s3-accelerate.amazonaws.com/ ``` ### Code to reproduce the bug ```python # I don't have a reproduction because in requires reddit to send back a 500. ``` ### My code example does not include the `Reddit()` initialization to prevent credential leakage. Yes ### This code has previously worked as intended. No ### Operating System/Environment Linux Mint 20.2 Cinnamon ### Python Version Python 3.8.10 ### PRAW Version 7.5.0 ### Prawcore Version 2.3.0 ### Anything else? _No response_
open
2021-12-12T00:49:40Z
2025-01-10T10:15:23Z
https://github.com/praw-dev/praw/issues/1825
[ "Bug" ]
endlisnis
15
torchbox/wagtail-grapple
graphql
398
Question: How to refer to type created by decorator (e.g. register_query_field)
Hello there. I'm wondering how to refer to a (GraphQL) type that was created by one of the decorators, e.g. by `register_query_field`. **Example** ```python @register_query_field("my_page") class MyPage(Page): pass ``` How do I e.g. set up a `QuerySetList` field on another model that references `MyPage`? This used to work: ```python my_pages = QuerySetList(graphene.NonNull(registry.pages.get(MyPage))) ``` However, since updating from `0.19.2` to `0.26.0`, `registry.pages` is empty when this runs. And it always felt a little weird, having to consult the registry like this. Is there a better way? Thank you!
open
2024-07-10T22:16:26Z
2024-07-10T22:32:42Z
https://github.com/torchbox/wagtail-grapple/issues/398
[]
PeterDekkers
0
aio-libs/aiohttp
asyncio
10,155
9 test failures
### Describe the bug Running the test suite of aiohttp 3.11.10, I see a few test failures as well as errors, for reasons unclear. ### To Reproduce Run the test suite with `pytest` in an environment having Python 3.10.7 and the following direct dependencies: ``` gunicorn@21.2.0 python-aiodns@3.2.0 python-aiohappyeyeballs@2.4.4 python-aiosignal@1.3.1 + python-async-generator@1.10 python-async-timeout@4.0.3 python-brotli@1.0.9 python-coverage@7.6.9 + python-cython@3.0.11 python-freezegun@1.2.2 python-frozenlist@1.5.0 python-multidict@5.2.0 + python-mypy-minimal@1.13.0 python-on-whales@0.74.0 python-propcache@0.2.1 python-pytest-codspeed@3.0.0 + python-pytest-cov@6.0.0 python-pytest-cython@0.3.1 python-pytest-mock@3.14.0 python-pytest-xdist@3.6.1 + python-pytest@8.3.3 python-re-assert@1.1.0 python-trustme@0.9.0 python-uvloop-bootstrap@0.21.0 + python-yarl@1.18.3 ``` ### Expected behavior All tests pass. ### Logs/tracebacks ```python-traceback TOTAL 37521 8794 77% ============================= slowest 10 durations ============================= 9.51s call tests/test_run_app.py::TestShutdown::test_shutdown_new_conn_rejected 4.01s call tests/test_run_app.py::TestShutdown::test_shutdown_pending_handler_responds 2.51s call tests/test_run_app.py::TestShutdown::test_shutdown_wait_for_handler 2.41s call tests/test_run_app.py::TestShutdown::test_shutdown_handler_cancellation_suppressed 2.01s call tests/test_client_functional.py::test_read_timeout_between_chunks[pyloop] 2.00s call tests/test_client_functional.py::test_set_cookies_max_age[pyloop] 1.61s call tests/test_run_app.py::TestShutdown::test_shutdown_timeout_handler 1.51s call tests/test_run_app.py::TestShutdown::test_shutdown_timeout_not_reached 1.06s call tests/test_benchmarks_web_urldispatcher.py::test_resolve_multiple_level_fixed_url_with_many_routes[pyloop] 1.01s call tests/test_client_functional.py::test_readline_error_on_conn_close[pyloop] =========================== short test summary info ============================ SKIPPED [1] tests/test_client_session.py:951: The check is applied in DEBUG mode only SKIPPED [1] tests/test_client_request.py:1207: Needs Task.cancelling() SKIPPED [1] tests/test_helpers.py:324: Python 3.11+ is required for .cancelling() SKIPPED [1] tests/test_helpers.py:343: Python 3.11+ is required for .cancelling() SKIPPED [1] tests/test_connector.py:3320: Proactor Event loop present only in Windows SKIPPED [1] tests/test_connector.py:3103: Proactor Event loop present only in Windows SKIPPED [1] tests/test_connector.py:3111: Proactor Event loop present only in Windows SKIPPED [1] tests/test_connector.py:3122: Proactor Event loop present only in Windows SKIPPED [1] tests/test_multipart_helpers.py:446: should raise decoding error: %82 is invalid for latin1 SKIPPED [1] tests/test_multipart_helpers.py:455: should raise decoding error: %E4 is invalid for utf-8 SKIPPED [1] tests/test_multipart_helpers.py:510: urllib.parse.unquote is tolerate to standalone % chars SKIPPED [1] tests/test_multipart_helpers.py:519: urllib.parse.unquote is tolerate to standalone % chars SKIPPED [1] tests/test_multipart_helpers.py:99: need more smart parser which respects quoted text SKIPPED [1] tests/test_urldispatch.py:1021: aiohttp folder is not placed in user's HOME SKIPPED [1] tests/test_web_app.py:365: The check is applied in DEBUG mode only SKIPPED [1] tests/test_resolver.py:404: aiodns <3.2.0 required SKIPPED [1] tests/test_web_functional.py:192: Needs Task.cancelling() SKIPPED [1] tests/test_web_runner.py:133: Proactor Event loop present only in Windows SKIPPED [1] tests/test_web_runner.py:121: Proactor Event loop present only in Windows XFAIL tests/autobahn/test_autobahn.py::test_client XFAIL tests/autobahn/test_autobahn.py::test_server XFAIL tests/test_connector.py::test_del_with_scheduled_cleanup[pyloop] XFAIL tests/test_client_functional.py::test_broken_connection[pyloop] XFAIL tests/test_http_parser.py::test_parse_unusual_request_line[c-parser-pyloop] - Regression test for Py parser. May match C behaviour later. XFAIL tests/test_http_parser.py::test_http_request_parser_utf8_request_line[c-parser-pyloop] - Regression test for Py parser. May match C behaviour later. XFAIL tests/test_http_parser.py::test_parse_uri_utf8[c-parser-pyloop] - reason: Not valid HTTP. Maybe update py-parser to reject later. FAILED tests/test_client_request.py::test_default_loop[pyloop] - DeprecationWarning: There is no current event loop FAILED tests/test_imports.py::test___all__ - ValueError: Pytest terminal summary report not found FAILED tests/test_imports.py::test_web___all__ - ValueError: Pytest terminal summary report not found FAILED tests/test_circular_imports.py::test_no_warnings[aiohttp._http_parser] - subprocess.CalledProcessError: Command '('/gnu/store/jjzbav1s0x8lqg1aal13irspwwblwdf4-python-wrapper-3.10.7/bin/python', '-W', 'error', '-W', "ignore:module 'sre_constants' is deprecated:DeprecationWarning:pkg_resources._vendor.pyparsing", '-W', 'ignore:Creating a LegacyVersion has been deprecated and will be removed in the next major release:DeprecationWarning:', '-W', 'ignore:pkg_resources is deprecated as an API:DeprecationWarning', '-c', 'import aiohttp._http_parser')' returned non-zero exit status 1. FAILED tests/test_circular_imports.py::test_no_warnings[aiohttp._http_writer] - subprocess.CalledProcessError: Command '('/gnu/store/jjzbav1s0x8lqg1aal13irspwwblwdf4-python-wrapper-3.10.7/bin/python', '-W', 'error', '-W', "ignore:module 'sre_constants' is deprecated:DeprecationWarning:pkg_resources._vendor.pyparsing", '-W', 'ignore:Creating a LegacyVersion has been deprecated and will be removed in the next major release:DeprecationWarning:', '-W', 'ignore:pkg_resources is deprecated as an API:DeprecationWarning', '-c', 'import aiohttp._http_writer')' returned non-zero exit status 1. FAILED tests/test_test_utils.py::test_testcase_no_app - Failed: remains unmatched: '*RuntimeError*' FAILED tests/test_circular_imports.py::test_no_warnings[aiohttp._websocket.mask] - subprocess.CalledProcessError: Command '('/gnu/store/jjzbav1s0x8lqg1aal13irspwwblwdf4-python-wrapper-3.10.7/bin/python', '-W', 'error', '-W', "ignore:module 'sre_constants' is deprecated:DeprecationWarning:pkg_resources._vendor.pyparsing", '-W', 'ignore:Creating a LegacyVersion has been deprecated and will be removed in the next major release:DeprecationWarning:', '-W', 'ignore:pkg_resources is deprecated as an API:DeprecationWarning', '-c', 'import aiohttp._websocket.mask')' returned non-zero exit status 1. FAILED tests/test_web_app.py::test_set_loop_default_loop - DeprecationWarning: There is no current event loop FAILED tests/test_streams.py::TestStreamReader::test_ctor_global_loop - DeprecationWarning: There is no current event loop ============ 9 failed, 3400 passed, 19 skipped, 7 xfailed in 41.68s ============ ``` ### Python Version ```console $ python --version 3.10.7 ``` ### aiohttp Version ```console $ python -m pip show aiohttp 3.11.10 ``` ### multidict Version ```console $ python -m pip show multidict 5.2.0 ``` ### propcache Version ```console $ python -m pip show propcache 0.2.1 ``` ### yarl Version ```console $ python -m pip show yarl 1.18.3 ``` ### OS GNU+Linux ### Related component Client ### Additional context Found while updating the `python-aiohttp` package on GNU Guix. ### Code of Conduct - [X] I agree to follow the aio-libs Code of Conduct
closed
2024-12-11T06:26:14Z
2025-02-28T01:01:42Z
https://github.com/aio-libs/aiohttp/issues/10155
[ "bug", "needs-info", "Stale" ]
Apteryks
4
donnemartin/system-design-primer
python
811
How much to creat a software to collect email, account on google, facebook, etc
We need to develop a software based on Google search engine and social media channels to collect visible email, account, and other information on web pages. Google search and map search are required to collect enough data, which is accurate and fast. Telephone and email collection Use keywords and various search instructions to collect as many relevant emails and accounts as possible on search engines and social media platforms, and freely select regions based on national cities 1. Search on Google search engine 2. On Facebook, INSTAGRAM, TWITTER, Relevant emails visible on social media platforms such as TIKTOK, LINKEDIN, YOUTUBE, PINTEREST, etc 3. Search for email and Facebook collection based on country and city on Google Maps 1. Collect the accounts and personal homepage of relevant members based on keywords, Company homepage 2. Collect the accounts of members in the live streaming room based on keywords, as personal owners Page, avatar 3. Collect the homepage and accounts of likes members based on the homepage link Number, avatar 4. Collect fan accounts, homepage, and avatars based on homepage links 5. Collect group members' accounts and personal homepage based on keywords, Avatar Instagram collection 1. Collect the accounts, homepage, and avatars of relevant members based on keywords 2. Collect fan accounts, homepage, and avatars from the user's homepage 3. Collect likes members' accounts, homepage, and avatars based on the user's homepage. WhatsApp group collection 1. Search through various search engines and social media channels based on keywords WS group 2. One click collection of Whatsapp for group members
open
2023-11-15T12:24:33Z
2024-06-23T17:39:50Z
https://github.com/donnemartin/system-design-primer/issues/811
[ "needs-review" ]
2023Andrew2024
1
pydata/pandas-datareader
pandas
15
COMPAT with 0.16.0
if someone would like to put all of the accumulated changes since @hayd pulled this from master as a PR back into pandas would be appreciated. I think a good idea would be to have `pandas-datareader`, the 0.1 (or whatever version), be released about the same time as 0.16.0 (target at end of feb). We'll put up an announcement in the pandas release about this, and about future availability..
closed
2015-01-25T22:22:03Z
2015-10-10T18:07:02Z
https://github.com/pydata/pandas-datareader/issues/15
[]
jreback
12
scrapy/scrapy
python
6,692
Slightly optimizing component fetch in Crawler class
## Summary I was reading scrapy code and found that method get_extension can be slightly optimized: https://github.com/scrapy/scrapy/blob/8c34e6d9a4994abda0059f76dbbdb64c2e8a9751/scrapy/crawler.py#L212C1-L229C1 If we can cache those lookup results so that if you call get_extension multiple times, you don’t re‑iterate over the list each time. **Raw code:** ``` class Crawler: def __init__(self, spidercls, settings=None, init_reactor=False): # ... existing code ... self._component_cache = {} # new cache @staticmethod def _get_component_from_list(component_class, components): for component in components: if isinstance(component, component_class): return component return None def get_extension(self, cls): if cls in self._component_cache: return self._component_cache[cls] if not self.extensions: raise RuntimeError( "Crawler.get_extension() can only be called after the " "extension manager has been created." ) result = self._get_component_from_list(cls, self.extensions.middlewares) # Cache the result (even if None) for subsequent lookups self._component_cache[cls] = result return result ```
open
2025-02-25T07:19:23Z
2025-02-25T10:25:39Z
https://github.com/scrapy/scrapy/issues/6692
[ "enhancement", "performance" ]
shivanshuzyte
1
koaning/scikit-lego
scikit-learn
471
[FEATURE] Time Series Meta Regressor
Heya! How about some meta model for time series? It can consist of two models, such as in the ZeroInflatedRegressor: One model to capture trend (and optionally seasonality) of the time series, another one for the rest. It's basically a special kind of decomposition model, similar to Prophet, but more general, simpler, and scikit-learn compatible. In the end you have prediction = trend (e.g. via linear regression) + seasonality (e.g. via RepeatingBasisFunction) + second_model output. Training procedure: (Input X, y) 1. first_model.fit(X, y) 2. Detrend (and deseasonalize) the training data via y' = y - first_model.predict(X). Some kind of stationary y' remains. 3. second_model.fit(X, y'). Prediction procedure: 1. Output first_model.predict(X) + second_model.predict(X). It's simple and similar to ZeroInflatedRegressor and works well together with your RepeatedBasisFunction. And I think it belongs here in contrast to sktime because sktime does not focus on exogenous regressors that much (do they even at all?). With this model, you could also use KNN tree-based approaches for time series forecasting without worrying about trends that the trees can't capture alone. In my opinion, modelling the trend directly is also better than differentiating y and then train a tree-based method with these labels because prediction errors propagate and increase into the future. Also, it's interesting also compared to Prophet because you can customize it much more. For example, there might be exogenous variables that influence the outcome positively (e.g. advertising spends), but there is no way to specify this in Prophet, as far as I know. What do you think? Best Robert
open
2021-06-04T12:54:51Z
2021-06-04T20:34:15Z
https://github.com/koaning/scikit-lego/issues/471
[ "enhancement" ]
Garve
6
JoeanAmier/TikTokDownloader
api
154
下载所有收藏作品时报错[1744] Failed to execute script 'main' due to unhandled exception!
选择批量下载收藏作品时报错,无法正常下载 已选择批量下载收藏作品模式 开始获取收藏数据 Traceback (most recent call last): File "main.py", line 343, in <module> File "main.py", line 321, in run File "main.py", line 217, in main_menu File "main.py", line 286, in compatible File "main.py", line 65, in inner File "main.py", line 226, in complete File "src\main_complete.py", line 904, in run File "src\main_complete.py", line 819, in collection_interactive File "src\main_complete.py", line 838, in _deal_collection_data File "src\DataAcquirer.py", line 1042, in run File "src\DataAcquirer.py", line 51, in inner File "src\DataAcquirer.py", line 1065, in _get_account_data File "src\DataAcquirer.py", line 152, in deal_item_data TypeError: 'NoneType' object is not subscriptable [1744] Failed to execute script 'main' due to unhandled exception!
open
2024-02-20T16:51:20Z
2024-02-25T04:34:28Z
https://github.com/JoeanAmier/TikTokDownloader/issues/154
[]
mortgoupil
3
microsoft/nni
data-science
5,759
Shivam
**Describe the bug**: **Environment**: - NNI version: - Training service (local|remote|pai|aml|etc): - Python version: - PyTorch version: - Cpu or cuda version: **Reproduce the problem** - Code|Example: - How to reproduce:
open
2024-03-19T13:07:51Z
2024-03-19T13:07:51Z
https://github.com/microsoft/nni/issues/5759
[]
shivaay9673
0
arogozhnikov/einops
tensorflow
101
Interested in how you implemented this
Thanks for the great work! Sorry I don't have enough experience to figure this out from the source code by myself, but I'd like to know how you implemented einops. If an operation has to be done in the backend framework in a few steps but can be done with einops in one step, do you first figure out the backend methods needed and then call them sequentially or do you manipulate the tensor directly in a more low-level way?
closed
2021-02-26T09:54:39Z
2021-03-01T05:48:10Z
https://github.com/arogozhnikov/einops/issues/101
[]
trisct
2
httpie/cli
rest-api
1,410
Python 3.11 test failures: Different enum reprs and different cookie order
## Checklist - [x] I've searched for similar issues. - [x] I'm using the latest version of HTTPie. --- ## Minimal reproduction code and steps The tests are failing with Python 3.11.0b3. 1. `git clone` httpie and `cd` into it, this is on the master branch @ 418b12bbd6072585118c06c5c4e17996d7f0b085 2. `python3.11 -m venv __venv__` and `. __venv__/bin/activate` 3. `pip install '.[test]'` 4. `pytest tests/` ## Current result ``` ============================= test session starts ============================== platform linux -- Python 3.11.0b3, pytest-7.1.2, pluggy-1.0.0 rootdir: .../httpie, configfile: pytest.ini plugins: mock-3.7.0, lazy-fixture-0.6.3, httpbin-1.0.2, forked-1.4.0, xdist-2.5.0 collected 1026 items tests/test_auth.py .......... [ 0%] tests/test_compress.py ....... [ 1%] tests/test_downloads.py ...... [ 2%] tests/test_errors.py .... [ 2%] tests/test_httpie.py ...........s............... [ 5%] tests/test_json.py . [ 5%] tests/test_auth.py .......... [ 6%] tests/test_compress.py ....... [ 7%] tests/test_downloads.py ...... [ 7%] tests/test_errors.py .... [ 7%] tests/test_httpie.py ...........s............... [ 10%] tests/test_json.py . [ 10%] tests/test_auth.py ......... [ 11%] tests/test_auth_plugins.py .... [ 11%] tests/test_binary.py ...... [ 12%] tests/test_cli.py ..............F....................... [ 16%] tests/test_cli_ui.py .... [ 16%] tests/test_cli_utils.py .. [ 16%] tests/test_compress.py .. [ 17%] tests/test_config.py ........s [ 17%] tests/test_cookie.py . [ 18%] tests/test_cookie_on_redirects.py ..................... [ 20%] tests/test_defaults.py ................. [ 21%] tests/test_downloads.py .................... [ 23%] tests/test_encoding.py ................................................. [ 28%] .... [ 28%] tests/test_errors.py .... [ 29%] tests/test_exit_status.py ......... [ 30%] tests/test_httpie.py ...................... [ 32%] tests/test_httpie_cli.py ..................................... [ 35%] tests/test_json.py ..................................................... [ 41%] ........................................................................ [ 48%] ........................................................................ [ 55%] ........................................................................ [ 62%] ....................... [ 64%] tests/test_meta.py ...... [ 64%] tests/test_offline.py ......... [ 65%] tests/test_output.py ......FF.FxXX...................................... [ 70%] ........................................................................ [ 77%] . [ 77%] tests/test_parser_schema.py . [ 77%] tests/test_plugins_cli.py ......FFFF..F.F [ 79%] tests/test_redirects.py ...x...... [ 80%] tests/test_regressions.py ... [ 80%] tests/test_sessions.py .........................FF.F.................... [ 85%] ............ [ 86%] tests/test_ssl.py .ss.....pytest-httpbin server hit an exception serving request: EOF occurred in violation of protocol (_ssl.c:992) attempting to ignore so the rest of the tests can run ............pytest-httpbin server hit an exception serving request: EOF occurred in violation of protocol (_ssl.c:992) attempting to ignore so the rest of the tests can run ... [ 88%] tests/test_stream.py ................ [ 90%] tests/test_tokens.py ................... [ 92%] tests/test_transport_plugin.py . [ 92%] tests/test_update_warnings.py ............ [ 93%] tests/test_uploads.py ........................... [ 96%] tests/test_windows.py s. [ 96%] tests/test_xml.py ................. [ 98%] tests/utils/matching/test_matching.py .................... [100%] =================================== FAILURES =================================== _______________________ test_url_colon_slash_slash_only ________________________ def test_url_colon_slash_slash_only(): r = http('://', tolerate_error_exit_status=True) > assert r.stderr.strip() == "http: error: InvalidURL: Invalid URL 'http://': No host supplied" E AssertionError: assert 'http: LogLev...host supplied' == 'http: error:...host supplied' E - http: error: InvalidURL: Invalid URL 'http://': No host supplied E ? ^^^^ E + http: LogLevel.ERROR: InvalidURL: Invalid URL 'http://': No host supplied E ? ++++ ^^^^^^^^^ tests/test_cli.py:192: AssertionError ----------------------------- Captured stderr call ----------------------------- http: LogLevel.ERROR: InvalidURL: Invalid URL 'http://': No host supplied _____________ TestQuietFlag.test_quiet_with_check_status_non_zero ______________ self = <tests.test_output.TestQuietFlag object at 0x7ff4dcbdfe90> httpbin = <pytest_httpbin.serve.Server object at 0x7ff4dcb5a4d0> def test_quiet_with_check_status_non_zero(self, httpbin): r = http( '--quiet', '--check-status', httpbin + '/status/500', tolerate_error_exit_status=True, ) > assert 'http: warning: HTTP 500' in r.stderr E AssertionError: assert 'http: warning: HTTP 500' in '\nhttp: LogLevel.WARNING: HTTP 500 INTERNAL SERVER ERROR\n\n\n' E + where '\nhttp: LogLevel.WARNING: HTTP 500 INTERNAL SERVER ERROR\n\n\n' = ''.stderr tests/test_output.py:69: AssertionError ----------------------------- Captured stderr call ----------------------------- 127.0.0.1 - - [07/Jun/2022 16:23:54] "GET /status/500 HTTP/1.1" 500 0 http: LogLevel.WARNING: HTTP 500 INTERNAL SERVER ERROR ___________ TestQuietFlag.test_quiet_with_check_status_non_zero_pipe ___________ self = <tests.test_output.TestQuietFlag object at 0x7ff4dcbe4650> httpbin = <pytest_httpbin.serve.Server object at 0x7ff4dcb5a4d0> def test_quiet_with_check_status_non_zero_pipe(self, httpbin): r = http( '--quiet', '--check-status', httpbin + '/status/500', tolerate_error_exit_status=True, env=MockEnvironment(stdout_isatty=False) ) > assert 'http: warning: HTTP 500' in r.stderr E AssertionError: assert 'http: warning: HTTP 500' in '\nhttp: LogLevel.WARNING: HTTP 500 INTERNAL SERVER ERROR\n\n\n' E + where '\nhttp: LogLevel.WARNING: HTTP 500 INTERNAL SERVER ERROR\n\n\n' = ''.stderr tests/test_output.py:77: AssertionError ----------------------------- Captured stderr call ----------------------------- 127.0.0.1 - - [07/Jun/2022 16:23:54] "GET /status/500 HTTP/1.1" 500 0 http: LogLevel.WARNING: HTTP 500 INTERNAL SERVER ERROR ________ TestQuietFlag.test_quiet_quiet_with_check_status_non_zero_pipe ________ self = <tests.test_output.TestQuietFlag object at 0x7ff4dcbe5550> httpbin = <pytest_httpbin.serve.Server object at 0x7ff4dcb5a4d0> def test_quiet_quiet_with_check_status_non_zero_pipe(self, httpbin): r = http( '--quiet', '--quiet', '--check-status', httpbin + '/status/500', tolerate_error_exit_status=True, env=MockEnvironment(stdout_isatty=False) ) > assert 'http: warning: HTTP 500' in r.stderr E AssertionError: assert 'http: warning: HTTP 500' in '\nhttp: LogLevel.WARNING: HTTP 500 INTERNAL SERVER ERROR\n\n\n' E + where '\nhttp: LogLevel.WARNING: HTTP 500 INTERNAL SERVER ERROR\n\n\n' = ''.stderr tests/test_output.py:92: AssertionError ----------------------------- Captured stderr call ----------------------------- 127.0.0.1 - - [07/Jun/2022 16:23:54] "GET /status/500 HTTP/1.1" 500 0 http: LogLevel.WARNING: HTTP 500 INTERNAL SERVER ERROR _________________________ test_plugins_uninstall[True] _________________________ interface = Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uninstall_True_0/interface'), environment=<...out': <tempfile._TemporaryFileWrapper object at 0x7ff4d5c78410>, 'stdout_encoding': 'utf-8', 'stdout_isatty': True}>) httpie_plugins_success = <function httpie_plugins_success.<locals>.runner at 0x7ff4d5cbe0c0> dummy_plugin = Plugin(interface=Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uninstall_True_0/interface...ue}>), name='httpie-4bef7587', version='1.0.0', entry_points=[EntryPoint(name='test', group='httpie.plugins.auth.v1')]) cli_mode = True @pytest.mark.requires_installation @pytest.mark.parametrize('cli_mode', [True, False]) def test_plugins_uninstall(interface, httpie_plugins_success, dummy_plugin, cli_mode): httpie_plugins_success('install', dummy_plugin.path, cli_mode=cli_mode) httpie_plugins_success('uninstall', dummy_plugin.name, cli_mode=cli_mode) > assert not interface.is_installed(dummy_plugin.name) E AssertionError: assert not True E + where True = <bound method Interface.is_installed of Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uni...ut': <tempfile._TemporaryFileWrapper object at 0x7ff4d5c78410>,\n 'stdout_encoding': 'utf-8',\n 'stdout_isatty': True}>)>('httpie-4bef7587') E + where <bound method Interface.is_installed of Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uni...ut': <tempfile._TemporaryFileWrapper object at 0x7ff4d5c78410>,\n 'stdout_encoding': 'utf-8',\n 'stdout_isatty': True}>)> = Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uninstall_True_0/interface'), environment=<...out': <tempfile._TemporaryFileWrapper object at 0x7ff4d5c78410>,\n 'stdout_encoding': 'utf-8',\n 'stdout_isatty': True}>).is_installed E + and 'httpie-4bef7587' = Plugin(interface=Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uninstall_True_0/interface...ue}>), name='httpie-4bef7587', version='1.0.0', entry_points=[EntryPoint(name='test', group='httpie.plugins.auth.v1')]).name tests/test_plugins_cli.py:59: AssertionError ________________________ test_plugins_uninstall[False] _________________________ interface = Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uninstall_False_0/interface'), environment=...out': <tempfile._TemporaryFileWrapper object at 0x7ff4d5c811d0>, 'stdout_encoding': 'utf-8', 'stdout_isatty': True}>) httpie_plugins_success = <function httpie_plugins_success.<locals>.runner at 0x7ff4d5cac220> dummy_plugin = Plugin(interface=Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uninstall_False_0/interfac...ue}>), name='httpie-300cc8fa', version='1.0.0', entry_points=[EntryPoint(name='test', group='httpie.plugins.auth.v1')]) cli_mode = False @pytest.mark.requires_installation @pytest.mark.parametrize('cli_mode', [True, False]) def test_plugins_uninstall(interface, httpie_plugins_success, dummy_plugin, cli_mode): httpie_plugins_success('install', dummy_plugin.path, cli_mode=cli_mode) httpie_plugins_success('uninstall', dummy_plugin.name, cli_mode=cli_mode) > assert not interface.is_installed(dummy_plugin.name) E AssertionError: assert not True E + where True = <bound method Interface.is_installed of Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uni...ut': <tempfile._TemporaryFileWrapper object at 0x7ff4d5c811d0>,\n 'stdout_encoding': 'utf-8',\n 'stdout_isatty': True}>)>('httpie-300cc8fa') E + where <bound method Interface.is_installed of Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uni...ut': <tempfile._TemporaryFileWrapper object at 0x7ff4d5c811d0>,\n 'stdout_encoding': 'utf-8',\n 'stdout_isatty': True}>)> = Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uninstall_False_0/interface'), environment=...out': <tempfile._TemporaryFileWrapper object at 0x7ff4d5c811d0>,\n 'stdout_encoding': 'utf-8',\n 'stdout_isatty': True}>).is_installed E + and 'httpie-300cc8fa' = Plugin(interface=Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uninstall_False_0/interfac...ue}>), name='httpie-300cc8fa', version='1.0.0', entry_points=[EntryPoint(name='test', group='httpie.plugins.auth.v1')]).name tests/test_plugins_cli.py:59: AssertionError _____________________ test_plugins_listing_after_uninstall _____________________ interface = Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_listing_after_uni0/interface'), environment...out': <tempfile._TemporaryFileWrapper object at 0x7ff4d5dbc410>, 'stdout_encoding': 'utf-8', 'stdout_isatty': True}>) httpie_plugins_success = <function httpie_plugins_success.<locals>.runner at 0x7ff4d5cd2200> dummy_plugin = Plugin(interface=Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_listing_after_uni0/interfa...ue}>), name='httpie-99a195f1', version='1.0.0', entry_points=[EntryPoint(name='test', group='httpie.plugins.auth.v1')]) @pytest.mark.requires_installation def test_plugins_listing_after_uninstall(interface, httpie_plugins_success, dummy_plugin): httpie_plugins_success('install', dummy_plugin.path) httpie_plugins_success('uninstall', dummy_plugin.name) data = parse_listing(httpie_plugins_success('list')) > assert len(data) == 0 E AssertionError: assert 1 == 0 E + where 1 = len({'httpie-99a195f1': {'entry_points': [{'group': 'httpie.plugins.auth.v1', 'name': 'test'}], 'version': '1.0.0'}}) tests/test_plugins_cli.py:68: AssertionError _______________________ test_plugins_uninstall_specific ________________________ interface = Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uninstall_specifi0/interface'), environment...out': <tempfile._TemporaryFileWrapper object at 0x7ff4d5cc8950>, 'stdout_encoding': 'utf-8', 'stdout_isatty': True}>) httpie_plugins_success = <function httpie_plugins_success.<locals>.runner at 0x7ff4d5cd3560> @pytest.mark.requires_installation def test_plugins_uninstall_specific(interface, httpie_plugins_success): new_plugin_1 = interface.make_dummy_plugin() new_plugin_2 = interface.make_dummy_plugin() target_plugin = interface.make_dummy_plugin() httpie_plugins_success('install', new_plugin_1.path, new_plugin_2.path, target_plugin.path) httpie_plugins_success('uninstall', target_plugin.name) assert interface.is_installed(new_plugin_1.name) assert interface.is_installed(new_plugin_2.name) > assert not interface.is_installed(target_plugin.name) E AssertionError: assert not True E + where True = <bound method Interface.is_installed of Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uni...ut': <tempfile._TemporaryFileWrapper object at 0x7ff4d5cc8950>,\n 'stdout_encoding': 'utf-8',\n 'stdout_isatty': True}>)>('httpie-79b6b9b0') E + where <bound method Interface.is_installed of Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uni...ut': <tempfile._TemporaryFileWrapper object at 0x7ff4d5cc8950>,\n 'stdout_encoding': 'utf-8',\n 'stdout_isatty': True}>)> = Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uninstall_specifi0/interface'), environment...out': <tempfile._TemporaryFileWrapper object at 0x7ff4d5cc8950>,\n 'stdout_encoding': 'utf-8',\n 'stdout_isatty': True}>).is_installed E + and 'httpie-79b6b9b0' = Plugin(interface=Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_uninstall_specifi0/interfa...ue}>), name='httpie-79b6b9b0', version='1.0.0', entry_points=[EntryPoint(name='test', group='httpie.plugins.auth.v1')]).name tests/test_plugins_cli.py:82: AssertionError ________________________ test_plugins_double_uninstall _________________________ httpie_plugins = <function httpie_plugins.<locals>.runner at 0x7ff4d5cfa160> httpie_plugins_success = <function httpie_plugins_success.<locals>.runner at 0x7ff4d5cfa3e0> dummy_plugin = Plugin(interface=Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_plugins_double_uninstall0/interfac...ue}>), name='httpie-6d157572', version='1.0.0', entry_points=[EntryPoint(name='test', group='httpie.plugins.auth.v1')]) @pytest.mark.requires_installation def test_plugins_double_uninstall(httpie_plugins, httpie_plugins_success, dummy_plugin): httpie_plugins_success("install", dummy_plugin.path) httpie_plugins_success("uninstall", dummy_plugin.name) result = httpie_plugins("uninstall", dummy_plugin.name) > assert result.exit_status == ExitStatus.ERROR E AssertionError: assert <ExitStatus.SUCCESS: 0> == <ExitStatus.ERROR: 1> E + where <ExitStatus.SUCCESS: 0> = 'Successfully uninstalled httpie-6d157572\n'.exit_status E + and <ExitStatus.ERROR: 1> = ExitStatus.ERROR tests/test_plugins_cli.py:113: AssertionError _____________________________ test_broken_plugins ______________________________ httpie_plugins = <function httpie_plugins.<locals>.runner at 0x7ff4d6d29300> httpie_plugins_success = <function httpie_plugins_success.<locals>.runner at 0x7ff4d76eb920> dummy_plugin = Plugin(interface=Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_broken_plugins0/interface'), envir...ue}>), name='httpie-8972797e', version='1.0.0', entry_points=[EntryPoint(name='test', group='httpie.plugins.auth.v1')]) broken_plugin = Plugin(interface=Interface(path=PosixPath('/tmp/pytest-of-.../pytest-31/test_broken_plugins0/interface'), envir...ue}>), name='httpie-4cd7d933', version='1.0.0', entry_points=[EntryPoint(name='test', group='httpie.plugins.auth.v1')]) @pytest.mark.requires_installation def test_broken_plugins(httpie_plugins, httpie_plugins_success, dummy_plugin, broken_plugin): httpie_plugins_success("install", dummy_plugin.path, broken_plugin.path) with pytest.warns( UserWarning, match=( f'While loading "{broken_plugin.name}", an error' ' occurred: broken plugin' ) ): data = parse_listing(httpie_plugins_success('list')) assert len(data) == 2 # We load before the uninstallation, so it will warn again. with pytest.warns(UserWarning): httpie_plugins_success("uninstall", broken_plugin.name) # No warning now, since it is uninstalled. data = parse_listing(httpie_plugins_success('list')) > assert len(data) == 1 E AssertionError: assert 2 == 1 E + where 2 = len({'httpie-4cd7d933': {'entry_points': [{'group': 'httpie.plugins.auth.v1', 'name': 'test'}], 'version': '1.0.0'}, 'httpie-8972797e': {'entry_points': [{'group': 'httpie.plugins.auth.v1', 'name': 'test'}], 'version': '1.0.0'}}) tests/test_plugins_cli.py:153: AssertionError _ TestCookieStorage.test_existing_and_new_cookies_sent_in_request[new=bar;chocolate=milk-new_cookies_dict1-chocolate=milk; cookie1=foo; cookie2=foo; new=bar] _ self = <tests.test_sessions.TestCookieStorage object at 0x7ff4dca91c10> new_cookies = 'new=bar;chocolate=milk' new_cookies_dict = {'chocolate': 'milk', 'new': 'bar'} expected = 'chocolate=milk; cookie1=foo; cookie2=foo; new=bar' httpbin = <pytest_httpbin.serve.Server object at 0x7ff4dcb5a4d0> @pytest.mark.parametrize( 'new_cookies, new_cookies_dict, expected', [( 'new=bar', {'new': 'bar'}, 'cookie1=foo; cookie2=foo; new=bar' ), ( 'new=bar;chocolate=milk', {'new': 'bar', 'chocolate': 'milk'}, 'chocolate=milk; cookie1=foo; cookie2=foo; new=bar' ), ( 'new=bar; chocolate=milk', {'new': 'bar', 'chocolate': 'milk'}, 'chocolate=milk; cookie1=foo; cookie2=foo; new=bar' ), ( 'new=bar;; chocolate=milk;;;', {'new': 'bar', 'chocolate': 'milk'}, 'cookie1=foo; cookie2=foo; new=bar' ), ( 'new=bar; chocolate=milk;;;', {'new': 'bar', 'chocolate': 'milk'}, 'chocolate=milk; cookie1=foo; cookie2=foo; new=bar' ) ] ) def test_existing_and_new_cookies_sent_in_request(self, new_cookies, new_cookies_dict, expected, httpbin): r = http( '--session', str(self.session_path), '--print=H', httpbin.url, 'Cookie:' + new_cookies, ) # Note: cookies in response are in alphabetical order > assert f'Cookie: {expected}' in r E AssertionError: assert 'Cookie: chocolate=milk; cookie1=foo; cookie2=foo; new=bar' in 'GET / HTTP/1.1\r\nAccept: */*\r\nAccept-Encoding: gzip, deflate, br\r\nConnection: keep-alive\r\nCookie: cookie1=foo; cookie2=foo; new=bar; chocolate=milk\r\nHost: 127.0.0.1:38085\r\nUser-Agent: HTTPie/3.2.1\r\n\r\n' tests/test_sessions.py:485: AssertionError ----------------------------- Captured stderr call ----------------------------- 127.0.0.1 - - [07/Jun/2022 16:24:20] "GET / HTTP/1.1" 200 12144 _ TestCookieStorage.test_existing_and_new_cookies_sent_in_request[new=bar; chocolate=milk-new_cookies_dict2-chocolate=milk; cookie1=foo; cookie2=foo; new=bar] _ self = <tests.test_sessions.TestCookieStorage object at 0x7ff4dca91e90> new_cookies = 'new=bar; chocolate=milk' new_cookies_dict = {'chocolate': 'milk', 'new': 'bar'} expected = 'chocolate=milk; cookie1=foo; cookie2=foo; new=bar' httpbin = <pytest_httpbin.serve.Server object at 0x7ff4dcb5a4d0> @pytest.mark.parametrize( 'new_cookies, new_cookies_dict, expected', [( 'new=bar', {'new': 'bar'}, 'cookie1=foo; cookie2=foo; new=bar' ), ( 'new=bar;chocolate=milk', {'new': 'bar', 'chocolate': 'milk'}, 'chocolate=milk; cookie1=foo; cookie2=foo; new=bar' ), ( 'new=bar; chocolate=milk', {'new': 'bar', 'chocolate': 'milk'}, 'chocolate=milk; cookie1=foo; cookie2=foo; new=bar' ), ( 'new=bar;; chocolate=milk;;;', {'new': 'bar', 'chocolate': 'milk'}, 'cookie1=foo; cookie2=foo; new=bar' ), ( 'new=bar; chocolate=milk;;;', {'new': 'bar', 'chocolate': 'milk'}, 'chocolate=milk; cookie1=foo; cookie2=foo; new=bar' ) ] ) def test_existing_and_new_cookies_sent_in_request(self, new_cookies, new_cookies_dict, expected, httpbin): r = http( '--session', str(self.session_path), '--print=H', httpbin.url, 'Cookie:' + new_cookies, ) # Note: cookies in response are in alphabetical order > assert f'Cookie: {expected}' in r E AssertionError: assert 'Cookie: chocolate=milk; cookie1=foo; cookie2=foo; new=bar' in 'GET / HTTP/1.1\r\nAccept: */*\r\nAccept-Encoding: gzip, deflate, br\r\nConnection: keep-alive\r\nCookie: cookie1=foo; cookie2=foo; new=bar; chocolate=milk\r\nHost: 127.0.0.1:38085\r\nUser-Agent: HTTPie/3.2.1\r\n\r\n' tests/test_sessions.py:485: AssertionError ----------------------------- Captured stderr call ----------------------------- 127.0.0.1 - - [07/Jun/2022 16:24:20] "GET / HTTP/1.1" 200 12144 _ TestCookieStorage.test_existing_and_new_cookies_sent_in_request[new=bar; chocolate=milk;;;-new_cookies_dict4-chocolate=milk; cookie1=foo; cookie2=foo; new=bar] _ self = <tests.test_sessions.TestCookieStorage object at 0x7ff4dca925d0> new_cookies = 'new=bar; chocolate=milk;;;' new_cookies_dict = {'chocolate': 'milk', 'new': 'bar'} expected = 'chocolate=milk; cookie1=foo; cookie2=foo; new=bar' httpbin = <pytest_httpbin.serve.Server object at 0x7ff4dcb5a4d0> @pytest.mark.parametrize( 'new_cookies, new_cookies_dict, expected', [( 'new=bar', {'new': 'bar'}, 'cookie1=foo; cookie2=foo; new=bar' ), ( 'new=bar;chocolate=milk', {'new': 'bar', 'chocolate': 'milk'}, 'chocolate=milk; cookie1=foo; cookie2=foo; new=bar' ), ( 'new=bar; chocolate=milk', {'new': 'bar', 'chocolate': 'milk'}, 'chocolate=milk; cookie1=foo; cookie2=foo; new=bar' ), ( 'new=bar;; chocolate=milk;;;', {'new': 'bar', 'chocolate': 'milk'}, 'cookie1=foo; cookie2=foo; new=bar' ), ( 'new=bar; chocolate=milk;;;', {'new': 'bar', 'chocolate': 'milk'}, 'chocolate=milk; cookie1=foo; cookie2=foo; new=bar' ) ] ) def test_existing_and_new_cookies_sent_in_request(self, new_cookies, new_cookies_dict, expected, httpbin): r = http( '--session', str(self.session_path), '--print=H', httpbin.url, 'Cookie:' + new_cookies, ) # Note: cookies in response are in alphabetical order > assert f'Cookie: {expected}' in r E AssertionError: assert 'Cookie: chocolate=milk; cookie1=foo; cookie2=foo; new=bar' in 'GET / HTTP/1.1\r\nAccept: */*\r\nAccept-Encoding: gzip, deflate, br\r\nConnection: keep-alive\r\nCookie: cookie1=foo; cookie2=foo; new=bar; chocolate=milk\r\nHost: 127.0.0.1:38085\r\nUser-Agent: HTTPie/3.2.1\r\n\r\n' tests/test_sessions.py:485: AssertionError ----------------------------- Captured stderr call ----------------------------- 127.0.0.1 - - [07/Jun/2022 16:24:20] "GET / HTTP/1.1" 200 12144 =============================== warnings summary =============================== ... =========================== short test summary info ============================ FAILED tests/test_cli.py::test_url_colon_slash_slash_only - AssertionError: a... FAILED tests/test_output.py::TestQuietFlag::test_quiet_with_check_status_non_zero FAILED tests/test_output.py::TestQuietFlag::test_quiet_with_check_status_non_zero_pipe FAILED tests/test_output.py::TestQuietFlag::test_quiet_quiet_with_check_status_non_zero_pipe FAILED tests/test_plugins_cli.py::test_plugins_uninstall[True] - AssertionErr... FAILED tests/test_plugins_cli.py::test_plugins_uninstall[False] - AssertionEr... FAILED tests/test_plugins_cli.py::test_plugins_listing_after_uninstall - Asse... FAILED tests/test_plugins_cli.py::test_plugins_uninstall_specific - Assertion... FAILED tests/test_plugins_cli.py::test_plugins_double_uninstall - AssertionEr... FAILED tests/test_plugins_cli.py::test_broken_plugins - AssertionError: asser... FAILED tests/test_sessions.py::TestCookieStorage::test_existing_and_new_cookies_sent_in_request[new=bar;chocolate=milk-new_cookies_dict1-chocolate=milk; cookie1=foo; cookie2=foo; new=bar] FAILED tests/test_sessions.py::TestCookieStorage::test_existing_and_new_cookies_sent_in_request[new=bar; chocolate=milk-new_cookies_dict2-chocolate=milk; cookie1=foo; cookie2=foo; new=bar] FAILED tests/test_sessions.py::TestCookieStorage::test_existing_and_new_cookies_sent_in_request[new=bar; chocolate=milk;;;-new_cookies_dict4-chocolate=milk; cookie1=foo; cookie2=foo; new=bar] = 13 failed, 1003 passed, 6 skipped, 2 xfailed, 2 xpassed, 351 warnings in 56.38s = ``` ## Expected result Tests pass. ## Debug output Please re-run the command with `--debug`, then copy the entire command & output and paste both below: Not relevant. ## Additional information, screenshots, or code examples We are about to update to Python 3.11 in Fedora 37 (the development version of Fedora). This is not yet blocking our users, but getting it sorted out will be eventually required. Fedora 37 Beta is to be released in September 2022. I believe that most of the test failures observed, if not all, are bad test expectations rather than actual problems in httpie.
closed
2022-06-07T14:30:54Z
2023-07-08T08:06:49Z
https://github.com/httpie/cli/issues/1410
[ "bug", "testing" ]
hroncok
1
deeppavlov/DeepPavlov
tensorflow
1,420
TorchModel typing for fname argument in save is str, but is used as pathlib.PosixPath
`fname` argument of TorchModel.save() is expected to be `str`, but is actually used as `pathlib.PosixPath` https://github.com/deepmipt/DeepPavlov/blob/master/deeppavlov/core/models/torch_model.py#L172 Support, please, both types of arguments or fix typing.
open
2021-03-24T15:11:01Z
2021-03-24T15:11:01Z
https://github.com/deeppavlov/DeepPavlov/issues/1420
[ "bug" ]
yurakuratov
0
marcomusy/vedo
numpy
824
cutWithMesh return a weird result
I am using `cutWithMesh` method, but the result is weird. ![image](https://user-images.githubusercontent.com/29271051/222913288-6429a57c-9bd1-4fc3-9e7b-d66dfd70a7f0.png) In the above figure, the left shows the original two polydata, and the right is the result. In the right result, there are some weird holes. My code is: ``` from vedo import * actor1 = Mesh('poly1.obj') actor2 = Mesh('poly2.obj') ret = actor1.clone().cutWithMesh(actor2) show(actor1, actor2, at=0, N=2) show(ret, at=1, interactive=True) ``` And the data is: [data.zip](https://github.com/marcomusy/vedo/files/10888918/data.zip) Is it a bug?
closed
2023-03-04T15:03:26Z
2023-03-05T14:30:39Z
https://github.com/marcomusy/vedo/issues/824
[]
zhang-qiang-github
4
modelscope/modelscope
nlp
932
Converting safetensors to modelscope format
How to convert standard safetensors model format to modelscope? In other words, how [this](https://huggingface.co/ali-vilab/text-to-video-ms-1.7b) was converted to [this](https://huggingface.co/ali-vilab/modelscope-damo-text-to-video-synthesis)?
closed
2024-07-28T13:28:00Z
2024-10-16T04:04:45Z
https://github.com/modelscope/modelscope/issues/932
[]
julkaztwittera
4
CorentinJ/Real-Time-Voice-Cloning
deep-learning
772
How to automize the input from a text file
Hello, than you for this amazing work. i'm currently working on automising the input of the text to synthesize via a text file without reloading the UI ? Is there a specific class or function where i can work to make this. Thank you
closed
2021-06-14T12:03:12Z
2021-06-14T12:21:43Z
https://github.com/CorentinJ/Real-Time-Voice-Cloning/issues/772
[]
Anas-eladili
0
keras-team/keras
data-science
20,795
Blackduck vulnerability reported in all the versions of Keras
In Blackduck scan ,Keras package is reported as vulnerable with CVE ID =BDSA-2025-0107. can you please let us know in which release this fix will be given and when is the release date. Issue description: Keras is vulnerable to arbitrary file write due to a flaw in the get_file function. This could allow an attacker to write arbitrary files to the user's machine by downloading a crafted tar file.
closed
2025-01-22T12:05:59Z
2025-03-07T05:28:38Z
https://github.com/keras-team/keras/issues/20795
[ "stat:contributions welcome", "stat:awaiting response from contributor", "stale", "type:Bug" ]
frnz123
5
newpanjing/simpleui
django
33
关于修改SIMPLEUI_LOGO自定义logo的问题
**bug描述** 简单的描述下遇到的bug: **重现步骤** 1.修改settings中的SIMPLEUI_LOGO的值,即图片的url 2.我将图片放在了项目包的一级目录下 3.SIMPLEUI_LOGO = os.path.join(BASE_DIR,'logo1.png') **环境** 1.操作系统:win10 2.python版本:py37 3.django版本:2.2.1 4.simpleui版本:2.0.6 **其他描述** 使用 Ctrl+选中url 的快捷方式可以打开图片,说明不是路径的问题 显示图片破裂应该是未找到图片? ![image](https://user-images.githubusercontent.com/45619625/57583118-ce694a00-74ff-11e9-9f4c-ec150c9ad004.png)我看到的效果如图 我又按照demo中使用图片的大小,60×60px,重新设置了logo的大小,还是无法正常显示
closed
2019-05-12T13:51:18Z
2019-05-13T08:19:04Z
https://github.com/newpanjing/simpleui/issues/33
[ "bug" ]
ShiningSYK
2
mwaskom/seaborn
data-science
3,352
Categorical scatter plots on symlog-scaled axis
Hi, On the current dev version (eb2b5a2) and matplotlib 3.7.1, consider the following code that uses `stripplot` to draw two points on unscaled and symlog-scaled yaxis: ```python import seaborn as sns import matplotlib.pyplot as plt x = [0.1,2] y = [0.1,5] fig, axs = plt.subplots(ncols=2) sns.stripplot(x=x, y=y, ax=axs[0]) axs[0].set_yscale("symlog", base=10, linthresh=1) axs[1].set_yscale("symlog", base=10, linthresh=1) sns.stripplot(x=x, y=y, ax=axs[1]) axs[0].set_ylim(0,10**4) axs[1].set_ylim(0,10**4) axs[0].set_title("stripplot on unscaled axis") axs[1].set_title("stripplot on symlog-scaled axis") ``` ![image](https://user-images.githubusercontent.com/13831112/235433829-a7497557-f628-424b-bdb2-466353ef70f3.png) The plot on the already-scaled yaxis contain values that were not provided. The plot changes (but still erroneous) if I set `linthresh` to something different (for example if using the linthresh default of 2). This also happens with `pointplot`. It works as expected with log-scaled axis or with pure matplotlib scatter calls. Couldn't reproduce using seaborn 0.12.2.
closed
2023-05-01T09:18:52Z
2023-08-20T22:08:34Z
https://github.com/mwaskom/seaborn/issues/3352
[ "bug", "mod:categorical" ]
MaozGelbart
0
vaexio/vaex
data-science
2,377
[BUG-REPORT] Vaex can't handle Enums being used to reference columns
**Description** ``` import vaex from enum import Enum, unique @unique class Col(str, Enum): id = "id" x = "x" df = vaex.from_dict({ Col.x: list(range(10)), Col.id: list(range(10, 20)), }) df ``` `ValueError: Unknown expression type: <class 'ast.Attribute'>` You need to do ``` df = vaex.from_dict({ Col.x.value: list(range(10)), Col.id.value: list(range(10, 20)), }) ``` Which shouldn't be because we defined the class to inherit from `str` first. **Software information** - Vaex version (`import vaex; vaex.__version__)`: - Vaex was installed via: pip / conda-forge / from source - OS: **Additional information** Please state any supplementary information or provide additional context for the problem (e.g. screenshots, data, etc..).
open
2023-06-26T14:36:36Z
2023-06-26T14:36:36Z
https://github.com/vaexio/vaex/issues/2377
[]
Ben-Epstein
0
kubeflow/katib
scikit-learn
2,466
TensorFlowEvent metrics collector doesn't find events written by torch / tensorboardX
### What happened? Trial code writes tensorboard summary using tensorboardX or torch.utils.tensorboard. Experiment configured to use TensorFlowEvent metrics collector. The metrics collector reports the metric is unavailable. ### What did you expect to happen? I expect the metrics collector to report the metric accurately. ### Environment Katib v0.17.0 ### Impacted by this bug? Give it a 👍 We prioritize the issues with most 👍
closed
2024-12-18T22:24:13Z
2025-02-15T00:59:38Z
https://github.com/kubeflow/katib/issues/2466
[ "kind/bug" ]
garymm
0
modin-project/modin
data-science
6,840
BUG: Using Concat to Merge Two GroupBy DataFrames
### Modin version checks - [X] I have checked that this issue has not already been reported. - [X] I have confirmed this bug exists on the latest released version of Modin. - [ ] I have confirmed this bug exists on the main branch of Modin. (In order to do this you can follow [this guide](https://modin.readthedocs.io/en/stable/getting_started/installation.html#installing-from-the-github-master-branch).) ### Reproducible Example ```python import modin.pandas as pd df1 = pd.DataFrame([['a', 1], ['b', 2], ['b', 4]], columns=['letter', 'number']) df1_g = df1.groupby('letter', as_index=False)['number'].agg('sum') df2 = pd.DataFrame([['a', 3], ['a', 4], ['b', 1]], columns=['letter', 'number']) df2_g = df2.groupby('letter', as_index=False)['number'].agg('sum') pd.concat([df1_g, df2_g]) ``` ### Issue Description I am trying to use 'concat' to combine two DataFrames that have undergone a 'groupby' operation, but I am encountering an error. As a new user of Modin, I'm not entirely sure if it's an issue with my usage or a bug in the software. I appreciate your help with this matter. ### Expected Behavior | letter | number | |--------|--------| | a | 1 | | b | 6 | | a | 7 | | b | 1 | ### Error Logs <details> ```python-traceback --------------------------------------------------------------------------- AttributeError Traceback (most recent call last) Cell In[1], [line 10](vscode-notebook-cell:?execution_count=1&line=10) [6](vscode-notebook-cell:?execution_count=1&line=6) df2 = pd.DataFrame([['a', 3], ['a', 4], ['b', 1]], [7](vscode-notebook-cell:?execution_count=1&line=7) columns=['letter', 'number']) [8](vscode-notebook-cell:?execution_count=1&line=8) df2_g = df2.groupby('letter', as_index=False)['number'].agg('sum') ---> [10](vscode-notebook-cell:?execution_count=1&line=10) pd.concat([df1_g, df2_g]) File [h:\miniconda3\envs\modin_env\Lib\site-packages\modin\logging\logger_decorator.py:129](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:129), in enable_logging.<locals>.decorator.<locals>.run_and_log(*args, **kwargs) [114](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:114) """ [115](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:115) Compute function with logging if Modin logging is enabled. [116](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:116) (...) [126](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:126) Any [127](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:127) """ [128](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:128) if LogMode.get() == "disable": --> [129](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:129) return obj(*args, **kwargs) [131](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:131) logger = get_logger() [132](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:132) logger_level = getattr(logger, log_level) File [h:\miniconda3\envs\modin_env\Lib\site-packages\modin\pandas\general.py:541](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:541), in concat(objs, axis, join, ignore_index, keys, levels, names, verify_integrity, sort, copy) [534](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:534) if len(list_of_objs) == 0: [535](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:535) return DataFrame( [536](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:536) index=input_list_of_objs[0].index.append( [537](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:537) [f.index for f in input_list_of_objs[1:]] [538](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:538) ) [539](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:539) ) --> [541](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:541) new_query_compiler = list_of_objs[0].concat( [542](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:542) axis, [543](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:543) list_of_objs[1:], [544](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:544) join=join, [545](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:545) join_axes=None, [546](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:546) ignore_index=ignore_index, [547](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:547) keys=None, [548](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:548) levels=None, [549](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:549) names=None, [550](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:550) verify_integrity=False, [551](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:551) copy=True, [552](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:552) sort=sort, [553](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:553) ) [554](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:554) result_df = DataFrame(query_compiler=new_query_compiler) [555](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/pandas/general.py:555) if new_idx is not None: File [h:\miniconda3\envs\modin_env\Lib\site-packages\modin\logging\logger_decorator.py:129](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:129), in enable_logging.<locals>.decorator.<locals>.run_and_log(*args, **kwargs) [114](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:114) """ [115](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:115) Compute function with logging if Modin logging is enabled. [116](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:116) (...) [126](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:126) Any [127](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:127) """ [128](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:128) if LogMode.get() == "disable": --> [129](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:129) return obj(*args, **kwargs) [131](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:131) logger = get_logger() [132](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:132) logger_level = getattr(logger, log_level) File [h:\miniconda3\envs\modin_env\Lib\site-packages\modin\core\storage_formats\pandas\query_compiler.py:384](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/storage_formats/pandas/query_compiler.py:384), in PandasQueryCompiler.concat(self, axis, other, **kwargs) [382](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/storage_formats/pandas/query_compiler.py:382) ignore_index = kwargs.get("ignore_index", False) [383](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/storage_formats/pandas/query_compiler.py:383) other_modin_frame = [o._modin_frame for o in other] --> [384](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/storage_formats/pandas/query_compiler.py:384) new_modin_frame = self._modin_frame.concat(axis, other_modin_frame, join, sort) [385](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/storage_formats/pandas/query_compiler.py:385) result = self.__constructor__(new_modin_frame) [386](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/storage_formats/pandas/query_compiler.py:386) if ignore_index: File [h:\miniconda3\envs\modin_env\Lib\site-packages\modin\logging\logger_decorator.py:129](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:129), in enable_logging.<locals>.decorator.<locals>.run_and_log(*args, **kwargs) [114](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:114) """ [115](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:115) Compute function with logging if Modin logging is enabled. [116](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:116) (...) [126](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:126) Any [127](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:127) """ [128](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:128) if LogMode.get() == "disable": --> [129](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:129) return obj(*args, **kwargs) [131](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:131) logger = get_logger() [132](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:132) logger_level = getattr(logger, log_level) File [h:\miniconda3\envs\modin_env\Lib\site-packages\modin\core\dataframe\pandas\dataframe\utils.py:501](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/dataframe/pandas/dataframe/utils.py:501), in lazy_metadata_decorator.<locals>.decorator.<locals>.run_f_on_minimally_updated_metadata(self, *args, **kwargs) [499](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/dataframe/pandas/dataframe/utils.py:499) elif apply_axis == "rows": [500](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/dataframe/pandas/dataframe/utils.py:500) obj._propagate_index_objs(axis=0) --> [501](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/dataframe/pandas/dataframe/utils.py:501) result = f(self, *args, **kwargs) [502](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/dataframe/pandas/dataframe/utils.py:502) if apply_axis is None and not transpose: [503](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/dataframe/pandas/dataframe/utils.py:503) result._deferred_index = self._deferred_index File [h:\miniconda3\envs\modin_env\Lib\site-packages\modin\core\dataframe\pandas\dataframe\dataframe.py:3690](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/dataframe/pandas/dataframe/dataframe.py:3690), in PandasDataframe.concat(self, axis, others, how, sort) [3688](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/dataframe/pandas/dataframe/dataframe.py:3688) new_widths = None [3689](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/dataframe/pandas/dataframe/dataframe.py:3689) break -> [3690](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/dataframe/pandas/dataframe/dataframe.py:3690) return self.__constructor__( [3691](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/dataframe/pandas/dataframe/dataframe.py:3691) new_partitions, new_index, new_columns, new_lengths, new_widths, new_dtypes [3692](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/dataframe/pandas/dataframe/dataframe.py:3692) ) File [h:\miniconda3\envs\modin_env\Lib\site-packages\modin\logging\logger_decorator.py:129](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:129), in enable_logging.<locals>.decorator.<locals>.run_and_log(*args, **kwargs) [114](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:114) """ [115](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:115) Compute function with logging if Modin logging is enabled. [116](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:116) (...) [126](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:126) Any [127](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/logging/logger_decorator.py:127) """ ... --> [182](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/dataframe/pandas/metadata/dtypes.py:182) self._cols_with_unknown_dtypes.extend(new_cols) [183](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/dataframe/pandas/metadata/dtypes.py:183) self._know_all_names = True [184](file:///H:/miniconda3/envs/modin_env/Lib/site-packages/modin/core/dataframe/pandas/metadata/dtypes.py:184) return self._columns_order AttributeError: 'RangeIndex' object has no attribute 'extend' ``` </details> ### Installed Versions <details> INSTALLED VERSIONS ------------------ commit : 47a9a4a294c75cd7b67f0fd7f95f846ed53fbafa python : 3.12.0.final.0 python-bits : 64 OS : Windows OS-release : 11 Version : 10.0.22621 machine : AMD64 processor : AMD64 Family 23 Model 113 Stepping 0, AuthenticAMD byteorder : little LC_ALL : None LANG : None LOCALE : Chinese (Traditional)_Taiwan.950 Modin dependencies ------------------ modin : 0.26.0 ray : None dask : 2023.12.1 distributed : 2023.12.1 hdk : None pandas dependencies ------------------- pandas : 2.1.4 numpy : 1.26.2 pytz : 2023.3.post1 dateutil : 2.8.2 setuptools : 68.2.2 pip : 23.3.1 Cython : None pytest : None hypothesis : None sphinx : None blosc : None feather : None xlsxwriter : None lxml.etree : None html5lib : None pymysql : None psycopg2 : None jinja2 : 3.1.2 IPython : 8.19.0 pandas_datareader : None bs4 : None bottleneck : None dataframe-api-compat: None fastparquet : None fsspec : 2023.12.2 gcsfs : None matplotlib : None numba : None numexpr : None odfpy : None openpyxl : None pandas_gbq : None pyarrow : None pyreadstat : None pyxlsb : None s3fs : None scipy : None sqlalchemy : None tables : None tabulate : None xarray : None xlrd : None zstandard : None tzdata : 2023.3 qtpy : None pyqt5 : None </details>
closed
2023-12-26T18:52:39Z
2024-01-08T13:24:04Z
https://github.com/modin-project/modin/issues/6840
[ "bug 🦗", "pandas concordance 🐼" ]
wsxqaza12
1
tflearn/tflearn
data-science
1,136
ValueError: Cannot feed value of shape (1, 46) for Tensor 'InputData/X:0', which has shape '(?, 4)'
How can i fix this error ? please help def bag_of_words(s, words): bag = [0 for _ in range(len(words))] s_words = nltk.word_tokenize(s) s_words = [stemmer.stem(word.lower()) for word in s_words] for se in s_words: for i, w in enumerate(words): if w == se: bag[i] = 1 return numpy.array(bag) def chat(): print("Hey ! I'm Mick The Mechanic! Having a problem with your vehicle ?") while True: inp = input("You:") if inp.lower() == "quit": break results = model.predict([bag_of_words(inp, words)]) results_index = numpy.argmax(results) tag = labels[results_index] for tg in data["intents"]: if tg['tag'] == tag: responses = tg['responses'] print(random.choice(responses)) chat() --------------------------------------------------------------------------- ValueError Traceback (most recent call last) <ipython-input-95-9b3ec892bb73> in <module> 35 36 ---> 37 chat() 38 39 <ipython-input-95-9b3ec892bb73> in chat() 22 break 23 ---> 24 results = model.predict([bag_of_words(inp, words)]) 25 results_index = numpy.argmax(results) 26 tag = labels[results_index] ~/anaconda3/lib/python3.7/site-packages/tflearn/models/dnn.py in predict(self, X) 255 """ 256 feed_dict = feed_dict_builder(X, None, self.inputs, None) --> 257 return self.predictor.predict(feed_dict) 258 259 def predict_label(self, X): ~/anaconda3/lib/python3.7/site-packages/tflearn/helpers/evaluator.py in predict(self, feed_dict) 67 prediction = [] 68 if len(self.tensors) == 1: ---> 69 return self.session.run(self.tensors[0], feed_dict=feed_dict) 70 else: 71 for output in self.tensors: ~/anaconda3/lib/python3.7/site-packages/tensorflow/python/client/session.py in run(self, fetches, feed_dict, options, run_metadata) 948 try: 949 result = self._run(None, fetches, feed_dict, options_ptr, --> 950 run_metadata_ptr) 951 if run_metadata: 952 proto_data = tf_session.TF_GetBuffer(run_metadata_ptr) ~/anaconda3/lib/python3.7/site-packages/tensorflow/python/client/session.py in _run(self, handle, fetches, feed_dict, options, run_metadata) 1147 'which has shape %r' % 1148 (np_val.shape, subfeed_t.name, -> 1149 str(subfeed_t.get_shape()))) 1150 if not self.graph.is_feedable(subfeed_t): 1151 raise ValueError('Tensor %s may not be fed.' % subfeed_t) ValueError: Cannot feed value of shape (1, 46) for Tensor 'InputData/X:0', which has shape '(?, 4)'
open
2019-09-10T08:08:42Z
2019-09-10T08:08:42Z
https://github.com/tflearn/tflearn/issues/1136
[]
karun12213
0
scrapy/scrapy
web-scraping
5,887
Integrating sqlalchemy into scrapy
<!-- Thanks for taking an interest in Scrapy! If you have a question that starts with "How to...", please see the Scrapy Community page: https://scrapy.org/community/. The GitHub issue tracker's purpose is to deal with bug reports and feature requests for the project itself. Keep in mind that by filing an issue, you are expected to comply with Scrapy's Code of Conduct, including treating everyone with respect: https://github.com/scrapy/scrapy/blob/master/CODE_OF_CONDUCT.md The following is a suggested template to structure your pull request, you can find more guidelines at https://doc.scrapy.org/en/latest/contributing.html#writing-patches and https://doc.scrapy.org/en/latest/contributing.html#submitting-patches --> ## Summary Consider integrating sqlalchemy into scrapy, such as allowing the Item class to support Object Relational Mapper ## Motivation If you want to use a structured database, you need to write sql statements by hand, manually create database tables, etc., which is very troublesome. ## Describe alternatives you've considered You can further refer to the structure of django, add the migration function of the database
closed
2023-04-06T11:50:44Z
2023-04-06T22:44:39Z
https://github.com/scrapy/scrapy/issues/5887
[]
keatonLiu
1
StackStorm/st2
automation
5,265
SET Type is not JSON serializable
## SUMMARY `.toSet()` yaql is no longer serialized to json by the workflow engine. Provide a quick summary of your bug report. While running a workflow with large input json I stumbled upon this serialization error for the workflow engine. ``` st2workflowengine.log:2021-05-12 17:49:46,697 140687301693512 ERROR workflows [-] [609c121c04f287f5ebf421a5] {'type': 'error', 'message': 'TypeError: Type is not JSON serializable ``` ### STACKSTORM VERSION st2 3.5dev (7af7122), on Python 3.6.8 ##### OS, environment, install method Centos 7 ## Steps to reproduce the problem Create a workflow with a yaql statement that outputs a set `.toSet()` ie ``` version: 1.0 tasks: valid: action: core.noop next: - when: <%succeeded()%> publish: - valid: <% list(alist).toSet() %> vars: - alist: [1,2,3,4] ``` ## Expected Results Workflow does not fail ## Actual Results ``` st2workflowengine.log:2021-05-12 17:49:46,697 140687301693512 ERROR workflows [-] [609c121c04f287f5ebf421a5] {'type': 'error', 'message': 'TypeError: Type is not JSON serializable set', 'task_id': 'valid_emails', 'route': 1} ``` workflow fails
closed
2021-05-12T18:14:42Z
2021-05-22T19:55:21Z
https://github.com/StackStorm/st2/issues/5265
[]
guzzijones
8
pytorch/pytorch
numpy
149,244
torch.sum_of_squares()
### 🚀 The feature, motivation and pitch It would be equivalent to `torch.linalg.vector_norm(..., ord=2) ** 2`. In my codebase, 14/18 of the norm calls have `** 2` immediately after. So the sum of squares is more common than the vector norm. This is slightly faster because `vector_norm` calls `sqrt`, which is extraneous if we undo it afterward. ### Alternatives `vector_norm(...) ** 2` is just one extra kernel call. ### Additional context If you care about abstract theory, this is because the sum of squares is more natural than the vector norm. They both measure the size of something, and size is linear in the square domain. Examples: reduction of grad norm across devices works with squares but not norms; variance is additive while standard deviation is not; the Pythagorean theorem adds squares of terms. cc @albanD
open
2025-03-15T01:45:28Z
2025-03-18T07:54:54Z
https://github.com/pytorch/pytorch/issues/149244
[ "triaged", "module: python frontend" ]
ad8e
4
gradio-app/gradio
data-science
10,484
Do Gradio 5 has gradio_imageslider natively now?
I mean like what this repo does https://github.com/pngwn/gradio-imageslider
closed
2025-02-02T10:30:40Z
2025-02-04T00:26:10Z
https://github.com/gradio-app/gradio/issues/10484
[ "bug" ]
FurkanGozukara
3
httpie/cli
rest-api
1,258
@maldito-dot
@maldito-dot _Originally posted by @maldito187 in https://github.com/httpie/httpie/pull/1256#pullrequestreview-842143225_
closed
2022-01-03T17:48:51Z
2022-01-03T21:47:01Z
https://github.com/httpie/cli/issues/1258
[]
luisuimi
0
dsdanielpark/Bard-API
nlp
285
Cookies refreshing very often and getting 429 all the time
Hey guys, I am trying to get gemini to work, but the cookies are refreshed too quickly for me to be able to even even type them in my code. Does anyone have working example I can try to follow :( I am using linux and am sending the requests from the UK.
closed
2024-02-20T16:27:04Z
2024-03-05T08:21:19Z
https://github.com/dsdanielpark/Bard-API/issues/285
[]
skalaydzhiyski
2
yihong0618/running_page
data-visualization
328
求Keep数据更新接口
因个人的Keep步行数据存在很多现在看来没有意义的自动记录,这样的数据大约有三千来条,手工删除实在费劲。我fork了作者的工具,实现可以查询到每一条步行记录,但是没有删除接口可用。看看作者这边有没有了解Keep的删除接口,或者给提供一种思路
closed
2022-10-21T01:28:48Z
2022-10-21T05:56:20Z
https://github.com/yihong0618/running_page/issues/328
[]
EugeneLau0
5
explosion/spaCy
deep-learning
13,264
Regex doesn't work if less than 3 characters?
<!-- NOTE: For questions or install related issues, please open a Discussion instead. --> ## How to reproduce the behaviour Taken and adjusted right from the docs: ```python import spacy from spacy.matcher import Matcher nlp = spacy.blank("en") matcher = Matcher(nlp.vocab, validate=True) pattern = [ { "TEXT": { "regex": r"4K" } } ] matcher.add("TV_RESOLUTION", [pattern]) doc = nlp("Sony 55 Inch 4K Ultra HD TV X90K Series:BRAVIA XR LED Smart Google TV, Dolby Vision HDR, Exclusive Features for PS 5 XR55X90K-2022 w/HT-A5000 5.1.2ch Dolby Atmos Sound Bar Surround Home Theater") res = matcher(doc) # res = [] ``` However if I add a `D` after `4K` in both strings, a match is found. Is there a minimal length restriction? ```python import spacy from spacy.matcher import Matcher nlp = spacy.blank("en") matcher = Matcher(nlp.vocab, validate=True) pattern = [ { "TEXT": { "regex": r"4KD" } } ] matcher.add("TV_RESOLUTION", [pattern]) doc = nlp("Sony 55 Inch 4KD Ultra HD TV X90K Series:BRAVIA XR LED Smart Google TV, Dolby Vision HDR, Exclusive Features for PS 5 XR55X90K-2022 w/HT-A5000 5.1.2ch Dolby Atmos Sound Bar Surround Home Theater") res = matcher(doc) # res = [[(11960903833032025891, 3, 4)]] ``` ## Your Environment <!-- Include details of your environment. You can also type `python -m spacy info --markdown` and copy-paste the result here.--> * Operating System: macOS * Python Version Used: 3.11.3 * spaCy Version Used: 3.7.2 * Environment Information: Nothing special
closed
2024-01-23T16:14:48Z
2024-02-23T00:05:21Z
https://github.com/explosion/spaCy/issues/13264
[ "feat / matcher" ]
SHxKM
3
sqlalchemy/sqlalchemy
sqlalchemy
10,342
Errors from use of hashlib.md5 for truncated names on FIPS-enabled systems
### Describe the bug On systems (RHEL) where FIPS is enabled, the md5 is disabled in openssl and calls to `hashlib.md5()` in python will fail with an error like: ``` # SNIP | File "/opt/app-root/lib64/python3.11/site-packages/sqlalchemy/sql/compiler.py", line 7507, in _truncate_and_render_maxlen_name | name = name[0 : max_ - 8] + "_" + util.md5_hex(name)[-4:] | ^^^^^^^^^^^^^^^^^^ | File "/opt/app-root/lib64/python3.11/site-packages/sqlalchemy/util/langhelpers.py", line 90, in md5_hex | m = hashlib.md5() | ^^^^^^^^^^^^^ | ValueError: [digital envelope routines] unsupported | | Application startup failed. Exiting. | Server stopped! ``` As mentioned in the [docs](https://docs.python.org/3/library/hashlib.html), for newer versions of python (3.9+) there is a `usedforsecurity` flag that can be set to `False` to avoid this error. ### Optional link from https://docs.sqlalchemy.org which documents the behavior that is expected _No response_ ### SQLAlchemy Version in Use 2.0.15 ### DBAPI (i.e. the database driver) psycopg2 ### Database Vendor and Major Version Postgresql 15 ### Python Version 3.11 ### Operating system Linux, OSX ### To Reproduce ```python # Create table with a constraint name or column name that exceeds a maximum length. class MyModel(DeclarativeBase): __tablename__ = "mytable" id: Mapped[int] = mapped_column(primary_key=True) very_very_very_very_very_very_very_very_very_very_long_name: Mapped[str] # 0123456789012345678901234567890123456789012345678901234567890123456789 # 1 2 3 4 5 6 7 ``` ### Error ``` # SNIP | File "/opt/app-root/lib64/python3.11/site-packages/sqlalchemy/sql/compiler.py", line 7507, in _truncate_and_render_maxlen_name | name = name[0 : max_ - 8] + "_" + util.md5_hex(name)[-4:] | ^^^^^^^^^^^^^^^^^^ | File "/opt/app-root/lib64/python3.11/site-packages/sqlalchemy/util/langhelpers.py", line 90, in md5_hex | m = hashlib.md5() | ^^^^^^^^^^^^^ | ValueError: [digital envelope routines] unsupported | | Application startup failed. Exiting. | Server stopped! ``` ### Additional context _No response_
closed
2023-09-12T14:38:50Z
2023-09-14T17:14:27Z
https://github.com/sqlalchemy/sqlalchemy/issues/10342
[ "bug", "high priority", "sql", "py3k", "near-term release" ]
hozn
16
ckan/ckan
api
8,682
Missing `humanize_entity_type` on submit button for `package/group_list.html`
## CKAN version 2.11 ## Describe the bug The listing of which groups a dataset belongs to uses `humanize_entity_type` for the page title and the "No groups" message, but not for the page submit button, which is still hardcoded to "Add to group" (or localised equivalent). ### Steps to reproduce - Create a plugin that overrides the `humanize_entity_type` helper to change "group" to "foo", similarly to https://github.com/ckan/ckan/blob/master/ckanext/example_humanizer/plugin.py - Log in, create a dataset, and go to the group listing page. - The page title will be "Foo" but the submit button will be "Add to group". ### Expected behavior All aspects of the group list page should be equally customisable by `humanize_entity_type`. ### Additional details
closed
2025-02-21T06:03:38Z
2025-02-27T15:57:58Z
https://github.com/ckan/ckan/issues/8682
[ "Good for Contribution" ]
ThrawnCA
1
biolab/orange3
data-visualization
7,031
Problem with wordcloud
<!-- Thanks for taking the time to report a bug! If you're raising an issue about an add-on (i.e., installed via Options > Add-ons), raise an issue in the relevant add-on's issue tracker instead. See: https://github.com/biolab?q=orange3 To fix the bug, we need to be able to reproduce it. Please answer the following questions to the best of your ability. --> **What's wrong?** <!-- Be specific, clear, and concise. Include screenshots if relevant. --> I use Orange through Anaconda, but I've already used version 3.38 downloaded directly from the program's website and had a similar problem. In the version with Anaconda, whenever I try to create a word cloud, Orange closes unexpectedly, generating the error message that I paste below. In version 3.38, which I downloaded directly from the program's website, the word cloud addon opened, but did not show the word cloud image. <!-- If you're getting an error message, copy it, and enclose it with three backticks (```). --> [Error.txt](https://github.com/user-attachments/files/18804024/Error.txt) **How can we reproduce the problem?** <!-- Upload a zip with the .ows file and data. --> <!-- Describe the steps (open this widget, click there, then add this...) --> It doesn't matter what type of database or file I'm working on in Orange, every time I open the word cloud the software closes. **What's your environment?** <!-- To find your Orange version, see "Help → About → Version" or `Orange.version.full_version` in code --> - Operating system: MacOs Big Sur - Orange version: 3.38 - How you installed Orange: Anaconda version 2.6.4
closed
2025-02-14T19:53:32Z
2025-03-07T09:47:36Z
https://github.com/biolab/orange3/issues/7031
[ "bug report" ]
uelramon
2
littlecodersh/ItChat
api
153
请问有使用ItChat后被微信封号的情况么?
closed
2016-11-16T02:26:10Z
2017-08-05T07:37:26Z
https://github.com/littlecodersh/ItChat/issues/153
[ "question" ]
spacelan
5
netbox-community/netbox
django
18,247
Disabled Cable Trace/Path Buttons in Device Interface View are invisible in Dark Mode
### Deployment Type Self-hosted ### Triage priority N/A ### NetBox Version v4.1.8 ### Python Version 3.10 ### Steps to Reproduce 1. Create a **Device** with One (or more) **Physical Interfaces** 2. Navigate to the **Interfaces** view 3. Toggle between Light and Dark Mode ### Expected Behavior ![image](https://github.com/user-attachments/assets/00de1ae4-4f42-4cc8-a12e-4ab92dd53bc6) Observe in **Light Mode**, these buttons are visible, but, **disabled**. ### Observed Behavior ![image](https://github.com/user-attachments/assets/aeb37b94-89a1-43ec-ac80-a37f8587531f) Observe in **Dark Mode**, the disabled buttons **are functionally invisible**, thus leading to a rendering-like issue.
closed
2024-12-17T06:01:23Z
2025-02-10T14:42:09Z
https://github.com/netbox-community/netbox/issues/18247
[ "type: bug", "status: accepted", "severity: low" ]
stuntguy3000
2
huggingface/transformers
python
36,544
`disable_compile` not honored as a kwarg in generate
### System Info - `transformers` version: 4.49.0 - Platform: Linux-6.8.0-49-generic-x86_64-with-glibc2.39 - Python version: 3.10.16 - Huggingface_hub version: 0.29.1 - Safetensors version: 0.5.3 - Accelerate version: 1.4.0 - Accelerate config: not found - DeepSpeed version: not installed - PyTorch version (GPU?): 2.6.0+cu124 (True) - Tensorflow version (GPU?): not installed (NA) - Flax version (CPU?/GPU?/TPU?): not installed (NA) - Jax version: not installed - JaxLib version: not installed - Using distributed or parallel set-up in script?: <fill in> - Using GPU in script?: <fill in> - GPU type: NVIDIA GeForce RTX 4090 ### Who can help? @gante, @SunMarc, @ArthurZucker ### Information - [ ] The official example scripts - [x] My own modified scripts ### Tasks - [ ] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...) - [ ] My own task or dataset (give details below) ### Reproduction The following snippet should disable `torch.compile`, note the use of `disable_compile` as a kwarg. From the documentation, it should replace the corresponding value in `generation_config`: ```python import os os.environ["TORCH_LOGS"]="+dynamo" from transformers import AutoTokenizer, AutoModelForCausalLM import torch model_id = "google/gemma-2-2b-it" device = "cuda:0" tokenizer = AutoTokenizer.from_pretrained(model_id) model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16).to(device) prompt = "<start_of_turn>user\nWrite a poem about the Kraken.<end_of_turn>\n<start_of_turn>model\n" inputs = tokenizer.encode(prompt, return_tensors="pt").to(device) outputs = model.generate(inputs, max_length=50, disable_compile=True) text = tokenizer.decode(outputs[0]) ``` But we can still see dynamo tracing calls when we run it. The reason appears to be [this line](https://github.com/huggingface/transformers/blob/66f29aaaf55c8fe0c3dbcd24beede2ca4effac56/src/transformers/generation/utils.py#L3219), which uses `self.generation_config` instead of `generation_config`. Note that this behaviour will be fixed by #36519 when it's merged. Alternatively, we could fix this issue first if that PR takes long to be approved. ### Expected behavior As discussed above.
closed
2025-03-04T17:48:15Z
2025-03-13T19:41:37Z
https://github.com/huggingface/transformers/issues/36544
[ "bug" ]
pcuenca
3
kennethreitz/responder
flask
375
SPA index.html static file references are not served after change
Hi, first of all, great project! I am running responder with Vue as frontend and Parcel as bundler, by serving the vue-generated ./static/index.html on the '/' route. During development i run: `parcel watch ./app/index.html -d ./static/ --public-url /static` to automatically update the static files when the Vue codebase are changed. When i try to force reload the page (without cached content) the updated static file are not served correctly and i receive this exception: `ERROR: Exception in ASGI application Traceback (most recent call last): File "/usr/lib/python3.7/site-packages/uvicorn/protocols/http/httptools_impl.py", line 375, in run_asgi result = await app(self.scope, self.receive, self.send) File "/usr/lib/python3.7/site-packages/uvicorn/middleware/asgi2.py", line 7, in __call__ await instance(receive, send) File "/usr/lib/python3.7/site-packages/starlette/middleware/wsgi.py", line 89, in __call__ await asyncio.wait_for(sender, None) File "/usr/lib/python3.7/asyncio/tasks.py", line 388, in wait_for return await fut File "/usr/lib/python3.7/site-packages/starlette/middleware/wsgi.py", line 104, in sender await send(message) File "/usr/lib/python3.7/site-packages/uvicorn/protocols/http/httptools_impl.py", line 504, in send raise RuntimeError("Response content shorter than Content-Length") RuntimeError: Response content shorter than Content-Length` Once the web server is restarted, the new files are served correctly. Any insight is appreciated.
closed
2019-07-23T22:38:47Z
2024-03-31T00:57:27Z
https://github.com/kennethreitz/responder/issues/375
[]
dsod
1
davidsandberg/facenet
computer-vision
1,167
No module named _pywrap_tensorflow_internal
Initially I had tensorflow 2.3 but due to module 'tensorflow' has no attribute 'GPUOptions' changed to tensorflow 1.7 and now iam getting the following error and I am unable to proceed with my proj.Though they have mentioned the link I am kind off lost.If anyone has any suggestions regarding the issue it would be much appreciated.Thank You in advance ImportError: No module named '_pywrap_tensorflow_internal' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "C:\Users\pnc\AppData\Local\Programs\Python\Python37\lib\site-packages\tensorflow\python\pywrap_tensorflow.py", line 58, in <module> from tensorflow.python.pywrap_tensorflow_internal import * File "C:\Users\pnc\AppData\Local\Programs\Python\Python37\lib\site-packages\tensorflow\python\pywrap_tensorflow_internal.py", line 28, in <module> _pywrap_tensorflow_internal = swig_import_helper() File "C:\Users\pnc\AppData\Local\Programs\Python\Python37\lib\site-packages\tensorflow\python\pywrap_tensorflow_internal.py", line 20, in swig_import_helper import _pywrap_tensorflow_internal ModuleNotFoundError: No module named '_pywrap_tensorflow_internal' Failed to load the native TensorFlow runtime. See https://www.tensorflow.org/install/install_sources#common_installation_problems for some common reasons and solutions. Include the entire stack trace above this error message when asking for help.
open
2020-08-22T05:15:45Z
2022-12-08T16:31:22Z
https://github.com/davidsandberg/facenet/issues/1167
[]
meghasn
2
pydata/xarray
pandas
9,914
Automatic Dask-Zarr chunk alignment on the to_zarr method
### Is your feature request related to a problem? In the time that I have used Xarray, I have seen many issues related to the alignment between the Dask and Zarr chunks, and most of them have been misunderstanding how to achieve a many-to-one relation between both chunks, for that reason, I would like to propose the addition of a new feature that allows aligning the chunks automatically, this should bring a significant reduction in the number of issues related to this problem and also simplify the way that Xarray, Dask, and Zarr interact from the user perspective. ### Describe the solution you'd like ### Add a new align chunks parameter. Pros: - It would not break the current behavior of the method. - It gives the users control over whether they want the automatic alignment. Cons: - It adds an extra parameter to a method that is already complex from my perspective. - We will have to add extra validations on the code and docs, for example, if the automatic chunk alignment is enabled and the synchronizer is also present, does it make sense to realign the chunks? or if the safe_chunks is enabled together with the automatic alignment, should we raise an error, or should we give priority to the chunk alignment? this also makes me think, what happens if someone sends a synchronizer and the safe_chunks is enabled? I have not seen this specified in the docs. - ### Drop the safe_chunks parameter and always align the chunks if it is necessary. Pros: - Reduce the number of parameters in the to_zarr method. - Simplify the user experience, now the users could ignore the difference between Dask and Zarr Chunks, which should reduce the number of issues reported on this topic. It is important to highlight that If the synchronizer is set, then the automatic alignment of the chunks should be disabled. I think this is a better workflow and prevents data corruption in all the scenarios, this means that synchronizer would work as a safe_chunks = False. - Delete the possibility of corrupting the data, which I think is more important than affecting the performance. Cons: - It would be a breaking change. - The automatic chunk alignment could increase the number of tasks in Dask or unexpectedly affect performance. A possible mitigation for this is to raise a warning indicating that the chunks were unaligned. I think option B is the best for a good part of the users, it prevents corruption of the data and simplifies the workflow, but probably there are some use cases that I'm not taking into consideration that would explain why the safe_chunks parameter was added in first place. Any feedback or different ideas are welcome ### Describe alternatives you've considered _No response_ ### Additional context _No response_
open
2024-12-21T17:45:08Z
2025-01-29T18:22:24Z
https://github.com/pydata/xarray/issues/9914
[ "enhancement" ]
josephnowak
3
suitenumerique/docs
django
626
Remove references to La Suite
- [ ] Remove the top right button with La Suite menu (the "gaufre") - [ ] In the GitHub repo, replace docs.numerique.gouv.fr in the description by a demo URL or a homepage for the project
open
2025-02-10T14:46:34Z
2025-02-19T07:26:02Z
https://github.com/suitenumerique/docs/issues/626
[ "reusability" ]
bzg
0
ccxt/ccxt
api
24,657
bitget, fetch_funding_rates(symbols=['ETH/USDT:USDT']), seems to return funding rate for all symbols
### Operating System Ubuntu 22.04.5 LTS ### Programming Languages _No response_ ### CCXT Version 4.4.43 ### Description _No response_ ### Code ```    import ccxt cex = ccxt.bitget() cex.load_markets() x = cex.fetch_funding_rates(symbols=['ETH/USDT:USDT']) print(x.keys()) ```
open
2024-12-25T09:26:45Z
2024-12-25T10:14:27Z
https://github.com/ccxt/ccxt/issues/24657
[]
williamyizhu
1
comfyanonymous/ComfyUI
pytorch
6,430
An error happened while trying to locate the files on the Hub and we cannot find the appropriate snapshot folder for the specified revision on the local disk. Please check your internet connection and try again.
### Expected Behavior not error ### Actual Behavior ![image](https://github.com/user-attachments/assets/89cc837b-7fd4-4ab0-9506-e3df94591650) ![image](https://github.com/user-attachments/assets/170fb936-94ec-41c3-a497-83d84480d8e4) ### Steps to Reproduce [生成动态图片.json](https://github.com/user-attachments/files/18386159/default.json) this is a workflow ### Debug Logs ```powershell # ComfyUI Error Report ## Error Details - **Node ID:** 42 - **Node Type:** MimicMotionGetPoses - **Exception Type:** huggingface_hub.utils._errors.LocalEntryNotFoundError - **Exception Message:** An error happened while trying to locate the files on the Hub and we cannot find the appropriate snapshot folder for the specified revision on the local disk. Please check your internet connection and try again. ## Stack Trace File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\execution.py", line 327, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\execution.py", line 202, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\execution.py", line 174, in _map_node_over_list process_inputs(input_dict, i) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\execution.py", line 163, in process_inputs results.append(getattr(obj, func)(**inputs)) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui-mimicmotionwrapper\nodes.py", line 393, in process snapshot_download(repo_id="yzd-v/DWPose", File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\utils\_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\_snapshot_download.py", line 236, in snapshot_download raise LocalEntryNotFoundError( ``` ## System Information - **ComfyUI Version:** v0.3.10-38-g4209edf4 - **Arguments:** D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\main.py --auto-launch --dont-upcast-attention --preview-method auto --use-quad-cross-attention --disable-smart-memory --disable-cuda-malloc - **OS:** nt - **Python Version:** 3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)] - **Embedded Python:** false - **PyTorch Version:** 2.3.1+cu118 ## Devices - **Name:** cuda:0 AMD Radeon RX 6600 : native - **Type:** cuda - **VRAM Total:** 8573157376 - **VRAM Free:** 8438939648 - **Torch VRAM Total:** 0 - **Torch VRAM Free:** 0 ## Logs ``` 2025-01-11T22:22:18.876013 - 2025-01-11T22:22:18.876013 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\aiohttp\connector.py", line 874, in _resolve_host 2025-01-11T22:22:18.877012 - 2025-01-11T22:22:18.877012 - addrs = await self._resolver.resolve(host, port, family=self._family)2025-01-11T22:22:18.877012 - 2025-01-11T22:22:18.877012 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\aiohttp\resolver.py", line 33, in resolve 2025-01-11T22:22:18.877012 - 2025-01-11T22:22:18.877012 - infos = await self._loop.getaddrinfo(2025-01-11T22:22:18.877012 - 2025-01-11T22:22:18.877012 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\asyncio\base_events.py", line 863, in getaddrinfo 2025-01-11T22:22:18.878012 - 2025-01-11T22:22:18.878012 - return await self.run_in_executor(2025-01-11T22:22:18.878012 - 2025-01-11T22:22:18.878012 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\concurrent\futures\thread.py", line 58, in run 2025-01-11T22:22:18.879012 - 2025-01-11T22:22:18.879012 - result = self.fn(*self.args, **self.kwargs)2025-01-11T22:22:18.879012 - 2025-01-11T22:22:18.879012 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\socket.py", line 955, in getaddrinfo 2025-01-11T22:22:18.879012 - 2025-01-11T22:22:18.879012 - for res in _socket.getaddrinfo(host, port, family, type, proto, flags):2025-01-11T22:22:18.879012 - 2025-01-11T22:22:18.879012 - socket2025-01-11T22:22:18.879012 - .2025-01-11T22:22:18.879012 - gaierror2025-01-11T22:22:18.880012 - : 2025-01-11T22:22:18.880012 - [Errno 11004] getaddrinfo failed2025-01-11T22:22:18.880012 - 2025-01-11T22:22:18.880012 - The above exception was the direct cause of the following exception: 2025-01-11T22:22:18.880012 - Traceback (most recent call last): 2025-01-11T22:22:18.880012 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\threading.py", line 1016, in _bootstrap_inner 2025-01-11T22:22:18.880012 - 2025-01-11T22:22:18.881013 - self.run()2025-01-11T22:22:18.881013 - 2025-01-11T22:22:18.881013 - File "<enhanced_experience vendors.sentry_sdk.integrations.threading>", line 99, in run 2025-01-11T22:22:18.882012 - File "<enhanced_experience vendors.sentry_sdk.integrations.threading>", line 94, in _run_old_run_func 2025-01-11T22:22:18.882012 - File "<enhanced_experience vendors.sentry_sdk.utils>", line 1649, in reraise 2025-01-11T22:22:18.883012 - File "<enhanced_experience vendors.sentry_sdk.integrations.threading>", line 92, in _run_old_run_func 2025-01-11T22:22:18.884012 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\threading.py", line 953, in run 2025-01-11T22:22:18.884012 - 2025-01-11T22:22:18.884012 - self._target(*self._args, **self._kwargs)2025-01-11T22:22:18.885033 - 2025-01-11T22:22:18.885033 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Manager\glob\manager_server.py", line 1408, in <lambda> 2025-01-11T22:22:18.885033 - 2025-01-11T22:22:18.885033 - threading.Thread(target=lambda: asyncio.run(default_cache_update())).start()2025-01-11T22:22:18.885033 - 2025-01-11T22:22:18.885033 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\asyncio\runners.py", line 44, in run 2025-01-11T22:22:18.886032 - 2025-01-11T22:22:18.886032 - return loop.run_until_complete(main)2025-01-11T22:22:18.886032 - 2025-01-11T22:22:18.886032 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\asyncio\base_events.py", line 649, in run_until_complete 2025-01-11T22:22:18.886032 - 2025-01-11T22:22:18.887031 - return future.result()2025-01-11T22:22:18.887031 - 2025-01-11T22:22:18.887031 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Manager\glob\manager_server.py", line 1400, in default_cache_update 2025-01-11T22:22:18.887031 - 2025-01-11T22:22:18.887031 - await asyncio.gather(a, b, c, d, e)2025-01-11T22:22:18.887031 - 2025-01-11T22:22:18.887031 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Manager\glob\manager_server.py", line 1387, in get_cache 2025-01-11T22:22:18.888032 - 2025-01-11T22:22:18.888032 - json_obj = await manager_util.get_data(uri, True)2025-01-11T22:22:18.888032 - 2025-01-11T22:22:18.888032 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Manager\glob\manager_util.py", line 116, in get_data 2025-01-11T22:22:18.888032 - 2025-01-11T22:22:18.889031 - async with session.get(uri, headers=headers) as resp:2025-01-11T22:22:18.889031 - 2025-01-11T22:22:18.889031 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\aiohttp\client.py", line 1141, in __aenter__ 2025-01-11T22:22:18.889031 - 2025-01-11T22:22:18.889031 - self._resp = await self._coro2025-01-11T22:22:18.889031 - 2025-01-11T22:22:18.889031 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\aiohttp\client.py", line 536, in _request 2025-01-11T22:22:18.890032 - 2025-01-11T22:22:18.890032 - conn = await self._connector.connect(2025-01-11T22:22:18.890032 - 2025-01-11T22:22:18.890032 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\aiohttp\connector.py", line 540, in connect 2025-01-11T22:22:18.890032 - 2025-01-11T22:22:18.890032 - proto = await self._create_connection(req, traces, timeout)2025-01-11T22:22:18.890032 - 2025-01-11T22:22:18.890032 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\aiohttp\connector.py", line 901, in _create_connection 2025-01-11T22:22:18.891032 - 2025-01-11T22:22:18.891032 - _, proto = await self._create_direct_connection(req, traces, timeout)2025-01-11T22:22:18.891032 - 2025-01-11T22:22:18.891032 - File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\aiohttp\connector.py", line 1169, in _create_direct_connection 2025-01-11T22:22:18.891032 - 2025-01-11T22:22:18.891032 - raise ClientConnectorError(req.connection_key, exc) from exc2025-01-11T22:22:18.891032 - 2025-01-11T22:22:18.891032 - aiohttp.client_exceptions2025-01-11T22:22:18.891032 - .2025-01-11T22:22:18.891032 - ClientConnectorError2025-01-11T22:22:18.891032 - : 2025-01-11T22:22:18.891032 - Cannot connect to host raw.githubusercontent.com:443 ssl:default [getaddrinfo failed]2025-01-11T22:22:18.892031 - 2025-01-11T22:22:19.768193 - [ReActor] - STATUS - Running v0.5.2-b1 in ComfyUI2025-01-11T22:22:19.768193 - 2025-01-11T22:22:20.130450 - Torch version: 2.3.1+cu1182025-01-11T22:22:20.130450 - 2025-01-11T22:22:20.302297 - D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui-sixgod_prompt2025-01-11T22:22:20.302297 - 2025-01-11T22:22:20.307759 - 找不到llama_cpp模块2025-01-11T22:22:20.307759 - 2025-01-11T22:22:20.377384 - (pysssss:WD14Tagger) [DEBUG] Available ORT providers: TensorrtExecutionProvider, CUDAExecutionProvider, CPUExecutionProvider2025-01-11T22:22:20.377384 - 2025-01-11T22:22:20.377384 - (pysssss:WD14Tagger) [DEBUG] Using ORT providers: CUDAExecutionProvider, CPUExecutionProvider2025-01-11T22:22:20.377384 - 2025-01-11T22:22:20.399980 - Workspace manager - Openning file hash dict2025-01-11T22:22:20.399980 - 2025-01-11T22:22:20.399980 - 🦄🦄Loading: Workspace Manager (V2.1.0)2025-01-11T22:22:20.399980 - 2025-01-11T22:22:20.573945 - ------------------------------------------2025-01-11T22:22:20.573945 - 2025-01-11T22:22:20.573945 - Comfyroll Studio v1.76 :  175 Nodes Loaded2025-01-11T22:22:20.573945 - 2025-01-11T22:22:20.573945 - ------------------------------------------2025-01-11T22:22:20.573945 - 2025-01-11T22:22:20.573945 - ** For changes, please see patch notes at https://github.com/Suzie1/ComfyUI_Comfyroll_CustomNodes/blob/main/Patch_Notes.md2025-01-11T22:22:20.573945 - 2025-01-11T22:22:20.573945 - ** For help, please see the wiki at https://github.com/Suzie1/ComfyUI_Comfyroll_CustomNodes/wiki2025-01-11T22:22:20.573945 - 2025-01-11T22:22:20.573945 - ------------------------------------------2025-01-11T22:22:20.573945 - 2025-01-11T22:22:20.579135 - [comfyui_controlnet_aux] | INFO -> Using ckpts path: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui_controlnet_aux\ckpts 2025-01-11T22:22:20.579135 - [comfyui_controlnet_aux] | INFO -> Using symlinks: False 2025-01-11T22:22:20.579135 - [comfyui_controlnet_aux] | INFO -> Using ort providers: ['CUDAExecutionProvider', 'DirectMLExecutionProvider', 'OpenVINOExecutionProvider', 'ROCMExecutionProvider', 'CPUExecutionProvider', 'CoreMLExecutionProvider'] 2025-01-11T22:22:20.592229 - DWPose: Onnxruntime with acceleration providers detected2025-01-11T22:22:20.592229 - 2025-01-11T22:22:20.608247 - ### [START] ComfyUI AlekPet Nodes v1.0.39 ###2025-01-11T22:22:20.608247 - 2025-01-11T22:22:22.308832 - Node -> ChatGLMNode: ChatGLM4TranslateCLIPTextEncodeNode, ChatGLM4TranslateTextNode, ChatGLM4InstructNode, ChatGLM4InstructMediaNode [Loading] 2025-01-11T22:22:22.309832 - 2025-01-11T22:22:22.312831 - Node -> ArgosTranslateNode: ArgosTranslateCLIPTextEncodeNode, ArgosTranslateTextNode [Loading] 2025-01-11T22:22:22.312831 - 2025-01-11T22:22:22.318339 - Node -> DeepTranslatorNode: DeepTranslatorCLIPTextEncodeNode, DeepTranslatorTextNode [Loading] 2025-01-11T22:22:22.318339 - 2025-01-11T22:22:22.320339 - Node -> GoogleTranslateNode: GoogleTranslateCLIPTextEncodeNode, GoogleTranslateTextNode [Loading] 2025-01-11T22:22:22.320339 - 2025-01-11T22:22:22.326433 - Node -> PoseNode: PoseNode [Loading] 2025-01-11T22:22:22.326433 - 2025-01-11T22:22:22.326433 - Node -> ExtrasNode: PreviewTextNode, HexToHueNode, ColorsCorrectNode [Loading] 2025-01-11T22:22:22.326433 - 2025-01-11T22:22:22.368967 - Node -> IDENode: IDENode [Loading] 2025-01-11T22:22:22.368967 - 2025-01-11T22:22:22.482689 - Node -> PainterNode: PainterNode [Loading] 2025-01-11T22:22:22.482689 - 2025-01-11T22:22:22.482689 - ### [END] ComfyUI AlekPet Nodes ###2025-01-11T22:22:22.482689 - 2025-01-11T22:22:22.752120 - Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\nodes.py", line 2106, in load_custom_node module_spec.loader.exec_module(module) File "<frozen importlib._bootstrap_external>", line 879, in exec_module File "<frozen importlib._bootstrap_external>", line 1016, in get_code File "<frozen importlib._bootstrap_external>", line 1073, in get_data FileNotFoundError: [Errno 2] No such file or directory: 'D:\\Tools\\StableDiffusion\\ComfyUI-aki-v1.4\\custom_nodes\\ComfyUI_examples\\__init__.py' 2025-01-11T22:22:22.752120 - Cannot import D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI_examples module for custom nodes: [Errno 2] No such file or directory: 'D:\\Tools\\StableDiffusion\\ComfyUI-aki-v1.4\\custom_nodes\\ComfyUI_examples\\__init__.py' 2025-01-11T22:22:23.115544 - FizzleDorf Custom Nodes: Loaded2025-01-11T22:22:23.115544 - 2025-01-11T22:22:23.305943 - Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\nodes.py", line 2106, in load_custom_node module_spec.loader.exec_module(module) File "<frozen importlib._bootstrap_external>", line 879, in exec_module File "<frozen importlib._bootstrap_external>", line 1016, in get_code File "<frozen importlib._bootstrap_external>", line 1073, in get_data FileNotFoundError: [Errno 2] No such file or directory: 'D:\\Tools\\StableDiffusion\\ComfyUI-aki-v1.4\\custom_nodes\\ECCV2022-RIFE\\__init__.py' 2025-01-11T22:22:23.305943 - Cannot import D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ECCV2022-RIFE module for custom nodes: [Errno 2] No such file or directory: 'D:\\Tools\\StableDiffusion\\ComfyUI-aki-v1.4\\custom_nodes\\ECCV2022-RIFE\\__init__.py' 2025-01-11T22:22:23.318180 - Efficiency Nodes: Attempting to add Control Net options to the 'HiRes-Fix Script' Node (comfyui_controlnet_aux add-on)...Success!2025-01-11T22:22:23.318180 - 2025-01-11T22:22:23.321181 - Patching UNetModel.forward2025-01-11T22:22:23.321181 - 2025-01-11T22:22:23.321181 - UNetModel.forward has been successfully patched.2025-01-11T22:22:23.321181 - 2025-01-11T22:22:23.339429 - [Power Noise Suite]: 🦚🦚🦚 Hrruuh-hhruh!!! 🦚🦚🦚2025-01-11T22:22:23.339429 - 2025-01-11T22:22:23.339429 - [Power Noise Suite]: Tamed 11 wild nodes.2025-01-11T22:22:23.339429 - 2025-01-11T22:22:23.352494 - 2025-01-11T22:22:23.352494 - [rgthree-comfy] Loaded 42 exciting nodes. 🎉2025-01-11T22:22:23.352494 - 2025-01-11T22:22:23.352494 - 2025-01-11T22:22:25.141376 - WAS Node Suite: OpenCV Python FFMPEG support is enabled2025-01-11T22:22:25.142376 - 2025-01-11T22:22:25.142376 - WAS Node Suite Warning: `ffmpeg_bin_path` is not set in `D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\was-node-suite-comfyui\was_suite_config.json` config file. Will attempt to use system ffmpeg binaries if available.2025-01-11T22:22:25.142376 - 2025-01-11T22:22:26.613135 - WAS Node Suite: Finished. Loaded 220 nodes successfully.2025-01-11T22:22:26.613135 - 2025-01-11T22:22:26.613135 - "The only way to do great work is to love what you do." - Steve Jobs 2025-01-11T22:22:26.613135 - 2025-01-11T22:22:26.617225 - Import times for custom nodes: 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Catcat 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\websocket_image_save.py 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\AIGODLIKE-ComfyUI-Translation 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\FreeU_Advanced 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ControlNet-LLLite-ComfyUI 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI_TiledKSampler 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\stability-ComfyUI-nodes 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui-lora-auto-trigger-words 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Pets 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-WD14-Tagger 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI_experiments 2025-01-11T22:22:26.618223 - 0.0 seconds (IMPORT FAILED): D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ECCV2022-RIFE 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\PowerNoiseSuite 2025-01-11T22:22:26.618223 - 0.0 seconds (IMPORT FAILED): D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI_examples 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI_IPAdapter_plus 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\images-grid-comfy-plugin 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Custom-Scripts 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui-sixgod_prompt 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI_UltimateSDUpscale 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\Derfuu_ComfyUI_ModdedNodes 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\OneButtonPrompt 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Advanced-ControlNet 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\rgthree-comfy 2025-01-11T22:22:26.618223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\efficiency-nodes-comfyui 2025-01-11T22:22:26.619223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui_essentials 2025-01-11T22:22:26.619223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui-mimicmotionwrapper 2025-01-11T22:22:26.619223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui-workspace-manager 2025-01-11T22:22:26.619223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-AnimateDiff-Evolved 2025-01-11T22:22:26.619223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui_controlnet_aux 2025-01-11T22:22:26.619223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-KJNodes 2025-01-11T22:22:26.619223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui-frame-interpolation 2025-01-11T22:22:26.619223 - 0.0 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Inspire-Pack 2025-01-11T22:22:26.619223 - 0.1 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui-videohelpersuite 2025-01-11T22:22:26.619223 - 0.1 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Crystools 2025-01-11T22:22:26.619223 - 0.2 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui_segment_anything 2025-01-11T22:22:26.619223 - 0.2 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Impact-Pack 2025-01-11T22:22:26.619223 - 0.2 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI_Comfyroll_CustomNodes 2025-01-11T22:22:26.619223 - 0.4 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI_FizzNodes 2025-01-11T22:22:26.619223 - 0.4 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Manager 2025-01-11T22:22:26.619223 - 0.4 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Marigold 2025-01-11T22:22:26.619223 - 1.3 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui-reactor-node 2025-01-11T22:22:26.619223 - 2.1 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI_Custom_Nodes_AlekPet 2025-01-11T22:22:26.619223 - 3.3 seconds: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\was-node-suite-comfyui 2025-01-11T22:22:26.619223 - 2025-01-11T22:22:26.633294 - Starting server 2025-01-11T22:22:26.634295 - To see the GUI go to: http://127.0.0.1:8188 2025-01-11T22:22:27.918366 - FETCH DATA from: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Manager\extension-node-map.json2025-01-11T22:22:27.918366 - 2025-01-11T22:22:27.924367 - [DONE]2025-01-11T22:22:27.924367 - 2025-01-11T22:22:28.100365 - FETCH DATA from: https://api.comfy.org/nodes?page=1&limit=10002025-01-11T22:22:28.100365 - 2025-01-11T22:22:31.869070 - [DONE]2025-01-11T22:22:31.869070 - 2025-01-11T22:22:31.921541 - [ComfyUI-Manager] default cache updated: https://api.comfy.org/nodes?page=1&limit=10002025-01-11T22:22:31.921541 - 2025-01-11T22:22:31.941187 - nightly_channel: 2025-01-11T22:22:31.941795 - 2025-01-11T22:22:31.941795 - https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/cache2025-01-11T22:22:31.941795 - 2025-01-11T22:22:31.941795 - FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json2025-01-11T22:22:31.941795 - 2025-01-11T22:22:31.943186 - [2025-01-11T22:22:31.943186 - ComfyUI-Manager2025-01-11T22:22:31.943186 - ]2025-01-11T22:22:31.943186 - Due to a network error, switching to local mode.2025-01-11T22:22:31.943186 - 2025-01-11T22:22:31.943186 - => custom-node-list.json2025-01-11T22:22:31.943186 - 2025-01-11T22:22:31.944256 - => Cannot connect to host raw.githubusercontent.com:2025-01-11T22:22:31.944256 - 4432025-01-11T22:22:31.944256 - ssl:default 2025-01-11T22:22:31.944256 - 2025-01-11T22:22:31.944256 - FETCH DATA from: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Manager\custom-node-list.json2025-01-11T22:22:31.944256 - 2025-01-11T22:22:31.949309 - [DONE]2025-01-11T22:22:31.949309 - 2025-01-11T22:23:34.928363 - got prompt 2025-01-11T22:23:36.180998 - Downloading yolo model to: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui-mimicmotionwrapper\models\DWPose2025-01-11T22:23:36.180998 - 2025-01-11T22:23:57.257622 - !!! Exception during processing !!! An error happened while trying to locate the files on the Hub and we cannot find the appropriate snapshot folder for the specified revision on the local disk. Please check your internet connection and try again. 2025-01-11T22:23:57.278960 - Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connection.py", line 174, in _new_conn conn = connection.create_connection( File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\util\connection.py", line 95, in create_connection raise err File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\util\connection.py", line 85, in create_connection sock.connect(sa) TimeoutError: [WinError 10060] 由于连接方在一段时间后没有正确答复或连接的主机没有反应,连接尝试失败。 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connectionpool.py", line 715, in urlopen httplib_response = self._make_request( File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connectionpool.py", line 404, in _make_request self._validate_conn(conn) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connectionpool.py", line 1058, in _validate_conn conn.connect() File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connection.py", line 363, in connect self.sock = conn = self._new_conn() File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connection.py", line 179, in _new_conn raise ConnectTimeoutError( urllib3.exceptions.ConnectTimeoutError: (<urllib3.connection.HTTPSConnection object at 0x0000018F90CCE890>, 'Connection to huggingface.co timed out. (connect timeout=None)') During handling of the above exception, another exception occurred: Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\requests\adapters.py", line 486, in send resp = conn.urlopen( File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connectionpool.py", line 799, in urlopen retries = retries.increment( File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\util\retry.py", line 592, in increment raise MaxRetryError(_pool, url, error or ResponseError(cause)) urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='huggingface.co', port=443): Max retries exceeded with url: /api/models/yzd-v/DWPose/revision/main (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x0000018F90CCE890>, 'Connection to huggingface.co timed out. (connect timeout=None)')) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\_snapshot_download.py", line 164, in snapshot_download repo_info = api.repo_info(repo_id=repo_id, repo_type=repo_type, revision=revision, token=token) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\utils\_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\hf_api.py", line 2491, in repo_info return method( File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\utils\_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\hf_api.py", line 2300, in model_info r = get_session().get(path, headers=headers, timeout=timeout, params=params) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\requests\sessions.py", line 602, in get return self.request("GET", url, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\requests\sessions.py", line 589, in request resp = self.send(prep, **send_kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\requests\sessions.py", line 703, in send r = adapter.send(request, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\utils\_http.py", line 66, in send return super().send(request, *args, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\requests\adapters.py", line 507, in send raise ConnectTimeout(e, request=request) requests.exceptions.ConnectTimeout: (MaxRetryError("HTTPSConnectionPool(host='huggingface.co', port=443): Max retries exceeded with url: /api/models/yzd-v/DWPose/revision/main (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x0000018F90CCE890>, 'Connection to huggingface.co timed out. (connect timeout=None)'))"), '(Request ID: dcc7982f-c1a8-49c5-966c-7a1eee6eebaf)') The above exception was the direct cause of the following exception: Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\execution.py", line 327, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\execution.py", line 202, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\execution.py", line 174, in _map_node_over_list process_inputs(input_dict, i) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\execution.py", line 163, in process_inputs results.append(getattr(obj, func)(**inputs)) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui-mimicmotionwrapper\nodes.py", line 393, in process snapshot_download(repo_id="yzd-v/DWPose", File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\utils\_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\_snapshot_download.py", line 236, in snapshot_download raise LocalEntryNotFoundError( huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the files on the Hub and we cannot find the appropriate snapshot folder for the specified revision on the local disk. Please check your internet connection and try again. 2025-01-11T22:23:57.279961 - Prompt executed in 22.34 seconds 2025-01-11T22:29:06.002774 - Exception in callback _ProactorBasePipeTransport._call_connection_lost(None) handle: <Handle _ProactorBasePipeTransport._call_connection_lost(None)> Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\asyncio\events.py", line 80, in _run self._context.run(self._callback, *self._args) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\asyncio\proactor_events.py", line 165, in _call_connection_lost self._sock.shutdown(socket.SHUT_RDWR) ConnectionResetError: [WinError 10054] 远程主机强迫关闭了一个现有的连接。 2025-01-11T22:29:09.577324 - FETCH DATA from: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Manager\extension-node-map.json2025-01-11T22:29:09.577324 - 2025-01-11T22:29:09.579985 - [DONE]2025-01-11T22:29:09.579985 - 2025-01-11T22:29:09.715106 - FETCH DATA from: https://api.comfy.org/nodes?page=1&limit=10002025-01-11T22:29:09.715106 - 2025-01-11T22:30:11.723150 - [DONE]2025-01-11T22:30:11.723150 - 2025-01-11T22:30:11.779866 - [ComfyUI-Manager] default cache updated: https://api.comfy.org/nodes?page=1&limit=10002025-01-11T22:30:11.779866 - 2025-01-11T22:30:11.800847 - nightly_channel: 2025-01-11T22:30:11.800847 - 2025-01-11T22:30:11.800847 - https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/cache2025-01-11T22:30:11.800847 - 2025-01-11T22:30:11.800847 - FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json2025-01-11T22:30:11.800847 - 2025-01-11T22:30:11.818138 - [2025-01-11T22:30:11.818138 - ComfyUI-Manager2025-01-11T22:30:11.818138 - ]2025-01-11T22:30:11.818138 - Due to a network error, switching to local mode.2025-01-11T22:30:11.818138 - 2025-01-11T22:30:11.818138 - => custom-node-list.json2025-01-11T22:30:11.818138 - 2025-01-11T22:30:11.818138 - => Cannot connect to host raw.githubusercontent.com:2025-01-11T22:30:11.818138 - 4432025-01-11T22:30:11.818138 - ssl:default 2025-01-11T22:30:11.818138 - 2025-01-11T22:30:11.818138 - FETCH DATA from: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\ComfyUI-Manager\custom-node-list.json2025-01-11T22:30:11.818138 - 2025-01-11T22:30:11.825147 - [DONE]2025-01-11T22:30:11.825147 - 2025-01-11T22:31:29.714024 - got prompt 2025-01-11T22:31:29.731470 - Downloading yolo model to: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui-mimicmotionwrapper\models\DWPose2025-01-11T22:31:29.731470 - 2025-01-11T22:31:50.802489 - !!! Exception during processing !!! An error happened while trying to locate the files on the Hub and we cannot find the appropriate snapshot folder for the specified revision on the local disk. Please check your internet connection and try again. 2025-01-11T22:31:50.803489 - Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connection.py", line 174, in _new_conn conn = connection.create_connection( File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\util\connection.py", line 95, in create_connection raise err File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\util\connection.py", line 85, in create_connection sock.connect(sa) TimeoutError: [WinError 10060] 由于连接方在一段时间后没有正确答复或连接的主机没有反应,连接尝试失败。 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connectionpool.py", line 715, in urlopen httplib_response = self._make_request( File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connectionpool.py", line 404, in _make_request self._validate_conn(conn) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connectionpool.py", line 1058, in _validate_conn conn.connect() File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connection.py", line 363, in connect self.sock = conn = self._new_conn() File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connection.py", line 179, in _new_conn raise ConnectTimeoutError( urllib3.exceptions.ConnectTimeoutError: (<urllib3.connection.HTTPSConnection object at 0x0000018F91086C50>, 'Connection to huggingface.co timed out. (connect timeout=None)') During handling of the above exception, another exception occurred: Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\requests\adapters.py", line 486, in send resp = conn.urlopen( File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connectionpool.py", line 799, in urlopen retries = retries.increment( File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\util\retry.py", line 592, in increment raise MaxRetryError(_pool, url, error or ResponseError(cause)) urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='huggingface.co', port=443): Max retries exceeded with url: /api/models/yzd-v/DWPose/revision/main (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x0000018F91086C50>, 'Connection to huggingface.co timed out. (connect timeout=None)')) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\_snapshot_download.py", line 164, in snapshot_download repo_info = api.repo_info(repo_id=repo_id, repo_type=repo_type, revision=revision, token=token) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\utils\_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\hf_api.py", line 2491, in repo_info return method( File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\utils\_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\hf_api.py", line 2300, in model_info r = get_session().get(path, headers=headers, timeout=timeout, params=params) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\requests\sessions.py", line 602, in get return self.request("GET", url, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\requests\sessions.py", line 589, in request resp = self.send(prep, **send_kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\requests\sessions.py", line 703, in send r = adapter.send(request, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\utils\_http.py", line 66, in send return super().send(request, *args, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\requests\adapters.py", line 507, in send raise ConnectTimeout(e, request=request) requests.exceptions.ConnectTimeout: (MaxRetryError("HTTPSConnectionPool(host='huggingface.co', port=443): Max retries exceeded with url: /api/models/yzd-v/DWPose/revision/main (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x0000018F91086C50>, 'Connection to huggingface.co timed out. (connect timeout=None)'))"), '(Request ID: 77a88aa9-3ff2-4540-9f2e-28464fb778e0)') The above exception was the direct cause of the following exception: Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\execution.py", line 327, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\execution.py", line 202, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\execution.py", line 174, in _map_node_over_list process_inputs(input_dict, i) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\execution.py", line 163, in process_inputs results.append(getattr(obj, func)(**inputs)) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui-mimicmotionwrapper\nodes.py", line 393, in process snapshot_download(repo_id="yzd-v/DWPose", File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\utils\_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\_snapshot_download.py", line 236, in snapshot_download raise LocalEntryNotFoundError( huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the files on the Hub and we cannot find the appropriate snapshot folder for the specified revision on the local disk. Please check your internet connection and try again. 2025-01-11T22:31:50.804489 - Prompt executed in 21.08 seconds 2025-01-11T22:35:28.302672 - got prompt 2025-01-11T22:35:28.320126 - Downloading yolo model to: D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui-mimicmotionwrapper\models\DWPose2025-01-11T22:35:28.320126 - 2025-01-11T22:35:49.368322 - !!! Exception during processing !!! An error happened while trying to locate the files on the Hub and we cannot find the appropriate snapshot folder for the specified revision on the local disk. Please check your internet connection and try again. 2025-01-11T22:35:49.368322 - Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connection.py", line 174, in _new_conn conn = connection.create_connection( File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\util\connection.py", line 95, in create_connection raise err File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\util\connection.py", line 85, in create_connection sock.connect(sa) TimeoutError: [WinError 10060] 由于连接方在一段时间后没有正确答复或连接的主机没有反应,连接尝试失败。 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connectionpool.py", line 715, in urlopen httplib_response = self._make_request( File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connectionpool.py", line 404, in _make_request self._validate_conn(conn) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connectionpool.py", line 1058, in _validate_conn conn.connect() File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connection.py", line 363, in connect self.sock = conn = self._new_conn() File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connection.py", line 179, in _new_conn raise ConnectTimeoutError( urllib3.exceptions.ConnectTimeoutError: (<urllib3.connection.HTTPSConnection object at 0x0000018F90898730>, 'Connection to huggingface.co timed out. (connect timeout=None)') During handling of the above exception, another exception occurred: Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\requests\adapters.py", line 486, in send resp = conn.urlopen( File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\connectionpool.py", line 799, in urlopen retries = retries.increment( File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\urllib3\util\retry.py", line 592, in increment raise MaxRetryError(_pool, url, error or ResponseError(cause)) urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='huggingface.co', port=443): Max retries exceeded with url: /api/models/yzd-v/DWPose/revision/main (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x0000018F90898730>, 'Connection to huggingface.co timed out. (connect timeout=None)')) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\_snapshot_download.py", line 164, in snapshot_download repo_info = api.repo_info(repo_id=repo_id, repo_type=repo_type, revision=revision, token=token) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\utils\_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\hf_api.py", line 2491, in repo_info return method( File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\utils\_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\hf_api.py", line 2300, in model_info r = get_session().get(path, headers=headers, timeout=timeout, params=params) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\requests\sessions.py", line 602, in get return self.request("GET", url, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\requests\sessions.py", line 589, in request resp = self.send(prep, **send_kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\requests\sessions.py", line 703, in send r = adapter.send(request, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\utils\_http.py", line 66, in send return super().send(request, *args, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\requests\adapters.py", line 507, in send raise ConnectTimeout(e, request=request) requests.exceptions.ConnectTimeout: (MaxRetryError("HTTPSConnectionPool(host='huggingface.co', port=443): Max retries exceeded with url: /api/models/yzd-v/DWPose/revision/main (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0x0000018F90898730>, 'Connection to huggingface.co timed out. (connect timeout=None)'))"), '(Request ID: 00461f6c-5c5d-45ee-9215-21f5d5a4fa7e)') The above exception was the direct cause of the following exception: Traceback (most recent call last): File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\execution.py", line 327, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\execution.py", line 202, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\execution.py", line 174, in _map_node_over_list process_inputs(input_dict, i) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\execution.py", line 163, in process_inputs results.append(getattr(obj, func)(**inputs)) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\custom_nodes\comfyui-mimicmotionwrapper\nodes.py", line 393, in process snapshot_download(repo_id="yzd-v/DWPose", File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\utils\_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "D:\Tools\StableDiffusion\ComfyUI-aki-v1.4\python\lib\site-packages\huggingface_hub\_snapshot_download.py", line 236, in snapshot_download raise LocalEntryNotFoundError( huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the files on the Hub and we cannot find the appropriate snapshot folder for the specified revision on the local disk. Please check your internet connection and try again. 2025-01-11T22:35:49.369825 - Prompt executed in 21.06 seconds ``` ## Attached Workflow Please make sure that workflow does not contain any sensitive information such as API keys or passwords. ``` {"last_node_id":76,"last_link_id":175,"nodes":[{"id":58,"type":"MimicMotionDecode","pos":[1195,247],"size":[255.466796875,78],"flags":{},"order":8,"mode":0,"inputs":[{"name":"mimic_pipeline","type":"MIMICPIPE","link":174,"slot_index":0,"label":"mimic_pipeline"},{"name":"samples","type":"LATENT","link":149,"label":"samples"}],"outputs":[{"name":"images","type":"IMAGE","links":[173],"slot_index":0,"shape":3,"label":"images"}],"properties":{"Node name for S&R":"MimicMotionDecode"},"widgets_values":[4]},{"id":62,"type":"RIFE VFI","pos":[1520,300],"size":[319.20001220703125,198],"flags":{},"order":9,"mode":0,"inputs":[{"name":"frames","type":"IMAGE","link":173,"slot_index":0,"label":"图像"},{"name":"optional_interpolation_states","type":"INTERPOLATION_STATES","link":null,"label":"插值规则(可选)","shape":7}],"outputs":[{"name":"IMAGE","type":"IMAGE","links":[156],"slot_index":0,"shape":3,"label":"图像"}],"properties":{"Node name for S&R":"RIFE VFI"},"widgets_values":["rife47.pth",10,2,true,true,1]},{"id":28,"type":"ImageResizeKJ","pos":[-101,442],"size":[315,242],"flags":{},"order":3,"mode":0,"inputs":[{"name":"image","type":"IMAGE","link":61,"label":"图像"},{"name":"get_image_size","type":"IMAGE","link":null,"label":"参考图像","shape":7},{"name":"width_input","type":"INT","link":null,"widget":{"name":"width_input"},"label":"宽度"},{"name":"height_input","type":"INT","link":null,"widget":{"name":"height_input"},"label":"高度"}],"outputs":[{"name":"IMAGE","type":"IMAGE","links":[110,147],"slot_index":0,"shape":3,"label":"图像"},{"name":"width","type":"INT","links":[88],"slot_index":1,"shape":3,"label":"宽度"},{"name":"height","type":"INT","links":[89],"slot_index":2,"shape":3,"label":"高度"}],"properties":{"Node name for S&R":"ImageResizeKJ"},"widgets_values":[256,456,"lanczos",true,64,0,0,"disabled"]},{"id":35,"type":"ImageResizeKJ","pos":[294,445],"size":[315,242],"flags":{},"order":4,"mode":0,"inputs":[{"name":"image","type":"IMAGE","link":86,"label":"图像"},{"name":"get_image_size","type":"IMAGE","link":null,"label":"参考图像","shape":7},{"name":"width_input","type":"INT","link":88,"widget":{"name":"width_input"},"label":"宽度"},{"name":"height_input","type":"INT","link":89,"widget":{"name":"height_input"},"label":"高度"}],"outputs":[{"name":"IMAGE","type":"IMAGE","links":[111],"slot_index":0,"shape":3,"label":"图像"},{"name":"width","type":"INT","links":null,"shape":3,"label":"宽度"},{"name":"height","type":"INT","links":null,"shape":3,"label":"高度"}],"properties":{"Node name for S&R":"ImageResizeKJ"},"widgets_values":[256,456,"lanczos",false,64,0,0,"disabled"]},{"id":57,"type":"MimicMotionSampler","pos":[1151,465],"size":[315,430],"flags":{},"order":7,"mode":0,"inputs":[{"name":"mimic_pipeline","type":"MIMICPIPE","link":175,"label":"mimic_pipeline"},{"name":"ref_image","type":"IMAGE","link":147,"label":"ref_image"},{"name":"pose_images","type":"IMAGE","link":148,"label":"pose_images"},{"name":"optional_scheduler","type":"DIFFUSERS_SCHEDULER","link":null,"label":"optional_scheduler","shape":7}],"outputs":[{"name":"samples","type":"LATENT","links":[149],"slot_index":0,"shape":3,"label":"samples"}],"title":"MimicMotion Sampler - B站 艾橘溪","properties":{"Node name for S&R":"MimicMotionSampler"},"widgets_values":[26,2,2,42,"fixed",15,0,16,6,false,1,0,1,1]},{"id":9,"type":"GetImageSizeAndCount","pos":[681,653],"size":[210,86],"flags":{},"order":6,"mode":0,"inputs":[{"name":"image","type":"IMAGE","link":112,"label":"图像"}],"outputs":[{"name":"image","type":"IMAGE","links":[148],"slot_index":0,"shape":3,"label":"图像"},{"name":"width","type":"INT","links":[],"slot_index":1,"shape":3,"label":"宽度"},{"name":"height","type":"INT","links":[],"slot_index":2,"shape":3,"label":"高度"},{"name":"count","type":"INT","links":null,"shape":3,"label":"数量"}],"properties":{"Node name for S&R":"GetImageSizeAndCount"},"widgets_values":[]},{"id":63,"type":"ImageScale","pos":[1534,565],"size":[315,130],"flags":{},"order":10,"mode":0,"inputs":[{"name":"image","type":"IMAGE","link":156,"slot_index":0,"label":"图像"}],"outputs":[{"name":"IMAGE","type":"IMAGE","links":[157],"slot_index":0,"shape":3,"label":"图像"}],"properties":{"Node name for S&R":"ImageScale"},"widgets_values":["nearest-exact",256,456,"disabled"]},{"id":74,"type":"DownloadAndLoadMimicMotionModel","pos":[734,444],"size":[315,106],"flags":{},"order":0,"mode":0,"inputs":[],"outputs":[{"name":"mimic_pipeline","type":"MIMICPIPE","links":[174,175],"slot_index":0,"shape":3,"label":"mimic_pipeline"}],"properties":{"Node name for S&R":"DownloadAndLoadMimicMotionModel"},"widgets_values":["MimicMotion-fp16.safetensors","fp16",false]},{"id":59,"type":"VHS_VideoCombine","pos":[1890,585],"size":[362.1622009277344,334],"flags":{},"order":11,"mode":0,"inputs":[{"name":"images","type":"IMAGE","link":157,"slot_index":0,"label":"图像"},{"name":"audio","type":"AUDIO","link":null,"label":"音频","shape":7},{"name":"meta_batch","type":"VHS_BatchManager","link":null,"label":"批次管理","shape":7},{"name":"vae","type":"VAE","link":null,"label":"vae","shape":7}],"outputs":[{"name":"Filenames","type":"VHS_FILENAMES","links":null,"shape":3,"label":"文件名"}],"properties":{"Node name for S&R":"VHS_VideoCombine"},"widgets_values":{"frame_rate":15,"loop_count":0,"filename_prefix":"MimicMotion/1536max","format":"video/h264-mp4","pix_fmt":"yuv420p","crf":19,"save_metadata":true,"trim_to_audio":false,"pingpong":false,"save_output":true,"videopreview":{"hidden":false,"paused":false,"params":{"filename":"1536max_00032.mp4","subfolder":"MimicMotion","type":"output","format":"video/h264-mp4","frame_rate":30}}}},{"id":3,"type":"LoadImage","pos":[-483,537],"size":[280.0934143066406,314],"flags":{},"order":1,"mode":0,"inputs":[],"outputs":[{"name":"IMAGE","type":"IMAGE","links":[61],"slot_index":0,"shape":3,"label":"图像"},{"name":"MASK","type":"MASK","links":null,"shape":3,"label":"遮罩"}],"properties":{"Node name for S&R":"LoadImage"},"widgets_values":["Snipaste_2025-01-07_14-22-47.png","image"]},{"id":5,"type":"VHS_LoadVideo","pos":[-109,761],"size":[235.1999969482422,658.3553466796875],"flags":{},"order":2,"mode":0,"inputs":[{"name":"meta_batch","type":"VHS_BatchManager","link":null,"label":"批次管理","shape":7},{"name":"vae","type":"VAE","link":null,"label":"vae","shape":7}],"outputs":[{"name":"IMAGE","type":"IMAGE","links":[86],"slot_index":0,"shape":3,"label":"图像"},{"name":"frame_count","type":"INT","links":null,"shape":3,"label":"帧计数"},{"name":"audio","type":"AUDIO","links":null,"shape":3,"label":"音频"},{"name":"video_info","type":"VHS_VIDEOINFO","links":null,"shape":3,"label":"视频信息"}],"properties":{"Node name for S&R":"VHS_LoadVideo"},"widgets_values":{"video":"抖音2025111-469578.mp4","force_rate":0,"force_size":"Disabled","custom_width":512,"custom_height":512,"frame_load_cap":30,"skip_first_frames":0,"select_every_nth":2,"choose video to upload":"image","videopreview":{"hidden":false,"paused":false,"params":{"frame_load_cap":30,"skip_first_frames":0,"force_rate":0,"filename":"抖音2025111-469578.mp4","type":"input","format":"video/mp4","select_every_nth":2}}}},{"id":42,"type":"MimicMotionGetPoses","pos":[283.9813537597656,759.572998046875],"size":[330,126],"flags":{},"order":5,"mode":0,"inputs":[{"name":"ref_image","type":"IMAGE","link":110,"label":"ref_image"},{"name":"pose_images","type":"IMAGE","link":111,"label":"pose_images"}],"outputs":[{"name":"poses_with_ref","type":"IMAGE","links":[112],"slot_index":0,"shape":3,"label":"poses_with_ref"},{"name":"pose_images","type":"IMAGE","links":[],"slot_index":1,"shape":3,"label":"pose_images"}],"properties":{"Node name for S&R":"MimicMotionGetPoses"},"widgets_values":[true,true,true]}],"links":[[61,3,0,28,0,"IMAGE"],[86,5,0,35,0,"IMAGE"],[88,28,1,35,2,"INT"],[89,28,2,35,3,"INT"],[110,28,0,42,0,"IMAGE"],[111,35,0,42,1,"IMAGE"],[112,42,0,9,0,"IMAGE"],[147,28,0,57,1,"IMAGE"],[148,9,0,57,2,"IMAGE"],[149,57,0,58,1,"LATENT"],[156,62,0,63,0,"IMAGE"],[157,63,0,59,0,"IMAGE"],[173,58,0,62,0,"IMAGE"],[174,74,0,58,0,"MIMICPIPE"],[175,74,0,57,0,"MIMICPIPE"]],"groups":[],"config":{},"extra":{"ds":{"scale":0.7247295000000635,"offset":[536.9277924284755,-143.65944374289103]},"info":{"name":"workflow","author":"","description":"","version":"1","created":"2024-07-24T10:32:39.441Z","modified":"2024-07-24T10:32:39.442Z","software":"ComfyUI"},"node_versions":{"comfyui-mimicmotionwrapper":"1.0.1","comfyui-frame-interpolation":"1.0.6","comfy-core":"v0.3.10-38-g4209edf4","comfyui-videohelpersuite":"1.4.5"},"VHS_latentpreview":false,"VHS_latentpreviewrate":0,"workspace_info":{"id":"ZRWpTpql4a7QSVdT1zWvh"}},"version":0.4} ``` ## Additional Context (Please add any additional context or steps to reproduce the error here) ``` ### Other _No response_
closed
2025-01-11T14:36:13Z
2025-01-23T11:22:00Z
https://github.com/comfyanonymous/ComfyUI/issues/6430
[ "Custom Nodes Bug" ]
huasio
6
pallets/flask
flask
4,880
Documentation enhancement request
I noticed that although there is a Code of Conduct file available to view in the main GitHub page, it does not seem to be prominently linked in the documentation here on GitHub. It is mentioned in the Discord server if someone chooses to join that, but I feel it could use a bit more visibility. I added a link to the Code of Conduct on both the README and CONTRIBUTING pages. I also updated some of the verbiage in the README sample program section, for clarity for newer users. Full disclosure, this was completed as part of a school project and this is my first Open Source contribution. If anything I submitted looks interesting enough to add, I would love for it to be accepted. Otherwise, no hard feelings and please just discard or reject this request. Thank you for your consideration.
closed
2022-11-27T05:27:17Z
2023-01-10T00:05:54Z
https://github.com/pallets/flask/issues/4880
[]
schmkelsosu
1
ivy-llc/ivy
numpy
27,947
Fix Ivy Failing Test: torch - elementwise.acos
closed
2024-01-17T20:46:12Z
2024-01-22T14:20:54Z
https://github.com/ivy-llc/ivy/issues/27947
[ "Sub Task" ]
samthakur587
0
writer/writer-framework
data-visualization
122
Roadmap
Hi, could you please keep the roadmap updated and make it more detailed because a lot of the change log in each version doesn't even appears on the roadmap. Thanks !
closed
2023-09-20T07:55:57Z
2023-09-21T11:33:51Z
https://github.com/writer/writer-framework/issues/122
[]
PasVegan
1
junyanz/pytorch-CycleGAN-and-pix2pix
computer-vision
1,308
test
test
closed
2021-08-25T02:31:52Z
2021-08-25T02:34:29Z
https://github.com/junyanz/pytorch-CycleGAN-and-pix2pix/issues/1308
[]
cena001plus
0
newpanjing/simpleui
django
64
跨关联关系查询不显示verbose_name值
![122](https://user-images.githubusercontent.com/9565630/58313113-67bd2800-7e3f-11e9-8456-575f266de6bc.png) list_filer 设置ManyToManyField字段查询框不显示verbose_name设置的值
closed
2019-05-24T08:18:32Z
2019-06-03T06:53:06Z
https://github.com/newpanjing/simpleui/issues/64
[ "bug" ]
yanlianhanlin
3
dunossauro/fastapi-do-zero
sqlalchemy
127
Erro em task format (Windows)
Oi, @dunossauro! Em **Configurando o ambiente de desenvolvimento** o comando **task format** está descrito como: `format = 'ruff check . --fix; ruff format .'` Mas ao rodar aqui no Windows ele acusou o seguinte erro: ``` error: unexpected argument '--fix;' found tip: a similar argument exists: '--fix' ``` Não sei se o problema é em alguma configuração na minha máquina, mas o comando é executado ao substituir o separador por **&** ou **&&**. Não entendi muito bem o motivo e se isto seria uma solução mesmo, pois o comando **task lint** também tem este mesmo separador.
closed
2024-04-23T01:56:52Z
2024-04-30T21:39:59Z
https://github.com/dunossauro/fastapi-do-zero/issues/127
[]
vitorfehelberg
1
harry0703/MoneyPrinterTurbo
automation
117
自定义文案,如果文字过多(200多个字),好像会导致语音过长报错
![image](https://github.com/harry0703/MoneyPrinterTurbo/assets/15190634/1a11133c-d579-4abc-a3c0-b492b3cc0517)
closed
2024-03-30T02:18:00Z
2024-04-03T08:46:17Z
https://github.com/harry0703/MoneyPrinterTurbo/issues/117
[ "enhancement" ]
youngphp
5
anselal/antminer-monitor
dash
135
V9 unit is GH/s
Hello V9 unit is GH/s !
closed
2018-10-26T08:52:28Z
2019-12-02T04:56:15Z
https://github.com/anselal/antminer-monitor/issues/135
[ ":dancing_men: duplicate" ]
Roddy1219
3
mljar/mljar-supervised
scikit-learn
301
Store keys only for trained models in mljar tuner
closed
2021-01-18T13:25:14Z
2021-01-21T11:57:35Z
https://github.com/mljar/mljar-supervised/issues/301
[ "bug" ]
pplonski
0
mlfoundations/open_clip
computer-vision
45
Error when training in DataParallel model.
**Hi, after updating to your most recent code, I got an error when training in single machine (8 GPUs) in DataParrallel model. I simply changed the flag args.dp = True and got the following error message:** miniconda3/envs/env37_amp/lib/python3.7/site-packages/torch/nn/parallel/_functions.py:64: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector. warnings.warn('Was asked to gather along dimension 0, but all ' 2022-03-18,06:20:51 | INFO | Start epoch 0 Traceback (most recent call last): File "CLIP_model/training/main.py", line 304, in <module> main() File "CLIP_model/training/main.py", line 243, in main train_one_epoch(model, data, epoch, optimizer, scaler, scheduler, args, writer) File "CLIP_model/training/train.py", line 149, in train_one_epoch total_loss = loss(image_features, text_features, logit_scale) File "miniconda3/envs/env37_amp/lib/python3.7/site-packages/torch/nn/modules/module.py", line 727, in _call_impl result = self.forward(*input, **kwargs) File "CLIP_model/training/train.py", line 97, in forward logits_per_image = logit_scale * image_features @ text_features.T RuntimeError: The size of tensor a (8) must match the size of tensor b (1024) at non-singleton dimension 1 **Code works good when turning args.dp = False and training on a single GPU. Thanks!**
closed
2022-03-17T22:32:05Z
2022-04-06T17:56:06Z
https://github.com/mlfoundations/open_clip/issues/45
[]
wenlinyao
9
dpgaspar/Flask-AppBuilder
rest-api
1,535
Using SQLA Filter `FilterContians`
https://github.com/dpgaspar/Flask-AppBuilder/blob/e21059fd0b8cf10c2eb98549e85e26103ef1de62/flask_appbuilder/models/sqla/filters.py#L114 ```NotImplementedError: <function ilike_op at 0x7fc150244510>``` getting this error on using `FilterContains` to check if a column which stores an array contains an array of users (string), contains a particular user. Is there something I am doing wrong?
closed
2020-12-28T17:48:23Z
2021-06-29T00:56:35Z
https://github.com/dpgaspar/Flask-AppBuilder/issues/1535
[ "stale" ]
kaushalvivek
2
mage-ai/mage-ai
data-science
5,065
[BUG] Clicking the copy icon for a secret does not actually copy secret name to clipboard (Windows 10)
### Mage version 0.9.70 ### Describe the bug Clicking the copy icon for a secret does not actually copy secret name to clipboard. Nothing happens when clicked. Version: 0.9.70 Server OS: Docker on Ubuntu Client OS: Windows 10 Browser: Chrome 124.0.6367.119 ### To reproduce _No response_ ### Expected behavior _No response_ ### Screenshots ![image](https://github.com/mage-ai/mage-ai/assets/17985189/9cf006ca-8316-4316-83a5-1688abe9108d) ### Operating system _No response_ ### Additional context _No response_
closed
2024-05-14T17:40:15Z
2024-05-15T19:54:18Z
https://github.com/mage-ai/mage-ai/issues/5065
[ "bug" ]
amlloyd
1
robinhood/faust
asyncio
380
Permission Denied for creating data directory
## Checklist - [X ] I have verified that the issue persists when using the `master` branch of Faust. ## Steps to reproduce I am currently trying to deploy a faust worker to Kubernetes. Here is my Dockerfile: `# Using lightweight alpine image FROM python:3.6-alpine # Installing packages RUN apk update RUN apk add --no-cache curl python3 pkgconfig snappy-dev python3-dev g++ gcc libxml2-dev libxslt-dev openssl-dev libffi-dev musl-dev make gcc # Defining working directory and adding source code WORKDIR /usr COPY requirements.txt ./ COPY src ./src COPY schemas ./schemas # Install API dependencies RUN pip3 install -r requirements.txt RUN chmod -R 755 . WORKDIR /usr/src CMD faust -A producer -l info worker` ## Expected behavior I expect the worker to run as it does locally ## Actual behavior This is the error I am getting in Kubernetes logs for this pod: `+ƒaµS† v1.7.0-+------------------------------------------------+ | id | producer | | transport | [URL('kafka://ovp.com:9093')] | | store | memory: | | web | http://localhost:6066/ | | log | -stderr- (info) | | pid | 1 | | hostname | faust-demo-producer-master-28-7bdbc9f9f6-qxbh5 | | platform | CPython 3.6.8 (Linux x86_64) | | drivers | | | transport | aiokafka=1.0.3 | | web | aiohttp=3.5.4 | | datadir | /usr/src/producer-data | | appdir | /usr/src/producer-data/v1 | +-------------+------------------------------------------------+ [2019-07-08 21:08:38,677: INFO]: [^Worker]: Starting... [2019-07-08 21:08:38,684: ERROR]: [^Worker]: Error: PermissionError(13, 'Permission denied') Traceback (most recent call last): File "/usr/local/lib/python3.6/site-packages/mode/worker.py", line 264, in execute_from_commandline self.loop.run_until_complete(self._starting_fut) File "/usr/local/lib/python3.6/asyncio/base_events.py", line 484, in run_until_complete return future.result() File "/usr/local/lib/python3.6/site-packages/mode/services.py", line 719, in start await self._default_start() File "/usr/local/lib/python3.6/site-packages/mode/services.py", line 726, in _default_start await self._actually_start() File "/usr/local/lib/python3.6/site-packages/mode/services.py", line 750, in _actually_start await child.maybe_start() File "/usr/local/lib/python3.6/site-packages/mode/services.py", line 778, in maybe_start await self.start() File "/usr/local/lib/python3.6/site-packages/mode/services.py", line 719, in start await self._default_start() File "/usr/local/lib/python3.6/site-packages/mode/services.py", line 726, in _default_start await self._actually_start() File "/usr/local/lib/python3.6/site-packages/mode/services.py", line 734, in _actually_start await self.on_first_start() File "/usr/local/lib/python3.6/site-packages/faust/app/base.py", line 562, in on_first_start self._create_directories() File "/usr/local/lib/python3.6/site-packages/faust/app/base.py", line 1607, in _create_directories self.conf.datadir.mkdir(exist_ok=True) File "/usr/local/lib/python3.6/pathlib.py", line 1248, in mkdir self._accessor.mkdir(self, mode) File "/usr/local/lib/python3.6/pathlib.py", line 387, in wrapped return strfunc(str(pathobj), *args) PermissionError: [Errno 13] Permission denied: 'producer-data'` Any ideas on why I continue to get this error? I have tried variations on the Dockerfile, such as creating the producer-data directory before hand but I am still getting the same error # Versions * Python version * Faust version * Operating system * Kafka version * RocksDB version (if applicable)
closed
2019-07-08T21:23:48Z
2024-04-17T15:24:28Z
https://github.com/robinhood/faust/issues/380
[]
milantx
2
pydantic/pydantic-ai
pydantic
738
Support for not providing a start_node
Currently, when running a Graph, the `start_node` is a required argument. However, `Nodes` are already provided in the initializer of the Graph. Would it be possible that, if there is no `start_node` specified in the run method, It automatically picks the first Node specified during the initialization of the Graph?
closed
2025-01-22T11:31:01Z
2025-01-22T15:58:47Z
https://github.com/pydantic/pydantic-ai/issues/738
[]
tpotjj
5
BeanieODM/beanie
asyncio
493
[BUG] using the Indexed option for a column does not insert data into DB nor does it create an Index
**Describe the bug** When using the Indexed option for a field, that field's data is not written to mongoDB when performing an `insert` operation. Everything works as intended when I created an index using the `IndexModel` object in the `indexes` option. Not only that, the `Indexed` option also does not create an index upon initializing a connection using `init_beanie`. **To Reproduce** ```python # in schema.py class User(Document): """User represents a User of the application.""" name: str = Field(max_length=128) email: EmailStr = Indexed(str, TEXT, unique=True) password: str = Field(max_length=128) date_created: dt.datetime date_updated: dt.datetime class Settings: name = "users" # in api.py async def add_user(): user = User(name="test", [email="testuser@gmail.com",](mailto:email="testuser@gmail.com",) password="123", date_created=dt.now(), date_updated=dt.now()) user_in_db = await user.insert() print(user_in_db) # this prints-- the email field is skipped altogether and no index is created either id=ObjectId('63ee080e0a11d533174eab69') revision_id=None name='test' password='123' date_created=datetime.datetime(2023, 2, 16, 16, 10, 14, 766522) date_updated=datetime.datetime(2023, 2, 16, 16, 10, 14, 766530) ``` **Expected behavior** Expected an index to be present in the mongoDB database instance and with the `email` field. Changing the above code to the below fixes the problem. ```python class User(Document): """User represents a User of the application.""" name: str = Field(max_length=128) email: EmailStr password: str = Field(max_length=128) date_created: dt.datetime date_updated: dt.datetime class Settings: name = "users" indexes = [ IndexModel([("email", TEXT)], name="unique_email", unique=True), ] ``` **Additional context** For reference, I am using Python 3.10.6, MongoDB atlas 5.0.14
closed
2023-02-17T03:07:16Z
2023-05-26T12:31:43Z
https://github.com/BeanieODM/beanie/issues/493
[ "bug" ]
dhruv-ahuja
3
deeppavlov/DeepPavlov
nlp
852
riseapi error
![image](https://user-images.githubusercontent.com/4029753/58387367-259a0f00-8016-11e9-8b5a-66c87b8b6590.png) Keep got error with rise api but command line say that is all good ![image](https://user-images.githubusercontent.com/4029753/58387373-406c8380-8016-11e9-97f9-91276a25a6d1.png) What could be wrong with that?
closed
2019-05-26T21:28:08Z
2019-05-28T22:01:27Z
https://github.com/deeppavlov/DeepPavlov/issues/852
[]
serhio7
2
tableau/server-client-python
rest-api
1,271
custom_view_item.shared always returns false
**Describe the bug** custom_view_item has a **shared** property, but it always shows False because we're not actually parsing that field in from_xml_element(). Adding this in from_xml_element() fixes the issue: `cv_item._shared = str(custom_view_xml.get("shared", None)).lower() == "true"`
closed
2023-08-08T21:51:40Z
2024-01-13T14:56:04Z
https://github.com/tableau/server-client-python/issues/1271
[ "bug", "fixed", "0.24" ]
ryanstryker
0
deepset-ai/haystack
pytorch
9,063
feat: add azure.identity.DefaultAzureCredential authentication to Azure components
**Is your feature request related to a problem? Please describe.** As in #8952 the same authentication mechanism should be applied to (not exhaustive list...): - `AzureOpenAIGenerator` - `AzureOpenAIChatGenerator` - `AzureOpenAIDocumentEmbedder` - `AzureOpenAITextEmbedder` - `AzureAISearchDocumentStore` I know that `AzureAISearchDocumentStore` is in the integrations repo. If you want I can open the same issue there... **Describe the solution you'd like** The same logic implemented for the AzureOpenAIGenerator (great work btw!) detailed in #8952 and implemented in #9012 to support authentication with different methods would be really useful **Describe alternatives you've considered** See #8952
closed
2025-03-19T08:52:39Z
2025-03-19T08:57:28Z
https://github.com/deepset-ai/haystack/issues/9063
[]
vmarchese
2
iMerica/dj-rest-auth
rest-api
622
"Partitioned" attribute needed for the set-cookie header
Hello, As many browsers are rejecting the set-cookie header when SAMESITE is set to none, it would be good if a new attribute "PARTITIONED" is created. Unless, the authentication under this set-up will be impossible in a near future. As far as I see, only Firefox is still supporting set-cookie headers without this parameter. Would it be possible to add this in the next major release?... as well as #584 . It would be extremely helpful. https://developers.google.com/privacy-sandbox/3pcd/chips I am sorry, I am new to Django, so I cannot send a PR (for now). But I am willing to help in any sort of way that I can. Thanks!
open
2024-04-22T16:11:56Z
2024-04-22T16:11:56Z
https://github.com/iMerica/dj-rest-auth/issues/622
[]
Parbelaez
0
graphql-python/graphene-django
graphql
1,542
`in` filters don't decode IDs
**Note: for support questions, please use stackoverflow**. This repository's issues are reserved for feature requests and bug reports. * **What is the current behavior?** Using django-filter for filtering an exact ID match functions well, while filtering for an id `in` some set of IDs does not. For AutoFields and ForeignKeys and the like, when specifying `"id": ["exact", "in"]` as part of a node's `filter_fields`, the `exact` filter works well, while the `in` filter raises exceptions resulting from not decoding global ids. * **If the current behavior is a bug, please provide the steps to reproduce and if possible a minimal demo of the problem** > via a github repo, https://repl.it or similar (you can use this template as a starting point: https://repl.it/@jkimbo/Graphene-Django-Example). If you'd like, I can setup a repl.it example, but for now I'll just illustrate here: ```python class Apple(models.Model): colour = models.CharField(max_length=7) class AppleType(DjangoObjectType): class Meta: model = Apple fields = ("id", "colour") filter_fields = {"id": ["exact", "in"]} interfaces = (graphene.relay.Node,) class Query(graphene.ObjectType): apples = DjangoFilterConnectionField(AppleType) ``` And then a query like ```graphql query getApples($id: ID) { apples(id: $id) { id colour } } ``` works fine, while ```graphql query getApplesIn($id__in: [ID]) { apples(id__in: $id__in) { id colour } } ``` does not. * **What is the expected behavior?** The latter query should return the objects that match the specified list of IDs. * **What is the motivation / use case for changing the behavior?** I think this is a bug/oversight. * **Please tell us about your environment:** - Version: 3.2.2 - Platform: macOS 15.3 (but I've also experienced this on manjaro, this is not OS specific) * **Other information** (e.g. detailed explanation, stacktraces, related issues, suggestions how to fix, links for us to have context, eg. stackoverflow) If this is the type of bug you would fix, I can provide a reproducible example.
open
2025-02-06T11:21:11Z
2025-03-10T09:27:15Z
https://github.com/graphql-python/graphene-django/issues/1542
[ "🐛bug" ]
DrumsnChocolate
0
microsoft/unilm
nlp
1,050
Where to find beit3.spm
Looking for the sentencepiece model for BEIT3, and can't find it anywhere in the docs or repo. Thanks!
closed
2023-04-02T07:48:52Z
2023-04-11T07:30:03Z
https://github.com/microsoft/unilm/issues/1050
[]
nathanodle
2
robinhood/faust
asyncio
42
Internal topics should be prefixed with app id
group_by doesn't use it.
closed
2017-11-07T04:18:16Z
2018-07-31T14:39:12Z
https://github.com/robinhood/faust/issues/42
[ "Issue Type: Enhancement" ]
vineetgoel
1
Miserlou/Zappa
flask
1,994
Drop support for Python versions < 3.6
Python 2.x is now end of life and will no longer receive updates. ## Context Python 2.x and 3.{4, 5} are either no longer supported or don't have official AWS Lambda runtimes. Some additional references in favor of this change: - [Python Clock - Lapsed](https://pythonclock.org/) - [Python 3 Statement](https://python3statement.org/) - [Pandas drops Python 2 support - July 2019](https://pandas.pydata.org/#v0-25-0-final-july-18-2019) - [NumPy drops Python 2 support - Proposed January 2019](https://numpy.org/neps/nep-0014-dropping-python2.7-proposal.html) - [Pytest drops Python 2 support](https://pytest.readthedocs.io/en/latest/py27-py34-deprecation.html) - [PEP-373 - 2.7 Lifecycle](https://www.python.org/dev/peps/pep-0373/) ## Possible Fix I propose we add a deprecation notice in the next patch for 0.49.x and increment the minor (`0.49.0` -> `0.50.0`), dropping support for Python versions `< 3.6` by setting python min requirement, removing six/compat support, and convert relevant code to Python 3.6+.
closed
2020-02-08T18:10:13Z
2020-02-24T13:27:30Z
https://github.com/Miserlou/Zappa/issues/1994
[]
mbeacom
1
miguelgrinberg/python-socketio
asyncio
1,179
Emit With "to" Definition Sometimes Doesn't Work
Hi! I am experiencing an issue in emitting calls from the server to the client. Just to clarify, when I am sending an emit from the client to the server and then, from the event handler on the server, I send an emit back to the client it always arrives at the client. **BUT** When I am trying to emit to the client outside of a client's emit event handler, the emit doesn't arrive at the client every time. After exploring the situation a little further, I've figured out that the session of the socket connection is not available to all the workers of the Flask application. When the emit works it is only when the socket connection is available to the worker that manages the request. Is there a way to solve it? I am running a Flask application with over 150 workers and preload settings. I expect that the socket session will be available to all workers of the application for the purpose of sending an emit from a different worker of the server to the client. Please Advise.
closed
2023-04-30T04:08:10Z
2023-04-30T09:35:33Z
https://github.com/miguelgrinberg/python-socketio/issues/1179
[]
yakirs57
0
OpenInterpreter/open-interpreter
python
603
So, now it wont run
### Describe the bug When i type intrepreter, it says this ![image](https://github.com/KillianLucas/open-interpreter/assets/86132722/f1c02d44-183a-4561-8639-7a62f7b789db) In terminal (specifically powershell) its just ![image](https://github.com/KillianLucas/open-interpreter/assets/86132722/2a596a32-1107-4f5f-93a8-db52b3ea6a42) ### Reproduce Just try to run it in terminal, python 3.11, 3.12 or anything really ### Expected behavior For it to run ### Screenshots _No response_ ### Open Interpreter version 0.1.7 ### Python version 3.12 ### Operating System name and version WIndows 11, latest stable update ### Additional context _No response_
closed
2023-10-08T00:36:10Z
2023-10-25T04:47:37Z
https://github.com/OpenInterpreter/open-interpreter/issues/603
[ "Bug" ]
literallytwo
4
kymatio/kymatio
numpy
980
Converted TFJS-model not working
I am currently trying to convert a simple Keras model with a Scattering1D Layer to TFJS. The conversion seems to work, but when i try to execute the model i get an error: Error: Failed to cast to unknown dtype null Has anyone successfully converted a model which is working in Javascript/Typescript? Any idea what i am doing wrong or what i could try to get it running? Steps to reproduce (im working in windows 10): create new conda env: - conda create -n tfjs_model python=3.8 tensorflow - activate tfjs_model - pip install kymatio tensorflowjs create a simple model with: ``` import tensorflow as tf from kymatio.keras import Scattering1D T = 2**10 J = 5 Q = 8 scattering = Scattering1D(J, Q=Q, max_order=2) model = tf.keras.models.Sequential() model.add(tf.keras.Input(shape=(T))) model.add(scattering) model.compile(metrics='accuracy', loss=None) model.save('scatter_transform') ``` The resulting model can be used in python as expected: ``` import tensorflow as tf model = tf.keras.models.load_model('scatter_transform') model.summary() data_tensor = tf.zeros([1, 1024], tf.float32) res = model.predict(data_tensor) print(res.shape) ``` Convert the model to a tfjs graph model by running: `tensorflowjs_converter scatter_transform scatter_transform_js` I am able to open the resulting model (model.json with one group-shard and a total size of ~2mb), but when i try to execute the model i get the error message: ``` C:\Scripts\typescript test\tensorflow_model\0_bugreport\node_modules\@tensorflow\tfjs-converter\dist\tf-converter.node.js:12600 throw new Error("Failed to cast to unknown dtype ".concat(dtype)); Error: Failed to cast to unknown dtype null at cast_ (C:\Scripts\typescript test\tensorflow_model\0_bugreport\node_modules\@tensorflow\tfjs-converter\dist\tf-converter.node.js:12600:15) at Object.cast__op [as cast] (C:\Scripts\typescript test\tensorflow_model\0_bugreport\node_modules\@tensorflow\tfjs-converter\dist\tf-converter.node.js:11981:29) at executeOp$1 (C:\Scripts\typescript test\tensorflow_model\0_bugreport\node_modules\@tensorflow\tfjs-converter\dist\tf-converter.node.js:29382:25) at C:\Scripts\typescript test\tensorflow_model\0_bugreport\node_modules\@tensorflow\tfjs-converter\dist\tf-converter.node.js:29492:50 at C:\Scripts\typescript test\tensorflow_model\0_bugreport\node_modules\@tensorflow\tfjs-core\dist\tf-core.node.js:4478:22 at Engine.scopedRun (C:\Scripts\typescript test\tensorflow_model\0_bugreport\node_modules\@tensorflow\tfjs-core\dist\tf-core.node.js:4488:23) at Engine.tidy (C:\Scripts\typescript test\tensorflow_model\0_bugreport\node_modules\@tensorflow\tfjs-core\dist\tf-core.node.js:4477:21) at tidy (C:\Scripts\typescript test\tensorflow_model\0_bugreport\node_modules\@tensorflow\tfjs-core\dist\tf-core.node.js:9102:19) at C:\Scripts\typescript test\tensorflow_model\0_bugreport\node_modules\@tensorflow\tfjs-converter\dist\tf-converter.node.js:29492:24 at executeOp (C:\Scripts\typescript test\tensorflow_model\0_bugreport\node_modules\@tensorflow\tfjs-converter\dist\tf-converter.node.js:29508:7) ``` JS-Code: ``` import * as tf from '@tensorflow/tfjs-node' console.log('testing keras model...') async function runModel() { const model_ = await tf.loadGraphModel('file://scatter_transform_js/model.json'); const data = tf.zeros([1, 1024], 'float32'); console.log(data.shape) //console.log(model_) const res = model_.execute(data) console.log((res).arraySync()) console.log('finished') } runModel() ``` with package.json: ``` { "name": "TFJS Scattering1D-Layer", "type": "module", "dependencies": { "@tensorflow/tfjs-node": "^4.1.0" } } ```
closed
2022-12-22T14:00:45Z
2023-02-07T12:11:23Z
https://github.com/kymatio/kymatio/issues/980
[]
martinknr
2
Lightning-AI/pytorch-lightning
machine-learning
20,464
A gracefull design to introduce third-party models as tool for validation
### Description & Motivation python3.10.12 + pytorch_lightning 2.4.0 I need a gracefull design to introduce third-party pretrained models for use during the validation steps. so that there is no such Error reported: ``` RuntimeError: It looks like your LightningModule has parameters that were not used in producing the loss returned by training_step. If this is intentional, you must enable the detection of unused parameters in DDP, .... ``` ### Pitch I am training a model which need other third-party pretrained model during validation. example: the third party model: ``` class PretrainedPicGen(torch.nn.Module): def __init__(self, pretrained_path): self.backbone = load_checkpoint(pretrained_path) def forward(self, to_validate): return self.backbone(to_validate) ``` And the lightning project I am training: ``` class MyModel(pl.LightningModule): def __init__(self, my_param, third_party_pretrained_path): .... self.pretrained_pic_gen = PretrainedPicGen(third_party_pretrained_path) self.validation_outs = [] .... def validation_step(self, batch, *args, **kwargs): validation_output = self.sample(....) self.validation_outputs.append({"vali_out": validation_output}) def on_validation_epoch_end(self) : # Here we use the third party model for post processing the validation out outputs = self.validation_outputs for i, output in enumerate(outputs): visible_output = self.pretrained_pic_gen(output) self.logger.experiment.add_image(f"validate/{i}", visible_output, self.global_step) ``` and the config file yaml: ``` model: class_path: myproject.MyModel init_args: my_param: 1234 third_party_pretrained_path: /path/to/third_party_pretrained ``` but When I run the training, there report the Error information as mentioned before: ``` RuntimeError: It looks like your LightningModule has parameters that were not used in producing the loss returned by training_step. If this is intentional, you must enable the detection of unused parameters in DDP, .... ``` And I think to config the `strategy=ddp_find_unused_parameters_true` may be not good solution, is there any gracefull design here? for example, support extra parameters in the `on_validation_epoch_end` callback and provide a gracefull third_party initialization supported in the config file. ### Alternatives _No response_ ### Additional context _No response_ cc @borda @tchaton @justusschock @awaelchli
open
2024-12-04T12:14:44Z
2024-12-05T13:54:34Z
https://github.com/Lightning-AI/pytorch-lightning/issues/20464
[ "feature", "design" ]
JohnHerry
1
pytorch/pytorch
python
149,227
Unexpected behaviour in using torch.nn.utils.rnn.pack_padded_sequence API
### 🐛 Describe the bug The test cases test_pack_padded_sequence_empty and test_pack_padded_sequence_zero_length expect a valid PackedSequence output for an empty tensor or a sequence with a zero length. This contradicts the underlying behaviour in the current implementation. ```python from torch.nn.utils.rnn import pack_padded_sequence def test_pack_padded_sequence_empty(self): # Test with an empty sequence sequences = torch.tensor([], dtype=torch.float32).reshape(0, 0) lengths = torch.tensor([], dtype=torch.int64) packed = pack_padded_sequence(sequences, lengths, batch_first=True, enforce_sorted=False) self.assertEqual(packed.data.numel(), 0) self.assertEqual(packed.batch_sizes.numel(), 0) def test_pack_padded_sequence_zero_length(self): # Test with a sequence of zero length sequences = torch.tensor([ [0, 0, 0, 0], [1, 2, 3, 0], [4, 5, 0, 0] ], dtype=torch.float32) lengths = torch.tensor([0, 3, 2]) packed = pack_padded_sequence(sequences, lengths, batch_first=True, enforce_sorted=False) self.assertEqual(packed.data.tolist(), [1, 4, 2, 5, 3]) self.assertEqual(packed.batch_sizes.tolist(), [2, 2, 1]) ``` ``` ====================================================================== ERROR: test_pack_padded_sequence_empty (__main__.TestPackPaddedSequence) ---------------------------------------------------------------------- Traceback (most recent call last): File "/home/user/projects/api_guided_testgen/out/bug_detect_gpt4o/exec/zero_shot/torch/torch.nn.utils.rnn.pack_padded_sequence.py", line 71, in test_pack_padded_sequence_empty packed = pack_padded_sequence(sequences, lengths, batch_first=True, enforce_sorted=False) File "/home/user/anaconda3/lib/python3.8/site-packages/torch/nn/utils/rnn.py", line 264, in pack_padded_sequence _VF._pack_padded_sequence(input, lengths, batch_first) RuntimeError: Cannot pack empty tensors. ====================================================================== ERROR: test_pack_padded_sequence_zero_length (__main__.TestPackPaddedSequence) ---------------------------------------------------------------------- Traceback (most recent call last): File "/home/user/projects/api_guided_testgen/out/bug_detect_gpt4o/exec/zero_shot/torch/torch.nn.utils.rnn.pack_padded_sequence.py", line 51, in test_pack_padded_sequence_zero_length packed = pack_padded_sequence(sequences, lengths, batch_first=True, enforce_sorted=False) File "/home/user/anaconda3/lib/python3.8/site-packages/torch/nn/utils/rnn.py", line 264, in pack_padded_sequence _VF._pack_padded_sequence(input, lengths, batch_first) RuntimeError: Length of all samples has to be greater than 0, but found an element in 'lengths' that is <= 0 ``` ### Versions Collecting environment information... PyTorch version: 2.5.0 Is debug build: False CUDA used to build PyTorch: None ROCM used to build PyTorch: N/A OS: Ubuntu 22.04.5 LTS (x86_64) GCC version: (Ubuntu 9.5.0-1ubuntu1~22.04) 9.5.0 Clang version: 14.0.0-1ubuntu1.1 CMake version: version 3.22.1 Libc version: glibc-2.35 Python version: 3.9.21 (main, Dec 11 2024, 16:24:11) [GCC 11.2.0] (64-bit runtime) Python platform: Linux-6.8.0-52-generic-x86_64-with-glibc2.35 Is CUDA available: False CUDA runtime version: No CUDA CUDA_MODULE_LOADING set to: N/A GPU models and configuration: No CUDA Nvidia driver version: No CUDA cuDNN version: No CUDA HIP runtime version: N/A MIOpen runtime version: N/A Is XNNPACK available: True CPU: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Address sizes: 39 bits physical, 48 bits virtual Byte Order: Little Endian CPU(s): 4 On-line CPU(s) list: 0-3 Vendor ID: GenuineIntel Model name: Intel(R) Xeon(R) E-2224G CPU @ 3.50GHz CPU family: 6 Model: 158 Thread(s) per core: 1 Core(s) per socket: 4 Socket(s): 1 Stepping: 10 CPU max MHz: 4700.0000 CPU min MHz: 800.0000 BogoMIPS: 6999.82 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch cpuid_fault epb pti ssbd ibrs ibpb stibp tpr_shadow flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid mpx rdseed adx smap clflushopt intel_pt xsaveopt xsavec xgetbv1 xsaves dtherm ida arat pln pts hwp hwp_notify hwp_act_window hwp_epp vnmi md_clear flush_l1d arch_capabilities Virtualization: VT-x L1d cache: 128 KiB (4 instances) L1i cache: 128 KiB (4 instances) L2 cache: 1 MiB (4 instances) L3 cache: 8 MiB (1 instance) NUMA node(s): 1 NUMA node0 CPU(s): 0-3 Vulnerability Gather data sampling: Mitigation; Microcode Vulnerability Itlb multihit: KVM: Mitigation: VMX disabled Vulnerability L1tf: Mitigation; PTE Inversion; VMX conditional cache flushes, SMT disabled Vulnerability Mds: Mitigation; Clear CPU buffers; SMT disabled Vulnerability Meltdown: Mitigation; PTI Vulnerability Mmio stale data: Mitigation; Clear CPU buffers; SMT disabled Vulnerability Reg file data sampling: Not affected Vulnerability Retbleed: Mitigation; IBRS Vulnerability Spec rstack overflow: Not affected Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization Vulnerability Spectre v2: Mitigation; IBRS; IBPB conditional; STIBP disabled; RSB filling; PBRSB-eIBRS Not affected; BHI Not affected Vulnerability Srbds: Mitigation; Microcode Vulnerability Tsx async abort: Mitigation; TSX disabled Versions of relevant libraries: [pip3] numpy==2.0.1 [pip3] torch==2.5.0 [pip3] torchaudio==2.5.0 [pip3] torchvision==0.20.0 [conda] blas 1.0 mkl [conda] cpuonly 2.0 0 pytorch [conda] ffmpeg 4.3 hf484d3e_0 pytorch [conda] libjpeg-turbo 2.0.0 h9bf148f_0 pytorch [conda] mkl 2023.1.0 h213fc3f_46344 [conda] mkl-service 2.4.0 py39h5eee18b_2 [conda] mkl_fft 1.3.11 py39h5eee18b_0 [conda] mkl_random 1.2.8 py39h1128e8f_0 [conda] numpy 2.0.1 py39h5f9d8c6_1 [conda] numpy-base 2.0.1 py39hb5e798b_1 [conda] pytorch 2.5.0 py3.9_cpu_0 pytorch [conda] pytorch-mutex 1.0 cpu pytorch [conda] torchaudio 2.5.0 py39_cpu pytorch [conda] torchvision 0.20.0 py39_cpu pytorch cc @mikaylagawarecki
open
2025-03-14T21:05:31Z
2025-03-17T15:25:26Z
https://github.com/pytorch/pytorch/issues/149227
[ "module: rnn", "triaged" ]
sjh0849
0
huggingface/diffusers
deep-learning
10,672
Please support callback_on_step_end for following pipelines
**Is your feature request related to a problem? Please describe.** Missing callback_on_step_end in these pipeline takes away the capability to show the progress in UI **Describe the solution you'd like.** Please support callback_on_step_end **Describe alternatives you've considered.** N.A. **Additional context.** 1. AuraFlowPipeline TypeError: AuraFlowPipeline.__call__() got an unexpected keyword argument 'callback_on_step_end' 2. LuminaText2ImgPipeline
closed
2025-01-28T16:26:56Z
2025-02-16T17:28:58Z
https://github.com/huggingface/diffusers/issues/10672
[ "good first issue", "help wanted", "contributions-welcome" ]
nitinmukesh
2
serengil/deepface
machine-learning
763
GPU too slow when first build pkl
I'm using laptop (i7-9850h, quadro t1000, 16gb ram) to demo app with flask. At first I planned to use pure windows but had an error in installing cuda, so I switched to wsl2. But when running with gpu in wsl2, the first time build pkl is too slower than build it in cpu. And after that (built pkl), gpu is faster than cpu, but not too much. My db_path includes 62 images, classified into 4 categories. I use Facenet512, mtcnn. I would really appreciate someone explaining why and maybe give me some solution for this problem Sorry about my English skills
closed
2023-05-31T18:36:24Z
2023-05-31T22:33:50Z
https://github.com/serengil/deepface/issues/763
[ "dependencies" ]
MiMi-Yup
1
bregman-arie/devops-exercises
python
4
Add questions on Grafana
closed
2019-10-04T15:54:33Z
2022-07-10T18:43:13Z
https://github.com/bregman-arie/devops-exercises/issues/4
[]
bregman-arie
5
geex-arts/django-jet
django
525
Logout error 405 using Jet, but logout normal without jet
Hello, this is my first time using Jet in django. And I immediately tried it on a new project, fresh, without installing any other packages first. I want to check the Jet interface first on my admin. Before I used Jet, I was able to login and logout normally (still with the standard Django admin). But after I activated Jet, the Django admin can login, with a nice view, but there's an error when logging out. There's a 405 error when logging out. Has anyone else experienced the same thing? Here's the code I wrote: Error Debug Code: [11/Feb/2024 05:28:39] "GET /jet/jsi18n/ HTTP/1.1" 200 3342 [11/Feb/2024 05:28:39] "GET /jet/jsi18n/ HTTP/1.1" 200 3342 Method Not Allowed (GET): /admin/logout/ Method Not Allowed: /admin/logout/ [11/Feb/2024 05:28:41] "GET /admin/logout/ HTTP/1.1" 405 0 File URLS.PY from django.contrib import admin from django.urls import path, include urlpatterns = [ path('jet/', include('jet.urls', 'jet')), # Django JET URLS path('jet/dashboard/', include('jet.dashboard.urls', 'jet-dashboard')), # Django JET dashboard URLS path('admin/', admin.site.urls), ] File Setting.py INSTALLED_APPS = [ 'jet.dashboard', 'jet', 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', ] MIDDLEWARE = [ 'django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', ] ROOT_URLCONF = 'testjet.urls' TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'DIRS': [], 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.debug', 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ], }, }, ] The rest of the code hasn't been changed because it's still a standard new project from Django. And I have tested this on two browsers, Firefox and the latest version of Edge. The result remains the same. Logout Error 405 if using Jet. But normal if without Jet.
open
2024-02-11T05:39:19Z
2024-02-11T05:39:19Z
https://github.com/geex-arts/django-jet/issues/525
[]
hapesurya
0
kornia/kornia
computer-vision
3,014
Test on Windows are failing because of no numpy.
LGTM. Test on Windows are failing because of no numpy. _Originally posted by @edgarriba in https://github.com/kornia/kornia/pull/3012#pullrequestreview-2289278611_
closed
2024-09-09T08:53:17Z
2024-09-09T23:44:12Z
https://github.com/kornia/kornia/issues/3014
[]
edgarriba
1
sammchardy/python-binance
api
706
help not worc close all orders
not worc my code in python close all orders python not expert to my - Python version: [e.g. 3.5] - OS: windows 10 - python-binance python_binance-0.7.9-py2.py3-none-any.whl from binance.client import Client from binance.websockets import BinanceSocketManager import time import math api_key = 'IIIIIIIIIIIIIIIIIIII' api_secret = 'IIIIIIIIIIIIIIIIIIII' client = Client(api_key, api_secret) ASSET = 'ZECUSDT' CURRENCY = 'ZEC' CRYPTOCURRENCY = 'USDT' def balance(symbol): balance = client.get_asset_balance(asset=symbol) balance = {'free': balance['free'], 'locked':balance['locked']} return balance def history(symbol): history = client.get_my_trades(symbol=symbol) return history def price(symbol): price = client.get_avg_price(symbol=symbol)['price'] return float(price) def order_market_buy(quantity): order = client.order_market_buy(symbol=ASSET, quantity=quantity) def order_market_sell(quantity): order = client.order_market_sell(symbol=ASSET, quantity=quantity) #!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! NOT WORC! def cancel_orders(self, **params): return self._delete('openOrders', True, data=params) ################# PRICE = price(ASSET) def main(FIRST_PRICE): print('СЕЙЧАС ЦЕНА',int(PRICE)) nteu = PRICE numbermu = 12 rezultae = nteu - numbermu okruglenie = round(rezultae, 2) print('ЗА СКОЛЬКО КУПИТЬ!', + okruglenie ) nteuz = PRICE numbermuz = 14 rezultaez = nteuz + numbermuz okrugleniez = round(rezultaez, 2) print('ЗА СКОЛЬКО ПРОДАТЬ!', + okrugleniez ) rtytrr = 170 print(okrugleniez) ###################################### #!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! NOT WORC! client.cancel_orders(symbol='ZECUSDT') ###################################### START_PRICE = price(ASSET) main(START_PRICE)
closed
2021-02-26T09:19:20Z
2021-02-26T11:35:33Z
https://github.com/sammchardy/python-binance/issues/706
[]
G34567
0
Lightning-AI/pytorch-lightning
pytorch
19,861
Validation dataloader is added to train dataloader after first epoch
### Bug description Validation dataloader is added to train dataloader after first epoch. What the f*ck ? ### What version are you seeing the problem on? v1.8 ### How to reproduce the bug ```python trainer.fit(model, train_dataloader, valid_dataloader) ``` ### Error messages and logs ``` # Error messages and logs here please ``` ewgtwag ### Environment <details> <summary>Current environment</summary> ``` #- Lightning Component (e.g. Trainer, LightningModule, LightningApp, LightningWork, LightningFlow): #- PyTorch Lightning Version (e.g., 1.5.0): #- Lightning App Version (e.g., 0.5.2): #- PyTorch Version (e.g., 2.0): #- Python version (e.g., 3.9): #- OS (e.g., Linux): #- CUDA/cuDNN version: #- GPU models and configuration: #- How you installed Lightning(`conda`, `pip`, source): #- Running environment of LightningApp (e.g. local, cloud): ``` </details> egewsg ### More info ![image](https://github.com/Lightning-AI/pytorch-lightning/assets/65295663/b61f4697-83c9-4eaa-ad60-13b2b9d85931)
closed
2024-05-12T15:55:37Z
2024-05-12T16:57:23Z
https://github.com/Lightning-AI/pytorch-lightning/issues/19861
[ "bug", "needs triage", "ver: 1.8.x" ]
AndreyStille
0
SALib/SALib
numpy
121
Add groups functionality for Sobol
Add a "groups" functionality for Sobol. @willu47 added a really nice feature for the Morris method to read groups from the parameter file, and it would be great to extend this to other methods that can handle groups. For Sobol, the biggest difference would be in the cross-sampling.
closed
2017-01-03T11:25:37Z
2017-01-03T17:54:09Z
https://github.com/SALib/SALib/issues/121
[ "enhancement" ]
willu47
1
microsoft/unilm
nlp
805
Model I am using is LayoutLMV2 large.
**Describe the bug** Model I am using is microsoft/layoutlmv2-large-uncased. I am following this link [https://towardsdatascience.com/fine-tuning-layoutlm-v2-for-invoice-recognition-91bf2546b19e](url) for fine tuning and inferencing LayoutLMV2 base uncased for document data extraction. The problem arises when using: I changed the model name present in the training.py, inferencing.py and preprocess.py from "layoutlmv2 base uncased" to "microsoft/layoutlmv2-large-uncased" in order to use the large model. A clear and concise description of what the bug is. After changing the name of the model in the 3 scripts, when I start training, as soon as the training starts, it stops abruptly. The output also shows ^C as if someone has stopped the training. But I am not terminating the training. **To Reproduce** Steps to reproduce the behavior: 1. Change the version of datasets by replacing that line by "!pip install -q datasets==2.2.2 seqeval". 2. In the 3 scripts (train.py, preprocess.py and inference.py) change the model name from "layoutlmv2 base uncased" to "microsoft/layoutlmv2-large-uncased" 3. Simply start the training (make sure the batch size is either 1 or 2 as Colab does not enough memory for larger batch size) **Expected behavior** A clear and concise description of what you expected to happen. As soon as the training starts, it will terminate itself and in the corresponding output you can see the ^C symbol as if we are terminatingthe training - Platform: Google Colab - Python version: default version used in the colab notebook/requirements.txt - PyTorch version (GPU?): default version used in the colab notebook/requirements.txt
open
2022-07-27T05:36:08Z
2022-07-27T05:53:04Z
https://github.com/microsoft/unilm/issues/805
[]
samyakai
0
apache/airflow
data-science
47,948
ExternalTaskSensor with mode='reschedule' raising module 'airflow.settings' has no attribute 'engine' error
### Apache Airflow version main (development) ### If "Other Airflow 2 version" selected, which one? _No response_ ### What happened? ExternalTaskSensor with mode='reschedule' raising module 'airflow.settings' has no attribute 'engine' error Traceback (most recent call last): File "/opt/airflow/airflow/sensors/base.py", line 206, in _validate_input_values if self.reschedule and _is_metadatabase_mysql(): File "/opt/airflow/airflow/sensors/base.py", line 60, in _is_metadatabase_mysql if settings.engine is None: AttributeError: module 'airflow.settings' has no attribute 'engine' ### What you think should happen instead? _No response_ ### How to reproduce Have the below Dag in dags folder and notice import error: ```python from airflow import DAG from airflow.providers.standard.operators.empty import EmptyOperator from airflow.sensors.external_task import ExternalTaskMarker, ExternalTaskSensor from pendulum import today start_date = today('UTC').add(days=-1) with DAG( dag_id="example_external_task_marker_parent", start_date=start_date, schedule=None, tags=["core"], ) as parent_dag: # [START howto_operator_external_task_marker] parent_task = ExternalTaskMarker( task_id="parent_task", external_dag_id="example_external_task_marker_child", external_task_id="child_task1", ) # [END howto_operator_external_task_marker] with DAG( dag_id="example_external_task_marker_child", start_date=start_date, schedule=None, tags=["core"], ) as child_dag: # [START howto_operator_external_task_sensor] child_task1 = ExternalTaskSensor( task_id="child_task1", external_dag_id=parent_dag.dag_id, external_task_id=parent_task.task_id, timeout=600, allowed_states=["success"], failed_states=["failed", "skipped"], mode="reschedule", ) # [END howto_operator_external_task_sensor] child_task2 = EmptyOperator(task_id="child_task2") child_task1 >> child_task2 ``` ### Operating System Linux ### Versions of Apache Airflow Providers _No response_ ### Deployment Other ### Deployment details _No response_ ### Anything else? _No response_ ### Are you willing to submit PR? - [ ] Yes I am willing to submit a PR! ### Code of Conduct - [x] I agree to follow this project's [Code of Conduct](https://github.com/apache/airflow/blob/main/CODE_OF_CONDUCT.md)
open
2025-03-19T11:01:00Z
2025-03-23T17:07:01Z
https://github.com/apache/airflow/issues/47948
[ "kind:bug", "priority:medium", "area:core", "affected_version:3.0.0beta" ]
atul-astronomer
1
Asabeneh/30-Days-Of-Python
numpy
543
Online playground
Great work. Thank you. I suggest adding links to run code samples in an online playground where users can run the code, edit it, save for later and share with others, etc. This are some examples: - Run functions and print output: [Run in playground](https://livecodes.io/?x=id/bzcdxxysijd) ``` python def greetings (name = 'Peter'): message = name + ', welcome to Python for Everyone!' return message print(greetings()) print(greetings('Asabeneh')) def generate_full_name (first_name = 'Asabeneh', last_name = 'Yetayeh'): space = ' ' full_name = first_name + space + last_name return full_name print(generate_full_name()) print(generate_full_name('David','Smith')) def calculate_age (birth_year,current_year = 2021): age = current_year - birth_year return age; print('Age: ', calculate_age(1821)) def weight_of_object (mass, gravity = 9.81): weight = str(mass * gravity)+ ' N' # the value has to be changed to string first return weight print('Weight of an object in Newtons: ', weight_of_object(100)) # 9.81 - average gravity on Earth's surface print('Weight of an object in Newtons: ', weight_of_object(100, 1.62)) # gravity on the surface of the Moon ``` - Formatting Date Output Using strftime: [Run in playground](https://livecodes.io/?x=id/7ue3z2srw8q) ```python from datetime import datetime new_year = datetime(2020, 1, 1) print(new_year) # 2020-01-01 00:00:00 day = new_year.day month = new_year.month year = new_year.year hour = new_year.hour minute = new_year.minute second = new_year.second print(day, month, year, hour, minute) #1 1 2020 0 0 print(f'{day}/{month}/{year}, {hour}:{minute}') # 1/1/2020, 0:0 ``` - Statistics [Run in playground](https://livecodes.io/?x=id/zhk8mwcky44) ```python import numpy as np from scipy import stats np_normal_dis = np.random.normal(5, 0.5, 1000) # mean, standard deviation, number of samples np_normal_dis ## min, max, mean, median, sd print('min: ', np.min(np_normal_dis)) print('max: ', np.max(np_normal_dis)) print('mean: ', np.mean(np_normal_dis)) print('median: ', np.median(np_normal_dis)) print('mode: ', stats.mode(np_normal_dis)) print('sd: ', np.std(np_normal_dis)) ``` - DataFrames [Run in playground](https://livecodes.io/?x=id/kqbjgzahnpv) ```python import pandas as pd import numpy as np data = [ {"Name": "Asabeneh", "Country":"Finland","City":"Helsinki"}, {"Name": "David", "Country":"UK","City":"London"}, {"Name": "John", "Country":"Sweden","City":"Stockholm"}] df = pd.DataFrame(data) print(df) ``` If you are interested, I can start a PR to add links for running code snippets in the playground. This playground is [LiveCodes](https://livecodes.io/), a [feature-rich](https://livecodes.io/docs/features/), [open-source](https://github.com/live-codes/livecodes), [client-side](https://livecodes.io/docs/why#client-side) code playground that supports [80+ languages/frameworks](https://livecodes.io/docs/why#language-support). The playground can be shared, exported (e.g. to GitHub gists), deployed (to GitHub Pages) and [embedded](https://livecodes.io/docs/features/embeds) in web pages using a powerful [SDK](https://livecodes.io/docs/sdk). LiveCodes runs completely on the client-side (with no backend). It uses [Brython](https://brython.info/) or [Pyodide](https://pyodide.org/) to run Python in the browser. App: https://livecodes.io/ Docs: https://livecodes.io/docs/ GitHub repo: https://github.com/live-codes/livecodes Disclosure: I'm the author of LiveCodes.
open
2024-07-03T04:25:40Z
2024-07-04T07:25:06Z
https://github.com/Asabeneh/30-Days-Of-Python/issues/543
[]
hatemhosny
2
docarray/docarray
fastapi
1,243
Make `search_field` default to empty string
`search_field` in `find()`, `text_search()` should default to `''`. Three reasons for that: - In some backends (concrete example: Weaviate), the field to perform similarity search on is known a priori (or at least during schema creation time), or even a reserved name that never changes. In those cases it does not make sense for the user to pass a `search_field` argument - If we default to `embedding` like we currently do, we make it simple for people that use our pre-defined documents (which all have an `embedding`) field, **but** we don't teach users the proper usage and flexibility that is required when they create their own schema. I think nudging users into that direction is a good thing - Why not `Optional` and default to `None`? I think for many backends specifying a `search_field` will still be required, and making it `Optional` would send a wrong message. To me, having it non-optional but defaulting to `''` strikes a good balance. But I could be convinced otherwise.
closed
2023-03-15T15:10:18Z
2023-03-17T07:52:27Z
https://github.com/docarray/docarray/issues/1243
[]
JohannesMessner
1
graphql-python/graphene
graphql
1,389
Tutorial mistakes
Hey :). Don't know where else to post this so... Found 2 mistakes in this tutorial: https://docs.graphene-python.org/projects/django/en/latest/tutorial-plain/ Where it says: "Now sync your database for the first time:" There is no manage.py in that directory. One must go back one directory. So, `cd ..` first, and then execute the `python manage.py` migrate command. Where it says: "Don’t forget to create & run migrations:" This will fail because in `class IngredientsConfig(AppConfig): ` we have `name = 'ingredients'`. The name is wrong. It should be `name = 'cookbook.ingredients'`
open
2021-11-13T13:06:58Z
2025-01-06T10:37:13Z
https://github.com/graphql-python/graphene/issues/1389
[ "🐛 bug" ]
Kyliathy
3
chainer/chainer
numpy
8,390
Flaky test: chainer_tests/distributions_tests/test_utils.py::ModifiedXLogX
From #8371 [Travis](https://travis-ci.org/chainer/chainer/jobs/607597850?utm_medium=notification&utm_source=github_status) ``` ______________________ TestModifiedXLogX_param_0_{dtype=float32, shape=(2, 3)}.test_backward_cpu _______________________ [2019-11-05 11:58:46] [2019-11-05 11:58:46] self = <chainer.testing._bundle.TestModifiedXLogX_param_0_{dtype=float32, shape=(2, 3)} testMethod=test_backward_cpu> [2019-11-05 11:58:46] [2019-11-05 11:58:46] def test_backward_cpu(self): [2019-11-05 11:58:46] > self.check_backward(self.x, self.gy) [2019-11-05 11:58:46] [2019-11-05 11:58:46] tests/chainer_tests/distributions_tests/test_utils.py:43: [2019-11-05 11:58:46] _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ [2019-11-05 11:58:46] tests/chainer_tests/distributions_tests/test_utils.py:40: in check_backward [2019-11-05 11:58:46] x_data, y_grad, **self.backward_options) [2019-11-05 11:58:46] ../../../virtualenv/python3.5.6/lib/python3.5/site-packages/chainer/gradient_check.py:944: in check_backward [2019-11-05 11:58:46] detect_nondifferentiable, is_immutable_params=False [2019-11-05 11:58:46] ../../../virtualenv/python3.5.6/lib/python3.5/site-packages/chainer/gradient_check.py:463: in run [2019-11-05 11:58:46] self._run() [2019-11-05 11:58:46] ../../../virtualenv/python3.5.6/lib/python3.5/site-packages/chainer/gradient_check.py:506: in _run [2019-11-05 11:58:46] self._compare_gradients(gx_numeric, gx_backward, directions) [2019-11-05 11:58:46] _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ [2019-11-05 11:58:46] [2019-11-05 11:58:46] self = <chainer.gradient_check._CheckBackward object at 0x7f8018f87f98>, gx_numeric = array(0.06594739) [2019-11-05 11:58:46] gx_backward = 0.07755014807363542 [2019-11-05 11:58:46] directions = (array([[-0.23456502, -0.76281573, 0.4189689 ], [2019-11-05 11:58:46] [-0.14629427, -0.31878032, -0.25403458]]),) [2019-11-05 11:58:46] [2019-11-05 11:58:46] def _compare_gradients(self, gx_numeric, gx_backward, directions): [2019-11-05 11:58:46] atol = self.atol [2019-11-05 11:58:46] rtol = self.rtol [2019-11-05 11:58:46] # Compare the gradients [2019-11-05 11:58:46] try: [2019-11-05 11:58:46] testing.assert_allclose( [2019-11-05 11:58:46] gx_numeric, gx_backward, atol=atol, rtol=rtol) [2019-11-05 11:58:46] except AssertionError as e: [2019-11-05 11:58:46] eps = self.eps [2019-11-05 11:58:46] xs = self.xs [2019-11-05 11:58:46] gys = self.gys [2019-11-05 11:58:46] f = six.StringIO() [2019-11-05 11:58:46] f.write('check_backward failed (eps={} atol={} rtol={})\n'.format( [2019-11-05 11:58:46] eps, atol, rtol)) [2019-11-05 11:58:46] for i, x in enumerate(xs): [2019-11-05 11:58:46] f.write('inputs[{}]:\n'.format(i)) [2019-11-05 11:58:46] f.write('{}\n'.format(x)) [2019-11-05 11:58:46] for i, gy in enumerate(gys): [2019-11-05 11:58:46] f.write('grad_outputs[{}]:\n'.format(i)) [2019-11-05 11:58:46] f.write('{}\n'.format(gy)) [2019-11-05 11:58:46] for i, d in enumerate(directions): [2019-11-05 11:58:46] f.write('directions[{}]:\n'.format(i)) [2019-11-05 11:58:46] f.write('{}\n'.format(d)) [2019-11-05 11:58:46] f.write('gradients (numeric): {}\n'.format(gx_numeric)) [2019-11-05 11:58:46] f.write('gradients (backward): {}\n'.format(gx_backward)) [2019-11-05 11:58:46] f.write('\n') [2019-11-05 11:58:46] f.write('x: numeric gradient, y: backward gradient') [2019-11-05 11:58:46] f.write(str(e)) [2019-11-05 11:58:46] > raise AssertionError(f.getvalue()) [2019-11-05 11:58:46] E AssertionError: Parameterized test failed. [2019-11-05 11:58:46] E [2019-11-05 11:58:46] E Base test method: TestModifiedXLogX.test_backward_cpu [2019-11-05 11:58:46] E Test parameters: [2019-11-05 11:58:46] E dtype: <class 'numpy.float32'> [2019-11-05 11:58:46] E shape: (2, 3) [2019-11-05 11:58:46] E [2019-11-05 11:58:46] E [2019-11-05 11:58:46] E (caused by) [2019-11-05 11:58:46] E AssertionError: check_backward failed (eps=1e-05 atol=0.01 rtol=0.01) [2019-11-05 11:58:46] E inputs[0]: [2019-11-05 11:58:46] E [[0.2653617 0.97955227 0.78702766] [2019-11-05 11:58:46] E [0.79942065 0.8388158 0.6125239 ]] [2019-11-05 11:58:46] E grad_outputs[0]: [2019-11-05 11:58:46] E [[-0.03339154 0.66924393 1.28142 ] [2019-11-05 11:58:46] E [ 0.7084703 -1.9439045 1.996255 ]] [2019-11-05 11:58:46] E directions[0]: [2019-11-05 11:58:46] E [[-0.23456502 -0.76281573 0.4189689 ] [2019-11-05 11:58:46] E [-0.14629427 -0.31878032 -0.25403458]] [2019-11-05 11:58:46] E gradients (numeric): 0.06594738676257705 [2019-11-05 11:58:46] E gradients (backward): 0.07755014807363542 [2019-11-05 11:58:46] E [2019-11-05 11:58:46] E x: numeric gradient, y: backward gradient [2019-11-05 11:58:46] E Not equal to tolerance rtol=0.01, atol=0.01 [2019-11-05 11:58:46] E [2019-11-05 11:58:46] E (mismatch 100.0%) [2019-11-05 11:58:46] E x: array(0.065947) [2019-11-05 11:58:46] E y: array(0.07755) [2019-11-05 11:58:46] E [2019-11-05 11:58:46] E assert_allclose failed: [2019-11-05 11:58:46] E shape: () () [2019-11-05 11:58:46] E dtype: float64 float64 [2019-11-05 11:58:46] E i: (0,) [2019-11-05 11:58:46] E x[i]: 0.06594738676257705 [2019-11-05 11:58:46] E y[i]: 0.07755014807363542 [2019-11-05 11:58:46] E relative error[i]: 0.1496162367097134 [2019-11-05 11:58:46] E absolute error[i]: 0.011602761311058363 [2019-11-05 11:58:46] E relative tolerance * |y[i]|: 0.0007755014807363542 [2019-11-05 11:58:46] E absolute tolerance: 0.01 [2019-11-05 11:58:46] E total tolerance: 0.010775501480736355 [2019-11-05 11:58:46] E x: 0.06594739 [2019-11-05 11:58:46] E y: 0.07755015 [2019-11-05 11:58:46] [2019-11-05 11:58:46] ../../../virtualenv/python3.5.6/lib/python3.5/site-packages/chainer/gradient_check.py:536: AssertionError ```
closed
2019-11-08T01:15:49Z
2019-11-28T10:44:30Z
https://github.com/chainer/chainer/issues/8390
[ "cat:test", "prio:high" ]
emcastillo
3
s3rius/FastAPI-template
asyncio
30
Add AWS cloud deployment.
We need to create an AWS configuration for cloud deploy. I guess, that terraform is a really good way to create infrastructure. In this thread we must choose the best way to deploy a cloud app. Currently I'm thinking about creating REST Gateway. But maybe someone would like to choose something else. This feature is planned for release/3.3.0.
open
2021-10-04T06:37:37Z
2023-11-21T01:18:18Z
https://github.com/s3rius/FastAPI-template/issues/30
[ "help wanted" ]
s3rius
11
Python3WebSpider/ProxyPool
flask
197
怎么解决python11下运行run,py后报ImportError: cannot import name 'ProxySchemeUnsupported' from 'requests.packages.urllib3.exceptions'的问题?
我在网上查了一下好像是Python3.6以后,在使用代理时,requests.get(url=url, headers=headers, proxies=…)中proxies的参数值发生了变化,3.6包括之前,proxies={‘https’: ‘127.0.0.1:8080’}或者proxies={‘http’: ‘127.0.0.1:8080’}即可,但是这样的字典类型并不适用于Python3.7及以上的版本。在Python3.7及以上版本,必须要在ip:port前面加上http://或者https://,不能去掉前面的http://或者https://,即Python3.7后必须使用proxies={‘http’: ‘http://127.0.0.1:8080’}或者proxies={‘https’: ‘https://127.0.0.1:8080’}。 我是初学者,实在不太懂在ProxyPool中哪里可以修改这个代码解决这个问题,望提供解决方案!
closed
2023-07-13T09:14:47Z
2023-07-13T13:58:37Z
https://github.com/Python3WebSpider/ProxyPool/issues/197
[ "bug" ]
2793525733
2
pydantic/pydantic
pydantic
11,041
Error message for return value in `validate_call` says "input"
### Initial Checks - [X] I confirm that I'm using Pydantic V2 ### Description I had a function like: ```py In [1]: from pydantic import ValidationError, validate_call ...: @validate_call(validate_return=True) ...: def abc(s: str, b: int) -> str ...: return 1 ``` and I ran it with: `abc('a', 1)` to get the following error: ```py In [2]: abc('a', 1) --------------------------------------------------------------------------- ValidationError Traceback (most recent call last) File ~/venv3.11/lib/python3.11/site-packages/pydantic/validate_call_decorator.py:60, in validate_call.<locals>.validate.<locals>.wrapper_function(*args, **kwargs) 58 @functools.wraps(function) 59 def wrapper_function(*args, **kwargs): ---> 60 return validate_call_wrapper(*args, **kwargs) File ~/venv3.11/lib/python3.11/site-packages/pydantic/_internal/_validate_call.py:98, in ValidateCallWrapper.__call__(self, *args, **kwargs) 96 res = self.__pydantic_validator__.validate_python(pydantic_core.ArgsKwargs(args, kwargs)) 97 if self.__return_pydantic_validator__: ---> 98 return self.__return_pydantic_validator__(res) 99 return res ValidationError: 1 validation error for abc Input should be a valid string [type=string_type, input_value=1, input_type=int] For further information visit https://errors.pydantic.dev/2.9/v/bytes_type ``` It took me waaay too long to figure out that this error which was saying "Input should be ..." actual meant "Return value should be ..." In addition, I feel the error got me even more confused because errors in arguments don't show the argument name currently - Related to: https://github.com/pydantic/pydantic/issues/6791 ### Example Code ```Python from pydantic import ValidationError, validate_call @validate_call(validate_return=True) def abc(s: str, b: int) -> str return 1 abc('a', 1) ``` ### Python, Pydantic & OS Version ```Text pydantic version: 2.9.2 pydantic-core version: 2.23.4 pydantic-core build: profile=release pgo=false install path: /home/ali/venv3.11/lib/python3.11/site-packages/pydantic python version: 3.11.9 (main, Aug 14 2024, 05:07:28) [Clang 18.1.8 ] platform: Linux-5.15.146.1-microsoft-standard-WSL2-x86_64-with-glibc2.39 related packages: typing_extensions-4.12.2 pyright-1.1.387 fastapi-0.115.4 commit: unknown ```
open
2024-12-03T16:22:29Z
2024-12-10T03:00:45Z
https://github.com/pydantic/pydantic/issues/11041
[ "bug V2", "pending" ]
AbdealiLoKo
1
netbox-community/netbox
django
18,770
Allow {module} variable in position field
### NetBox version v4.2.4 ### Feature type New functionality ### Proposed functionality Allow the {module} variable to be used in the Position field. Currently the text is not parsed and {module} remains as text. The position of the module then has to be manually updated after it has been instantiated into a module within a device. ![Image](https://github.com/user-attachments/assets/59a195cf-d298-4d33-a63a-92b02227aa32) ### Use case When modules have nested modules such as PSUs, it would be useful to have the parent modules position passed down to the child module so that interfaces and ports can be renamed based on their parents position. ### Database changes _No response_ ### External dependencies _No response_
closed
2025-02-28T11:28:38Z
2025-02-28T14:39:39Z
https://github.com/netbox-community/netbox/issues/18770
[ "type: feature" ]
mattg66
1
localstack/localstack
python
11,871
bug: Empty nullable fields return as true when queried from RDS Postgres
### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior When i return an object from a table with a nullable column, if that records value is null... the RDS Data API appears to return TRUE. ### Expected Behavior The column should be set to NULL. ### How are you starting LocalStack? With the `localstack` script ### Steps To Reproduce #### How are you starting localstack (e.g., `bin/localstack` command, arguments, or `docker-compose.yml`) DISABLE_CORS_CHECKS=1 DEBUG=1 localstack start #### Client commands (e.g., AWS SDK code snippet, or sequence of "awslocal" commands) Add a table such as this to Postgres v11 or v13 on LocalStack. ``` CREATE TABLE IF NOT EXISTS my_table ( my_table_id VARCHAR(36) PRIMARY KEY, will_return_ok TEXT NOT NULL, will_return_true_if_empty: TEXT[], ); ``` Add data where will_return_true_if_empty=[], will_return_true_if_empty=['text'], will_return_true_if_empty=NULL Call this query 'SELECT * FROM my_table;' ``` export const executeRdsStatement = async ( queryString: string, secretArn: string, dbClusterArn: string, ): Promise<DatabaseResponse[]> => { const command = new ExecuteStatementCommand({ database: "root" /* required */, formatRecordsAs: "JSON", includeResultMetadata: true, resourceArn: dbClusterArn, secretArn: secretArn, sql: queryString /* required */, }); const result = await rdsdataservice.send(command); if (result.formattedRecords) { return JSON.parse(result.formattedRecords) as DatabaseResponse[]; } return []; }; ``` Observe that the formatted records have TRUE for the null value record. ### Environment ```markdown - OS: MacOS - LocalStack: LocalStack version: 3.8.1 LocalStack Docker image sha: sha256:ede1832a84149228365c9a431c526da8cc004f3a199d6fc20fc4e3f739b9dd06 LocalStack build date: 2024-07-16 LocalStack build git hash: c4457f6 ``` ### Anything else? Current workaround is the inspect the table schema and map the column value back. This does not occur in my prod/dev deploys in AWS. The database, when queried directly has the correct values. ``` const getTableSchema = async ( secretArn: string, dbClusterArn: string, tableName: string, ) => { const schemaQuery = ` SELECT column_name, data_type, is_nullable FROM information_schema.columns WHERE table_name = $1 ORDER BY ordinal_position; `; const command = new ExecuteStatementCommand({ database: "root" /* required */, formatRecordsAs: "JSON", includeResultMetadata: true, parameters: [{ name: "1", value: { stringValue: tableName } }], resourceArn: dbClusterArn, secretArn: secretArn, sql: schemaQuery, }); const response = await rdsdataservice.send(command); if (response.formattedRecords) { return JSON.parse(response.formattedRecords); } return []; }; ```
closed
2024-11-19T04:00:06Z
2024-11-28T18:42:16Z
https://github.com/localstack/localstack/issues/11871
[ "type: bug", "aws:rds", "status: backlog" ]
Neuroforge
3
onnx/onnx
scikit-learn
6,477
shape inferrence about dynamic axis
I'm trying to export the pytorch model into onnx without dynamic axis and infer the shape with shape_inference. But the inferred model still have unkown dimensions. How to ensure all the dimensions in onnx is fixed? Thank you. ![微信截图_20241021124448](https://github.com/user-attachments/assets/1cea6db5-3b8b-4220-8123-ea0461fb5cc2) ![微信截图_20241021125032](https://github.com/user-attachments/assets/85a20849-f20d-4e3e-a74e-f7938a4e5ade) The code is as follows: ``` model = torchvision.models.detection.retinanet_resnet50_fpn(pretrained=True) model.eval() dummy_input = torch.randn(1, 3, 800, 800) onnx_model_path = "retinanet.onnx" torch.onnx.export(model, dummy_input, onnx_model_path, export_params=True, opset_version=12, do_constant_folding=True, input_names=['input'], output_names=['output'], ) model = onnx.load(onnx_model_path) inferred_model = onnx.shape_inference.infer_shapes(model) onnx.save(inferred_model, onnx_model_path) ```
open
2024-10-21T04:51:44Z
2024-10-24T13:27:09Z
https://github.com/onnx/onnx/issues/6477
[ "question" ]
ChipmunkKeeper
6
plotly/dash-table
dash
236
Embedded links support
Now it seems not possible to insert `dcc.Link` inside a table cell. It produces error: `Objects are not valid as a React child (found: object with keys {props, type, namespace}).` The use case: I have a bunch of time-series data and would like to map a page with analytics for each series, and each page is referred from the aggregated table located at the main page (multi-page Dash).
closed
2018-11-09T08:11:16Z
2018-11-09T13:14:38Z
https://github.com/plotly/dash-table/issues/236
[]
azhuchkov
1
Lightning-AI/pytorch-lightning
deep-learning
19,686
`dirpath` in `OnExceptionCheckpoint` should default to the same directory as `ModelCheckpoint`
### Description & Motivation `ModelCheckpoint` is a very handy features and `OnExceptionCheckpoint` is nice to have when you want to interrupt a training or there is a bug. However there is no default value for `dirpath` in `OnExceptionCheckpoint`, it would be nice if it defaulted to the same directory as `ModelCheckpoint` which is evaluated at runtime and under the logging directory. ### Pitch _No response_ ### Alternatives _No response_ ### Additional context _No response_
open
2024-03-22T16:54:29Z
2024-03-22T16:55:10Z
https://github.com/Lightning-AI/pytorch-lightning/issues/19686
[ "feature", "needs triage" ]
cajoek
0
ResidentMario/missingno
data-visualization
54
README is too crowded - reorganize into a separate GitHub pages docs site
I suggest keeping the README short and to the point: badges showing status etc. of the package, what is the purpose of the package, how to install, 1-2 basic code examples, and contributing information. More examples, function API docs, and detailed descriptions can be moved to the docs site. I also suggest adding a gallery to the docs, [similar to Seaborn](https://seaborn.pydata.org/examples/index.html).
closed
2018-01-31T03:08:53Z
2020-04-09T21:14:51Z
https://github.com/ResidentMario/missingno/issues/54
[]
rhiever
1
kizniche/Mycodo
automation
1,394
Mycodo is not accessible and Raspberry Pi3 B+ after less than 12 hours
### Versions: - Mycodo Version: Mycodo v8.16.0 - Raspberry Pi Version: 3B+ - Raspbian OS Version: Debian GNU/Linux 12 Release: 12 last one of this moring with Pi imager Hello, since a week I have installed Mycodo on a Pi3 B+ with a hat UPS and a Hat relay Pi 6 from Sbcomponnets. It has a usb SSD disk connected (ORICO SSD M.2 NVME 940 256Go USB 3.1 Gen2) A SHT31 temperature sensor and a LCD 4x20 both on I2C with a logic converter to use 3V and 5V together. The make measures and activate the relay connected to a fridge with minimum functionning time of 120 seconds. Temperature target of the PID is 15°C I reinstall all, Rpi OS this morning and Mycodo with InfluxDb1 instead of InfluxDB2. But I still have the same error. In DEBUG logs I can see that at an undetermined time the daemon restart (12:53 am) After that the PID no more functions correctly, and the regulation is stopped. I observe that after this event, I cannot connect to Rpi with SSH because Ip address is no more answering to pings. I think that with your experience you can understand what I'm doing baddly. But I cannot find any rational explanantion by myself. I join all the lines of the Mycodo logs frome this morning 9:30 AM to 12:53am when the error came back. Thank's for your help. With best regards. Pierre [errors Mycodo Influxdbx1.txt](https://github.com/user-attachments/files/17249290/errors.Mycodo.Influxdbx1.txt)
closed
2024-10-03T19:47:02Z
2025-01-08T19:44:26Z
https://github.com/kizniche/Mycodo/issues/1394
[]
Pierrevictor67
3
jpadilla/django-rest-framework-jwt
django
82
Notes on RSASSA-PKCS1-v1_5 Should be Updates
> For the RSASSA-PKCS1-v1_5 algorithms, the "secret" argument in jwt.encode is supposed to be a private RSA key as imported with Crypto.PublicKey.RSA.importKey. Likewise, the "secret" argument in jwt.decode is supposed to be the public RSA key imported with the same method. 1. It isn't clear what library has `Crypto.PublicKey.RSA.importKey` 2. Probably want to use `cryptography` anyways 3. `pyjwt` handles strings now in some cases so no special action is needed,
closed
2015-02-21T20:59:49Z
2015-02-22T01:23:03Z
https://github.com/jpadilla/django-rest-framework-jwt/issues/82
[]
cancan101
1
Significant-Gravitas/AutoGPT
python
8,779
Implement (real-time) datasource for agent status
* Part of #8774 * Needed for real-time agent status indication in #8775 ### Requirements * Shares updates of agent status in real time * Part of WebSockets API * The same info needs to be added to the `GET /graphs` REST endpoint * Events: * Agent status updated * Running (≥ 1 graph execution active) * Inactive (no graph executions active, no webhooks or schedule attached) * Scheduled (schedule attached) * Listening (webhook attached) * Error (not sure how to determine this)
open
2024-11-26T17:51:20Z
2025-03-20T19:25:46Z
https://github.com/Significant-Gravitas/AutoGPT/issues/8779
[ "API", "platform/backend" ]
Pwuts
0
aminalaee/sqladmin
sqlalchemy
539
"Create" form fails when model has `meta` field
### Checklist - [X] The bug is reproducible against the latest release or `master`. - [X] There are no similar issues or pull requests to fix it yet. ### Describe the bug When I include a model which has a `meta` field, this clashes with the `meta` field of the form object from wtforms. I cannot immediately say whether this should be fixed in sqladmin or in wtforms (I can open the issue over there, if it's more appropriate), nor I could find an easy workaround. Any advice is welcome. ### Steps to reproduce the bug 1. Install sqladmin (for instance from the current main, with `git clone` and then `pip install -e .`). This also installs wtforms (version 3.0.1). 2. Install fastapi (e.g. with `pip install fastapi[all]`, which is currently fetching version 0.99.1). 3. Prepare the following `main.py` script ```python from sqlalchemy import Column, Integer, String, create_engine from sqlalchemy.ext.declarative import declarative_base from fastapi import FastAPI from sqladmin import Admin, ModelView Base = declarative_base() engine = create_engine("sqlite:///example.db", connect_args={"check_same_thread": False}) class User(Base): __tablename__ = "users" id = Column(Integer, primary_key=True) name = Column(String) meta = Column(String) # NOTE: there is a `meta` field Base.metadata.create_all(engine) # Create tables app = FastAPI() admin = Admin(app, engine) class UserAdmin(ModelView, model=User): column_exclude_list = [User.meta] admin.add_view(UserAdmin) ``` 5. Run the app with `uvicorn main:app`, and browse to `http://127.0.0.1:8000/admin/user/list`. This page correctly shows an empty list. 6. Click the "New user button". The browser will simply show a "Internal server error message", while the console will have a more detailed error traceback. ### Expected behavior A model with a `meta` field can be handled by sqladmin+wtforms. ### Actual behavior A model with a `meta` field cannot be handled by sqladmin+wtforms, due to the name clash with the `meta` attribute of the form object. ### Debugging material When following the steps described above, the traceback reads ```python traceback [...] INFO: 127.0.0.1:35254 - "GET /user/create HTTP/1.1" 500 Internal Server Error ERROR: Exception in ASGI application Traceback (most recent call last): File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/uvicorn/protocols/http/httptools_impl.py", line 435, in run_asgi result = await app( # type: ignore[func-returns-value] File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/uvicorn/middleware/proxy_headers.py", line 78, in __call__ return await self.app(scope, receive, send) File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/fastapi/applications.py", line 290, in __call__ await super().__call__(scope, receive, send) File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/starlette/applications.py", line 122, in __call__ await self.middleware_stack(scope, receive, send) File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/starlette/middleware/errors.py", line 184, in __call__ raise exc File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/starlette/middleware/errors.py", line 162, in __call__ await self.app(scope, receive, _send) File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/starlette/middleware/exceptions.py", line 79, in __call__ raise exc File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/starlette/middleware/exceptions.py", line 68, in __call__ await self.app(scope, receive, sender) File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/fastapi/middleware/asyncexitstack.py", line 20, in __call__ raise e File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/fastapi/middleware/asyncexitstack.py", line 17, in __call__ await self.app(scope, receive, send) File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/starlette/routing.py", line 718, in __call__ await route.handle(scope, receive, send) File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/starlette/routing.py", line 443, in handle await self.app(scope, receive, send) File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/starlette/applications.py", line 122, in __call__ await self.middleware_stack(scope, receive, send) File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/starlette/middleware/errors.py", line 184, in __call__ raise exc File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/starlette/middleware/errors.py", line 162, in __call__ await self.app(scope, receive, _send) File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/starlette/middleware/exceptions.py", line 79, in __call__ raise exc File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/starlette/middleware/exceptions.py", line 68, in __call__ await self.app(scope, receive, sender) File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/starlette/routing.py", line 718, in __call__ await route.handle(scope, receive, send) File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/starlette/routing.py", line 276, in handle await self.app(scope, receive, send) File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/starlette/routing.py", line 66, in app response = await func(request) File "/somewhere/sqladmin/sqladmin/authentication.py", line 62, in wrapper_decorator return await func(*args, **kwargs) File "/somewhere/sqladmin/sqladmin/application.py", line 497, in create form = Form(form_data) File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/wtforms/form.py", line 208, in __call__ return type.__call__(cls, *args, **kwargs) File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/wtforms/form.py", line 286, in __init__ self.process(formdata, obj, data=data, **kwargs) File "/somewhere/sqladmin/venv/lib/python3.10/site-packages/wtforms/form.py", line 106, in process formdata = self.meta.wrap_formdata(self, formdata) AttributeError: 'StringField' object has no attribute 'wrap_formdata' ``` ### Environment Ubuntu 22, Python 3.10.6, sqladmin 978cdd6587c7fc8345076393ae6774cddbdcd48c (that is, version 0.13.0). ### Additional context The wtforms `BaseForm` object is defined in https://github.com/wtforms/wtforms/blob/master/src/wtforms/form.py, and it clearly includes a `meta` attribute ```python class BaseForm: """ Base Form Class. Provides core behaviour like field construction, validation, and data and error proxying. """ def __init__(self, fields, prefix="", meta=_default_meta): """ :param fields: A dict or sequence of 2-tuples of partially-constructed fields. :param prefix: If provided, all fields will have their name prefixed with the value. :param meta: A meta instance which is used for configuration and customization of WTForms behaviors. """ ```
closed
2023-07-05T07:40:13Z
2023-07-06T07:36:57Z
https://github.com/aminalaee/sqladmin/issues/539
[]
tcompa
2
d2l-ai/d2l-en
tensorflow
2,636
failed to access https://discuss.d2l.ai/
![Image](https://github.com/user-attachments/assets/7282fb67-1e69-425f-9493-1d86ba522cdf) have tried to delete cookie cache, but didn't work related issue https://github.com/d2l-ai/d2l-en/issues/2581
open
2025-02-28T11:08:17Z
2025-03-10T03:28:04Z
https://github.com/d2l-ai/d2l-en/issues/2636
[]
domonnss
4