repo
stringclasses
32 values
instance_id
stringlengths
13
37
base_commit
stringlengths
40
40
patch
stringlengths
1
1.89M
test_patch
stringclasses
1 value
problem_statement
stringlengths
304
69k
hints_text
stringlengths
0
246k
created_at
stringlengths
20
20
version
stringclasses
1 value
FAIL_TO_PASS
stringclasses
1 value
PASS_TO_PASS
stringclasses
1 value
environment_setup_commit
stringclasses
1 value
traceback
stringlengths
64
23.4k
__index_level_0__
int64
29
19k
conda/conda
conda__conda-2915
deaccea600d7b80cbcc939f018a5fdfe2a066967
diff --git a/conda/egg_info.py b/conda/egg_info.py --- a/conda/egg_info.py +++ b/conda/egg_info.py @@ -15,6 +15,7 @@ from .misc import rel_path + def get_site_packages_dir(installed_pkgs): for info in itervalues(installed_pkgs): if info['name'] == 'python': diff --git a/conda/exceptions.py b/conda/exceptions.py --- a/conda/exceptions.py +++ b/conda/exceptions.py @@ -8,6 +8,3 @@ class InvalidInstruction(CondaException): def __init__(self, instruction, *args, **kwargs): msg = "No handler for instruction: %r" % instruction super(InvalidInstruction, self).__init__(msg, *args, **kwargs) - -class LockError(RuntimeError, CondaException): - pass diff --git a/conda/lock.py b/conda/lock.py --- a/conda/lock.py +++ b/conda/lock.py @@ -17,11 +17,11 @@ """ from __future__ import absolute_import, division, print_function -import logging import os -import time - -from .exceptions import LockError +import logging +from os.path import join +import glob +from time import sleep LOCKFN = '.conda_lock' @@ -33,13 +33,15 @@ class Locked(object): """ Context manager to handle locks. """ - def __init__(self, path, retries=10): + def __init__(self, path): self.path = path self.end = "-" + str(os.getpid()) - self.lock_path = os.path.join(self.path, LOCKFN + self.end) - self.retries = retries + self.lock_path = join(self.path, LOCKFN + self.end) + self.pattern = join(self.path, LOCKFN + '-*') + self.remove = True def __enter__(self): + retries = 10 # Keep the string "LOCKERROR" in this string so that external # programs can look for it. lockstr = ("""\ @@ -48,24 +50,33 @@ def __enter__(self): If you are sure that conda is not running, remove it and try again. You can also use: $ conda clean --lock\n""") sleeptime = 1 - - for _ in range(self.retries): - if os.path.isdir(self.lock_path): - stdoutlog.info(lockstr % self.lock_path) + files = None + while retries: + files = glob.glob(self.pattern) + if files and not files[0].endswith(self.end): + stdoutlog.info(lockstr % str(files)) stdoutlog.info("Sleeping for %s seconds\n" % sleeptime) - - time.sleep(sleeptime) + sleep(sleeptime) sleeptime *= 2 + retries -= 1 else: - os.makedirs(self.lock_path) - return self + break + else: + stdoutlog.error("Exceeded max retries, giving up") + raise RuntimeError(lockstr % str(files)) - stdoutlog.error("Exceeded max retries, giving up") - raise LockError(lockstr % self.lock_path) + if not files: + try: + os.makedirs(self.lock_path) + except OSError: + pass + else: # PID lock already here --- someone else will remove it. + self.remove = False def __exit__(self, exc_type, exc_value, traceback): - try: - os.rmdir(self.lock_path) - os.rmdir(self.path) - except OSError: - pass + if self.remove: + for path in self.lock_path, self.path: + try: + os.rmdir(path) + except OSError: + pass
[Regression] Conda create environment fails on lock if root environment is not under user control This issue is introduced in Conda 4.1.0 (Conda 4.0.8 works fine). ``` $ conda create -n root2 python=2 [123/1811] Fetching package metadata ....... Solving package specifications ............. Package plan for installation in environment /home/frol/.conda/envs/root2: The following NEW packages will be INSTALLED: openssl: 1.0.2h-1 (soft-link) pip: 8.1.2-py27_0 (soft-link) python: 2.7.11-0 (soft-link) readline: 6.2-2 (soft-link) setuptools: 23.0.0-py27_0 (soft-link) sqlite: 3.13.0-0 (soft-link) tk: 8.5.18-0 (soft-link) wheel: 0.29.0-py27_0 (soft-link) zlib: 1.2.8-3 (soft-link) Proceed ([y]/n)? Linking packages ... An unexpected error has occurred, please consider sending the following traceback to the conda GitHub issue tracker at: https://github.com/conda/conda/issues Include the output of the command 'conda info' in your report. Traceback (most recent call last): File "/usr/local/miniconda/bin/conda", line 6, in <module> sys.exit(main()) File "/usr/local/miniconda/lib/python2.7/site-packages/conda/cli/main.py", line 120, in main args_func(args, p) File "/usr/local/miniconda/lib/python2.7/site-packages/conda/cli/main.py", line 127, in args_func args.func(args, p) File "/usr/local/miniconda/lib/python2.7/site-packages/conda/cli/main_create.py", line 57, in execute install(args, parser, 'create') File "/usr/local/miniconda/lib/python2.7/site-packages/conda/cli/install.py", line 407, in install execute_actions(actions, index, verbose=not args.quiet) File "/usr/local/miniconda/lib/python2.7/site-packages/conda/plan.py", line 566, in execute_actions inst.execute_instructions(plan, index, verbose) File "/usr/local/miniconda/lib/python2.7/site-packages/conda/instructions.py", line 137, in execute_instructions cmd(state, arg) File "/usr/local/miniconda/lib/python2.7/site-packages/conda/instructions.py", line 80, in LINK_CMD link(state['prefix'], dist, lt, index=state['index'], shortcuts=shortcuts) File "/usr/local/miniconda/lib/python2.7/site-packages/conda/install.py", line 1035, in link with Locked(prefix), Locked(pkgs_dir): File "/usr/local/miniconda/lib/python2.7/site-packages/conda/lock.py", line 60, in __enter__ os.makedirs(self.lock_path) File "/usr/local/miniconda/lib/python2.7/os.py", line 157, in makedirs mkdir(name, mode) OSError: [Errno 13] Permission denied: '/usr/local/miniconda/pkgs/.conda_lock-949' ``` `/usr/local/miniconda/` is a system-wide installation of miniconda, so obviously, users cannot create lock files there. P.S. I have a dream that updating conda software won't break things on every release...
It seems that I cannot even do `source activate ...` as a regular user now. It just hangs. Here is what I get when I interrupt it with `^C`: ``` $ source activate root2 ^CTraceback (most recent call last): File "/usr/local/miniconda/bin/conda", line 6, in <module> sys.exit(main()) File "/usr/local/miniconda/lib/python2.7/site-packages/conda/cli/main.py", line 48, in main activate.main() File "/usr/local/miniconda/lib/python2.7/site-packages/conda/cli/activate.py", line 121, in main path = get_path(shelldict) File "/usr/local/miniconda/lib/python2.7/site-packages/conda/cli/activate.py", line 108, in get_path return run_in(shelldict["printpath"], shelldict)[0] File "/usr/local/miniconda/lib/python2.7/site-packages/conda/utils.py", line 174, in run_in stdout, stderr = p.communicate() File "/usr/local/miniconda/lib/python2.7/subprocess.py", line 799, in communicate return self._communicate(input) File "/usr/local/miniconda/lib/python2.7/subprocess.py", line 1409, in _communicate stdout, stderr = self._communicate_with_poll(input) File "/usr/local/miniconda/lib/python2.7/subprocess.py", line 1463, in _communicate_with_poll ready = poller.poll() KeyboardInterrupt ``` It also seems like there is no way to pin Conda version. Installing any package to the root env tries to update Conda to the latest version: ``` $ conda install python Using Anaconda Cloud api site https://api.anaconda.org Fetching package metadata: .... Solving package specifications: ......... Package plan for installation in environment /usr/local/miniconda: The following packages will be UPDATED: conda: 4.0.8-py27_0 --> 4.1.0-py27_0 conda-env: 2.4.5-py27_0 --> 2.5.0-py27_0 Proceed ([y]/n)? ``` You can now pin 4.0.9. ``` conda install conda=4.0.9 conda config --set auto_update_conda false ``` `auto_update_conda` setting will be added to 4.1.1 also, which is coming out tonight or tomorrow morning. Freaking lock on the whole package cache. We're going to get rid of that soon. I can't really tell what initially tripped here. What are the permissions on `/usr/local/miniconda/pkgs/.conda_lock-949`? Are they yours, or root, or does it even exist? Ohhh I think I get it, from your issue title. I actually thought this was already broken anyway. I plan on having 4.2 out in a couple weeks now, and should be a simple fix at that point. Ok with staying on 4.0.9 for this use case until then? @kalefranz It seems that I don't have options here but wait. I use Docker containers and this bug doesn't bother me that much. However, I would say that it is not a minor regression to postpone it to the next release. (From my experience, next release will break things in another way, so people will stuck with 4.0.*) In the last several months our coverage has gone from ~48% to almost 70%. The code base is still far more fragile and brittle than I'd like it to be, but we're making progress I think. Reverting #2320 fixed the regression. However, it seems that it just fails silently at locking, but at least it works in non-concurrent scenarios. cc @alanhdu @frol: Yeah, that's about right. Before #2320, conda would just swallow all `OsError`s (including `PermissionError`s). For now, we could add a `try ... except` around https://github.com/conda/conda/blob/master/conda/lock.py#L60, catch a `PermissionError` (or whatever the equivalent Python 2 error is), and try to do something smart (or just silently fail... depends how important that lock actually is). I am seeing this issue as well @kalefranz After reading this, I thought I would be able to downgrade conda to 4.0.9. However after doing that, I am still unable to activate centrally administered conda environments as a user. Is there a prior version of conda you would reccommend? Or did the 4.1.3 version leave something in my miniconda install that is causing the problem? Do I need to re-install miniconda2 from scratch? Should I just wait for this to be fixed before proceeding with trying to build a central conda install for our users? @davidslac There are two options you may try: 1. Downgrade conda-env together with conda: ``` bash $ conda install conda=4.0.9 'conda-env<2.5' ``` 2. Patch (revert changes) conda lock in Conda 4.1.x: ``` bash $ curl -o "/usr/local/miniconda/lib/python"*"/site-packages/conda/lock.py" \ "https://raw.githubusercontent.com/conda/conda/9428ad0b76be55e8070e04dd577c96e7dab571e0/conda/lock.py" ``` Thanks! I tried both, but it sill did not work. It's probably something I'm overlooking on my part - after the first failure I deleted the lock.pyc, but still no luck. I'll just hope for a fix in the future. I suspect though, that central administration is not as standard of a use case, so one is more likely to run into problems. It may be we should just provide a channel of our packages to our users and let them administer their own software stacks, give them some environments we know work. best, David On 06/27/16 10:58, Vlad Frolov wrote: > @davidslac https://github.com/davidslac There are two options you > may try: > > 1. > > ``` > Downgrade conda-env: > > $ conda install'conda-env<2.5' > ``` > > 2. > > ``` > Patch (revert changes) conda lock: > > $ curl -o"/usr/local/miniconda/lib/python"*"/site-packages/conda/lock.py" \ > "https://raw.githubusercontent.com/conda/conda/9428ad0b76be55e8070e04dd577c96e7dab571e0/conda/lock.py" > ```
2016-06-29T17:16:25Z
[]
[]
Traceback (most recent call last): File "/usr/local/miniconda/bin/conda", line 6, in <module> sys.exit(main()) File "/usr/local/miniconda/lib/python2.7/site-packages/conda/cli/main.py", line 120, in main args_func(args, p) File "/usr/local/miniconda/lib/python2.7/site-packages/conda/cli/main.py", line 127, in args_func args.func(args, p) File "/usr/local/miniconda/lib/python2.7/site-packages/conda/cli/main_create.py", line 57, in execute install(args, parser, 'create') File "/usr/local/miniconda/lib/python2.7/site-packages/conda/cli/install.py", line 407, in install execute_actions(actions, index, verbose=not args.quiet) File "/usr/local/miniconda/lib/python2.7/site-packages/conda/plan.py", line 566, in execute_actions inst.execute_instructions(plan, index, verbose) File "/usr/local/miniconda/lib/python2.7/site-packages/conda/instructions.py", line 137, in execute_instructions cmd(state, arg) File "/usr/local/miniconda/lib/python2.7/site-packages/conda/instructions.py", line 80, in LINK_CMD link(state['prefix'], dist, lt, index=state['index'], shortcuts=shortcuts) File "/usr/local/miniconda/lib/python2.7/site-packages/conda/install.py", line 1035, in link with Locked(prefix), Locked(pkgs_dir): File "/usr/local/miniconda/lib/python2.7/site-packages/conda/lock.py", line 60, in __enter__ os.makedirs(self.lock_path) File "/usr/local/miniconda/lib/python2.7/os.py", line 157, in makedirs mkdir(name, mode) OSError: [Errno 13] Permission denied: '/usr/local/miniconda/pkgs/.conda_lock-949'
3,981
conda/conda
conda__conda-3257
83f397d3d596dedc93e0160104dad46c6bfdb7ef
diff --git a/conda/utils.py b/conda/utils.py --- a/conda/utils.py +++ b/conda/utils.py @@ -313,6 +313,12 @@ def human_bytes(n): "sh.exe": dict( msys2_shell_base, exe="sh.exe", ), + "zsh.exe": dict( + msys2_shell_base, exe="zsh.exe", + ), + "zsh": dict( + msys2_shell_base, exe="zsh", + ), } else:
Zsh.exe not supported on MSYS2 The following error is reported in a MSYS2 zsh shell: ``` ➜ dotfiles git:(master) ✗ source activate py35_32 Traceback (most recent call last): File "C:\Miniconda3\Scripts\conda-script.py", line 5, in <module> sys.exit(main()) File "C:\Miniconda3\lib\site-packages\conda\cli\main.py", line 48, in main activate.main() File "C:\Miniconda3\lib\site-packages\conda\cli\activate.py", line 105, in main shelldict = shells[shell] KeyError: 'zsh.exe' ```
2016-08-09T15:16:41Z
[]
[]
Traceback (most recent call last): File "C:\Miniconda3\Scripts\conda-script.py", line 5, in <module> sys.exit(main()) File "C:\Miniconda3\lib\site-packages\conda\cli\main.py", line 48, in main activate.main() File "C:\Miniconda3\lib\site-packages\conda\cli\activate.py", line 105, in main shelldict = shells[shell] KeyError: 'zsh.exe'
3,990
conda/conda
conda__conda-3633
2d850dc7ce6ded5c7573ce470dca81517a6f9b94
diff --git a/conda/common/configuration.py b/conda/common/configuration.py --- a/conda/common/configuration.py +++ b/conda/common/configuration.py @@ -23,7 +23,7 @@ from itertools import chain from logging import getLogger from os import environ, stat -from os.path import join +from os.path import join, basename from stat import S_IFDIR, S_IFMT, S_IFREG try: @@ -343,7 +343,7 @@ def load_file_configs(search_path): # returns an ordered map of filepath and dict of raw parameter objects def _file_yaml_loader(fullpath): - assert fullpath.endswith(".yml") or fullpath.endswith("condarc"), fullpath + assert fullpath.endswith((".yml", ".yaml")) or "condarc" in basename(fullpath), fullpath yield fullpath, YamlRawParameter.make_raw_parameters_from_file(fullpath) def _dir_yaml_loader(fullpath):
BUG: CONDARC env var broken in latest conda After upgrading to conda version 4.2.7 the CONDARC env var no longer supports filenames of any format. It appears to only support filenames that end with .yml or .condarc. This is a functionality regression bug. Can this be fixed immediately!? **conda info** Current conda install: ``` platform : osx-64 conda version : 4.2.7 conda is private : False conda-env version : 4.2.7 conda-build version : 1.21.3 python version : 2.7.12.final.0 requests version : 2.9.1 root environment : /opt/anaconda (writable) default environment : /opt/anaconda envs directories : /opt/anaconda/envs package cache : /opt/anaconda/pkgs channel URLs : https://repo.continuum.io/pkgs/free/osx-64/ https://repo.continuum.io/pkgs/free/noarch/ https://repo.continuum.io/pkgs/pro/osx-64/ https://repo.continuum.io/pkgs/pro/noarch/ https://conda.anaconda.org/r/osx-64/ https://conda.anaconda.org/r/noarch/ config file : /Users/jhull/.condarc offline mode : False ``` **export CONDARC=~/.condarc.cloud** **conda info** Traceback (most recent call last): File "/opt/anaconda/bin/conda", line 4, in <module> import conda.cli File "/opt/anaconda/lib/python2.7/site-packages/conda/cli/**init**.py", line 8, in <module> from .main import main # NOQA File "/opt/anaconda/lib/python2.7/site-packages/conda/cli/main.py", line 46, in <module> from ..base.context import context File "/opt/anaconda/lib/python2.7/site-packages/conda/base/context.py", line 252, in <module> context = Context(SEARCH_PATH, conda, None) File "/opt/anaconda/lib/python2.7/site-packages/conda/common/configuration.py", line 692, in **init** self._add_search_path(search_path) File "/opt/anaconda/lib/python2.7/site-packages/conda/common/configuration.py", line 699, in _add_search_path return self._add_raw_data(load_file_configs(search_path)) File "/opt/anaconda/lib/python2.7/site-packages/conda/common/configuration.py", line 371, in load_file_configs raw_data = odict(kv for kv in chain.from_iterable(load_paths)) File "/opt/anaconda/lib/python2.7/collections.py", line 69, in __init__ self.__update(_args, *_kwds) File "/opt/anaconda/lib/python2.7/_abcoll.py", line 571, in update for key, value in other: File "/opt/anaconda/lib/python2.7/site-packages/conda/common/configuration.py", line 371, in <genexpr> raw_data = odict(kv for kv in chain.from_iterable(load_paths)) File "/opt/anaconda/lib/python2.7/site-packages/conda/common/configuration.py", line 346, in _file_yaml_loader **assert fullpath.endswith(".yml") or fullpath.endswith("condarc"), fullpath AssertionError: /Users/jhull/.condarc.cloud**
Added to my list to talk through with @mcg1969 tomorrow.
2016-10-14T16:07:18Z
[]
[]
Traceback (most recent call last): File "/opt/anaconda/bin/conda", line 4, in <module> import conda.cli File "/opt/anaconda/lib/python2.7/site-packages/conda/cli/**init**.py", line 8, in <module> from .main import main # NOQA File "/opt/anaconda/lib/python2.7/site-packages/conda/cli/main.py", line 46, in <module> from ..base.context import context File "/opt/anaconda/lib/python2.7/site-packages/conda/base/context.py", line 252, in <module> context = Context(SEARCH_PATH, conda, None) File "/opt/anaconda/lib/python2.7/site-packages/conda/common/configuration.py", line 692, in **init** self._add_search_path(search_path) File "/opt/anaconda/lib/python2.7/site-packages/conda/common/configuration.py", line 699, in _add_search_path return self._add_raw_data(load_file_configs(search_path)) File "/opt/anaconda/lib/python2.7/site-packages/conda/common/configuration.py", line 371, in load_file_configs raw_data = odict(kv for kv in chain.from_iterable(load_paths)) File "/opt/anaconda/lib/python2.7/collections.py", line 69, in __init__ self.__update(_args, *_kwds) File "/opt/anaconda/lib/python2.7/_abcoll.py", line 571, in update for key, value in other: File "/opt/anaconda/lib/python2.7/site-packages/conda/common/configuration.py", line 371, in <genexpr> raw_data = odict(kv for kv in chain.from_iterable(load_paths)) File "/opt/anaconda/lib/python2.7/site-packages/conda/common/configuration.py", line 346, in _file_yaml_loader **assert fullpath.endswith(".yml") or fullpath.endswith("condarc"), fullpath AssertionError: /Users/jhull/.condarc.cloud**
4,012
conda/conda
conda__conda-4774
14dd0efb8e2116ed07d6cca01b8f63f3ac04cb66
diff --git a/conda/common/configuration.py b/conda/common/configuration.py --- a/conda/common/configuration.py +++ b/conda/common/configuration.py @@ -17,7 +17,7 @@ from __future__ import absolute_import, division, print_function, unicode_literals from abc import ABCMeta, abstractmethod -from collections import Mapping, defaultdict +from collections import Mapping, Sequence, defaultdict from glob import glob from itertools import chain from logging import getLogger @@ -35,7 +35,7 @@ from .._vendor.auxlib.collection import AttrDict, first, frozendict, last, make_immutable from .._vendor.auxlib.exceptions import ThisShouldNeverHappenError from .._vendor.auxlib.path import expand -from .._vendor.auxlib.type_coercion import TypeCoercionError, typify_data_structure +from .._vendor.auxlib.type_coercion import TypeCoercionError, typify, typify_data_structure try: from cytoolz.dicttoolz import merge @@ -561,7 +561,6 @@ def __init__(self, element_type, default=(), aliases=(), validation=None, def collect_errors(self, instance, value, source="<<merged>>"): errors = super(SequenceParameter, self).collect_errors(instance, value) - element_type = self._element_type for idx, element in enumerate(value): if not isinstance(element, element_type): @@ -616,9 +615,10 @@ def repr_raw(self, raw_parameter): def _get_all_matches(self, instance): # this is necessary to handle argparse `action="append"`, which can't be set to a # default value of NULL - matches, multikey_exceptions = super(SequenceParameter, self)._get_all_matches(instance) + # it also config settings like `channels: ~` + matches, exceptions = super(SequenceParameter, self)._get_all_matches(instance) matches = tuple(m for m in matches if m._raw_value is not None) - return matches, multikey_exceptions + return matches, exceptions class MapParameter(Parameter): @@ -647,6 +647,7 @@ def collect_errors(self, instance, value, source="<<merged>>"): errors.extend(InvalidElementTypeError(self.name, val, source, type(val), element_type, key) for key, val in iteritems(value) if not isinstance(val, element_type)) + return errors def _merge(self, matches): @@ -676,6 +677,12 @@ def repr_raw(self, raw_parameter): self._str_format_flag(valueflag))) return '\n'.join(lines) + def _get_all_matches(self, instance): + # it also config settings like `proxy_servers: ~` + matches, exceptions = super(MapParameter, self)._get_all_matches(instance) + matches = tuple(m for m in matches if m._raw_value is not None) + return matches, exceptions + class ConfigurationType(type): """metaclass for Configuration"""
Error when installing r-essential on windows 7 When I try to install the r-essentials, I get the following error: $ C:\ProgramData\Anaconda3\Scripts\conda-script.py install -c r r-essentials=1.5.2 Traceback (most recent call last): File "C:\ProgramData\Anaconda3\lib\site-packages\conda\exceptions.py", line 617, in conda_exception_handler return_value = func(*args, **kwargs) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\cli\main.py", line 137, in _main exit_code = args.func(args, p) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\cli\main_install.py", line 80, in execute install(args, parser, 'install') File "C:\ProgramData\Anaconda3\lib\site-packages\conda\cli\install.py", line 118, in install context.validate_configuration() File "C:\ProgramData\Anaconda3\lib\site-packages\conda\common\configuration.py", line 830, in validate_configuration raise_errors(tuple(chain.from_iterable((errors, post_errors)))) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\common\configuration.py", line 828, in for name in self.parameter_names) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\common\configuration.py", line 821, in _collect_validation_error func(*args, **kwargs) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\common\configuration.py", line 446, in get result = typify_data_structure(self._merge(matches) if matches else self.default, File "C:\ProgramData\Anaconda3\lib\site-packages\conda\common\configuration.py", line 662, in _merge for match in relevant_matches) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\common\configuration.py", line 662, in for match in relevant_matches) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\common\compat.py", line 72, in iteritems return iter(d.items(**kw)) AttributeError: 'NoneType' object has no attribute 'items' Current conda install: platform : win-64 conda version : 4.3.8 conda is private : False conda-env version : 4.3.8 conda-build version : not installed python version : 3.6.0.final.0 requests version : 2.12.4 root environment : C:\ProgramData\Anaconda3 (writable) default environment : C:\ProgramData\Anaconda3 envs directories : C:\ProgramData\Anaconda3\envs package cache : C:\ProgramData\Anaconda3\pkgs channel URLs : https://conda.anaconda.org/r/win-64 https://conda.anaconda.org/r/noarch https://repo.continuum.io/pkgs/free/win-64 https://repo.continuum.io/pkgs/free/noarch https://repo.continuum.io/pkgs/r/win-64 https://repo.continuum.io/pkgs/r/noarch https://repo.continuum.io/pkgs/pro/win-64 https://repo.continuum.io/pkgs/pro/noarch https://repo.continuum.io/pkgs/msys2/win-64 https://repo.continuum.io/pkgs/msys2/noarch config file : C:\Users\xxx.condarc offline mode : False user-agent : conda/4.3.8 requests/2.12.4 CPython/3.6.0 Windows/7 Windows/6.1.7601 If somebody could help me with this, that would be great. Thanks
Latest conda version is 4.3.13. `conda update conda` Thanks. I had an error in the .condarc file. It seems, that the above error occurs when another error message is going to be created. Oh yeah. Syntax errors in yaml configuration files don't go through conda's normal error handler because it hasn't been initialized yet. Can you give me what you actually had as contents for the bad condarc file? I'll write a regression test for it...
2017-03-01T15:28:01Z
[]
[]
Traceback (most recent call last): File "C:\ProgramData\Anaconda3\lib\site-packages\conda\exceptions.py", line 617, in conda_exception_handler return_value = func(*args, **kwargs) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\cli\main.py", line 137, in _main exit_code = args.func(args, p) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\cli\main_install.py", line 80, in execute install(args, parser, 'install') File "C:\ProgramData\Anaconda3\lib\site-packages\conda\cli\install.py", line 118, in install context.validate_configuration() File "C:\ProgramData\Anaconda3\lib\site-packages\conda\common\configuration.py", line 830, in validate_configuration raise_errors(tuple(chain.from_iterable((errors, post_errors)))) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\common\configuration.py", line 828, in for name in self.parameter_names) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\common\configuration.py", line 821, in _collect_validation_error func(*args, **kwargs) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\common\configuration.py", line 446, in get result = typify_data_structure(self._merge(matches) if matches else self.default, File "C:\ProgramData\Anaconda3\lib\site-packages\conda\common\configuration.py", line 662, in _merge for match in relevant_matches) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\common\configuration.py", line 662, in for match in relevant_matches) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\common\compat.py", line 72, in iteritems return iter(d.items(**kw)) AttributeError: 'NoneType' object has no attribute 'items' Current conda install:
4,058
conda/conda
conda__conda-5090
ec9c9999ea57d3351d5c5c688baacc982c3daee2
diff --git a/conda/cli/activate.py b/conda/cli/activate.py --- a/conda/cli/activate.py +++ b/conda/cli/activate.py @@ -45,7 +45,7 @@ def help(command, shell): def prefix_from_arg(arg, shelldict): - from conda.base.context import context, locate_prefix_by_name + from ..base.context import context, locate_prefix_by_name 'Returns a platform-native path' # MSYS2 converts Unix paths to Windows paths with unix seps # so we must check for the drive identifier too. @@ -114,8 +114,8 @@ def get_activate_path(shelldict): def main(): - from conda.base.constants import ROOT_ENV_NAME - from conda.utils import shells + from ..base.constants import ROOT_ENV_NAME + from ..utils import shells if '-h' in sys.argv or '--help' in sys.argv: # all execution paths sys.exit at end. help(sys.argv[1], sys.argv[2]) @@ -163,9 +163,9 @@ def main(): # Make sure an env always has the conda symlink try: - from conda.base.context import context - import conda.install - conda.install.symlink_conda(prefix, context.root_prefix, shell) + from ..base.context import context + from ..install import symlink_conda + symlink_conda(prefix, context.root_prefix, shell) except (IOError, OSError) as e: if e.errno == errno.EPERM or e.errno == errno.EACCES: msg = ("Cannot activate environment {0}.\n" @@ -176,7 +176,7 @@ def main(): sys.exit(0) # raise CondaSystemExit elif sys.argv[1] == '..changeps1': - from conda.base.context import context + from ..base.context import context path = int(context.changeps1) else: diff --git a/conda/cli/common.py b/conda/cli/common.py --- a/conda/cli/common.py +++ b/conda/cli/common.py @@ -88,7 +88,7 @@ def __init__(self, prefix, parsed_args, **kwargs): def _get_items(self): # TODO: Include .tar.bz2 files for local installs. - from conda.core.index import get_index + from ..core.index import get_index args = self.parsed_args call_dict = dict(channel_urls=args.channel or (), use_cache=True, @@ -106,7 +106,7 @@ def __init__(self, prefix, parsed_args, **kwargs): @memoize def _get_items(self): - from conda.core.linked_data import linked + from ..core.linked_data import linked packages = linked(context.prefix_w_legacy_search) return [dist.quad[0] for dist in packages] @@ -573,7 +573,7 @@ def get_index_trap(*args, **kwargs): Retrieves the package index, but traps exceptions and reports them as JSON if necessary. """ - from conda.core.index import get_index + from ..core.index import get_index kwargs.pop('json', None) return get_index(*args, **kwargs) @@ -603,7 +603,7 @@ def stdout_json_success(success=True, **kwargs): def handle_envs_list(acc, output=True): - from conda import misc + from .. import misc if output: print("# conda environments:") diff --git a/conda/cli/install.py b/conda/cli/install.py --- a/conda/cli/install.py +++ b/conda/cli/install.py @@ -377,7 +377,7 @@ def install(args, parser, command='install'): def check_write(command, prefix, json=False): if inroot_notwritable(prefix): - from conda.cli.help import root_read_only + from .help import root_read_only root_read_only(command, prefix, json=json) diff --git a/conda/cli/main.py b/conda/cli/main.py --- a/conda/cli/main.py +++ b/conda/cli/main.py @@ -102,7 +102,7 @@ def _main(*args): if "remove" in module: imported.configure_parser(sub_parsers, name='uninstall') - from conda.cli.find_commands import find_commands + from .find_commands import find_commands def completer(prefix, **kwargs): return [i for i in list(sub_parsers.choices) + find_commands() diff --git a/conda/cli/main_info.py b/conda/cli/main_info.py --- a/conda/cli/main_info.py +++ b/conda/cli/main_info.py @@ -17,9 +17,6 @@ from .common import add_parser_json, add_parser_offline, arg2spec, handle_envs_list, stdout_json from ..common.compat import itervalues, on_win, iteritems -from ..common.url import mask_anaconda_token -from ..config import rc_path, sys_rc_path, user_rc_path -from ..models.channel import prioritize_channels log = getLogger(__name__) @@ -80,12 +77,12 @@ def configure_parser(sub_parsers): p.set_defaults(func=execute) -python_re = re.compile('python\d\.\d') def get_user_site(): site_dirs = [] try: if not on_win: if exists(expanduser('~/.local/lib')): + python_re = re.compile('python\d\.\d') for path in listdir(expanduser('~/.local/lib/')): if python_re.match(path): site_dirs.append("~/.local/lib/%s" % path) @@ -107,12 +104,13 @@ def get_user_site(): SKIP_FIELDS = IGNORE_FIELDS | {'name', 'version', 'build', 'build_number', 'channel', 'schannel', 'size', 'fn', 'depends'} + def dump_record(pkg): return {k: v for k, v in iteritems(pkg.dump()) if k not in IGNORE_FIELDS} def pretty_package(dist, pkg): - from conda.utils import human_bytes + from ..utils import human_bytes pkg = dump_record(pkg) d = OrderedDict([ @@ -137,58 +135,48 @@ def pretty_package(dist, pkg): for dep in pkg['depends']: print(' %s' % dep) -def execute(args, parser): - import os - from os.path import dirname - - import conda - from conda.base.context import context - from conda.models.channel import offline_keep - from conda.resolve import Resolve - from conda.api import get_index - from conda.connection import user_agent - if args.root: - if context.json: - stdout_json({'root_prefix': context.root_prefix}) - else: - print(context.root_prefix) - return +def print_package_info(packages): + from ..api import get_index + from ..base.context import context + from ..resolve import Resolve + index = get_index() + r = Resolve(index) + if context.json: + stdout_json({ + package: [dump_record(r.index[d]) + for d in r.get_dists_for_spec(arg2spec(package))] + for package in packages + }) + else: + for package in packages: + for dist in r.get_dists_for_spec(arg2spec(package)): + pretty_package(dist, r.index[dist]) - if args.packages: - index = get_index() - r = Resolve(index) - if context.json: - stdout_json({ - package: [dump_record(r.index[d]) - for d in r.get_dists_for_spec(arg2spec(package))] - for package in args.packages - }) - else: - for package in args.packages: - for dist in r.get_dists_for_spec(arg2spec(package)): - pretty_package(dist, r.index[dist]) - return - options = 'envs', 'system', 'license' +def get_info_dict(system=False): + from .. import CONDA_PACKAGE_ROOT, __version__ as conda_version + from ..base.context import context + from ..common.url import mask_anaconda_token + from ..config import rc_path, sys_rc_path, user_rc_path + from ..connection import user_agent + from ..models.channel import offline_keep, prioritize_channels try: - from conda.install import linked_data + from ..install import linked_data root_pkgs = linked_data(context.root_prefix) except: root_pkgs = None try: - import requests - requests_version = requests.__version__ + from requests import __version__ as requests_version except ImportError: requests_version = "could not import" except Exception as e: - requests_version = "Error %s" % e + requests_version = "Error %r" % e try: - import conda_env - conda_env_version = conda_env.__version__ + from conda_env import __version__ as conda_env_version except: try: cenv = [p for p in itervalues(root_pkgs) if p['name'] == 'conda-env'] @@ -205,16 +193,7 @@ def execute(args, parser): else: conda_build_version = conda_build.__version__ - channels = context.channels - - if args.unsafe_channels: - if not context.json: - print("\n".join(channels)) - else: - print(json.dumps({"channels": channels})) - return 0 - - channels = list(prioritize_channels(channels).keys()) + channels = list(prioritize_channels(context.channels).keys()) if not context.json: channels = [c + ('' if offline_keep(c) else ' (offline)') for c in channels] @@ -222,7 +201,7 @@ def execute(args, parser): info_dict = dict( platform=context.subdir, - conda_version=conda.__version__, + conda_version=conda_version, conda_env_version=conda_env_version, conda_build_version=conda_build_version, root_prefix=context.root_prefix, @@ -242,54 +221,13 @@ def execute(args, parser): python_version='.'.join(map(str, sys.version_info)), requests_version=requests_version, user_agent=user_agent, + conda_location=CONDA_PACKAGE_ROOT, ) if not on_win: info_dict['UID'] = os.geteuid() info_dict['GID'] = os.getegid() - if args.all or context.json: - for option in options: - setattr(args, option, True) - - if (args.all or all(not getattr(args, opt) for opt in options)) and not context.json: - for key in 'pkgs_dirs', 'envs_dirs', 'channels': - info_dict['_' + key] = ('\n' + 26 * ' ').join(info_dict[key]) - info_dict['_rtwro'] = ('writable' if info_dict['root_writable'] else - 'read only') - print("""\ -Current conda install: - - platform : %(platform)s - conda version : %(conda_version)s - conda is private : %(conda_private)s - conda-env version : %(conda_env_version)s - conda-build version : %(conda_build_version)s - python version : %(python_version)s - requests version : %(requests_version)s - root environment : %(root_prefix)s (%(_rtwro)s) - default environment : %(default_prefix)s - envs directories : %(_envs_dirs)s - package cache : %(_pkgs_dirs)s - channel URLs : %(_channels)s - config file : %(rc_path)s - offline mode : %(offline)s - user-agent : %(user_agent)s\ -""" % info_dict) - - if not on_win: - print("""\ - UID:GID : %(UID)s:%(GID)s -""" % info_dict) - else: - print() - - if args.envs: - handle_envs_list(info_dict['envs'], not context.json) - - if args.system: - from conda.cli.find_commands import find_commands, find_executable - - site_dirs = get_user_site() + if system: evars = ['PATH', 'PYTHONPATH', 'PYTHONHOME', 'CONDA_DEFAULT_ENV', 'CIO_TEST', 'CONDA_ENVS_PATH'] @@ -298,20 +236,99 @@ def execute(args, parser): elif context.platform == 'osx': evars.append('DYLD_LIBRARY_PATH') + info_dict.update({ + 'sys.version': sys.version, + 'sys.prefix': sys.prefix, + 'sys.executable': sys.executable, + 'site_dirs': get_user_site(), + 'env_vars': {ev: os.getenv(ev, '<not set>') for ev in evars}, + }) + + return info_dict + + +def get_main_info_str(info_dict): + for key in 'pkgs_dirs', 'envs_dirs', 'channels': + info_dict['_' + key] = ('\n' + 26 * ' ').join(info_dict[key]) + info_dict['_rtwro'] = ('writable' if info_dict['root_writable'] else 'read only') + + builder = [] + builder.append("""\ + Current conda install: + + platform : %(platform)s + conda version : %(conda_version)s + conda is private : %(conda_private)s + conda-env version : %(conda_env_version)s + conda-build version : %(conda_build_version)s + python version : %(python_version)s + requests version : %(requests_version)s + root environment : %(root_prefix)s (%(_rtwro)s) + default environment : %(default_prefix)s + envs directories : %(_envs_dirs)s + package cache : %(_pkgs_dirs)s + channel URLs : %(_channels)s + config file : %(rc_path)s + offline mode : %(offline)s + user-agent : %(user_agent)s\ + """ % info_dict) + + if not on_win: + builder.append("""\ + UID:GID : %(UID)s:%(GID)s + """ % info_dict) + else: + builder.append("") + + return '\n'.join(builder) + + +def execute(args, parser): + from ..base.context import context + + if args.root: if context.json: - info_dict['sys.version'] = sys.version - info_dict['sys.prefix'] = sys.prefix - info_dict['sys.executable'] = sys.executable - info_dict['site_dirs'] = get_user_site() - info_dict['env_vars'] = {ev: os.getenv(ev, '<not set>') for ev in evars} + stdout_json({'root_prefix': context.root_prefix}) + else: + print(context.root_prefix) + return + + if args.packages: + print_package_info(args.packages) + return + + if args.unsafe_channels: + if not context.json: + print("\n".join(context.channels)) else: + print(json.dumps({"channels": context.channels})) + return 0 + + options = 'envs', 'system', 'license' + + if args.all or context.json: + for option in options: + setattr(args, option, True) + + info_dict = get_info_dict(args.system) + + if (args.all or all(not getattr(args, opt) for opt in options)) and not context.json: + print(get_main_info_str(info_dict)) + + if args.envs: + handle_envs_list(info_dict['envs'], not context.json) + + if args.system: + if not context.json: + from .find_commands import find_commands, find_executable print("sys.version: %s..." % (sys.version[:40])) print("sys.prefix: %s" % sys.prefix) print("sys.executable: %s" % sys.executable) - print("conda location: %s" % dirname(conda.__file__)) + print("conda location: %s" % info_dict['conda_location']) for cmd in sorted(set(find_commands() + ['build'])): print("conda-%s: %s" % (cmd, find_executable('conda-' + cmd))) print("user site dirs: ", end='') + site_dirs = get_user_site() if site_dirs: print(site_dirs[0]) else: @@ -320,8 +337,8 @@ def execute(args, parser): print(' %s' % site_dir) print() - for ev in sorted(evars): - print("%s: %s" % (ev, os.getenv(ev, '<not set>'))) + for name, value in sorted(iteritems(info_dict['env_vars'])): + print("%s: %s" % (name, value)) print() if args.license and not context.json: diff --git a/conda/cli/main_list.py b/conda/cli/main_list.py --- a/conda/cli/main_list.py +++ b/conda/cli/main_list.py @@ -200,7 +200,7 @@ def execute(args, parser): regex = r'^%s$' % regex if args.revisions: - from conda.history import History + from ..history import History h = History(prefix) if isfile(h.path): if not context.json: diff --git a/conda/cli/main_remove.py b/conda/cli/main_remove.py --- a/conda/cli/main_remove.py +++ b/conda/cli/main_remove.py @@ -109,10 +109,11 @@ def configure_parser(sub_parsers, name='remove'): def execute(args, parser): - import conda.plan as plan - import conda.instructions as inst - from conda.gateways.disk.delete import rm_rf - from conda.core.linked_data import linked_data + from ..core.linked_data import linked_data + from ..gateways.disk.delete import rm_rf + from ..instructions import PREFIX + from ..plan import (add_unlink, display_actions, execute_actions, is_root_prefix, + nothing_to_do, remove_actions) if not (args.all or args.package_names): raise CondaValueError('no package names supplied,\n' @@ -137,22 +138,22 @@ def execute(args, parser): specs = None if args.features: specs = ['@' + f for f in set(args.package_names)] - actions = plan.remove_actions(prefix, specs, index, pinned=args.pinned) + actions = remove_actions(prefix, specs, index, pinned=args.pinned) action_groups = actions, elif args.all: - if plan.is_root_prefix(prefix): + if is_root_prefix(prefix): raise CondaEnvironmentError('cannot remove root environment,\n' ' add -n NAME or -p PREFIX option') - actions = {inst.PREFIX: prefix} + actions = {PREFIX: prefix} for dist in sorted(iterkeys(index)): - plan.add_unlink(actions, dist) + add_unlink(actions, dist) action_groups = actions, else: specs = specs_from_args(args.package_names) r = Resolve(index) prefix_spec_map = create_prefix_spec_map_with_deps(r, specs, prefix) - if (context.conda_in_root and plan.is_root_prefix(prefix) and names_in_specs( + if (context.conda_in_root and is_root_prefix(prefix) and names_in_specs( ROOT_NO_RM, specs) and not args.force): raise CondaEnvironmentError('cannot remove %s from root environment' % ', '.join(ROOT_NO_RM)) @@ -160,12 +161,12 @@ def execute(args, parser): for prfx, spcs in iteritems(prefix_spec_map): index = linked_data(prfx) index = {dist: info for dist, info in iteritems(index)} - actions.append(plan.remove_actions(prfx, list(spcs), index=index, force=args.force, - pinned=args.pinned)) + actions.append(remove_actions(prfx, list(spcs), index=index, force=args.force, + pinned=args.pinned)) action_groups = tuple(actions) delete_trash() - if any(plan.nothing_to_do(actions) for actions in action_groups): + if any(nothing_to_do(actions) for actions in action_groups): if args.all: print("\nRemove all packages in environment %s:\n" % prefix, file=sys.stderr) if not context.json: @@ -184,7 +185,7 @@ def execute(args, parser): if not context.json: print() print("Package plan for package removal in environment %s:" % action["PREFIX"]) - plan.display_actions(action, index) + display_actions(action, index) if context.json and args.dry_run: stdout_json({ @@ -200,9 +201,9 @@ def execute(args, parser): for actions in action_groups: if context.json and not context.quiet: with json_progress_bars(): - plan.execute_actions(actions, index, verbose=not context.quiet) + execute_actions(actions, index, verbose=not context.quiet) else: - plan.execute_actions(actions, index, verbose=not context.quiet) + execute_actions(actions, index, verbose=not context.quiet) if specs: try: with open(join(prefix, 'conda-meta', 'history'), 'a') as f: diff --git a/conda/cli/main_search.py b/conda/cli/main_search.py --- a/conda/cli/main_search.py +++ b/conda/cli/main_search.py @@ -128,7 +128,7 @@ def execute(args, parser): def execute_search(args, parser): import re - from conda.resolve import Resolve + from ..resolve import Resolve if args.reverse_dependency: if not args.regex: diff --git a/conda/config.py b/conda/config.py --- a/conda/config.py +++ b/conda/config.py @@ -9,7 +9,7 @@ import sys from os.path import abspath, expanduser, isfile, join -from conda.base.context import context, non_x86_linux_machines +from .base.context import context, non_x86_linux_machines non_x86_linux_machines = non_x86_linux_machines @@ -77,7 +77,7 @@ def get_local_urls(): - from conda.models.channel import get_conda_build_local_url + from .models.channel import get_conda_build_local_url return get_conda_build_local_url() or [] diff --git a/conda/connection.py b/conda/connection.py --- a/conda/connection.py +++ b/conda/connection.py @@ -142,7 +142,7 @@ def add_binstar_token(url): for binstar_url, token in iteritems(read_binstar_tokens()): if clean_url.startswith(binstar_url): log.debug("Adding anaconda token for url <%s>", clean_url) - from conda.models.channel import Channel + from .models.channel import Channel channel = Channel(clean_url) channel.token = token return channel.url(with_credentials=True) diff --git a/conda/exceptions.py b/conda/exceptions.py --- a/conda/exceptions.py +++ b/conda/exceptions.py @@ -547,7 +547,7 @@ def __init__(self, invalid_spec): def print_conda_exception(exception): - from conda.base.context import context + from .base.context import context stdoutlogger = getLogger('stdout') stderrlogger = getLogger('stderr') @@ -559,22 +559,6 @@ def print_conda_exception(exception): stderrlogger.info("\n\n%r", exception) -def get_info(): - from conda.cli import conda_argparse - from conda.cli.main_info import configure_parser - from shlex import split - from conda.common.io import captured - - p = conda_argparse.ArgumentParser() - sub_parsers = p.add_subparsers(metavar='command', dest='cmd') - configure_parser(sub_parsers) - - args = p.parse_args(split("info")) - with captured() as c: - args.func(args, p) - return c.stdout, c.stderr - - def print_unexpected_error_message(e): # bomb = "\U0001F4A3 " # explosion = "\U0001F4A5 " @@ -584,9 +568,9 @@ def print_unexpected_error_message(e): stderrlogger = getLogger('stderr') - from conda.base.context import context + from .base.context import context if context.json: - from conda.cli.common import stdout_json + from .cli.common import stdout_json stdout_json(dict(error=traceback)) else: message = """\ @@ -600,9 +584,8 @@ def print_unexpected_error_message(e): stderrlogger.info(message) command = ' '.join(sys.argv) if ' info' not in command: - # get and print `conda info` - info_stdout, info_stderr = get_info() - stderrlogger.info(info_stdout if info_stdout else info_stderr) + from .cli.main_info import get_info_dict, get_main_info_str + stderrlogger.info(get_main_info_str(get_info_dict())) stderrlogger.info("`$ {0}`".format(command)) stderrlogger.info('\n') stderrlogger.info('\n'.join(' ' + line for line in traceback.splitlines())) @@ -633,7 +616,7 @@ def handle_exception(e): if isinstance(e, CondaExitZero): return 0 elif isinstance(e, CondaError): - from conda.base.context import context + from .base.context import context if context.debug or context.verbosity > 0: print_unexpected_error_message(e) else: diff --git a/conda/exports.py b/conda/exports.py --- a/conda/exports.py +++ b/conda/exports.py @@ -79,7 +79,7 @@ import conda.base.context # NOQA -from conda.base.context import get_prefix as context_get_prefix, non_x86_linux_machines # NOQA +from .base.context import get_prefix as context_get_prefix, non_x86_linux_machines # NOQA non_x86_linux_machines = non_x86_linux_machines from ._vendor.auxlib.entity import EntityEncoder # NOQA diff --git a/conda/gateways/adapters/ftp.py b/conda/gateways/adapters/ftp.py --- a/conda/gateways/adapters/ftp.py +++ b/conda/gateways/adapters/ftp.py @@ -17,21 +17,21 @@ # limitations under the License. from __future__ import absolute_import, division, print_function, unicode_literals +from base64 import b64decode import cgi import ftplib -import os -from base64 import b64decode from io import BytesIO -from conda.gateways.logging import getLogger +import os + from requests import Response from requests.adapters import BaseAdapter from requests.hooks import dispatch_hook +from ..logging import getLogger from ...common.compat import StringIO from ...common.url import urlparse from ...exceptions import AuthenticationError - log = getLogger(__name__) diff --git a/conda/gateways/adapters/localfs.py b/conda/gateways/adapters/localfs.py --- a/conda/gateways/adapters/localfs.py +++ b/conda/gateways/adapters/localfs.py @@ -12,7 +12,7 @@ from tempfile import SpooledTemporaryFile from ...common.compat import ensure_binary -from conda.common.path import url_to_path +from ...common.path import url_to_path log = getLogger(__name__) diff --git a/conda/gateways/download.py b/conda/gateways/download.py --- a/conda/gateways/download.py +++ b/conda/gateways/download.py @@ -2,16 +2,16 @@ from __future__ import absolute_import, division, print_function, unicode_literals import hashlib -from logging import getLogger, DEBUG -from os.path import exists, basename +from logging import DEBUG, getLogger +from os.path import basename, exists from threading import Lock import warnings -from conda._vendor.auxlib.logz import stringify from requests.exceptions import ConnectionError, HTTPError, SSLError from .. import CondaError from .._vendor.auxlib.ish import dals +from .._vendor.auxlib.logz import stringify from ..base.context import context from ..connection import CondaSession from ..exceptions import BasicClobberError, CondaHTTPError, MD5MismatchError, maybe_raise diff --git a/conda/install.py b/conda/install.py --- a/conda/install.py +++ b/conda/install.py @@ -60,7 +60,7 @@ def win_conda_bat_redirect(src, dst, shell): Works of course only with callable files, e.g. `.bat` or `.exe` files. """ - from conda.utils import shells + from .utils import shells try: makedirs(dirname(dst)) except OSError as exc: # Python >2.5
Unexpected Error related to urllib3 Using Anaconda2-4.3.1-Linux-x86_64 in Linux 14.04.5 with Python 2.7. Anytime I use any conda command I get the following error: ```` >>$ conda -h An unexpected error has occurred. Please consider posting the following information to the conda GitHub issue tracker at: https://github.com/conda/conda/issues Traceback (most recent call last): File "/home/vmrguser/anaconda2/bin/conda", line 6, in <module> sys.exit(conda.cli.main()) File "/home/vmrguser/anaconda2/lib/python2.7/site-packages/conda/cli/main.py", line 164, in main return conda_exception_handler(_main, *args) File "/home/vmrguser/anaconda2/lib/python2.7/site-packages/conda/exceptions.py", line 589, in conda_exception_handler print_unexpected_error_message(e) File "/home/vmrguser/anaconda2/lib/python2.7/site-packages/conda/exceptions.py", line 543, in print_unexpected_error_message info_stdout, info_stderr = get_info() File "/home/vmrguser/anaconda2/lib/python2.7/site-packages/conda/exceptions.py", line 513, in get_info args.func(args, p) File "/home/vmrguser/anaconda2/lib/python2.7/site-packages/conda/cli/main_info.py", line 148, in execute from conda.api import get_index File "/home/vmrguser/anaconda2/lib/python2.7/site-packages/conda/api.py", line 1, in <module> from .core.index import get_index File "/home/vmrguser/anaconda2/lib/python2.7/site-packages/conda/core/index.py", line 9, in <module> from .repodata import collect_all_repodata File "/home/vmrguser/anaconda2/lib/python2.7/site-packages/conda/core/repodata.py", line 21, in <module> from requests.packages.urllib3.exceptions import InsecureRequestWarning ImportError: No module named packages.urllib3.exceptions ```` I have already tried to install/upgrade urllib3, requests, and boxsdk with pip, but to no avail. Anyone know how to fix this?
I download Anaconda(python 2.7 & ubuntu 64 bit ) and has the same problem. I use Anaconda2-4.1.1-Linux-x86_64 in ubuntu 14.04 64 bit , and this problem not occurs
2017-04-21T04:16:02Z
[]
[]
Traceback (most recent call last): File "/home/vmrguser/anaconda2/bin/conda", line 6, in <module> sys.exit(conda.cli.main()) File "/home/vmrguser/anaconda2/lib/python2.7/site-packages/conda/cli/main.py", line 164, in main return conda_exception_handler(_main, *args) File "/home/vmrguser/anaconda2/lib/python2.7/site-packages/conda/exceptions.py", line 589, in conda_exception_handler print_unexpected_error_message(e) File "/home/vmrguser/anaconda2/lib/python2.7/site-packages/conda/exceptions.py", line 543, in print_unexpected_error_message info_stdout, info_stderr = get_info() File "/home/vmrguser/anaconda2/lib/python2.7/site-packages/conda/exceptions.py", line 513, in get_info args.func(args, p) File "/home/vmrguser/anaconda2/lib/python2.7/site-packages/conda/cli/main_info.py", line 148, in execute from conda.api import get_index File "/home/vmrguser/anaconda2/lib/python2.7/site-packages/conda/api.py", line 1, in <module> from .core.index import get_index File "/home/vmrguser/anaconda2/lib/python2.7/site-packages/conda/core/index.py", line 9, in <module> from .repodata import collect_all_repodata File "/home/vmrguser/anaconda2/lib/python2.7/site-packages/conda/core/repodata.py", line 21, in <module> from requests.packages.urllib3.exceptions import InsecureRequestWarning ImportError: No module named packages.urllib3.exceptions
4,075
conda/conda
conda__conda-5232
0eddaa3089822cec0394b69b3c6a9ac294afafef
diff --git a/conda/cli/main_config.py b/conda/cli/main_config.py --- a/conda/cli/main_config.py +++ b/conda/cli/main_config.py @@ -423,8 +423,12 @@ def clean_element_type(element_types): # config.rc_keys if not args.get: - with open(rc_path, 'w') as rc: - rc.write(yaml_dump(rc_config)) + try: + with open(rc_path, 'w') as rc: + rc.write(yaml_dump(rc_config)) + except (IOError, OSError) as e: + raise CondaError('Cannot write to condarc file at %s\n' + 'Caused by %r' % (rc_path, e)) if context.json: stdout_json_success(
conda config stack trace when can't write config file This situation should be handled nicer. `conda config` doesn't have permission to write the config file. Thanks. ``` An unexpected error has occurred, please consider sending the following traceback to the conda GitHub issue tracker at: https://github.com/conda/conda/issues Include the output of the command 'conda info' in your report. Traceback (most recent call last): File "/opt/anaconda/bin/conda", line 5, in <module> sys.exit(main()) File "/opt/anaconda/lib/python2.7/site-packages/conda/cli/main.py", line 179, in main args.func(args, p) File "/opt/anaconda/lib/python2.7/site-packages/conda/cli/main_config.py", line 339, in execute with open(rc_path, 'w') as rc: IOError: [Errno 13] Permission denied: '/opt/anaconda/.condarc' ``` <!--- @huboard:{"order":9.781875224740546e-29,"custom_state":""} -->
2017-05-05T02:31:42Z
[]
[]
Traceback (most recent call last): File "/opt/anaconda/bin/conda", line 5, in <module> sys.exit(main()) File "/opt/anaconda/lib/python2.7/site-packages/conda/cli/main.py", line 179, in main args.func(args, p) File "/opt/anaconda/lib/python2.7/site-packages/conda/cli/main_config.py", line 339, in execute with open(rc_path, 'w') as rc: IOError: [Errno 13] Permission denied: '/opt/anaconda/.condarc'
4,106
conda/conda
conda__conda-5357
79f86d4cf31e6275cf3b201e76b43a95ecbb9888
diff --git a/conda/cli/conda_argparse.py b/conda/cli/conda_argparse.py --- a/conda/cli/conda_argparse.py +++ b/conda/cli/conda_argparse.py @@ -121,7 +121,7 @@ def error(self, message): else: argument = None if argument and argument.dest == "cmd": - m = re.compile(r"invalid choice: '([\w\-]+)'").match(exc.message) + m = re.compile(r"invalid choice: u?'([\w\-]+)'").match(exc.message) if m: cmd = m.group(1) executable = find_executable('conda-' + cmd)
latest 4.3.x (4.3.18-37-g79f86d4c) not picking up conda-build subcommands From conda-build's test suite: ``` ________________________________ test_skeleton_pypi ________________________________ Traceback (most recent call last): File "/home/dev/code/conda-build/tests/test_published_examples.py", line 15, in test_skeleton_pypi check_call_env(cmd.split()) File "/home/dev/code/conda-build/conda_build/utils.py", line 670, in check_call_env return _func_defaulting_env_to_os_environ(subprocess.check_call, *popenargs, **kwargs) File "/home/dev/code/conda-build/conda_build/utils.py", line 666, in _func_defaulting_env_to_os_environ return func(_args, **kwargs) File "/opt/miniconda/lib/python2.7/subprocess.py", line 541, in check_call raise CalledProcessError(retcode, cmd) CalledProcessError: Command '['conda', 'skeleton', 'pypi', 'pyinstrument']' returned non-zero exit status 2 ------------------------------- Captured stderr call ------------------------------- usage: conda [-h] [-V] command ... conda: error: argument command: invalid choice: u'skeleton' (choose from u'info', u'help', u'list', u'search', u'create', u'install', u'update', u'upgrade', u'remove', u'uninstall', u'config', u'clean', u'package') ``` This seems to happen only with python 2.7, not 3.6: https://travis-ci.org/conda/conda-build/builds/232848688
Am able to reproduce ``` kfranz@0283:~/continuum/conda-build *(no branch) ❯ /conda/bin/conda info Current conda install: platform : osx-64 conda version : 4.3.18.post37+79f86d4cf conda is private : False conda-env version : 4.3.18.post37+79f86d4cf conda-build version : 3.0.0rc0 python version : 2.7.12.final.0 requests version : 2.12.0 root environment : /conda (writable) default environment : /conda envs directories : /conda/envs /Users/kfranz/.conda/envs package cache : /conda/pkgs /Users/kfranz/.conda/pkgs channel URLs : https://repo.continuum.io/pkgs/free/osx-64 https://repo.continuum.io/pkgs/free/noarch https://repo.continuum.io/pkgs/r/osx-64 https://repo.continuum.io/pkgs/r/noarch https://repo.continuum.io/pkgs/pro/osx-64 https://repo.continuum.io/pkgs/pro/noarch config file : /Users/kfranz/.condarc netrc file : None offline mode : False user-agent : conda/4.3.18.post37+79f86d4cf requests/2.12.0 CPython/2.7.12 Darwin/16.5.0 OSX/10.12.4 UID:GID : 502:20 kfranz@0283:~/continuum/conda-build *(no branch) ❯ /conda/bin/conda skeleton -h usage: conda [-h] [-V] command ... conda: error: argument command: invalid choice: u'skeleton' (choose from u'info', u'help', u'list', u'search', u'create', u'install', u'update', u'upgrade', u'remove', u'uninstall', u'config', u'clean', u'package') ``` This could be my fault, checking.
2017-05-18T13:09:00Z
[]
[]
Traceback (most recent call last): File "/home/dev/code/conda-build/tests/test_published_examples.py", line 15, in test_skeleton_pypi check_call_env(cmd.split()) File "/home/dev/code/conda-build/conda_build/utils.py", line 670, in check_call_env return _func_defaulting_env_to_os_environ(subprocess.check_call, *popenargs, **kwargs) File "/home/dev/code/conda-build/conda_build/utils.py", line 666, in _func_defaulting_env_to_os_environ return func(_args, **kwargs) File "/opt/miniconda/lib/python2.7/subprocess.py", line 541, in check_call raise CalledProcessError(retcode, cmd) CalledProcessError: Command '['conda', 'skeleton', 'pypi', 'pyinstrument']' returned non-zero exit status 2
4,136
conda/conda
conda__conda-6442
5c0cf2afb1c2d0b1dddd7e396cf13f4bfaecf4d6
diff --git a/conda/models/match_spec.py b/conda/models/match_spec.py --- a/conda/models/match_spec.py +++ b/conda/models/match_spec.py @@ -349,10 +349,8 @@ def _make(field_name, value): matcher = value elif field_name in _implementors: matcher = _implementors[field_name](value) - elif text_type(value): - matcher = StrMatch(value) else: - raise NotImplementedError() + matcher = StrMatch(text_type(value)) return field_name, matcher
conda 4.4.0rc2 NotImplementedError on 'conda list' "command": "/home/rob/miniconda3/bin/conda list" ``` Traceback (most recent call last): File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/exceptions.py\", line 683, in __call__ return func(*args, **kwargs) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/cli/main.py\", line 78, in _main exit_code = do_call(args, p) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/cli/conda_argparse.py\", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/cli/main_list.py\", line 150, in execute show_channel_urls=context.show_channel_urls) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/cli/main_list.py\", line 90, in print_packages show_channel_urls=show_channel_urls) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/cli/main_list.py\", line 52, in list_packages info = is_linked(prefix, dist) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/core/linked_data.py\", line 166, in is_linked elif MatchSpec(dist).match(prefix_record): File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/models/match_spec.py\", line 67, in __call__ return super(MatchSpecType, cls).__call__(**parsed) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/models/match_spec.py\", line 178, in __init__ self._match_components = self._build_components(**kwargs) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/models/match_spec.py\", line 359, in _build_components return frozendict(_make(key, value) for key, value in iteritems(kwargs)) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/models/match_spec.py\", line 359, in <genexpr> return frozendict(_make(key, value) for key, value in iteritems(kwargs)) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/models/match_spec.py\", line 355, in _make raise NotImplementedError() NotImplementedError ```
2017-12-12T18:06:40Z
[]
[]
Traceback (most recent call last): File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/exceptions.py\", line 683, in __call__ return func(*args, **kwargs) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/cli/main.py\", line 78, in _main exit_code = do_call(args, p) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/cli/conda_argparse.py\", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/cli/main_list.py\", line 150, in execute show_channel_urls=context.show_channel_urls) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/cli/main_list.py\", line 90, in print_packages show_channel_urls=show_channel_urls) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/cli/main_list.py\", line 52, in list_packages info = is_linked(prefix, dist) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/core/linked_data.py\", line 166, in is_linked elif MatchSpec(dist).match(prefix_record): File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/models/match_spec.py\", line 67, in __call__ return super(MatchSpecType, cls).__call__(**parsed) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/models/match_spec.py\", line 178, in __init__ self._match_components = self._build_components(**kwargs) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/models/match_spec.py\", line 359, in _build_components return frozendict(_make(key, value) for key, value in iteritems(kwargs)) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/models/match_spec.py\", line 359, in <genexpr> return frozendict(_make(key, value) for key, value in iteritems(kwargs)) File \"/home/rob/miniconda3/lib/python3.6/site-packages/conda/models/match_spec.py\", line 355, in _make raise NotImplementedError() NotImplementedError
4,218
conda/conda
conda__conda-6511
eba59e3d6ba486b6e9a39864b1daacbc7f99f719
diff --git a/conda/base/context.py b/conda/base/context.py --- a/conda/base/context.py +++ b/conda/base/context.py @@ -982,7 +982,7 @@ def _first_writable_envs_dir(): return envs_dir from ..exceptions import NotWritableError - raise NotWritableError(context.envs_dirs[0]) + raise NotWritableError(context.envs_dirs[0], None) # backward compatibility for conda-build diff --git a/conda/exceptions.py b/conda/exceptions.py --- a/conda/exceptions.py +++ b/conda/exceptions.py @@ -591,12 +591,13 @@ def __init__(self, message): super(SafetyError, self).__init__(message) -class NotWritableError(CondaError): +class NotWritableError(CondaError, OSError): - def __init__(self, path): - kwargs = { + def __init__(self, path, errno, **kwargs): + kwargs.update({ 'path': path, - } + 'errno': errno, + }) if on_win: message = dals(""" The current user does not have write permissions to a required path. diff --git a/conda/gateways/disk/update.py b/conda/gateways/disk/update.py --- a/conda/gateways/disk/update.py +++ b/conda/gateways/disk/update.py @@ -12,6 +12,7 @@ from .link import lexists from ...common.compat import on_win from ...common.path import expand +from ...exceptions import NotWritableError log = getLogger(__name__) @@ -62,29 +63,32 @@ def touch(path, mkdir=False, sudo_safe=False): # True if the file did not exist but was created # False if the file already existed # raises: permissions errors such as EPERM and EACCES - path = expand(path) - log.trace("touching path %s", path) - if lexists(path): - utime(path, None) - return True - else: - dirpath = dirname(path) - if not isdir(dirpath) and mkdir: - if sudo_safe: - mkdir_p_sudo_safe(dirpath) - else: - mkdir_p(dirpath) - else: - assert isdir(dirname(path)) - try: - fh = open(path, 'a') - except: - raise + try: + path = expand(path) + log.trace("touching path %s", path) + if lexists(path): + utime(path, None) + return True else: - fh.close() - if sudo_safe and not on_win and os.environ.get('SUDO_UID') is not None: - uid = int(os.environ['SUDO_UID']) - gid = int(os.environ.get('SUDO_GID', -1)) - log.trace("chowning %s:%s %s", uid, gid, path) - os.chown(path, uid, gid) - return False + dirpath = dirname(path) + if not isdir(dirpath) and mkdir: + if sudo_safe: + mkdir_p_sudo_safe(dirpath) + else: + mkdir_p(dirpath) + else: + assert isdir(dirname(path)) + try: + fh = open(path, 'a') + except: + raise + else: + fh.close() + if sudo_safe and not on_win and os.environ.get('SUDO_UID') is not None: + uid = int(os.environ['SUDO_UID']) + gid = int(os.environ.get('SUDO_GID', -1)) + log.trace("chowning %s:%s %s", uid, gid, path) + os.chown(path, uid, gid) + return False + except (IOError, OSError) as e: + raise NotWritableError(path, e.errno, caused_by=e)
PermissionError writing to environments.txt "command": "conda install jupyter_core --yes" ``` Traceback (most recent call last): File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/exceptions.py\", line 721, in __call__ return func(*args, **kwargs) File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/cli/main.py\", line 78, in _main exit_code = do_call(args, p) File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/cli/conda_argparse.py\", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/cli/main_install.py\", line 11, in execute install(args, parser, 'install') File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/cli/install.py\", line 239, in install handle_txn(progressive_fetch_extract, unlink_link_transaction, prefix, args, newenv) File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/cli/install.py\", line 269, in handle_txn unlink_link_transaction.execute() File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/core/link.py\", line 220, in execute self.verify() File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/common/io.py\", line 402, in decorated return f(*args, **kwds) File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/core/link.py\", line 207, in verify exceptions = self._verify(self.prefix_setups, self.prefix_action_groups) File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/core/link.py\", line 460, in _verify cls._verify_transaction_level(prefix_setups), File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/core/link.py\", line 455, in <genexpr> exceptions = tuple(exc for exc in concatv( File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/core/link.py\", line 323, in _verify_individual_level error_result = axn.verify() File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/core/path_actions.py\", line 870, in verify touch(USER_ENVIRONMENTS_TXT_FILE, mkdir=True, sudo_safe=True) File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/gateways/disk/update.py\", line 89, in touch os.chown(path, uid, gid) PermissionError: [Errno 1] Operation not permitted: '/home/nbcommon/.conda/environments.txt' ```
2017-12-21T20:40:03Z
[]
[]
Traceback (most recent call last): File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/exceptions.py\", line 721, in __call__ return func(*args, **kwargs) File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/cli/main.py\", line 78, in _main exit_code = do_call(args, p) File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/cli/conda_argparse.py\", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/cli/main_install.py\", line 11, in execute install(args, parser, 'install') File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/cli/install.py\", line 239, in install handle_txn(progressive_fetch_extract, unlink_link_transaction, prefix, args, newenv) File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/cli/install.py\", line 269, in handle_txn unlink_link_transaction.execute() File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/core/link.py\", line 220, in execute self.verify() File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/common/io.py\", line 402, in decorated return f(*args, **kwds) File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/core/link.py\", line 207, in verify exceptions = self._verify(self.prefix_setups, self.prefix_action_groups) File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/core/link.py\", line 460, in _verify cls._verify_transaction_level(prefix_setups), File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/core/link.py\", line 455, in <genexpr> exceptions = tuple(exc for exc in concatv( File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/core/link.py\", line 323, in _verify_individual_level error_result = axn.verify() File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/core/path_actions.py\", line 870, in verify touch(USER_ENVIRONMENTS_TXT_FILE, mkdir=True, sudo_safe=True) File \"/home/nbcommon/anaconda3_420/lib/python3.5/site-packages/conda/gateways/disk/update.py\", line 89, in touch os.chown(path, uid, gid) PermissionError: [Errno 1] Operation not permitted: '/home/nbcommon/.conda/environments.txt'
4,228
conda/conda
conda__conda-6525
09912b87b21b0a66d638e65561ed126677846a91
diff --git a/conda/core/solve.py b/conda/core/solve.py --- a/conda/core/solve.py +++ b/conda/core/solve.py @@ -241,7 +241,7 @@ def solve_final_state(self, deps_modifier=NULL, prune=NULL, ignore_pinned=NULL, the output of 'conda info' and 'conda list' for the active environment, along with the command you invoked that resulted in this error. matches_for_spec: %s - """) % matches_for_spec) + """) % dashlist((text_type(s) for s in matches_for_spec), indent=4)) target_dist = matches_for_spec[0] if deps_modifier == DepsModifier.FREEZE_INSTALLED: new_spec = MatchSpec(index[target_dist])
TypeError: not all arguments converted during string formatting "command": "C:\\Users\\court\\Anaconda3\\Scripts\\conda install matplotlib" "user_agent": "conda/4.4.1 requests/2.18.4 CPython/3.6.3 Windows/10 Windows/10.0.16299" ``` Traceback (most recent call last): File \"C:\\Users\\court\\Anaconda3\\lib\\site-packages\\conda\\exceptions.py\", line 721, in __call__ return func(*args, **kwargs) File \"C:\\Users\\court\\Anaconda3\\lib\\site-packages\\conda\\cli\\main.py\", line 78, in _main exit_code = do_call(args, p) File \"C:\\Users\\court\\Anaconda3\\lib\\site-packages\\conda\\cli\\conda_argparse.py\", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File \"C:\\Users\\court\\Anaconda3\\lib\\site-packages\\conda\\cli\\main_install.py\", line 11, in execute install(args, parser, 'install') File \"C:\\Users\\court\\Anaconda3\\lib\\site-packages\\conda\\cli\\install.py\", line 220, in install force_reinstall=context.force, File \"C:\\Users\\court\\Anaconda3\\lib\\site-packages\\conda\\core\\solve.py\", line 503, in solve_for_transaction force_remove, force_reinstall) File \"C:\\Users\\court\\Anaconda3\\lib\\site-packages\\conda\\core\\solve.py\", line 436, in solve_for_diff final_precs = self.solve_final_state(deps_modifier, prune, ignore_pinned, force_remove) File \"C:\\Users\\court\\Anaconda3\\lib\\site-packages\\conda\\core\\solve.py\", line 244, in solve_final_state \"\"\") % matches_for_spec) TypeError: not all arguments converted during string formatting ```
2017-12-22T03:34:59Z
[]
[]
Traceback (most recent call last): File \"C:\\Users\\court\\Anaconda3\\lib\\site-packages\\conda\\exceptions.py\", line 721, in __call__ return func(*args, **kwargs) File \"C:\\Users\\court\\Anaconda3\\lib\\site-packages\\conda\\cli\\main.py\", line 78, in _main exit_code = do_call(args, p) File \"C:\\Users\\court\\Anaconda3\\lib\\site-packages\\conda\\cli\\conda_argparse.py\", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File \"C:\\Users\\court\\Anaconda3\\lib\\site-packages\\conda\\cli\\main_install.py\", line 11, in execute install(args, parser, 'install') File \"C:\\Users\\court\\Anaconda3\\lib\\site-packages\\conda\\cli\\install.py\", line 220, in install force_reinstall=context.force, File \"C:\\Users\\court\\Anaconda3\\lib\\site-packages\\conda\\core\\solve.py\", line 503, in solve_for_transaction force_remove, force_reinstall) File \"C:\\Users\\court\\Anaconda3\\lib\\site-packages\\conda\\core\\solve.py\", line 436, in solve_for_diff final_precs = self.solve_final_state(deps_modifier, prune, ignore_pinned, force_remove) File \"C:\\Users\\court\\Anaconda3\\lib\\site-packages\\conda\\core\\solve.py\", line 244, in solve_final_state \"\"\") % matches_for_spec) TypeError: not all arguments converted during string formatting
4,232
conda/conda
conda__conda-6540
6c5d2bab095f9a2a4f846515800dbcf38f2fac39
diff --git a/conda/common/io.py b/conda/common/io.py --- a/conda/common/io.py +++ b/conda/common/io.py @@ -5,6 +5,7 @@ from concurrent.futures import ThreadPoolExecutor from contextlib import contextmanager from enum import Enum +from errno import EPIPE, ESHUTDOWN from functools import wraps from itertools import cycle import json @@ -326,11 +327,17 @@ def spinner(message=None, enabled=True, json=False): if json: pass else: - if exception_raised: - sys.stdout.write("failed\n") - else: - sys.stdout.write("done\n") - sys.stdout.flush() + try: + if exception_raised: + sys.stdout.write("failed\n") + else: + sys.stdout.write("done\n") + sys.stdout.flush() + except (IOError, OSError) as e: + # Ignore BrokenPipeError and errors related to stdout or stderr being + # closed by a downstream program. + if e.errno not in (EPIPE, ESHUTDOWN): + raise class ProgressBar(object): @@ -368,12 +375,18 @@ def finish(self): self.update_to(1) def close(self): - if self.json: - sys.stdout.write('{"fetch":"%s","finished":true,"maxval":1,"progress":1}\n\0' - % self.description) - sys.stdout.flush() - elif self.enabled: - self.pbar.close() + try: + if self.json: + sys.stdout.write('{"fetch":"%s","finished":true,"maxval":1,"progress":1}\n\0' + % self.description) + sys.stdout.flush() + elif self.enabled: + self.pbar.close() + except (IOError, OSError) as e: + # Ignore BrokenPipeError and errors related to stdout or stderr being + # closed by a downstream program. + if e.errno not in (EPIPE, ESHUTDOWN): + raise self.enabled = False
BrokenPipeError "command": "conda create -p ./untitled -y python=3.6" "user_agent": "conda/4.4.2 requests/2.18.4 CPython/3.6.3 Linux/4.10.0-28-generic ubuntu/16.04 glibc/2.23" ``` Traceback (most recent call last): File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/exceptions.py\", line 722, in __call__ return func(*args, **kwargs) File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/cli/main.py\", line 78, in _main exit_code = do_call(args, p) File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/cli/conda_argparse.py\", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/cli/main_create.py\", line 11, in execute install(args, parser, 'create') File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/cli/install.py\", line 239, in install handle_txn(progressive_fetch_extract, unlink_link_transaction, prefix, args, newenv) File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/cli/install.py\", line 265, in handle_txn progressive_fetch_extract.execute() File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/core/package_cache.py\", line 543, in execute exc = self._execute_actions(prec_or_spec, prec_actions) File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/core/package_cache.py\", line 595, in _execute_actions progress_bar.close() File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/common/io.py\", line 376, in close self.pbar.close() File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/_vendor/tqdm/_tqdm.py\", line 1054, in close self.bar_format, self.postfix, self.unit_divisor)) File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/_vendor/tqdm/_tqdm.py\", line 201, in print_status fp_write('\\r' + s + (' ' * max(last_len[0] - len_s, 0))) File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/_vendor/tqdm/_tqdm.py\", line 195, in fp_write fp_flush() BrokenPipeError: [Errno 32] Broken pipe ```
2017-12-22T18:08:33Z
[]
[]
Traceback (most recent call last): File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/exceptions.py\", line 722, in __call__ return func(*args, **kwargs) File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/cli/main.py\", line 78, in _main exit_code = do_call(args, p) File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/cli/conda_argparse.py\", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/cli/main_create.py\", line 11, in execute install(args, parser, 'create') File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/cli/install.py\", line 239, in install handle_txn(progressive_fetch_extract, unlink_link_transaction, prefix, args, newenv) File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/cli/install.py\", line 265, in handle_txn progressive_fetch_extract.execute() File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/core/package_cache.py\", line 543, in execute exc = self._execute_actions(prec_or_spec, prec_actions) File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/core/package_cache.py\", line 595, in _execute_actions progress_bar.close() File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/common/io.py\", line 376, in close self.pbar.close() File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/_vendor/tqdm/_tqdm.py\", line 1054, in close self.bar_format, self.postfix, self.unit_divisor)) File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/_vendor/tqdm/_tqdm.py\", line 201, in print_status fp_write('\\r' + s + (' ' * max(last_len[0] - len_s, 0))) File \"/home/jon/anaconda3/lib/python3.6/site-packages/conda/_vendor/tqdm/_tqdm.py\", line 195, in fp_write fp_flush() BrokenPipeError: [Errno 32] Broken pipe
4,236
conda/conda
conda__conda-6542
84d8f425252dafadcaeed5a834300e058456881f
diff --git a/conda/core/envs_manager.py b/conda/core/envs_manager.py --- a/conda/core/envs_manager.py +++ b/conda/core/envs_manager.py @@ -7,7 +7,7 @@ from ..base.constants import ROOT_ENV_NAME from ..base.context import context -from ..common.compat import on_win +from ..common.compat import ensure_text_type, on_win, open from ..common.path import expand, paths_equal from ..gateways.disk.read import yield_lines from ..gateways.disk.test import is_conda_environment @@ -30,7 +30,7 @@ def register_env(location): return with open(USER_ENVIRONMENTS_TXT_FILE, 'a') as fh: - fh.write(location) + fh.write(ensure_text_type(location)) fh.write('\n') diff --git a/conda/gateways/disk/read.py b/conda/gateways/disk/read.py --- a/conda/gateways/disk/read.py +++ b/conda/gateways/disk/read.py @@ -19,7 +19,7 @@ from ..._vendor.auxlib.collection import first from ..._vendor.auxlib.ish import dals from ...base.constants import PREFIX_PLACEHOLDER -from ...common.compat import ensure_text_type +from ...common.compat import ensure_text_type, open from ...exceptions import CondaUpgradeError, CondaVerificationError, PathNotFoundError from ...models.channel import Channel from ...models.enums import FileMode, PathType
python 2 can't read unicode in environments.txt on a conda install operation, during the transaction ``` Traceback (most recent call last): File "/Users/kfranz/continuum/conda/conda/core/link.py", line 535, in _execute_actions action.execute() File "/Users/kfranz/continuum/conda/conda/core/path_actions.py", line 876, in execute register_env(self.target_prefix) File "/Users/kfranz/continuum/conda/conda/core/envs_manager.py", line 28, in register_env if location in yield_lines(USER_ENVIRONMENTS_TXT_FILE): File "/Users/kfranz/continuum/conda/conda/gateways/disk/read.py", line 49, in yield_lines if not line or line.startswith('#'): UnicodeDecodeError: 'ascii' codec can't decode byte 0xc3 in position 54: ordinal not in range(128) ```
Part of my `~/.conda/environments.txt` file that's causing python 2 to choke: ``` /usr/local/Cellar/python3/3.5.2_1/Frameworks/Python.framework/Versions/3.5/envs/shapely=1.6.0 /var/folders/cp/7r2s_s593j7_cpdtxxsmct880000gp/T/26b3 çêôñáß /var/folders/cp/7r2s_s593j7_cpdtxxsmct880000gp/T/8eba áçêßôñ /Users/kfranz/miniconda/envs/py2 /conda3 /Users/kfranz/miniconda/envs/_test2 ```
2017-12-22T18:45:25Z
[]
[]
Traceback (most recent call last): File "/Users/kfranz/continuum/conda/conda/core/link.py", line 535, in _execute_actions action.execute() File "/Users/kfranz/continuum/conda/conda/core/path_actions.py", line 876, in execute register_env(self.target_prefix) File "/Users/kfranz/continuum/conda/conda/core/envs_manager.py", line 28, in register_env if location in yield_lines(USER_ENVIRONMENTS_TXT_FILE): File "/Users/kfranz/continuum/conda/conda/gateways/disk/read.py", line 49, in yield_lines if not line or line.startswith('#'): UnicodeDecodeError: 'ascii' codec can't decode byte 0xc3 in position 54: ordinal not in range(128)
4,237
conda/conda
conda__conda-6558
2060458e0c546a3c70fa041e738cb73b41c4b0d3
diff --git a/conda/egg_info.py b/conda/egg_info.py --- a/conda/egg_info.py +++ b/conda/egg_info.py @@ -29,7 +29,7 @@ def get_site_packages_dir(installed_pkgs): def get_egg_info_files(sp_dir): - for fn in os.listdir(sp_dir): + for fn in (isdir(sp_dir) and os.listdir(sp_dir) or ()): if fn.endswith('.egg-link'): with open(join(sp_dir, fn), 'r') as reader: for egg in get_egg_info_files(reader.readline().strip()):
conda list FileNotFoundError ``` "command": "/Users/hideki/anaconda/bin/conda list" "user_agent": "conda/4.4.3 requests/2.14.2 CPython/3.5.2 Darwin/16.7.0 OSX/10.12.6" Traceback (most recent call last): File \"/Users/hideki/anaconda/lib/python3.5/site-packages/conda/exceptions.py\", line 722, in __call__ return func(*args, **kwargs) File \"/Users/hideki/anaconda/lib/python3.5/site-packages/conda/cli/main.py\", line 78, in _main exit_code = do_call(args, p) File \"/Users/hideki/anaconda/lib/python3.5/site-packages/conda/cli/conda_argparse.py\", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File \"/Users/hideki/anaconda/lib/python3.5/site-packages/conda/cli/main_list.py\", line 150, in execute show_channel_urls=context.show_channel_urls) File \"/Users/hideki/anaconda/lib/python3.5/site-packages/conda/cli/main_list.py\", line 85, in print_packages other_python = get_egg_info(prefix) File \"/Users/hideki/anaconda/lib/python3.5/site-packages/conda/egg_info.py\", line 86, in get_egg_info for path in get_egg_info_files(join(prefix, sp_dir)): File \"/Users/hideki/anaconda/lib/python3.5/site-packages/conda/egg_info.py\", line 35, in get_egg_info_files for egg in get_egg_info_files(reader.readline().strip()): File \"/Users/hideki/anaconda/lib/python3.5/site-packages/conda/egg_info.py\", line 32, in get_egg_info_files for fn in os.listdir(sp_dir): FileNotFoundError: [Errno 2] No such file or directory: '/Users/hideki/darkflow' ```
2017-12-23T07:04:44Z
[]
[]
Traceback (most recent call last): File \"/Users/hideki/anaconda/lib/python3.5/site-packages/conda/exceptions.py\", line 722, in __call__ return func(*args, **kwargs) File \"/Users/hideki/anaconda/lib/python3.5/site-packages/conda/cli/main.py\", line 78, in _main exit_code = do_call(args, p) File \"/Users/hideki/anaconda/lib/python3.5/site-packages/conda/cli/conda_argparse.py\", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File \"/Users/hideki/anaconda/lib/python3.5/site-packages/conda/cli/main_list.py\", line 150, in execute show_channel_urls=context.show_channel_urls) File \"/Users/hideki/anaconda/lib/python3.5/site-packages/conda/cli/main_list.py\", line 85, in print_packages other_python = get_egg_info(prefix) File \"/Users/hideki/anaconda/lib/python3.5/site-packages/conda/egg_info.py\", line 86, in get_egg_info for path in get_egg_info_files(join(prefix, sp_dir)): File \"/Users/hideki/anaconda/lib/python3.5/site-packages/conda/egg_info.py\", line 35, in get_egg_info_files for egg in get_egg_info_files(reader.readline().strip()): File \"/Users/hideki/anaconda/lib/python3.5/site-packages/conda/egg_info.py\", line 32, in get_egg_info_files for fn in os.listdir(sp_dir): FileNotFoundError: [Errno 2] No such file or directory: '/Users/hideki/darkflow'
4,240
conda/conda
conda__conda-6573
fbf3c0dac0667293d8926031119fb0798344d848
diff --git a/conda/core/repodata.py b/conda/core/repodata.py --- a/conda/core/repodata.py +++ b/conda/core/repodata.py @@ -4,7 +4,7 @@ import bz2 from collections import defaultdict from contextlib import closing -from errno import ENODEV +from errno import EACCES, ENODEV, EPERM from genericpath import getmtime, isfile import hashlib import json @@ -26,7 +26,7 @@ with_metaclass) from ..common.url import join_url, maybe_unquote from ..core.package_cache import PackageCache -from ..exceptions import CondaDependencyError, CondaHTTPError, CondaIndexError +from ..exceptions import CondaDependencyError, CondaHTTPError, CondaIndexError, NotWritableError from ..gateways.connection import (ConnectionError, HTTPError, InsecureRequestWarning, InvalidSchema, SSLError) from ..gateways.connection.session import CondaSession @@ -219,8 +219,14 @@ def _load(self): else: if not isdir(dirname(self.cache_path_json)): mkdir_p(dirname(self.cache_path_json)) - with open(self.cache_path_json, 'w') as fh: - fh.write(raw_repodata_str or '{}') + try: + with open(self.cache_path_json, 'w') as fh: + fh.write(raw_repodata_str or '{}') + except (IOError, OSError) as e: + if e.errno in (EACCES, EPERM): + raise NotWritableError(self.cache_path_json, e.errno, caused_by=e) + else: + raise _internal_state = self._process_raw_repodata_str(raw_repodata_str) self._internal_state = _internal_state self._pickle_me()
cached repodata.json PermissionsError ``` error: PermissionError(13, 'Permission denied') command: /Users/apple/.conda/envs/cafe/bin/conda uninstall protobuf user_agent: conda/4.4.3 requests/2.18.4 CPython/3.6.1 Darwin/16.1.0 OSX/10.12.1 _messageid: -9223372036845375609 _messagetime: 1514148478933 / 2017-12-24 14:47:58 _receipttime: 1514148478933 / 2017-12-24 14:47:58 Traceback (most recent call last): File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/exceptions.py", line 722, in __call__ return func(*args, **kwargs) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/cli/main.py", line 78, in _main exit_code = do_call(args, p) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/cli/conda_argparse.py", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/cli/main_remove.py", line 82, in execute txn = solver.solve_for_transaction(force_remove=args.force) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/solve.py", line 501, in solve_for_transaction force_remove, force_reinstall) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/solve.py", line 434, in solve_for_diff final_precs = self.solve_final_state(deps_modifier, prune, ignore_pinned, force_remove) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/solve.py", line 178, in solve_final_state index, r = self._prepare(prepared_specs) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/solve.py", line 557, in _prepare self.subdirs, prepared_specs) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/index.py", line 214, in get_reduced_index new_records = query_all(spec) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/index.py", line 187, in query_all return tuple(concat(future.result() for future in as_completed(futures))) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/repodata.py", line 99, in query self.load() File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/repodata.py", line 143, in load _internal_state = self._load() File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/repodata.py", line 222, in _load with open(self.cache_path_json, 'w') as fh: PermissionError: [Errno 13] Permission denied: '/Users/apple/miniconda3/pkgs/cache/809318c1.json' ```
2017-12-24T23:17:48Z
[]
[]
Traceback (most recent call last): File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/exceptions.py", line 722, in __call__ return func(*args, **kwargs) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/cli/main.py", line 78, in _main exit_code = do_call(args, p) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/cli/conda_argparse.py", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/cli/main_remove.py", line 82, in execute txn = solver.solve_for_transaction(force_remove=args.force) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/solve.py", line 501, in solve_for_transaction force_remove, force_reinstall) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/solve.py", line 434, in solve_for_diff final_precs = self.solve_final_state(deps_modifier, prune, ignore_pinned, force_remove) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/solve.py", line 178, in solve_final_state index, r = self._prepare(prepared_specs) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/solve.py", line 557, in _prepare self.subdirs, prepared_specs) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/index.py", line 214, in get_reduced_index new_records = query_all(spec) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/index.py", line 187, in query_all return tuple(concat(future.result() for future in as_completed(futures))) File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/repodata.py", line 99, in query self.load() File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/repodata.py", line 143, in load _internal_state = self._load() File "/Users/apple/miniconda3/lib/python3.6/site-packages/conda/core/repodata.py", line 222, in _load with open(self.cache_path_json, 'w') as fh: PermissionError: [Errno 13] Permission denied: '/Users/apple/miniconda3/pkgs/cache/809318c1.json'
4,246
conda/conda
conda__conda-6589
fbf3c0dac0667293d8926031119fb0798344d848
diff --git a/conda/base/context.py b/conda/base/context.py --- a/conda/base/context.py +++ b/conda/base/context.py @@ -1,6 +1,7 @@ # -*- coding: utf-8 -*- from __future__ import absolute_import, division, print_function, unicode_literals +from errno import ENOENT from logging import getLogger import os from os.path import (abspath, basename, dirname, expanduser, isdir, isfile, join, normpath, @@ -30,6 +31,16 @@ except ImportError: # pragma: no cover from .._vendor.toolz.itertoolz import concat, concatv, unique +try: + os.getcwd() +except (IOError, OSError) as e: + if e.errno == ENOENT: + # FileNotFoundError can occur when cwd has been deleted out from underneath the process. + # To resolve #6584, let's go with setting cwd to sys.prefix, and see how far we get. + os.chdir(sys.prefix) + else: + raise + log = getLogger(__name__) _platform_map = {
os.getcwd() FileNotFoundError ``` error: FileNotFoundError(2, 'No such file or directory') command: /root/anaconda3/bin/conda shell.posix deactivate user_agent: conda/4.4.4 requests/2.18.4 CPython/3.6.3 Linux/4.13.0-kali1-amd64 kali/2017.3 glibc/2.25 _messageid: -9223372036844975611 _messagetime: 1514233339489 / 2017-12-25 14:22:19 Traceback (most recent call last): File "/root/anaconda3/lib/python3.6/site-packages/conda/cli/main.py", line 110, in main from ..activate import main as activator_main File "/root/anaconda3/lib/python3.6/site-packages/conda/activate.py", line 12, in <module> context.__init__() # oOn import, context does not include SEARCH_PATH. This line fixes that. File "/root/anaconda3/lib/python3.6/site-packages/conda/base/context.py", line 214, in __init__ argparse_args=argparse_args) File "/root/anaconda3/lib/python3.6/site-packages/conda/common/configuration.py", line 723, in __init__ self._set_search_path(search_path) File "/root/anaconda3/lib/python3.6/site-packages/conda/common/configuration.py", line 729, in _set_search_path self._set_raw_data(load_file_configs(search_path)) File "/root/anaconda3/lib/python3.6/site-packages/conda/common/configuration.py", line 366, in load_file_configs expanded_paths = tuple(expand(path) for path in search_path) File "/root/anaconda3/lib/python3.6/site-packages/conda/common/configuration.py", line 366, in <genexpr> expanded_paths = tuple(expand(path) for path in search_path) File "/root/anaconda3/lib/python3.6/site-packages/conda/common/path.py", line 49, in expand return abspath(expanduser(expandvars(path))) File "/root/anaconda3/lib/python3.6/posixpath.py", line 374, in abspath cwd = os.getcwd() FileNotFoundError: [Errno 2] No such file or directory ```
``` error: FileNotFoundError(2, 'No such file or directory') command: /Users/limxing/anaconda3/bin/conda install iris user_agent: conda/4.4.4 requests/2.18.4 CPython/3.6.4 Darwin/17.3.0 OSX/10.13.2 _messageid: -9223372036844475580 _messagetime: 1514289512434 / 2017-12-26 05:58:32 Traceback (most recent call last): File "/Users/limxing/anaconda3/lib/python3.6/site-packages/conda/exceptions.py", line 724, in __call__ return func(*args, **kwargs) File "/Users/limxing/anaconda3/lib/python3.6/site-packages/conda/cli/main.py", line 74, in _main context.__init__(argparse_args=args) File "/Users/limxing/anaconda3/lib/python3.6/site-packages/conda/base/context.py", line 214, in __init__ argparse_args=argparse_args) File "/Users/limxing/anaconda3/lib/python3.6/site-packages/conda/common/configuration.py", line 723, in __init__ self._set_search_path(search_path) File "/Users/limxing/anaconda3/lib/python3.6/site-packages/conda/common/configuration.py", line 729, in _set_search_path self._set_raw_data(load_file_configs(search_path)) File "/Users/limxing/anaconda3/lib/python3.6/site-packages/conda/common/configuration.py", line 366, in load_file_configs expanded_paths = tuple(expand(path) for path in search_path) File "/Users/limxing/anaconda3/lib/python3.6/site-packages/conda/common/configuration.py", line 366, in <genexpr> expanded_paths = tuple(expand(path) for path in search_path) File "/Users/limxing/anaconda3/lib/python3.6/site-packages/conda/common/path.py", line 49, in expand return abspath(expanduser(expandvars(path))) File "/Users/limxing/anaconda3/lib/python3.6/posixpath.py", line 374, in abspath cwd = os.getcwd() FileNotFoundError: [Errno 2] No such file or directory ``` ``` error: FileNotFoundError(2, 'No such file or directory') command: /home/anti/miniconda3/bin/conda update --all user_agent: conda/4.4.4 requests/2.18.4 CPython/3.6.4 Linux/3.10.0-693.11.1.el7.x86_64 centos/7 glibc/2.17 _messageid: -9223372036844375596 _messagetime: 1514305043219 / 2017-12-26 10:17:23 Traceback (most recent call last): File "/home/anti/miniconda3/lib/python3.6/site-packages/conda/exceptions.py", line 724, in __call__ return func(*args, **kwargs) File "/home/anti/miniconda3/lib/python3.6/site-packages/conda/cli/main.py", line 74, in _main context.__init__(argparse_args=args) File "/home/anti/miniconda3/lib/python3.6/site-packages/conda/base/context.py", line 214, in __init__ argparse_args=argparse_args) File "/home/anti/miniconda3/lib/python3.6/site-packages/conda/common/configuration.py", line 723, in __init__ self._set_search_path(search_path) File "/home/anti/miniconda3/lib/python3.6/site-packages/conda/common/configuration.py", line 729, in _set_search_path self._set_raw_data(load_file_configs(search_path)) File "/home/anti/miniconda3/lib/python3.6/site-packages/conda/common/configuration.py", line 366, in load_file_configs expanded_paths = tuple(expand(path) for path in search_path) File "/home/anti/miniconda3/lib/python3.6/site-packages/conda/common/configuration.py", line 366, in <genexpr> expanded_paths = tuple(expand(path) for path in search_path) File "/home/anti/miniconda3/lib/python3.6/site-packages/conda/common/path.py", line 49, in expand return abspath(expanduser(expandvars(path))) File "/home/anti/miniconda3/lib/python3.6/posixpath.py", line 374, in abspath cwd = os.getcwd() FileNotFoundError: [Errno 2] No such file or directory ``` https://stackoverflow.com/questions/23889636/error-while-executing-os-getcwd
2017-12-27T20:26:53Z
[]
[]
Traceback (most recent call last): File "/root/anaconda3/lib/python3.6/site-packages/conda/cli/main.py", line 110, in main from ..activate import main as activator_main File "/root/anaconda3/lib/python3.6/site-packages/conda/activate.py", line 12, in <module> context.__init__() # oOn import, context does not include SEARCH_PATH. This line fixes that. File "/root/anaconda3/lib/python3.6/site-packages/conda/base/context.py", line 214, in __init__ argparse_args=argparse_args) File "/root/anaconda3/lib/python3.6/site-packages/conda/common/configuration.py", line 723, in __init__ self._set_search_path(search_path) File "/root/anaconda3/lib/python3.6/site-packages/conda/common/configuration.py", line 729, in _set_search_path self._set_raw_data(load_file_configs(search_path)) File "/root/anaconda3/lib/python3.6/site-packages/conda/common/configuration.py", line 366, in load_file_configs expanded_paths = tuple(expand(path) for path in search_path) File "/root/anaconda3/lib/python3.6/site-packages/conda/common/configuration.py", line 366, in <genexpr> expanded_paths = tuple(expand(path) for path in search_path) File "/root/anaconda3/lib/python3.6/site-packages/conda/common/path.py", line 49, in expand return abspath(expanduser(expandvars(path))) File "/root/anaconda3/lib/python3.6/posixpath.py", line 374, in abspath cwd = os.getcwd() FileNotFoundError: [Errno 2] No such file or directory
4,249
conda/conda
conda__conda-6616
f9f6c2195ae457c899883c49846d9292d3776ce8
diff --git a/conda/cli/common.py b/conda/cli/common.py --- a/conda/cli/common.py +++ b/conda/cli/common.py @@ -7,6 +7,7 @@ from .._vendor.auxlib.ish import dals from ..base.constants import ROOT_ENV_NAME from ..base.context import context +from ..common.io import swallow_broken_pipe from ..common.path import paths_equal from ..models.match_spec import MatchSpec @@ -155,6 +156,7 @@ def disp_features(features): return '' +@swallow_broken_pipe def stdout_json(d): import json from .._vendor.auxlib.entity import EntityEncoder diff --git a/conda/cli/main_search.py b/conda/cli/main_search.py --- a/conda/cli/main_search.py +++ b/conda/cli/main_search.py @@ -2,12 +2,11 @@ from __future__ import absolute_import, division, print_function, unicode_literals from collections import defaultdict -import sys from .install import calculate_channel_urls from ..base.context import context from ..cli.common import stdout_json -from ..common.io import spinner +from ..common.io import Spinner from ..compat import text_type from ..core.repodata import SubdirData from ..models.match_spec import MatchSpec @@ -25,7 +24,7 @@ def execute(args, parser): else: subdirs = context.subdirs - with spinner("Loading channels", not context.verbosity and not context.quiet, context.json): + with Spinner("Loading channels", not context.verbosity and not context.quiet, context.json): spec_channel = spec.get_exact_value('channel') channel_urls = (spec_channel,) if spec_channel else context.channels @@ -66,8 +65,7 @@ def execute(args, parser): record.build, record.schannel, )) - sys.stdout.write('\n'.join(builder)) - sys.stdout.write('\n') + print('\n'.join(builder)) def pretty_record(record): @@ -95,5 +93,4 @@ def push_line(display_name, attr_name): push_line("md5", "md5") builder.append("%-12s: %s" % ("dependencies", dashlist(record.depends))) builder.append('\n') - sys.stdout.write('\n'.join(builder)) - sys.stdout.write('\n') + print('\n'.join(builder)) diff --git a/conda/common/io.py b/conda/common/io.py --- a/conda/common/io.py +++ b/conda/common/io.py @@ -32,6 +32,39 @@ _FORMATTER = Formatter("%(levelname)s %(name)s:%(funcName)s(%(lineno)d): %(message)s") +class ContextDecorator(object): + """Base class for a context manager class (implementing __enter__() and __exit__()) that also + makes it a decorator. + """ + + # TODO: figure out how to improve this pattern so e.g. swallow_broken_pipe doesn't have to be instantiated # NOQA + + def __call__(self, f): + @wraps(f) + def decorated(*args, **kwds): + with self: + return f(*args, **kwds) + return decorated + + +class SwallowBrokenPipe(ContextDecorator): + # Ignore BrokenPipeError and errors related to stdout or stderr being + # closed by a downstream program. + + def __enter__(self): + pass + + def __exit__(self, exc_type, exc_val, exc_tb): + if (exc_val + and isinstance(exc_val, EnvironmentError) + and getattr(exc_val, 'errno', None) + and exc_val.errno in (EPIPE, ESHUTDOWN)): + return True + + +swallow_broken_pipe = SwallowBrokenPipe() + + class CaptureTarget(Enum): """Constants used for contextmanager captured. @@ -262,20 +295,35 @@ def interrupt(signum, frame): class Spinner(object): + """ + Args: + message (str): + A message to prefix the spinner with. The string ': ' is automatically appended. + enabled (bool): + If False, usage is a no-op. + json (bool): + If True, will not output non-json to stdout. + + """ + # spinner_cycle = cycle("⠋⠙⠹⠸⠼⠴⠦⠧⠇⠏") spinner_cycle = cycle('/-\\|') - def __init__(self, enable_spin=True): + def __init__(self, message, enabled=True, json=False): + self.message = message + self.enabled = enabled + self.json = json + self._stop_running = Event() self._spinner_thread = Thread(target=self._start_spinning) self._indicator_length = len(next(self.spinner_cycle)) + 1 self.fh = sys.stdout - self.show_spin = enable_spin and hasattr(self.fh, "isatty") and self.fh.isatty() + self.show_spin = enabled and not json and hasattr(self.fh, "isatty") and self.fh.isatty() def start(self): if self.show_spin: self._spinner_thread.start() - else: + elif not self.json: self.fh.write("...working... ") self.fh.flush() @@ -283,60 +331,37 @@ def stop(self): if self.show_spin: self._stop_running.set() self._spinner_thread.join() + self.show_spin = False def _start_spinning(self): - while not self._stop_running.is_set(): - self.fh.write(next(self.spinner_cycle) + ' ') - self.fh.flush() - sleep(0.10) - self.fh.write('\b' * self._indicator_length) + try: + while not self._stop_running.is_set(): + self.fh.write(next(self.spinner_cycle) + ' ') + self.fh.flush() + sleep(0.10) + self.fh.write('\b' * self._indicator_length) + except EnvironmentError as e: + if e.errno in (EPIPE, ESHUTDOWN): + self.stop() + else: + raise + @swallow_broken_pipe + def __enter__(self): + if not self.json: + sys.stdout.write("%s: " % self.message) + sys.stdout.flush() + self.start() -@contextmanager -def spinner(message=None, enabled=True, json=False): - """ - Args: - message (str, optional): - An optional message to prefix the spinner with. - If given, ': ' are automatically added. - enabled (bool): - If False, usage is a no-op. - json (bool): - If True, will not output non-json to stdout. - """ - sp = Spinner(enabled) - exception_raised = False - try: - if message: - if json: - pass - else: - sys.stdout.write("%s: " % message) + def __exit__(self, exc_type, exc_val, exc_tb): + self.stop() + if not self.json: + with swallow_broken_pipe: + if exc_type or exc_val: + sys.stdout.write("failed\n") + else: + sys.stdout.write("done\n") sys.stdout.flush() - if not json: - sp.start() - yield - except: - exception_raised = True - raise - finally: - if not json: - sp.stop() - if message: - if json: - pass - else: - try: - if exception_raised: - sys.stdout.write("failed\n") - else: - sys.stdout.write("done\n") - sys.stdout.flush() - except (IOError, OSError) as e: - # Ignore BrokenPipeError and errors related to stdout or stderr being - # closed by a downstream program. - if e.errno not in (EPIPE, ESHUTDOWN): - raise class ProgressBar(object): @@ -361,32 +386,38 @@ def __init__(self, description, enabled=True, json=False): pass elif enabled: bar_format = "{desc}{bar} | {percentage:3.0f}% " - self.pbar = tqdm(desc=description, bar_format=bar_format, ascii=True, total=1) + try: + self.pbar = tqdm(desc=description, bar_format=bar_format, ascii=True, total=1) + except EnvironmentError as e: + if e.errno in (EPIPE, ESHUTDOWN): + self.enabled = False + else: + raise def update_to(self, fraction): - if self.json: - sys.stdout.write('{"fetch":"%s","finished":false,"maxval":1,"progress":%f}\n\0' - % (self.description, fraction)) - elif self.enabled: - self.pbar.update(fraction - self.pbar.n) + try: + if self.json and self.enabled: + sys.stdout.write('{"fetch":"%s","finished":false,"maxval":1,"progress":%f}\n\0' + % (self.description, fraction)) + elif self.enabled: + self.pbar.update(fraction - self.pbar.n) + except EnvironmentError as e: + if e.errno in (EPIPE, ESHUTDOWN): + self.enabled = False + else: + raise def finish(self): self.update_to(1) + @swallow_broken_pipe def close(self): - try: - if self.json: - sys.stdout.write('{"fetch":"%s","finished":true,"maxval":1,"progress":1}\n\0' - % self.description) - sys.stdout.flush() - elif self.enabled: - self.pbar.close() - except (IOError, OSError) as e: - # Ignore BrokenPipeError and errors related to stdout or stderr being - # closed by a downstream program. - if e.errno not in (EPIPE, ESHUTDOWN): - raise - self.enabled = False + if self.enabled and self.json: + sys.stdout.write('{"fetch":"%s","finished":true,"maxval":1,"progress":1}\n\0' + % self.description) + sys.stdout.flush() + elif self.enabled: + self.pbar.close() class ThreadLimitedThreadPoolExecutor(ThreadPoolExecutor): @@ -433,15 +464,6 @@ def submit(self, fn, *args, **kwargs): as_completed = as_completed # anchored here for import by other modules -class ContextDecorator(object): - def __call__(self, f): - @wraps(f) - def decorated(*args, **kwds): - with self: - return f(*args, **kwds) - return decorated - - class time_recorder(ContextDecorator): # pragma: no cover start_time = None record_file = expand(join('~', '.conda', 'instrumentation-record.csv')) diff --git a/conda/core/link.py b/conda/core/link.py --- a/conda/core/link.py +++ b/conda/core/link.py @@ -24,7 +24,7 @@ from ..base.constants import SafetyChecks from ..base.context import context from ..common.compat import ensure_text_type, iteritems, itervalues, odict, on_win -from ..common.io import spinner, time_recorder +from ..common.io import Spinner, time_recorder from ..common.path import (explode_directories, get_all_directories, get_major_minor_version, get_python_site_packages_short_path) from ..common.signals import signal_handler @@ -181,7 +181,7 @@ def prepare(self): self.transaction_context = {} - with spinner("Preparing transaction", not context.verbosity and not context.quiet, + with Spinner("Preparing transaction", not context.verbosity and not context.quiet, context.json): for stp in itervalues(self.prefix_setups): grps = self._prepare(self.transaction_context, stp.target_prefix, @@ -202,7 +202,7 @@ def verify(self): self._verified = True return - with spinner("Verifying transaction", not context.verbosity and not context.quiet, + with Spinner("Verifying transaction", not context.verbosity and not context.quiet, context.json): exceptions = self._verify(self.prefix_setups, self.prefix_action_groups) if exceptions: @@ -466,7 +466,7 @@ def _execute(cls, all_action_groups): with signal_handler(conda_signal_handler), time_recorder("unlink_link_execute"): pkg_idx = 0 try: - with spinner("Executing transaction", not context.verbosity and not context.quiet, + with Spinner("Executing transaction", not context.verbosity and not context.quiet, context.json): for pkg_idx, axngroup in enumerate(all_action_groups): cls._execute_actions(pkg_idx, axngroup) @@ -483,7 +483,7 @@ def _execute(cls, all_action_groups): # reverse all executed packages except the one that failed rollback_excs = [] if context.rollback_enabled: - with spinner("Rolling back transaction", + with Spinner("Rolling back transaction", not context.verbosity and not context.quiet, context.json): failed_pkg_idx = pkg_idx reverse_actions = reversed(tuple(enumerate( diff --git a/conda/core/solve.py b/conda/core/solve.py --- a/conda/core/solve.py +++ b/conda/core/solve.py @@ -18,7 +18,7 @@ from ..base.context import context from ..common.compat import iteritems, itervalues, odict, string_types, text_type from ..common.constants import NULL -from ..common.io import spinner +from ..common.io import Spinner from ..common.path import get_major_minor_version, paths_equal from ..exceptions import PackagesNotFoundError from ..gateways.logging import TRACE @@ -492,7 +492,7 @@ def solve_for_transaction(self, deps_modifier=NULL, prune=NULL, ignore_pinned=NU UnlinkLinkTransaction: """ - with spinner("Solving environment", not context.verbosity and not context.quiet, + with Spinner("Solving environment", not context.verbosity and not context.quiet, context.json): if self.prefix == context.root_prefix and context.enable_private_envs: # This path has the ability to generate a multi-prefix transaction. The basic logic @@ -516,7 +516,7 @@ def solve_for_transaction(self, deps_modifier=NULL, prune=NULL, ignore_pinned=NU ) if conda_newer_records: latest_version = conda_newer_records[-1].version - sys.stderr.write(dedent(""" + print(dedent(""" ==> WARNING: A newer version of conda exists. <== current version: %s @@ -526,8 +526,7 @@ def solve_for_transaction(self, deps_modifier=NULL, prune=NULL, ignore_pinned=NU $ conda update -n base conda - - """) % (CONDA_VERSION, latest_version)) + """) % (CONDA_VERSION, latest_version), file=sys.stderr) return UnlinkLinkTransaction(stp)
More BrokenPipeError ``` error: BrokenPipeError(32, 'Broken pipe') command: /home/ubuntu/anaconda3/envs/tensorflow_p36/bin/conda install -p /home/ubuntu/anaconda3/envs/tensorflow_p36 tensorflow-gpu -y user_agent: conda/4.4.4 requests/2.18.4 CPython/3.6.3 Linux/4.4.0-1044-aws ubuntu/16.04 glibc/2.23 _messageid: -9223372036845075603 _messagetime: 1514219033248 / 2017-12-25 10:23:53 Traceback (most recent call last): File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/exceptions.py", line 724, in __call__ return func(*args, **kwargs) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/cli/main.py", line 78, in _main exit_code = do_call(args, p) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/cli/conda_argparse.py", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/cli/main_install.py", line 11, in execute install(args, parser, 'install') File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/cli/install.py", line 255, in install handle_txn(progressive_fetch_extract, unlink_link_transaction, prefix, args, newenv) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/cli/install.py", line 281, in handle_txn progressive_fetch_extract.execute() File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/core/package_cache.py", line 575, in execute exc = self._execute_actions(prec_or_spec, prec_actions) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/core/package_cache.py", line 590, in _execute_actions progress_bar = ProgressBar(desc, not context.verbosity and not context.quiet, context.json) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/common/io.py", line 365, in __init__ self.pbar = tqdm(desc=description, bar_format=bar_format, ascii=True, total=1) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/_vendor/tqdm/_tqdm.py", line 767, in __init__ self.postfix, unit_divisor)) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/_vendor/tqdm/_tqdm.py", line 201, in print_status fp_write('\r' + s + (' ' * max(last_len[0] - len_s, 0))) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/_vendor/tqdm/_tqdm.py", line 195, in fp_write fp_flush() BrokenPipeError: [Errno 32] Broken pipe ```
``` error: BrokenPipeError(32, 'Broken pipe') command: /Users/Rex/anaconda3/bin/conda install --yes --json --force-pscheck --prefix /Users/Rex/anaconda3/envs/test spyder==3.2.5 user_agent: conda/4.4.4 requests/2.18.4 CPython/3.6.3 Darwin/17.3.0 OSX/10.13.2 _messageid: -9223372036843575583 _messagetime: 1514436380331 / 2017-12-27 22:46:20 Traceback (most recent call last): File "/Users/Rex/anaconda3/lib/python3.6/site-packages/conda/exceptions.py", line 724, in __call__ return func(*args, **kwargs) File "/Users/Rex/anaconda3/lib/python3.6/site-packages/conda/cli/main.py", line 78, in _main exit_code = do_call(args, p) File "/Users/Rex/anaconda3/lib/python3.6/site-packages/conda/cli/conda_argparse.py", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File "/Users/Rex/anaconda3/lib/python3.6/site-packages/conda/cli/main_install.py", line 11, in execute install(args, parser, 'install') File "/Users/Rex/anaconda3/lib/python3.6/site-packages/conda/cli/install.py", line 255, in install handle_txn(progressive_fetch_extract, unlink_link_transaction, prefix, args, newenv) File "/Users/Rex/anaconda3/lib/python3.6/site-packages/conda/cli/install.py", line 296, in handle_txn common.stdout_json_success(prefix=prefix, actions=actions) File "/Users/Rex/anaconda3/lib/python3.6/site-packages/conda/cli/common.py", line 168, in stdout_json_success stdout_json(result) File "/Users/Rex/anaconda3/lib/python3.6/site-packages/conda/cli/common.py", line 161, in stdout_json json.dump(d, sys.stdout, indent=2, sort_keys=True, cls=EntityEncoder) File "/Users/Rex/anaconda3/lib/python3.6/json/__init__.py", line 180, in dump fp.write(chunk) BrokenPipeError: [Errno 32] Broken pipe ```
2017-12-30T23:28:33Z
[]
[]
Traceback (most recent call last): File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/exceptions.py", line 724, in __call__ return func(*args, **kwargs) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/cli/main.py", line 78, in _main exit_code = do_call(args, p) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/cli/conda_argparse.py", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/cli/main_install.py", line 11, in execute install(args, parser, 'install') File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/cli/install.py", line 255, in install handle_txn(progressive_fetch_extract, unlink_link_transaction, prefix, args, newenv) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/cli/install.py", line 281, in handle_txn progressive_fetch_extract.execute() File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/core/package_cache.py", line 575, in execute exc = self._execute_actions(prec_or_spec, prec_actions) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/core/package_cache.py", line 590, in _execute_actions progress_bar = ProgressBar(desc, not context.verbosity and not context.quiet, context.json) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/common/io.py", line 365, in __init__ self.pbar = tqdm(desc=description, bar_format=bar_format, ascii=True, total=1) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/_vendor/tqdm/_tqdm.py", line 767, in __init__ self.postfix, unit_divisor)) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/_vendor/tqdm/_tqdm.py", line 201, in print_status fp_write('\r' + s + (' ' * max(last_len[0] - len_s, 0))) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/conda/_vendor/tqdm/_tqdm.py", line 195, in fp_write fp_flush() BrokenPipeError: [Errno 32] Broken pipe
4,253
conda/conda
conda__conda-6619
114e4b3cfb0fae3d566fd08a0c0c1d7e24c19891
diff --git a/conda/cli/main_config.py b/conda/cli/main_config.py --- a/conda/cli/main_config.py +++ b/conda/cli/main_config.py @@ -11,11 +11,12 @@ from .. import CondaError from .._vendor.auxlib.entity import EntityEncoder +from ..base.constants import PathConflict, SafetyChecks from ..base.context import context, sys_rc_path, user_rc_path from ..common.compat import isiterable, iteritems, itervalues, string_types, text_type from ..common.configuration import pretty_list, pretty_map from ..common.io import timeout -from ..common.serialize import yaml_dump, yaml_load +from ..common.serialize import yaml, yaml_dump, yaml_load try: from cytoolz.itertoolz import concat, groupby @@ -309,6 +310,17 @@ def execute_config(args, parser): # config.rc_keys if not args.get: + + # Add representers for enums. + # Because a representer cannot be added for the base Enum class (it must be added for + # each specific Enum subclass), and because of import rules), I don't know of a better + # location to do this. + def enum_representer(dumper, data): + return dumper.represent_str(str(data)) + + yaml.representer.RoundTripRepresenter.add_representer(SafetyChecks, enum_representer) + yaml.representer.RoundTripRepresenter.add_representer(PathConflict, enum_representer) + try: with open(rc_path, 'w') as rc: rc.write(yaml_dump(rc_config))
RepresenterError('cannot represent an object: prevent',) ``` error: RepresenterError('cannot represent an object: prevent',) command: C:\tools\Anaconda3\Scripts\conda config --set path_conflict prevent user_agent: conda/4.4.4 requests/2.18.4 CPython/3.6.3 Windows/10 Windows/10.0.16299 _messageid: -9223372036843225525 _messagetime: 1514570427640 / 2017-12-29 12:00:27 Traceback (most recent call last): File "C:\tools\Anaconda3\lib\site-packages\conda\exceptions.py", line 724, in __call__ return func(*args, **kwargs) File "C:\tools\Anaconda3\lib\site-packages\conda\cli\main.py", line 78, in _main exit_code = do_call(args, p) File "C:\tools\Anaconda3\lib\site-packages\conda\cli\conda_argparse.py", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File "C:\tools\Anaconda3\lib\site-packages\conda\cli\main_config.py", line 29, in execute execute_config(args, parser) File "C:\tools\Anaconda3\lib\site-packages\conda\cli\main_config.py", line 312, in execute_config rc.write(yaml_dump(rc_config)) File "C:\tools\Anaconda3\lib\site-packages\conda\common\serialize.py", line 81, in yaml_dump indent=2) File "C:\tools\Anaconda3\lib\site-packages\ruamel_yaml\main.py", line 254, in dump version=version, tags=tags, block_seq_indent=block_seq_indent) File "C:\tools\Anaconda3\lib\site-packages\ruamel_yaml\main.py", line 224, in dump_all dumper.represent(data) File "C:\tools\Anaconda3\lib\site-packages\ruamel_yaml\representer.py", line 49, in represent node = self.represent_data(data) File "C:\tools\Anaconda3\lib\site-packages\ruamel_yaml\representer.py", line 83, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "C:\tools\Anaconda3\lib\site-packages\ruamel_yaml\representer.py", line 852, in represent_dict return self.represent_mapping(tag, data) File "C:\tools\Anaconda3\lib\site-packages\ruamel_yaml\representer.py", line 698, in represent_mapping node_value = self.represent_data(item_value) File "C:\tools\Anaconda3\lib\site-packages\ruamel_yaml\representer.py", line 93, in represent_data node = self.yaml_representers[None](self, data) File "C:\tools\Anaconda3\lib\site-packages\ruamel_yaml\representer.py", line 324, in represent_undefined raise RepresenterError("cannot represent an object: %s" % data) ruamel_yaml.representer.RepresenterError: cannot represent an object: prevent ```
2018-01-01T00:01:33Z
[]
[]
Traceback (most recent call last): File "C:\tools\Anaconda3\lib\site-packages\conda\exceptions.py", line 724, in __call__ return func(*args, **kwargs) File "C:\tools\Anaconda3\lib\site-packages\conda\cli\main.py", line 78, in _main exit_code = do_call(args, p) File "C:\tools\Anaconda3\lib\site-packages\conda\cli\conda_argparse.py", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File "C:\tools\Anaconda3\lib\site-packages\conda\cli\main_config.py", line 29, in execute execute_config(args, parser) File "C:\tools\Anaconda3\lib\site-packages\conda\cli\main_config.py", line 312, in execute_config rc.write(yaml_dump(rc_config)) File "C:\tools\Anaconda3\lib\site-packages\conda\common\serialize.py", line 81, in yaml_dump indent=2) File "C:\tools\Anaconda3\lib\site-packages\ruamel_yaml\main.py", line 254, in dump version=version, tags=tags, block_seq_indent=block_seq_indent) File "C:\tools\Anaconda3\lib\site-packages\ruamel_yaml\main.py", line 224, in dump_all dumper.represent(data) File "C:\tools\Anaconda3\lib\site-packages\ruamel_yaml\representer.py", line 49, in represent node = self.represent_data(data) File "C:\tools\Anaconda3\lib\site-packages\ruamel_yaml\representer.py", line 83, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "C:\tools\Anaconda3\lib\site-packages\ruamel_yaml\representer.py", line 852, in represent_dict return self.represent_mapping(tag, data) File "C:\tools\Anaconda3\lib\site-packages\ruamel_yaml\representer.py", line 698, in represent_mapping node_value = self.represent_data(item_value) File "C:\tools\Anaconda3\lib\site-packages\ruamel_yaml\representer.py", line 93, in represent_data node = self.yaml_representers[None](self, data) File "C:\tools\Anaconda3\lib\site-packages\ruamel_yaml\representer.py", line 324, in represent_undefined raise RepresenterError("cannot represent an object: %s" % data) ruamel_yaml.representer.RepresenterError: cannot represent an object: prevent
4,254
conda/conda
conda__conda-6652
114e4b3cfb0fae3d566fd08a0c0c1d7e24c19891
diff --git a/conda/activate.py b/conda/activate.py --- a/conda/activate.py +++ b/conda/activate.py @@ -283,7 +283,8 @@ def build_activate(self, env_name_or_prefix): def build_deactivate(self): # query environment old_conda_shlvl = int(os.getenv('CONDA_SHLVL', 0)) - if old_conda_shlvl <= 0: + old_conda_prefix = os.getenv('CONDA_PREFIX', None) + if old_conda_shlvl <= 0 or old_conda_prefix is None: return { 'unset_vars': (), 'set_vars': {}, @@ -291,7 +292,6 @@ def build_deactivate(self): 'deactivate_scripts': (), 'activate_scripts': (), } - old_conda_prefix = os.environ['CONDA_PREFIX'] deactivate_scripts = self._get_deactivate_scripts(old_conda_prefix) new_conda_shlvl = old_conda_shlvl - 1 @@ -548,12 +548,12 @@ def main(argv=None): activator_args = argv[2:] activator = Activator(shell, activator_args) try: - sys.stdout.write(activator.execute()) + print(activator.execute(), end='') return 0 except Exception as e: from . import CondaError if isinstance(e, CondaError): - sys.stderr.write(text_type(e)) + print(text_type(e), file=sys.stderr) return e.return_code else: raise
KeyError(u'CONDA_PREFIX',) ``` error: KeyError(u'CONDA_PREFIX',) command: D:\Anaconda2\Scripts\conda shell.cmd.exe deactivate deactivate user_agent: conda/4.4.4 requests/2.18.4 CPython/2.7.14 Windows/10 Windows/10.0.16299 _messageid: -9223372036842425555 _messagetime: 1514685063692 / 2017-12-30 19:51:03 Traceback (most recent call last): File "D:\Anaconda2\lib\site-packages\conda\cli\main.py", line 111, in main return activator_main() File "D:\Anaconda2\lib\site-packages\conda\activate.py", line 529, in main sys.stdout.write(activator.execute()) File "D:\Anaconda2\lib\site-packages\conda\activate.py", line 149, in execute return getattr(self, self.command)() File "D:\Anaconda2\lib\site-packages\conda\activate.py", line 139, in deactivate return self._finalize(self._yield_commands(self.build_deactivate()), File "D:\Anaconda2\lib\site-packages\conda\activate.py", line 288, in build_deactivate old_conda_prefix = os.environ['CONDA_PREFIX'] File "D:\Anaconda2\lib\os.py", line 425, in __getitem__ return self.data[key.upper()] KeyError: u'CONDA_PREFIX' ```
2018-01-04T20:35:44Z
[]
[]
Traceback (most recent call last): File "D:\Anaconda2\lib\site-packages\conda\cli\main.py", line 111, in main return activator_main() File "D:\Anaconda2\lib\site-packages\conda\activate.py", line 529, in main sys.stdout.write(activator.execute()) File "D:\Anaconda2\lib\site-packages\conda\activate.py", line 149, in execute return getattr(self, self.command)() File "D:\Anaconda2\lib\site-packages\conda\activate.py", line 139, in deactivate return self._finalize(self._yield_commands(self.build_deactivate()), File "D:\Anaconda2\lib\site-packages\conda\activate.py", line 288, in build_deactivate old_conda_prefix = os.environ['CONDA_PREFIX'] File "D:\Anaconda2\lib\os.py", line 425, in __getitem__ return self.data[key.upper()] KeyError: u'CONDA_PREFIX'
4,258
conda/conda
conda__conda-6653
4d963e1da043f3eb8b0430263a6185db134f33df
diff --git a/conda/common/io.py b/conda/common/io.py --- a/conda/common/io.py +++ b/conda/common/io.py @@ -2,7 +2,8 @@ from __future__ import absolute_import, division, print_function, unicode_literals from collections import defaultdict -from concurrent.futures import ThreadPoolExecutor +from concurrent.futures import ThreadPoolExecutor, _base, as_completed +from concurrent.futures.thread import _WorkItem from contextlib import contextmanager from enum import Enum from errno import EPIPE, ESHUTDOWN @@ -11,7 +12,6 @@ import json import logging from logging import CRITICAL, Formatter, NOTSET, StreamHandler, WARN, getLogger -from math import floor import os from os.path import dirname, isdir, isfile, join import signal @@ -35,7 +35,7 @@ class CaptureTarget(Enum): """Constants used for contextmanager captured. - Used similarily like the constants PIPE, STDOUT for stdlib's subprocess.Popen. + Used similarly like the constants PIPE, STDOUT for stdlib's subprocess.Popen. """ STRING = -1 STDOUT = -2 @@ -303,7 +303,6 @@ def spinner(message=None, enabled=True, json=False): If False, usage is a no-op. json (bool): If True, will not output non-json to stdout. - """ sp = Spinner(enabled) exception_raised = False @@ -390,21 +389,48 @@ def close(self): self.enabled = False -@contextmanager -def backdown_thread_pool(max_workers=10): - """Tries to create an executor with max_workers, but will back down ultimately to a single - thread of the OS decides you can't have more than one. - """ - try: - yield ThreadPoolExecutor(max_workers) - except RuntimeError as e: # pragma: no cover - # RuntimeError is thrown if number of threads are limited by OS - log.debug(repr(e)) - try: - yield ThreadPoolExecutor(floor(max_workers / 2)) - except RuntimeError as e: - log.debug(repr(e)) - yield ThreadPoolExecutor(1) +class ThreadLimitedThreadPoolExecutor(ThreadPoolExecutor): + + def __init__(self, max_workers=10): + super(ThreadLimitedThreadPoolExecutor, self).__init__(max_workers) + + def submit(self, fn, *args, **kwargs): + """ + This is an exact reimplementation of the `submit()` method on the parent class, except + with an added `try/except` around `self._adjust_thread_count()`. So long as there is at + least one living thread, this thread pool will not throw an exception if threads cannot + be expanded to `max_workers`. + + In the implementation, we use "protected" attributes from concurrent.futures (`_base` + and `_WorkItem`). Consider vendoring the whole concurrent.futures library + as an alternative to these protected imports. + + https://github.com/agronholm/pythonfutures/blob/3.2.0/concurrent/futures/thread.py#L121-L131 # NOQA + https://github.com/python/cpython/blob/v3.6.4/Lib/concurrent/futures/thread.py#L114-L124 + """ + with self._shutdown_lock: + if self._shutdown: + raise RuntimeError('cannot schedule new futures after shutdown') + + f = _base.Future() + w = _WorkItem(f, fn, args, kwargs) + + self._work_queue.put(w) + try: + self._adjust_thread_count() + except RuntimeError: + # RuntimeError: can't start new thread + # See https://github.com/conda/conda/issues/6624 + if len(self._threads) > 0: + # It's ok to not be able to start new threads if we already have at least + # one thread alive. + pass + else: + raise + return f + + +as_completed = as_completed # anchored here for import by other modules class ContextDecorator(object): @@ -427,6 +453,7 @@ def __enter__(self): enabled = os.environ.get('CONDA_INSTRUMENTATION_ENABLED') if enabled and boolify(enabled): self.start_time = time() + return self def __exit__(self, exc_type, exc_val, exc_tb): if self.start_time: diff --git a/conda/core/index.py b/conda/core/index.py --- a/conda/core/index.py +++ b/conda/core/index.py @@ -1,7 +1,6 @@ # -*- coding: utf-8 -*- from __future__ import absolute_import, division, print_function, unicode_literals -from concurrent.futures import as_completed from itertools import chain from logging import getLogger @@ -11,7 +10,7 @@ from .._vendor.boltons.setutils import IndexedSet from ..base.context import context from ..common.compat import iteritems, itervalues -from ..common.io import backdown_thread_pool, time_recorder +from ..common.io import ThreadLimitedThreadPoolExecutor, as_completed, time_recorder from ..exceptions import OperationNotAllowed from ..models.channel import Channel, all_channel_urls from ..models.dist import Dist @@ -162,7 +161,7 @@ def get_reduced_index(prefix, channels, subdirs, specs): # keep_specs.append(spec) # consolidated_specs.update(keep_specs) - with backdown_thread_pool() as executor: + with ThreadLimitedThreadPoolExecutor() as executor: channel_urls = all_channel_urls(channels, subdirs=subdirs) check_whitelist(channel_urls) diff --git a/conda/core/repodata.py b/conda/core/repodata.py --- a/conda/core/repodata.py +++ b/conda/core/repodata.py @@ -23,6 +23,7 @@ from ..base.context import context from ..common.compat import (ensure_binary, ensure_text_type, ensure_unicode, text_type, with_metaclass) +from ..common.io import ThreadLimitedThreadPoolExecutor, as_completed from ..common.url import join_url, maybe_unquote from ..core.package_cache import PackageCache from ..exceptions import CondaDependencyError, CondaHTTPError, CondaIndexError, NotWritableError @@ -75,23 +76,14 @@ class SubdirData(object): @staticmethod def query_all(channels, subdirs, package_ref_or_match_spec): + from .index import check_whitelist # TODO: fix in-line import channel_urls = all_channel_urls(channels, subdirs=subdirs) - - executor = None - try: - from concurrent.futures import ThreadPoolExecutor, as_completed - executor = ThreadPoolExecutor(10) + check_whitelist(channel_urls) + with ThreadLimitedThreadPoolExecutor() as executor: futures = (executor.submit( SubdirData(Channel(url)).query, package_ref_or_match_spec ) for url in channel_urls) return tuple(concat(future.result() for future in as_completed(futures))) - except RuntimeError as e: # pragma: no cover - # concurrent.futures is only available in Python >= 3.2 or if futures is installed - # RuntimeError is thrown if number of threads are limited by OS - raise - finally: - if executor: - executor.shutdown(wait=True) def query(self, package_ref_or_match_spec): if not self._loaded:
RuntimeError: can't start new thread ``` error: RuntimeError("generator didn't stop after throw()",) command: /home/mholekev/miniconda3/bin/conda install --yes numpy scipy matplotlib jupyter git pandas user_agent: conda/4.4.4 requests/2.18.4 CPython/3.6.3 Linux/2.6.32-696.10.3.el6.x86_64 centos/6.7 glibc/2.12 _messageid: -9223372036842075552 _messagetime: 1514766990662 / 2017-12-31 18:36:30 Traceback (most recent call last): File "/home/mholekev/miniconda3/lib/python3.6/site-packages/conda/common/io.py", line 399, in backdown_thread_pool yield ThreadPoolExecutor(max_workers) File "/home/mholekev/miniconda3/lib/python3.6/site-packages/conda/core/index.py", line 214, in get_reduced_index new_records = query_all(spec) File "/home/mholekev/miniconda3/lib/python3.6/site-packages/conda/core/index.py", line 187, in query_all return tuple(concat(future.result() for future in as_completed(futures))) File "/home/mholekev/miniconda3/lib/python3.6/site-packages/conda/core/index.py", line 187, in <genexpr> return tuple(concat(future.result() for future in as_completed(futures))) File "/home/mholekev/miniconda3/lib/python3.6/concurrent/futures/_base.py", line 217, in as_completed fs = set(fs) File "/home/mholekev/miniconda3/lib/python3.6/site-packages/conda/core/index.py", line 186, in <genexpr> futures = (executor.submit(sd.query, spec) for sd in subdir_datas) File "/home/mholekev/miniconda3/lib/python3.6/concurrent/futures/thread.py", line 123, in submit self._adjust_thread_count() File "/home/mholekev/miniconda3/lib/python3.6/concurrent/futures/thread.py", line 142, in _adjust_thread_count t.start() File "/home/mholekev/miniconda3/lib/python3.6/threading.py", line 846, in start _start_new_thread(self._bootstrap, ()) RuntimeError: can't start new thread ```
``` error: RuntimeError("can't start new thread",) command: /ahg/regevdata/users/subraman/software/miniconda3/bin/conda create -n test user_agent: conda/4.4.4 requests/2.18.4 CPython/3.6.3 Linux/2.6.32-696.16.1.el6.x86_64 rhel/6.9 glibc/2.12 _messageid: -9223372036842425559 _messagetime: 1514681157023 / 2017-12-30 18:45:57 Traceback (most recent call last): File "/ahg/regevdata/users/subraman/software/miniconda3/lib/python3.6/site-packages/conda/exceptions.py", line 724, in __call__ return func(*args, **kwargs) File "/ahg/regevdata/users/subraman/software/miniconda3/lib/python3.6/site-packages/conda/cli/main.py", line 78, in _main exit_code = do_call(args, p) File "/ahg/regevdata/users/subraman/software/miniconda3/lib/python3.6/site-packages/conda/cli/conda_argparse.py", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File "/ahg/regevdata/users/subraman/software/miniconda3/lib/python3.6/site-packages/conda/cli/main_create.py", line 11, in execute install(args, parser, 'create') File "/ahg/regevdata/users/subraman/software/miniconda3/lib/python3.6/site-packages/conda/cli/install.py", line 236, in install force_reinstall=context.force, File "/ahg/regevdata/users/subraman/software/miniconda3/lib/python3.6/site-packages/conda/core/solve.py", line 514, in solve_for_transaction SubdirData.query_all(self.channels, self.subdirs, conda_newer_spec), File "/ahg/regevdata/users/subraman/software/miniconda3/lib/python3.6/site-packages/conda/core/repodata.py", line 88, in query_all return tuple(concat(future.result() for future in as_completed(futures))) File "/ahg/regevdata/users/subraman/software/miniconda3/lib/python3.6/site-packages/conda/core/repodata.py", line 88, in <genexpr> return tuple(concat(future.result() for future in as_completed(futures))) File "/ahg/regevdata/users/subraman/software/miniconda3/lib/python3.6/concurrent/futures/_base.py", line 217, in as_completed fs = set(fs) File "/ahg/regevdata/users/subraman/software/miniconda3/lib/python3.6/site-packages/conda/core/repodata.py", line 87, in <genexpr> ) for url in channel_urls) File "/ahg/regevdata/users/subraman/software/miniconda3/lib/python3.6/concurrent/futures/thread.py", line 123, in submit self._adjust_thread_count() File "/ahg/regevdata/users/subraman/software/miniconda3/lib/python3.6/concurrent/futures/thread.py", line 142, in _adjust_thread_count t.start() File "/ahg/regevdata/users/subraman/software/miniconda3/lib/python3.6/threading.py", line 846, in start _start_new_thread(self._bootstrap, ()) RuntimeError: can't start new thread ```
2018-01-04T21:02:35Z
[]
[]
Traceback (most recent call last): File "/home/mholekev/miniconda3/lib/python3.6/site-packages/conda/common/io.py", line 399, in backdown_thread_pool yield ThreadPoolExecutor(max_workers) File "/home/mholekev/miniconda3/lib/python3.6/site-packages/conda/core/index.py", line 214, in get_reduced_index new_records = query_all(spec) File "/home/mholekev/miniconda3/lib/python3.6/site-packages/conda/core/index.py", line 187, in query_all return tuple(concat(future.result() for future in as_completed(futures))) File "/home/mholekev/miniconda3/lib/python3.6/site-packages/conda/core/index.py", line 187, in <genexpr> return tuple(concat(future.result() for future in as_completed(futures))) File "/home/mholekev/miniconda3/lib/python3.6/concurrent/futures/_base.py", line 217, in as_completed fs = set(fs) File "/home/mholekev/miniconda3/lib/python3.6/site-packages/conda/core/index.py", line 186, in <genexpr> futures = (executor.submit(sd.query, spec) for sd in subdir_datas) File "/home/mholekev/miniconda3/lib/python3.6/concurrent/futures/thread.py", line 123, in submit self._adjust_thread_count() File "/home/mholekev/miniconda3/lib/python3.6/concurrent/futures/thread.py", line 142, in _adjust_thread_count t.start() File "/home/mholekev/miniconda3/lib/python3.6/threading.py", line 846, in start _start_new_thread(self._bootstrap, ()) RuntimeError: can't start new thread
4,259
conda/conda
conda__conda-6655
114e4b3cfb0fae3d566fd08a0c0c1d7e24c19891
diff --git a/conda/activate.py b/conda/activate.py --- a/conda/activate.py +++ b/conda/activate.py @@ -413,24 +413,30 @@ def _replace_prefix_in_path(self, old_prefix, new_prefix, starting_path_dirs=Non else: path_list = list(starting_path_dirs) if on_win: # pragma: unix no cover - # windows has a nasty habit of adding extra Library\bin directories - prefix_dirs = tuple(self._get_path_dirs(old_prefix)) - try: - first_idx = path_list.index(prefix_dirs[0]) - except ValueError: - first_idx = 0 + if old_prefix is not None: + # windows has a nasty habit of adding extra Library\bin directories + prefix_dirs = tuple(self._get_path_dirs(old_prefix)) + try: + first_idx = path_list.index(prefix_dirs[0]) + except ValueError: + first_idx = 0 + else: + last_idx = path_list.index(prefix_dirs[-1]) + del path_list[first_idx:last_idx+1] else: - last_idx = path_list.index(prefix_dirs[-1]) - del path_list[first_idx:last_idx+1] + first_idx = 0 if new_prefix is not None: path_list[first_idx:first_idx] = list(self._get_path_dirs(new_prefix)) else: - try: - idx = path_list.index(join(old_prefix, 'bin')) - except ValueError: - idx = 0 + if old_prefix is not None: + try: + idx = path_list.index(join(old_prefix, 'bin')) + except ValueError: + idx = 0 + else: + del path_list[idx] else: - del path_list[idx] + idx = 0 if new_prefix is not None: path_list.insert(idx, join(new_prefix, 'bin')) return self.path_conversion(path_list)
AttributeError: 'NoneType' object has no attribute 'endswith' ``` error: AttributeError("'NoneType' object has no attribute 'endswith'",) command: /home/wang/anaconda2/bin/conda shell.posix activate py2 user_agent: conda/4.4.4 requests/2.18.4 CPython/2.7.13 Linux/3.19.0-42-generic ubuntu/14.04 glibc/2.19 _messageid: -9223372036842325539 _messagetime: 1514696872936 / 2017-12-30 23:07:52 Traceback (most recent call last): File "/home/wang/anaconda2/lib/python2.7/site-packages/conda/cli/main.py", line 111, in main return activator_main() File "/home/wang/anaconda2/lib/python2.7/site-packages/conda/activate.py", line 529, in main sys.stdout.write(activator.execute()) File "/home/wang/anaconda2/lib/python2.7/site-packages/conda/activate.py", line 149, in execute return getattr(self, self.command)() File "/home/wang/anaconda2/lib/python2.7/site-packages/conda/activate.py", line 135, in activate return self._finalize(self._yield_commands(self.build_activate(self.env_name_or_prefix)), File "/home/wang/anaconda2/lib/python2.7/site-packages/conda/activate.py", line 247, in build_activate new_path = self.pathsep_join(self._replace_prefix_in_path(old_conda_prefix, prefix)) File "/home/wang/anaconda2/lib/python2.7/site-packages/conda/activate.py", line 411, in _replace_prefix_in_path idx = path_list.index(join(old_prefix, 'bin')) File "/home/wang/anaconda2/lib/python2.7/posixpath.py", line 70, in join elif path == '' or path.endswith('/'): AttributeError: 'NoneType' object has no attribute 'endswith' ```
2018-01-04T21:22:38Z
[]
[]
Traceback (most recent call last): File "/home/wang/anaconda2/lib/python2.7/site-packages/conda/cli/main.py", line 111, in main return activator_main() File "/home/wang/anaconda2/lib/python2.7/site-packages/conda/activate.py", line 529, in main sys.stdout.write(activator.execute()) File "/home/wang/anaconda2/lib/python2.7/site-packages/conda/activate.py", line 149, in execute return getattr(self, self.command)() File "/home/wang/anaconda2/lib/python2.7/site-packages/conda/activate.py", line 135, in activate return self._finalize(self._yield_commands(self.build_activate(self.env_name_or_prefix)), File "/home/wang/anaconda2/lib/python2.7/site-packages/conda/activate.py", line 247, in build_activate new_path = self.pathsep_join(self._replace_prefix_in_path(old_conda_prefix, prefix)) File "/home/wang/anaconda2/lib/python2.7/site-packages/conda/activate.py", line 411, in _replace_prefix_in_path idx = path_list.index(join(old_prefix, 'bin')) File "/home/wang/anaconda2/lib/python2.7/posixpath.py", line 70, in join elif path == '' or path.endswith('/'): AttributeError: 'NoneType' object has no attribute 'endswith'
4,260
conda/conda
conda__conda-6656
114e4b3cfb0fae3d566fd08a0c0c1d7e24c19891
diff --git a/conda/base/context.py b/conda/base/context.py --- a/conda/base/context.py +++ b/conda/base/context.py @@ -968,6 +968,14 @@ def determine_target_prefix(ctx, args=None): except AttributeError: prefix_path = None + if prefix_name is not None and not prefix_name.strip(): # pragma: no cover + from ..exceptions import ArgumentError + raise ArgumentError("Argument --name requires a value.") + + if prefix_path is not None and not prefix_path.strip(): # pragma: no cover + from ..exceptions import ArgumentError + raise ArgumentError("Argument --prefix requires a value.") + if prefix_name is None and prefix_path is None: return ctx.default_prefix elif prefix_path is not None:
conda update --name AssertionError ``` error: AssertionError() command: /Users/cpthgli/.anyenv/envs/pyenv/versions/default/bin/conda update --name --all --yes user_agent: conda/4.4.4 requests/2.18.4 CPython/3.6.4 Darwin/17.3.0 OSX/10.13.2 _messageid: -9223372036843525540 _messagetime: 1514462637935 / 2017-12-28 06:03:57 Traceback (most recent call last): File "/Users/cpthgli/.anyenv/envs/pyenv/versions/anaconda3-5.0.1/lib/python3.6/site-packages/conda/exceptions.py", line 724, in __call__ return func(*args, **kwargs) File "/Users/cpthgli/.anyenv/envs/pyenv/versions/anaconda3-5.0.1/lib/python3.6/site-packages/conda/cli/main.py", line 78, in _main exit_code = do_call(args, p) File "/Users/cpthgli/.anyenv/envs/pyenv/versions/anaconda3-5.0.1/lib/python3.6/site-packages/conda/cli/conda_argparse.py", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File "/Users/cpthgli/.anyenv/envs/pyenv/versions/anaconda3-5.0.1/lib/python3.6/site-packages/conda/cli/main_update.py", line 14, in execute install(args, parser, 'update') File "/Users/cpthgli/.anyenv/envs/pyenv/versions/anaconda3-5.0.1/lib/python3.6/site-packages/conda/cli/install.py", line 140, in install prefix = context.target_prefix File "/Users/cpthgli/.anyenv/envs/pyenv/versions/anaconda3-5.0.1/lib/python3.6/site-packages/conda/base/context.py", line 454, in target_prefix return determine_target_prefix(self) File "/Users/cpthgli/.anyenv/envs/pyenv/versions/anaconda3-5.0.1/lib/python3.6/site-packages/conda/base/context.py", line 974, in determine_target_prefix return locate_prefix_by_name(prefix_name) File "/Users/cpthgli/.anyenv/envs/pyenv/versions/anaconda3-5.0.1/lib/python3.6/site-packages/conda/base/context.py", line 923, in locate_prefix_by_name assert name AssertionError ```
2018-01-04T21:31:32Z
[]
[]
Traceback (most recent call last): File "/Users/cpthgli/.anyenv/envs/pyenv/versions/anaconda3-5.0.1/lib/python3.6/site-packages/conda/exceptions.py", line 724, in __call__ return func(*args, **kwargs) File "/Users/cpthgli/.anyenv/envs/pyenv/versions/anaconda3-5.0.1/lib/python3.6/site-packages/conda/cli/main.py", line 78, in _main exit_code = do_call(args, p) File "/Users/cpthgli/.anyenv/envs/pyenv/versions/anaconda3-5.0.1/lib/python3.6/site-packages/conda/cli/conda_argparse.py", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File "/Users/cpthgli/.anyenv/envs/pyenv/versions/anaconda3-5.0.1/lib/python3.6/site-packages/conda/cli/main_update.py", line 14, in execute install(args, parser, 'update') File "/Users/cpthgli/.anyenv/envs/pyenv/versions/anaconda3-5.0.1/lib/python3.6/site-packages/conda/cli/install.py", line 140, in install prefix = context.target_prefix File "/Users/cpthgli/.anyenv/envs/pyenv/versions/anaconda3-5.0.1/lib/python3.6/site-packages/conda/base/context.py", line 454, in target_prefix return determine_target_prefix(self) File "/Users/cpthgli/.anyenv/envs/pyenv/versions/anaconda3-5.0.1/lib/python3.6/site-packages/conda/base/context.py", line 974, in determine_target_prefix return locate_prefix_by_name(prefix_name) File "/Users/cpthgli/.anyenv/envs/pyenv/versions/anaconda3-5.0.1/lib/python3.6/site-packages/conda/base/context.py", line 923, in locate_prefix_by_name assert name AssertionError
4,261
conda/conda
conda__conda-6694
4d963e1da043f3eb8b0430263a6185db134f33df
diff --git a/conda/core/solve.py b/conda/core/solve.py --- a/conda/core/solve.py +++ b/conda/core/solve.py @@ -380,9 +380,9 @@ def solve_final_state(self, deps_modifier=NULL, prune=NULL, ignore_pinned=NULL, if ascendant_name not in specs_to_add_names: update_names.add(ascendant_name) grouped_specs = groupby(lambda s: s.name in update_names, final_environment_specs) - new_final_environment_specs = set(grouped_specs[False]) + new_final_environment_specs = set(grouped_specs.get(False, ())) update_specs = set(MatchSpec(spec.name, optional=spec.optional) - for spec in grouped_specs[True]) + for spec in grouped_specs.get(True, ())) final_environment_specs = new_final_environment_specs | update_specs solution = r.solve(final_environment_specs)
KeyError with '--update-deps' ``` error: KeyError(True,) command: /home/jgrnt/miniconda/bin/conda install --update-deps python= numpy nose scipy matplotlib pandas pytest h5py mkl mkl-service user_agent: conda/4.4.7 requests/2.18.4 CPython/3.6.3 Linux/4.14.4-gentoo gentoo/2.4.1 glibc/2.25 _messageid: -9223372036837525416 _messagetime: 1515513689937 / 2018-01-09 10:01:29 Traceback (most recent call last): File "/home/jgrnt/miniconda/lib/python3.6/site-packages/conda/exceptions.py", line 743, in __call__ return func(*args, **kwargs) File "/home/jgrnt/miniconda/lib/python3.6/site-packages/conda/cli/main.py", line 78, in _main exit_code = do_call(args, p) File "/home/jgrnt/miniconda/lib/python3.6/site-packages/conda/cli/conda_argparse.py", line 76, in do_call exit_code = getattr(module, func_name)(args, parser) File "/home/jgrnt/miniconda/lib/python3.6/site-packages/conda/cli/main_install.py", line 11, in execute install(args, parser, 'install') File "/home/jgrnt/miniconda/lib/python3.6/site-packages/conda/cli/install.py", line 236, in install force_reinstall=context.force, File "/home/jgrnt/miniconda/lib/python3.6/site-packages/conda/core/solve.py", line 504, in solve_for_transaction force_remove, force_reinstall) File "/home/jgrnt/miniconda/lib/python3.6/site-packages/conda/core/solve.py", line 437, in solve_for_diff final_precs = self.solve_final_state(deps_modifier, prune, ignore_pinned, force_remove) File "/home/jgrnt/miniconda/lib/python3.6/site-packages/conda/core/solve.py", line 385, in solve_final_state for spec in grouped_specs[True]) KeyError: True ```
2018-01-09T17:28:03Z
[]
[]
Traceback (most recent call last): File "/home/jgrnt/miniconda/lib/python3.6/site-packages/conda/exceptions.py", line 743, in __call__ return func(*args, **kwargs) File "/home/jgrnt/miniconda/lib/python3.6/site-packages/conda/cli/main.py", line 78, in _main exit_code = do_call(args, p) File "/home/jgrnt/miniconda/lib/python3.6/site-packages/conda/cli/conda_argparse.py", line 76, in do_call exit_code = getattr(module, func_name)(args, parser) File "/home/jgrnt/miniconda/lib/python3.6/site-packages/conda/cli/main_install.py", line 11, in execute install(args, parser, 'install') File "/home/jgrnt/miniconda/lib/python3.6/site-packages/conda/cli/install.py", line 236, in install force_reinstall=context.force, File "/home/jgrnt/miniconda/lib/python3.6/site-packages/conda/core/solve.py", line 504, in solve_for_transaction force_remove, force_reinstall) File "/home/jgrnt/miniconda/lib/python3.6/site-packages/conda/core/solve.py", line 437, in solve_for_diff final_precs = self.solve_final_state(deps_modifier, prune, ignore_pinned, force_remove) File "/home/jgrnt/miniconda/lib/python3.6/site-packages/conda/core/solve.py", line 385, in solve_final_state for spec in grouped_specs[True]) KeyError: True
4,268
conda/conda
conda__conda-6724
c23e57dfb935b5852eefb4e04187caf82b4e0b78
diff --git a/conda/plan.py b/conda/plan.py --- a/conda/plan.py +++ b/conda/plan.py @@ -351,7 +351,6 @@ def revert_actions(prefix, revision=-1, index=None): # TODO: If revision raise a revision error, should always go back to a safe revision # change h = History(prefix) - h.update() user_requested_specs = itervalues(h.get_requested_specs_map()) try: state = h.get_state(revision) @@ -360,7 +359,7 @@ def revert_actions(prefix, revision=-1, index=None): curr = h.get_state() if state == curr: - return {} # TODO: return txn with nothing_to_do + return UnlinkLinkTransaction() _supplement_index_with_prefix(index, prefix) r = Resolve(index) @@ -371,11 +370,6 @@ def revert_actions(prefix, revision=-1, index=None): link_dists = tuple(d for d in state if not is_linked(prefix, d)) unlink_dists = set(curr) - set(state) - # dists = (Dist(s) for s in state) - # actions = ensure_linked_actions(dists, prefix) - # for dist in curr - state: - # add_unlink(actions, Dist(dist)) - # check whether it is a safe revision for dist in concatv(link_dists, unlink_dists): if dist not in index:
Cannot roll back with 'conda install --revision' ``` error: AttributeError("'dict' object has no attribute 'get_pfe'",) command: C:\Program Files\Anaconda3\Scripts\conda install --revision 4 user_agent: conda/4.4.4 requests/2.18.4 CPython/3.5.2 Windows/8.1 Windows/6.3.9600 _messageid: -9223372036843575560 _messagetime: 1514442489595 / 2017-12-28 00:28:09 Traceback (most recent call last): File "C:\Program Files\Anaconda3\lib\site-packages\conda\exceptions.py", line 724, in __call__ return func(*args, **kwargs) File "C:\Program Files\Anaconda3\lib\site-packages\conda\cli\main.py", line 78, in _main exit_code = do_call(args, p) File "C:\Program Files\Anaconda3\lib\site-packages\conda\cli\conda_argparse.py", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File "C:\Program Files\Anaconda3\lib\site-packages\conda\cli\main_install.py", line 11, in execute install(args, parser, 'install') File "C:\Program Files\Anaconda3\lib\site-packages\conda\cli\install.py", line 232, in install progressive_fetch_extract = unlink_link_transaction.get_pfe() AttributeError: 'dict' object has no attribute 'get_pfe' ```
2018-01-13T05:54:39Z
[]
[]
Traceback (most recent call last): File "C:\Program Files\Anaconda3\lib\site-packages\conda\exceptions.py", line 724, in __call__ return func(*args, **kwargs) File "C:\Program Files\Anaconda3\lib\site-packages\conda\cli\main.py", line 78, in _main exit_code = do_call(args, p) File "C:\Program Files\Anaconda3\lib\site-packages\conda\cli\conda_argparse.py", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File "C:\Program Files\Anaconda3\lib\site-packages\conda\cli\main_install.py", line 11, in execute install(args, parser, 'install') File "C:\Program Files\Anaconda3\lib\site-packages\conda\cli\install.py", line 232, in install progressive_fetch_extract = unlink_link_transaction.get_pfe() AttributeError: 'dict' object has no attribute 'get_pfe'
4,274
conda/conda
conda__conda-6764
fa6526110300f241a41610894fbce34ff4341fdd
diff --git a/conda/exceptions.py b/conda/exceptions.py --- a/conda/exceptions.py +++ b/conda/exceptions.py @@ -582,6 +582,21 @@ def __init__(self, message): super(CondaUpgradeError, self).__init__(msg) +class CaseInsensitiveFileSystemError(CondaError): + def __init__(self, package_location, extract_location, **kwargs): + message = dals(""" + Cannot extract package to a case-insensitive file system. + package location: %(package_location)s + extract location: %(extract_location)s + """) + super(CaseInsensitiveFileSystemError, self).__init__( + message, + package_location=package_location, + extract_location=extract_location, + **kwargs + ) + + class CondaVerificationError(CondaError): def __init__(self, message): super(CondaVerificationError, self).__init__(message) diff --git a/conda/gateways/disk/create.py b/conda/gateways/disk/create.py --- a/conda/gateways/disk/create.py +++ b/conda/gateways/disk/create.py @@ -1,7 +1,7 @@ # -*- coding: utf-8 -*- from __future__ import absolute_import, division, print_function, unicode_literals -from errno import EACCES, EPERM +from errno import EACCES, ELOOP, EPERM from io import open from logging import getLogger import os @@ -24,7 +24,8 @@ from ...common.compat import ensure_binary, on_win from ...common.path import ensure_pad, expand, win_path_double_escape, win_path_ok from ...common.serialize import json_dump -from ...exceptions import BasicClobberError, CondaOSError, maybe_raise +from ...exceptions import (BasicClobberError, CaseInsensitiveFileSystemError, CondaOSError, + maybe_raise) from ...models.enums import FileMode, LinkType log = getLogger(__name__) @@ -142,7 +143,17 @@ def members_with_progress(): progress_update_callback(q / num_members) yield member - t.extractall(path=destination_directory, members=members_with_progress()) + try: + t.extractall(path=destination_directory, members=members_with_progress()) + except EnvironmentError as e: + if e.errno == ELOOP: + raise CaseInsensitiveFileSystemError( + package_location=tarball_full_path, + extract_location=destination_directory, + caused_by=e, + ) + else: + raise if sys.platform.startswith('linux') and os.getuid() == 0: # When extracting as root, tarfile will by restore ownership
OSError: Too many levels of symlinks This happened in a conda render operation, but does not appear to be readily reproducible - perhaps just a fluke, but something to watch out for if it happens often enough. ``` Traceback (most recent call last): File "/Users/msarahan/miniconda3/bin/c3i", line 11, in <module> load_entry_point('conda-concourse-ci', 'console_scripts', 'c3i')() File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/cli.py", line 123, in main execute.submit_one_off(**args.__dict__) File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/execute.py", line 647, in submit_one_off **kwargs) File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/execute.py", line 508, in compute_builds variant_config_files=kw.get('variant_config_files', [])) File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/execute.py", line 72, in collect_tasks config=config) File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/compute_build_graph.py", line 379, in construct_graph recipes_dir, config=config) File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/compute_build_graph.py", line 206, in add_recipe_to_graph rendered = _get_or_render_metadata(recipe_dir, worker, config=config) File "/Users/msarahan/miniconda3/lib/python3.6/site-packages/conda/utils.py", line 38, in __call__ return self.func(*args, **kw) File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/compute_build_graph.py", line 199, in _get_or_render_metadata bypass_env_check=True, config=config) File "/Users/msarahan/code/conda-build/conda_build/api.py", line 49, in render permit_undefined_jinja=not finalize): File "/Users/msarahan/code/conda-build/conda_build/metadata.py", line 1781, in get_output_metadata_set permit_unsatisfiable_variants=permit_unsatisfiable_variants) File "/Users/msarahan/code/conda-build/conda_build/metadata.py", line 645, in finalize_outputs_pass fm = finalize_metadata(om, permit_unsatisfiable_variants=permit_unsatisfiable_variants) File "/Users/msarahan/code/conda-build/conda_build/render.py", line 387, in finalize_metadata exclude_pattern) File "/Users/msarahan/code/conda-build/conda_build/render.py", line 306, in add_upstream_pins permit_unsatisfiable_variants, exclude_pattern) File "/Users/msarahan/code/conda-build/conda_build/render.py", line 289, in _read_upstream_pin_files extra_run_specs = get_upstream_pins(m, actions, env) File "/Users/msarahan/code/conda-build/conda_build/render.py", line 251, in get_upstream_pins pfe.execute() File "/Users/msarahan/miniconda3/lib/python3.6/site-packages/conda/core/package_cache.py", line 494, in execute self._execute_action(action) File "/Users/msarahan/miniconda3/lib/python3.6/site-packages/conda/core/package_cache.py", line 515, in _execute_action raise CondaMultiError(exceptions) conda.CondaMultiError: OSError(62, 'Too many levels of symbolic links') OSError(62, 'Too many levels of symbolic links') OSError(62, 'Too many levels of symbolic links') ```
In an auto-upload report we got something similar. Maybe more details here. "command": "/Users/boris/bin/miniconda/bin/conda install clangxx_osx-64" "user_agent": "conda/4.4.1 requests/2.18.4 CPython/2.7.14 Darwin/17.3.0 OSX/10.13.2" ``` Traceback (most recent call last): File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/exceptions.py\", line 721, in __call__ return func(*args, **kwargs) File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/cli/main.py\", line 78, in _main exit_code = do_call(args, p) File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/cli/conda_argparse.py\", line 75, in do_call exit_code = getattr(module, func_name)(args, parser) File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/cli/main_install.py\", line 11, in execute install(args, parser, 'install') File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/cli/install.py\", line 239, in install handle_txn(progressive_fetch_extract, unlink_link_transaction, prefix, args, newenv) File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/cli/install.py\", line 256, in handle_txn unlink_link_transaction.display_actions(progressive_fetch_extract) File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/core/link.py\", line 706, in display_actions legacy_action_groups = self.make_legacy_action_groups(pfe) File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/core/link.py\", line 689, in make_legacy_action_groups pfe.prepare() File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/common/io.py\", line 402, in decorated return f(*args, **kwds) File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/core/package_cache.py\", line 508, in prepare for prec in self.link_precs) File \"/Users/boris/bin/miniconda/lib/python2.7/_abcoll.py\", line 571, in update for key, value in other: File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/core/package_cache.py\", line 508, in <genexpr> for prec in self.link_precs) File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/core/package_cache.py\", line 402, in make_actions_for_record ), None) File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/core/package_cache.py\", line 399, in <genexpr> pcrec for pcrec in concat(PackageCache(pkgs_dir).query(pref_or_spec) File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/core/package_cache.py\", line 400, in <genexpr> for pkgs_dir in context.pkgs_dirs) File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/core/package_cache.py\", line 116, in query return (pcrec for pcrec in itervalues(self._package_cache_records) if pcrec == param) File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/core/package_cache.py\", line 209, in _package_cache_records self.load() File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/core/package_cache.py\", line 88, in load package_cache_record = self._make_single_record(base_name) File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/core/package_cache.py\", line 294, in _make_single_record extract_tarball(package_tarball_full_path, extracted_package_dir) File \"/Users/boris/bin/miniconda/lib/python2.7/site-packages/conda/gateways/disk/create.py\", line 145, in extract_tarball t.extractall(path=destination_directory, members=members_with_progress()) File \"/Users/boris/bin/miniconda/lib/python2.7/tarfile.py\", line 2081, in extractall self.extract(tarinfo, path) File \"/Users/boris/bin/miniconda/lib/python2.7/tarfile.py\", line 2118, in extract self._extract_member(tarinfo, os.path.join(path, tarinfo.name)) File \"/Users/boris/bin/miniconda/lib/python2.7/tarfile.py\", line 2194, in _extract_member self.makefile(tarinfo, targetpath) File \"/Users/boris/bin/miniconda/lib/python2.7/tarfile.py\", line 2234, in makefile with bltn_open(targetpath, \"wb\") as target: IOError: [Errno 62] Too many levels of symbolic links: u'/Users/boris/bin/miniconda/pkgs/openssl-1.0.2l-h077ae2c_5/ssl/man/man3/mdc2.3' ``` Only `defaults` in the operation's channel list. I'm running into the exact error above when trying to use the undocumented `CONDA_SUBDIR` setting on a fresh anaconda install. System: python 3.6.3 anaconda 5.0.1 macOS 10.13.2 All other config options (channels, environment, etc) are default. I am able to replicate the error from the root environment of a clean install with the following command: ``` CONDA_SUBDIR=linux-64 \ $HOME/anaconda3/bin/conda install \ --yes --mkdir --json -vv \ --prefix $HOME/temp_conda_env \ python=3.6 flask pyyaml yaml pandas nomkl ``` Traceback: https://gist.github.com/cscanlin/25fa05575918752252a071f352a7395a For reference, I am trying to use `CONDA_SUBDIR` in an attempt at picking up the torch on conda support for Zappa: https://github.com/Miserlou/Zappa/pull/108 Please let me know if I can provide additional details, thanks! Hey Chris, really appreciate someone digging into conda + Zappa. It’s been on my list for a *long* time, and I just haven’t had the bandwidth. Going to try to look into this particular issue later today. Sent from my iPhone > On Jan 21, 2018, at 9:34 PM, Chris Scanlin <notifications@github.com> wrote: > > I'm running into the exact error above when trying to use the undocumented CONDA_SUBDIR setting on a fresh anaconda install. > > System: > python 3.6.3 > anaconda 5.0.1 > macOS 10.13.2 > > All other config options (channels, environment, etc) are default. > > I am able to replicate the error from the root environment of a clean install with the following command: > > CONDA_SUBDIR=linux-64 \ > $HOME/anaconda3/bin/conda install \ > --yes --mkdir --json -vv \ > --prefix $HOME/temp_conda_env \ > python=3.6 flask pyyaml yaml pandas nomkl > Traceback: https://gist.github.com/cscanlin/25fa05575918752252a071f352a7395a > > For reference, I am trying to use CONDA_SUBDIR in an attempt at picking up the torch on conda support for Zappa: Miserlou/Zappa#108 > > Please let me know if I can provide additional details, thanks! > > — > You are receiving this because you commented. > Reply to this email directly, view it on GitHub, or mute the thread. > Issue here is with trying to use python's `tarfile.extractall()` on the linux-64 openssl package, while on macOS. Relevant stack trace is ``` Traceback (most recent call last): File "/Users/kfranz/continuum/conda/conda/exceptions.py", line 751, in __call__ return func(*args, **kwargs) File "/Users/kfranz/continuum/conda/conda/cli/main.py", line 78, in _main exit_code = do_call(args, p) File "/Users/kfranz/continuum/conda/conda/cli/conda_argparse.py", line 76, in do_call exit_code = getattr(module, func_name)(args, parser) File "/Users/kfranz/continuum/conda/conda/cli/main_install.py", line 11, in execute install(args, parser, 'install') File "/Users/kfranz/continuum/conda/conda/cli/install.py", line 255, in install handle_txn(progressive_fetch_extract, unlink_link_transaction, prefix, args, newenv) File "/Users/kfranz/continuum/conda/conda/cli/install.py", line 272, in handle_txn unlink_link_transaction.display_actions(progressive_fetch_extract) File "/Users/kfranz/continuum/conda/conda/core/link.py", line 706, in display_actions legacy_action_groups = self.make_legacy_action_groups(pfe) File "/Users/kfranz/continuum/conda/conda/core/link.py", line 689, in make_legacy_action_groups pfe.prepare() File "/Users/kfranz/continuum/conda/conda/common/io.py", line 415, in decorated return f(*args, **kwds) File "/Users/kfranz/continuum/conda/conda/core/package_cache.py", line 540, in prepare for prec in self.link_precs) File "/Users/kfranz/continuum/conda/conda/core/package_cache.py", line 540, in <genexpr> for prec in self.link_precs) File "/Users/kfranz/continuum/conda/conda/core/package_cache.py", line 434, in make_actions_for_record ), None) File "/Users/kfranz/continuum/conda/conda/core/package_cache.py", line 431, in <genexpr> pcrec for pcrec in concat(PackageCache(pkgs_dir).query(pref_or_spec) File "/Users/kfranz/continuum/conda/conda/core/package_cache.py", line 432, in <genexpr> for pkgs_dir in context.pkgs_dirs) File "/Users/kfranz/continuum/conda/conda/core/package_cache.py", line 118, in query return (pcrec for pcrec in itervalues(self._package_cache_records) if pcrec == param) File "/Users/kfranz/continuum/conda/conda/core/package_cache.py", line 211, in _package_cache_records self.load() File "/Users/kfranz/continuum/conda/conda/core/package_cache.py", line 90, in load package_cache_record = self._make_single_record(base_name) File "/Users/kfranz/continuum/conda/conda/core/package_cache.py", line 308, in _make_single_record extract_tarball(package_tarball_full_path, extracted_package_dir) File "/Users/kfranz/continuum/conda/conda/gateways/disk/create.py", line 145, in extract_tarball t.extractall(path=destination_directory, members=members_with_progress()) File "/usr/local/Cellar/python3/3.5.2_1/Frameworks/Python.framework/Versions/3.5/lib/python3.5/tarfile.py", line 1996, in extractall numeric_owner=numeric_owner) File "/usr/local/Cellar/python3/3.5.2_1/Frameworks/Python.framework/Versions/3.5/lib/python3.5/tarfile.py", line 2038, in extract numeric_owner=numeric_owner) File "/usr/local/Cellar/python3/3.5.2_1/Frameworks/Python.framework/Versions/3.5/lib/python3.5/tarfile.py", line 2108, in _extract_member self.makefile(tarinfo, targetpath) File "/usr/local/Cellar/python3/3.5.2_1/Frameworks/Python.framework/Versions/3.5/lib/python3.5/tarfile.py", line 2148, in makefile with bltn_open(targetpath, "wb") as target: OSError: [Errno 62] Too many levels of symbolic links: '/Users/kfranz/linux-64/openssl-1.0.2n-hb7f436b_0/ssl/man/man3/mdc2.3' ``` The linux-64 package has ``` ❯ ls -al /Users/kfranz/linux-64/openssl-1.0.2n-hb7f436b_0/ssl/man/man3/MDC* lrwxr-xr-x 1 kfranz staff 6B Jan 22 13:23 /Users/kfranz/linux-64/openssl-1.0.2n-hb7f436b_0/ssl/man/man3/MDC2.3@ -> mdc2.3 lrwxr-xr-x 1 kfranz staff 6B Jan 22 13:23 /Users/kfranz/linux-64/openssl-1.0.2n-hb7f436b_0/ssl/man/man3/MDC2_Final.3@ -> mdc2.3 lrwxr-xr-x 1 kfranz staff 6B Jan 22 13:23 /Users/kfranz/linux-64/openssl-1.0.2n-hb7f436b_0/ssl/man/man3/MDC2_Init.3@ -> mdc2.3 lrwxr-xr-x 1 kfranz staff 6B Jan 22 13:23 /Users/kfranz/linux-64/openssl-1.0.2n-hb7f436b_0/ssl/man/man3/MDC2_Update.3@ -> mdc2.3 ``` Specifically the problem is with ssl/man/man3/MDC2.3@ -> mdc2.3 So probably something going on with case sensitivity differences in the file system here. Still figuring out what the correct fix is...
2018-01-22T20:43:42Z
[]
[]
Traceback (most recent call last): File "/Users/msarahan/miniconda3/bin/c3i", line 11, in <module> load_entry_point('conda-concourse-ci', 'console_scripts', 'c3i')() File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/cli.py", line 123, in main execute.submit_one_off(**args.__dict__) File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/execute.py", line 647, in submit_one_off **kwargs) File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/execute.py", line 508, in compute_builds variant_config_files=kw.get('variant_config_files', [])) File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/execute.py", line 72, in collect_tasks config=config) File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/compute_build_graph.py", line 379, in construct_graph recipes_dir, config=config) File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/compute_build_graph.py", line 206, in add_recipe_to_graph rendered = _get_or_render_metadata(recipe_dir, worker, config=config) File "/Users/msarahan/miniconda3/lib/python3.6/site-packages/conda/utils.py", line 38, in __call__ return self.func(*args, **kw) File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/compute_build_graph.py", line 199, in _get_or_render_metadata bypass_env_check=True, config=config) File "/Users/msarahan/code/conda-build/conda_build/api.py", line 49, in render permit_undefined_jinja=not finalize): File "/Users/msarahan/code/conda-build/conda_build/metadata.py", line 1781, in get_output_metadata_set permit_unsatisfiable_variants=permit_unsatisfiable_variants) File "/Users/msarahan/code/conda-build/conda_build/metadata.py", line 645, in finalize_outputs_pass fm = finalize_metadata(om, permit_unsatisfiable_variants=permit_unsatisfiable_variants) File "/Users/msarahan/code/conda-build/conda_build/render.py", line 387, in finalize_metadata exclude_pattern) File "/Users/msarahan/code/conda-build/conda_build/render.py", line 306, in add_upstream_pins permit_unsatisfiable_variants, exclude_pattern) File "/Users/msarahan/code/conda-build/conda_build/render.py", line 289, in _read_upstream_pin_files extra_run_specs = get_upstream_pins(m, actions, env) File "/Users/msarahan/code/conda-build/conda_build/render.py", line 251, in get_upstream_pins pfe.execute() File "/Users/msarahan/miniconda3/lib/python3.6/site-packages/conda/core/package_cache.py", line 494, in execute self._execute_action(action) File "/Users/msarahan/miniconda3/lib/python3.6/site-packages/conda/core/package_cache.py", line 515, in _execute_action raise CondaMultiError(exceptions) conda.CondaMultiError: OSError(62, 'Too many levels of symbolic links')
4,279
conda/conda
conda__conda-6776
c1af02bb4fcfb8c962bb52d040ba705174345889
diff --git a/conda/exceptions.py b/conda/exceptions.py --- a/conda/exceptions.py +++ b/conda/exceptions.py @@ -592,6 +592,12 @@ def __init__(self, message): super(SafetyError, self).__init__(message) +class CondaMemoryError(MemoryError, CondaError): + def __init__(self, caused_by, **kwargs): + message = "The conda process ran out of memory. Increase system memory and/or try again." + super(CondaMemoryError, self).__init__(message, caused_by=caused_by, **kwargs) + + class NotWritableError(CondaError, OSError): def __init__(self, path, errno, **kwargs): @@ -784,6 +790,8 @@ def handle_exception(self, exc_val, exc_tb): if isinstance(exc_val, EnvironmentError): if getattr(exc_val, 'errno', None) == ENOSPC: return self.handle_application_exception(NoSpaceLeftError(exc_val), exc_tb) + if isinstance(exc_val, MemoryError): + return self.handle_application_exception(CondaMemoryError(exc_val), exc_tb) if isinstance(exc_val, KeyboardInterrupt): self._print_conda_exception(CondaError("KeyboardInterrupt"), _format_exc()) return 1
MemoryError ``` error: MemoryError() command: C:\Users\Jason\Anaconda3\Scripts\conda install rise user_agent: conda/4.4.7 requests/2.18.4 CPython/3.6.3 Windows/10 Windows/10.0.16299 _messageid: -9223372036837525402 _messagetime: 1515517517366 / 2018-01-09 11:05:17 MemoryError During handling of the above exception, another exception occurred: MemoryError During handling of the above exception, another exception occurred: Traceback (most recent call last): File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\exceptions.py", line 743, in __call__ return func(*args, **kwargs) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\cli\main.py", line 78, in _main exit_code = do_call(args, p) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\cli\conda_argparse.py", line 76, in do_call exit_code = getattr(module, func_name)(args, parser) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\cli\main_install.py", line 11, in execute install(args, parser, 'install') File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\cli\install.py", line 236, in install force_reinstall=context.force, File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\core\solve.py", line 504, in solve_for_transaction force_remove, force_reinstall) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\core\solve.py", line 437, in solve_for_diff final_precs = self.solve_final_state(deps_modifier, prune, ignore_pinned, force_remove) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\core\solve.py", line 145, in solve_final_state solution = tuple(Dist(d) for d in prefix_data.iter_records()) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\core\linked_data.py", line 90, in iter_records return itervalues(self._prefix_records) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\core\linked_data.py", line 110, in _prefix_records return self.__prefix_records or self.load() or self.__prefix_records File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\core\linked_data.py", line 48, in load self._load_single_record(meta_file) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\core\linked_data.py", line 116, in _load_single_record prefix_record = PrefixRecord(**json_data) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 718, in __call__ instance = super(EntityType, cls).__call__(*args, **kwargs) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 735, in __init__ setattr(self, key, kwargs[key]) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 427, in __set__ instance.__dict__[self.name] = self.validate(instance, self.box(instance, instance.__class__, val)) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 666, in box return self._type(**val) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 718, in __call__ instance = super(EntityType, cls).__call__(*args, **kwargs) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 735, in __init__ setattr(self, key, kwargs[key]) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 427, in __set__ instance.__dict__[self.name] = self.validate(instance, self.box(instance, instance.__class__, val)) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 590, in box return self._type(v if isinstance(v, et) else et(**v) for v in val) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 590, in <genexpr> return self._type(v if isinstance(v, et) else et(**v) for v in val) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 718, in __call__ instance = super(EntityType, cls).__call__(*args, **kwargs) MemoryError ```
2018-01-24T11:31:59Z
[]
[]
Traceback (most recent call last): File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\exceptions.py", line 743, in __call__ return func(*args, **kwargs) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\cli\main.py", line 78, in _main exit_code = do_call(args, p) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\cli\conda_argparse.py", line 76, in do_call exit_code = getattr(module, func_name)(args, parser) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\cli\main_install.py", line 11, in execute install(args, parser, 'install') File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\cli\install.py", line 236, in install force_reinstall=context.force, File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\core\solve.py", line 504, in solve_for_transaction force_remove, force_reinstall) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\core\solve.py", line 437, in solve_for_diff final_precs = self.solve_final_state(deps_modifier, prune, ignore_pinned, force_remove) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\core\solve.py", line 145, in solve_final_state solution = tuple(Dist(d) for d in prefix_data.iter_records()) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\core\linked_data.py", line 90, in iter_records return itervalues(self._prefix_records) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\core\linked_data.py", line 110, in _prefix_records return self.__prefix_records or self.load() or self.__prefix_records File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\core\linked_data.py", line 48, in load self._load_single_record(meta_file) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\core\linked_data.py", line 116, in _load_single_record prefix_record = PrefixRecord(**json_data) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 718, in __call__ instance = super(EntityType, cls).__call__(*args, **kwargs) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 735, in __init__ setattr(self, key, kwargs[key]) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 427, in __set__ instance.__dict__[self.name] = self.validate(instance, self.box(instance, instance.__class__, val)) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 666, in box return self._type(**val) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 718, in __call__ instance = super(EntityType, cls).__call__(*args, **kwargs) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 735, in __init__ setattr(self, key, kwargs[key]) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 427, in __set__ instance.__dict__[self.name] = self.validate(instance, self.box(instance, instance.__class__, val)) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 590, in box return self._type(v if isinstance(v, et) else et(**v) for v in val) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 590, in <genexpr> return self._type(v if isinstance(v, et) else et(**v) for v in val) File "C:\Users\Jason\Anaconda3\lib\site-packages\conda\_vendor\auxlib\entity.py", line 718, in __call__ instance = super(EntityType, cls).__call__(*args, **kwargs) MemoryError
4,283
conda/conda
conda__conda-6777
d503c26e0a9165f21b758238cdf08c3f1b81d20f
diff --git a/conda/activate.py b/conda/activate.py --- a/conda/activate.py +++ b/conda/activate.py @@ -43,6 +43,12 @@ def __init__(self, shell, arguments=None): self.shell = shell self._raw_arguments = arguments + if PY2: + self.environ = {ensure_fs_path_encoding(k): ensure_fs_path_encoding(v) + for k, v in iteritems(os.environ)} + else: + self.environ = os.environ.copy() + if shell == 'posix': self.pathsep_join = ':'.join self.path_conversion = native_path_to_unix @@ -215,13 +221,13 @@ def build_activate(self, env_name_or_prefix): prefix = normpath(prefix) # query environment - old_conda_shlvl = int(os.getenv('CONDA_SHLVL', 0)) - old_conda_prefix = os.getenv('CONDA_PREFIX') + old_conda_shlvl = int(self.environ.get('CONDA_SHLVL', 0)) + old_conda_prefix = self.environ.get('CONDA_PREFIX') max_shlvl = context.max_shlvl if old_conda_prefix == prefix: return self.build_reactivate() - if os.getenv('CONDA_PREFIX_%s' % (old_conda_shlvl-1)) == prefix: + if self.environ.get('CONDA_PREFIX_%s' % (old_conda_shlvl-1)) == prefix: # in this case, user is attempting to activate the previous environment, # i.e. step back down return self.build_deactivate() @@ -282,8 +288,8 @@ def build_activate(self, env_name_or_prefix): def build_deactivate(self): # query environment - old_conda_shlvl = int(os.getenv('CONDA_SHLVL', 0)) - old_conda_prefix = os.getenv('CONDA_PREFIX', None) + old_conda_shlvl = int(self.environ.get('CONDA_SHLVL', 0)) + old_conda_prefix = self.environ.get('CONDA_PREFIX', None) if old_conda_shlvl <= 0 or old_conda_prefix is None: return { 'unset_vars': (), @@ -314,7 +320,7 @@ def build_deactivate(self): } activate_scripts = () else: - new_prefix = os.getenv('CONDA_PREFIX_%d' % new_conda_shlvl) + new_prefix = self.environ.get('CONDA_PREFIX_%d' % new_conda_shlvl) conda_default_env = self._default_env(new_prefix) conda_prompt_modifier = self._prompt_modifier(conda_default_env) @@ -341,8 +347,8 @@ def build_deactivate(self): } def build_reactivate(self): - conda_prefix = os.environ.get('CONDA_PREFIX') - conda_shlvl = int(os.environ.get('CONDA_SHLVL', -1)) + conda_prefix = self.environ.get('CONDA_PREFIX') + conda_shlvl = int(self.environ.get('CONDA_SHLVL', -1)) if not conda_prefix or conda_shlvl < 1: # no active environment, so cannot reactivate; do nothing return { @@ -352,7 +358,7 @@ def build_reactivate(self): 'deactivate_scripts': (), 'activate_scripts': (), } - conda_default_env = os.environ.get('CONDA_DEFAULT_ENV', self._default_env(conda_prefix)) + conda_default_env = self.environ.get('CONDA_DEFAULT_ENV', self._default_env(conda_prefix)) # environment variables are set only to aid transition from conda 4.3 to conda 4.4 return { 'unset_vars': (), @@ -366,7 +372,7 @@ def build_reactivate(self): } def _get_starting_path_list(self): - path = os.environ['PATH'] + path = self.environ['PATH'] if on_win: # On Windows, the Anaconda Python interpreter prepends sys.prefix\Library\bin on # startup. It's a hack that allows users to avoid using the correct activation @@ -446,8 +452,8 @@ def _update_prompt(self, set_vars, conda_prompt_modifier): return if self.shell == 'posix': - ps1 = os.environ.get('PS1', '') - current_prompt_modifier = os.environ.get('CONDA_PROMPT_MODIFIER') + ps1 = self.environ.get('PS1', '') + current_prompt_modifier = self.environ.get('CONDA_PROMPT_MODIFIER') if current_prompt_modifier: ps1 = re.sub(re.escape(current_prompt_modifier), r'', ps1) # Because we're using single-quotes to set shell variables, we need to handle the @@ -458,8 +464,8 @@ def _update_prompt(self, set_vars, conda_prompt_modifier): 'PS1': conda_prompt_modifier + ps1, }) elif self.shell == 'csh': - prompt = os.environ.get('prompt', '') - current_prompt_modifier = os.environ.get('CONDA_PROMPT_MODIFIER') + prompt = self.environ.get('prompt', '') + current_prompt_modifier = self.environ.get('CONDA_PROMPT_MODIFIER') if current_prompt_modifier: prompt = re.sub(re.escape(current_prompt_modifier), r'', prompt) set_vars.update({ @@ -498,6 +504,13 @@ def ensure_binary(value): return value +def ensure_fs_path_encoding(value): + try: + return value.decode(FILESYSTEM_ENCODING) + except AttributeError: + return value + + def native_path_to_unix(paths): # pragma: unix no cover # on windows, uses cygpath to convert windows native paths to posix paths if not on_win: @@ -532,6 +545,7 @@ def path_identity(paths): on_win = bool(sys.platform == "win32") PY2 = sys.version_info[0] == 2 +FILESYSTEM_ENCODING = sys.getfilesystemencoding() if PY2: # pragma: py3 no cover string_types = basestring, # NOQA text_type = unicode # NOQA
source activate env error **I'm submitting a...** - [x] bug report - [ ] feature request ### Current Behavior when i try to activate a conda env,this doesn't work for me the output is: ``` Traceback (most recent call last): File "/anaconda/lib/python2.7/site-packages/conda/gateways/logging.py", line 64, in emit msg = self.format(record) File "/anaconda/lib/python2.7/logging/__init__.py", line 734, in format return fmt.format(record) File "/anaconda/lib/python2.7/logging/__init__.py", line 465, in format record.message = record.getMessage() File "/anaconda/lib/python2.7/logging/__init__.py", line 329, in getMessage msg = msg % self.args File "/anaconda/lib/python2.7/site-packages/conda/__init__.py", line 43, in __repr__ return '%s: %s' % (self.__class__.__name__, text_type(self)) File "/anaconda/lib/python2.7/site-packages/conda/__init__.py", line 47, in __str__ return text_type(self.message % self._kwargs) ValueError: unsupported format character '{' (0x7b) at index 513 Logged from file exceptions.py, line 724 ``` ### Steps to Reproduce i have regularly updated conda to the lateset version recently, and i installed pipenv.i know that cannot work with a conda-created env, but i suspect this may lead to this error then this error came ### Expected Behavior <!-- What do you think should happen? --> ##### `conda info` <!-- between the ticks below, paste the output of 'conda info' --> ``` active environment : None shell level : 0 user config file : /Users/bubu/.condarc populated config files : /Users/bubu/.condarc conda version : 4.4.7 conda-build version : 3.2.1 python version : 2.7.13.final.0 base environment : /anaconda (writable) channel URLs : https://conda.anaconda.org/anaconda-fusion/osx-64 https://conda.anaconda.org/anaconda-fusion/noarch https://repo.continuum.io/pkgs/main/osx-64 https://repo.continuum.io/pkgs/main/noarch https://repo.continuum.io/pkgs/free/osx-64 https://repo.continuum.io/pkgs/free/noarch https://repo.continuum.io/pkgs/r/osx-64 https://repo.continuum.io/pkgs/r/noarch https://repo.continuum.io/pkgs/pro/osx-64 https://repo.continuum.io/pkgs/pro/noarch package cache : /anaconda/pkgs /Users/bubu/.conda/pkgs envs directories : /anaconda/envs /Users/bubu/.conda/envs platform : osx-64 user-agent : conda/4.4.7 requests/2.18.4 CPython/2.7.13 Darwin/17.3.0 OSX/10.13.2 UID:GID : 501:20 netrc file : None offline mode : False ``` ##### `conda config --show-sources` <!-- between the ticks below, paste the output of 'conda config --show-sources' --> ``` ==> /Users/bubu/.condarc <== ssl_verify: True channels: - anaconda-fusion - defaults ``` ##### `conda list --show-channel-urls` <!-- between the ticks below, paste the output of 'conda list --show-channel-urls' --> ``` # packages in environment at /anaconda: # _license 1.1 py27_1 defaults alabaster 0.7.9 py27_0 defaults amqp 1.4.9 <pip> anaconda custom py27_0 defaults anaconda-client 1.6.3 py27_0 defaults anaconda-navigator 1.5.2 py27_0 defaults anaconda-project 0.4.1 py27_0 defaults ansible 2.4.1.0 <pip> apipkg 1.4 <pip> appdirs 1.4.3 <pip> appnope 0.1.0 py27_0 defaults appscript 1.0.1 py27_0 defaults argcomplete 1.0.0 py27_1 defaults argon2-cffi 16.3.0 <pip> astroid 1.4.9 py27_0 defaults astropy 1.3.3 np112py27_0 defaults attrs 17.2.0 <pip> autobahn 17.7.1 <pip> Automat 0.6.0 <pip> babel 2.3.4 py27_0 defaults backports 1.0 py27_0 defaults backports.weakref 1.0rc1 <pip> backports_abc 0.5 py27_0 defaults bcrypt 3.1.4 <pip> beautifulsoup4 4.5.3 py27_0 defaults billiard 3.3.0.23 <pip> binaryornot 0.4.3 <pip> bitarray 0.8.1 py27_0 defaults blaze 0.10.1 py27_0 defaults bleach 2.0.0 <pip> bokeh 0.12.4 py27_0 defaults boto 2.45.0 py27_0 defaults bottle 0.12.13 <pip> bottle-sqlite 0.1.3 <pip> bottleneck 1.2.1 np112py27_0 defaults ca-certificates 2017.08.26 ha1e5d58_0 defaults cache-client 0.1.0 <pip> cdecimal 2.3 py27_2 defaults celery 3.1.23 <pip> certifi 2017.11.5 py27hfa9a1c4_0 defaults certifi 2017.4.17 <pip> cffi 1.9.1 py27_0 defaults Chameleon 3.1 <pip> chardet 2.3.0 py27_0 defaults chardet 3.0.3 <pip> check-manifest 0.35 <pip> chest 0.2.3 py27_0 defaults click 6.7 py27_0 defaults click-default-group 1.2 <pip> cloudpickle 0.2.2 py27_0 defaults clyent 1.2.2 py27_0 defaults colorama 0.3.7 py27_0 defaults conda 4.4.7 py27_0 defaults conda-build 3.2.1 py27_0 defaults conda-env 2.6.0 h36134e3_0 defaults conda-verify 2.0.0 py27_0 defaults configobj 5.0.6 py27_0 defaults configparser 3.5.0 py27_0 defaults constantly 15.1.0 <pip> contextlib2 0.5.4 py27_0 defaults cookiecutter 1.5.1 <pip> crockford 0.0.2 <pip> cryptography 1.7.1 py27_0 defaults cssselect 1.0.1 <pip> curl 7.52.1 0 defaults cvxopt 1.1.9 <pip> cycler 0.10.0 py27_0 defaults cython 0.25.2 py27_0 defaults cytoolz 0.8.2 py27_0 defaults dask 0.13.0 py27_0 defaults datashape 0.5.4 py27_0 defaults decorator 4.0.11 py27_0 defaults defusedxml 0.5.0 <pip> devpi 2.2.0 <pip> devpi-client 3.0.0 <pip> devpi-common 3.1.0 <pip> devpi-server 4.3.0 <pip> devpi-web 3.2.0 <pip> dill 0.2.5 py27_0 defaults Django 1.10.5 <pip> django-extensions 1.7.9 <pip> django-filter 1.0.1 <pip> django-rest-swagger 2.1.1 <pip> django-silk 0.7.3 <pip> djangorestframework 3.5.4 <pip> djangorestframework-bulk 0.2.1 <pip> dlib 19.4.0 <pip> docopt 0.6.2 <pip> docutils 0.13.1 py27_0 defaults dpkt 1.9.1 <pip> dry-rest-permissions 0.1.8 <pip> entrypoints 0.2.2 py27_0 defaults entrypoints 0.2.2 <pip> enum34 1.1.6 py27_0 defaults environment-kernels 1.1 <pip> et_xmlfile 1.0.1 py27_0 defaults etu-tools 0.3.0 <pip> execnet 1.4.1 <pip> factory-boy 2.8.1 <pip> fake-useragent 0.1.7 <pip> Faker 0.7.12 <pip> falcon 1.2.0 <pip> fastcache 1.0.2 py27_1 defaults filelock 2.0.7 py27_0 defaults fire 0.1.2 <pip> first 2.0.1 <pip> flake8 3.3.0 <pip> Flask 0.3 <pip> flask 0.12 py27_0 defaults flask-cors 3.0.2 py27_0 defaults Flask-Login 0.4.0 <pip> freetype 2.5.5 2 defaults funcsigs 1.0.2 py27_0 defaults functools32 3.2.3.2 py27_0 defaults future 0.16.0 <pip> futures 3.0.5 py27_0 defaults GB2260 0.4.1 <pip> gdbn 0.1 <pip> get_terminal_size 1.0.0 py27_0 defaults gevent 1.2.2 <pip> gevent 1.2.1 py27_0 defaults Glances 2.9.1 <pip> glob2 0.6 py27h55c9705_0 defaults gnumpy 0.2 <pip> greenlet 0.4.12 <pip> greenlet 0.4.11 py27_0 defaults grequests 0.3.0 <pip> grin 1.2.1 py27_3 defaults gunicorn 19.7.1 <pip> gym 0.9.3 <pip> h5py 2.7.0 np112py27_0 defaults hdf5 1.8.17 1 defaults heapdict 1.0.0 py27_1 defaults hkdf 0.0.3 <pip> holidays 0.8.1 <pip> html5lib 0.999999999 <pip> humanize 0.5.1 <pip> hupper 1.0 <pip> hyperlink 17.3.0 <pip> icu 54.1 0 defaults idna 2.6 <pip> idna 2.2 py27_0 defaults imagesize 0.7.1 py27_0 defaults incremental 17.5.0 <pip> ipaddress 1.0.18 py27_0 defaults ipykernel 4.5.2 py27_0 defaults ipykernel 4.6.1 <pip> ipython 5.1.0 py27_1 defaults ipython 5.4.1 <pip> ipython_genutils 0.1.0 py27_0 defaults ipywidgets 6.0.0 <pip> ipywidgets 5.2.2 py27_1 defaults isbnlib 3.7.2 <pip> isort 4.2.5 py27_0 defaults itchat 1.3.10 <pip> itsdangerous 0.24 py27_0 defaults jbig 2.1 0 defaults jdcal 1.3 py27_0 defaults jedi 0.10.2 <pip> jedi 0.9.0 py27_1 defaults jieba 0.39 <pip> jinja2 2.9.4 py27_0 defaults jinja2-time 0.2.0 <pip> joblib 0.11 <pip> jpeg 9b 0 defaults js2xml 0.3.1 <pip> json-lines 0.3.1 <pip> jsonschema 2.5.1 py27_0 defaults jsonschema 2.6.0 <pip> jupyter 1.0.0 py27_3 defaults jupyter-client 5.0.1 <pip> jupyter-core 4.3.0 <pip> jupyter_client 4.4.0 py27_0 defaults jupyter_console 5.0.0 py27_0 defaults jupyter_core 4.2.1 py27_0 defaults kii-client 0.1.3 <pip> kombu 3.0.37 <pip> lazy-object-proxy 1.2.2 py27_0 defaults libiconv 1.14 0 defaults libpng 1.6.27 0 defaults libtiff 4.0.6 3 defaults libxml2 2.9.4 0 defaults libxslt 1.1.29 0 defaults line-profiler 2.0 <pip> llvmlite 0.18.0 py27_0 defaults locket 0.2.0 py27_1 defaults lxml 3.7.2 py27_0 defaults magic-wormhole 0.10.2 <pip> Markdown 2.2.0 <pip> markupsafe 0.23 py27_2 defaults matplotlib 2.0.2 np112py27_0 defaults mccabe 0.6.1 <pip> meld3 1.0.2 <pip> mistune 0.7.4 <pip> mistune 0.7.3 py27_1 defaults mkl 2017.0.1 0 defaults mkl-service 1.1.2 py27_3 defaults more-itertools 3.2.0 <pip> mpmath 0.19 py27_1 defaults multipledispatch 0.4.9 py27_0 defaults mysql 0.0.1 <pip> MySQL-python 1.2.5 <pip> nbconvert 4.2.0 py27_0 defaults nbconvert 5.2.1 <pip> nbformat 4.2.0 py27_0 defaults nbformat 4.3.0 <pip> ncmbot 0.1.6 <pip> networkx 1.11 py27_0 defaults nlib 0.6 <pip> nltk 3.2.2 py27_0 defaults nolearn 0.5b1 <pip> nose 1.3.7 py27_1 defaults notebook 5.0.0 <pip> notebook 4.3.1 py27_0 defaults numba 0.33.0 np112py27_0 defaults numexpr 2.6.2 np112py27_0 defaults numpy 1.12.1 py27_0 defaults numpydoc 0.6.0 py27_0 defaults odo 0.5.0 py27_1 defaults olefile 0.44 <pip> openapi-codec 1.3.2 <pip> opencv-python 3.2.0.7 <pip> openface 0.2.1 <pip> openpyxl 2.4.1 py27_0 defaults openssl 1.0.2n hdbc3d79_0 defaults org-client 0.4.12 <pip> packaging 16.8 <pip> pandas 0.17.1 <pip> pandas 0.20.2 np112py27_0 defaults pandocfilters 1.4.1 <pip> paramiko 2.3.1 <pip> parse-accept-language 0.1.2 <pip> parsel 1.2.0 <pip> partd 0.3.7 py27_0 defaults passlib 1.7.1 <pip> PasteDeploy 1.5.2 <pip> path.py 10.0 py27_0 defaults pathlib 1.0.1 <pip> pathlib2 2.2.0 py27_0 defaults patsy 0.4.1 py27_0 defaults pendulum 1.0.2 <pip> pep8 1.7.0 py27_0 defaults petl 1.1.1 <pip> pew 1.1.2 <pip> pexpect 4.2.1 py27_0 defaults pickleshare 0.7.4 py27_0 defaults pigar 0.7.0 <pip> pillow 4.0.0 py27_0 defaults pip 9.0.1 py27h1567d89_4 defaults pip-tools 1.9.0 <pip> pipreqs 0.4.9 <pip> pkginfo 1.4.1 <pip> pkginfo 1.4.1 py27_0 defaults pluggy 0.4.0 <pip> ply 3.9 py27_0 defaults poyo 0.4.1 <pip> pprofile 1.11.0 <pip> profiling 0.1.3 <pip> progressbar 2.3 <pip> prompt_toolkit 1.0.9 py27_0 defaults protobuf 3.3.0 <pip> psutil 5.0.1 py27_0 defaults psycopg2 2.7.1 <pip> ptpython 0.39 <pip> ptyprocess 0.5.1 py27_0 defaults py 1.4.32 py27_0 defaults py-heat 0.0.2 <pip> py-heat-magic 0.0.2 <pip> pyasn1 0.1.9 py27_0 defaults pyasn1-modules 0.0.10 <pip> pyaudio 0.2.7 py27_0 defaults pycodestyle 2.3.1 <pip> pycosat 0.6.3 py27h6c51c7e_0 defaults pycparser 2.17 py27_0 defaults pycrypto 2.6.1 py27_4 defaults pycurl 7.43.0 py27_2 defaults PyDispatcher 2.0.5 <pip> pyflakes 1.5.0 py27_0 defaults pyglet 1.2.4 <pip> pygments 2.1.3 py27_0 defaults pylint 1.6.4 py27_1 defaults pymongo 3.5.1 <pip> PyMySQL 0.7.11 <pip> PyNaCl 1.1.2 <pip> pyopenssl 16.2.0 py27_0 defaults pyparsing 2.1.4 py27_0 defaults pyparsing 2.2.0 <pip> pypng 0.0.18 <pip> PyQRCode 1.2.1 <pip> pyqt 5.6.0 py27_1 defaults pyquery 1.3.0 <pip> pyramid 1.8.3 <pip> pyramid-chameleon 0.3 <pip> pyresttest 1.7.2.dev0 <pip> pyspider 0.3.9 <pip> pytables 3.3.0 np112py27_0 defaults pytest 3.0.5 py27_0 defaults python 2.7.13 0 defaults python-cjson 1.2.1 <pip> python-dateutil 2.6.0 py27_0 defaults python-mimeparse 1.6.0 <pip> python.app 1.2 py27_4 defaults pythonpy 0.4.11 <pip> pytz 2016.10 py27_0 defaults pywavelets 0.5.2 np112py27_0 defaults pyyaml 3.12 py27_0 defaults pyzmq 16.0.2 py27_0 defaults qiniu 7.2.0 <pip> qt 5.6.2 0 defaults qtawesome 0.4.3 py27_0 defaults qtconsole 4.2.1 py27_1 defaults qtmodern 0.1.4 <pip> qtpy 1.2.1 py27_0 defaults QtPy 1.3.1 <pip> queuelib 1.4.2 <pip> raven 6.0.0 <pip> readline 6.2 2 defaults readme-renderer 17.2 <pip> redis 3.2.0 0 defaults redis-py 2.10.5 py27_0 defaults repoze.lru 0.6 <pip> requestium 0.1.9 <pip> requests 2.12.4 py27_0 defaults requests 2.18.4 <pip> requests-file 1.4.2 <pip> rope 0.9.4 py27_1 defaults ruamel_yaml 0.11.14 py27_1 defaults scandir 1.4 py27_0 defaults scikit-image 0.13.0 np112py27_0 defaults scikit-learn 0.18.2 np112py27_0 defaults scikit-learn 0.17.1 <pip> scipy 0.19.1 np112py27_0 defaults scipy 0.16.1 <pip> Scrapy 1.4.0 <pip> seaborn 0.7.1 py27_0 defaults selenium 3.8.0 <pip> serpy 0.1.1 <pip> service-identity 17.0.0 <pip> setuptools 38.4.0 py27_0 defaults setuptools 38.4.0 <pip> shutilwhich 1.1.0 <pip> simplegeneric 0.8.1 py27_1 defaults simplejson 3.11.1 <pip> singledispatch 3.4.0.3 py27_0 defaults sip 4.18 py27_0 defaults six 1.10.0 py27_0 defaults sklearn 0.0 <pip> slimit 0.8.1 <pip> snowballstemmer 1.2.1 py27_0 defaults sockjs-tornado 1.0.3 py27_0 defaults spake2 0.7 <pip> speechpy 1.1 <pip> sphinx 1.5.1 py27_0 defaults spyder 3.1.2 py27_0 defaults sqlalchemy 1.1.5 py27_0 defaults sqlite 3.13.0 0 defaults ssl_match_hostname 3.4.0.2 py27_1 defaults sso-auth-drf 1.4.2 <pip> statsmodels 0.8.0 np112py27_0 defaults subprocess32 3.2.7 py27_0 defaults supervisor 3.3.2 <pip> sympy 1.0 py27_0 defaults tblib 1.3.2 <pip> tensorflow 1.2.0 <pip> terminado 0.6 py27_0 defaults terminaltables 3.1.0 <pip> testpath 0.3.1 <pip> textrank4zh 0.3 <pip> tflearn 0.3.2 <pip> tk 8.5.18 0 defaults tldextract 2.2.0 <pip> toolz 0.8.2 py27_0 defaults tornado 4.4.2 py27_0 defaults tornado 4.5.1 <pip> tox 2.7.0 <pip> tqdm 4.15.0 <pip> traitlets 4.3.1 py27_0 defaults translationstring 1.3 <pip> Twisted 17.5.0 <pip> txaio 2.8.1 <pip> txtorcon 0.19.3 <pip> u-msgpack-python 2.4.1 <pip> unicodecsv 0.14.1 py27_0 defaults urllib3 1.22 <pip> urwid 1.3.1 <pip> valuedispatch 0.0.1 <pip> venusian 1.1.0 <pip> virtualenv 15.1.0 <pip> virtualenv-clone 0.2.6 <pip> w3lib 1.18.0 <pip> waitress 1.0.2 <pip> wcwidth 0.1.7 py27_0 defaults web.py 0.38 <pip> webencodings 0.5.1 <pip> WebOb 1.7.2 <pip> weibo 0.2.2 <pip> Werkzeug 0.12.1 <pip> werkzeug 0.11.15 py27_0 defaults wheel 0.29.0 py27_0 defaults whichcraft 0.4.1 <pip> Whoosh 2.7.4 <pip> widgetsnbextension 2.0.0 <pip> widgetsnbextension 1.2.6 py27_0 defaults wrapt 1.10.8 py27_0 defaults WsgiDAV 2.2.4 <pip> xlrd 1.0.0 py27_0 defaults xlsxwriter 0.9.6 py27_0 defaults xlwings 0.10.2 py27_0 defaults xlwt 1.2.0 py27_0 defaults xz 5.2.2 1 defaults yaml 0.1.6 0 defaults yarg 0.1.9 <pip> zlib 1.2.8 3 defaults zoo-client 0.7.8 <pip> zope.deprecation 4.2.0 <pip> zope.interface 4.4.1 <pip> ```
Have the same issue after an update The same issue +1 The same issue +1 Can someone who is having this problem please give me the output of CONDA_VERBOSE=3 conda shell.posix activate base For people having this problem, please give me step-by-step setup and commands I can use to reproduce this myself. ``` → source activate py3 Traceback (most recent call last): File "/Users/sherbein/opt/miniconda2/lib/python2.7/site-packages/conda/gateways/logging.py", line 64, in emit msg = self.format(record) File "/Users/sherbein/opt/miniconda2/lib/python2.7/logging/__init__.py", line 734, in format return fmt.format(record) File "/Users/sherbein/opt/miniconda2/lib/python2.7/logging/__init__.py", line 465, in format record.message = record.getMessage() File "/Users/sherbein/opt/miniconda2/lib/python2.7/logging/__init__.py", line 329, in getMessage msg = msg % self.args File "/Users/sherbein/opt/miniconda2/lib/python2.7/site-packages/conda/__init__.py", line 43, in __repr__ return '%s: %s' % (self.__class__.__name__, text_type(self)) File "/Users/sherbein/opt/miniconda2/lib/python2.7/site-packages/conda/__init__.py", line 47, in __str__ return text_type(self.message % self._kwargs) ValueError: unsupported format character '{' (0x7b) at index 447 Logged from file exceptions.py, line 724 ``` ``` → CONDA_VERBOSE=3 conda shell.posix activate base PS1='(base) ' export CONDA_DEFAULT_ENV='base' export CONDA_PREFIX='/Users/sherbein/opt/miniconda2' export CONDA_PROMPT_MODIFIER='(base) ' export CONDA_PYTHON_EXE='/Users/sherbein/opt/miniconda2/bin/python' export CONDA_SHLVL='1' export PATH='/Users/sherbein/opt/miniconda2/bin:/Users/sherbein/opt/miniconda2/bin:/Users/sherbein/.cargo/bin:/usr/local/bin:/usr/local/texlive/2016/bin/x86_64-darwin:/usr/local/sbin:/Users/sherbein/.dotfiles/bin:/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:/Applications/VMware Fusion.app/Contents/Public:/Library/TeX/texbin:/opt/X11/bin:/Applications/Wireshark.app/Contents/MacOS:/Users/sherbein/.antigen/bundles/robbyrussell/oh-my-zsh/lib:/Users/sherbein/.antigen/bundles/robbyrussell/oh-my-zsh/plugins/git:/Users/sherbein/.antigen/bundles/robbyrussell/oh-my-zsh/plugins/command-not-found:/Users/sherbein/.antigen/bundles/zsh-users/zsh-syntax-highlighting' ``` FWIW, I get the exact same error when I try and use `conda activate py3` (both before and after removing conda from my PATH and sourcing `conda.sh` in `etc/profile.d`) @SteVwonder: Can you show the output of ``` /Users/sherbein/opt/miniconda2/bin/python -c 'import conda; conda.CondaError.__str__ = lambda self: "\n".join(map(conda.text_type, [self.message, self._kwargs])); from conda.cli.main import main; main("conda", "shell.posix", "activate", "py3")' ``` ? ``` → /Users/sherbein/opt/miniconda2/bin/python -c 'import conda; conda.CondaError.__str__ = lambda self: "\n".join(map(conda.text_type, [self.message, self._kwargs])); from conda.cli.main import main; main("conda", "shell.posix", "activate", "py3")' # >>>>>>>>>>>>>>>>>>>>>> ERROR REPORT <<<<<<<<<<<<<<<<<<<<<< Traceback (most recent call last): File "/Users/sherbein/opt/miniconda2/lib/python2.7/site-packages/conda/cli/main.py", line 111, in main return activator_main() File "/Users/sherbein/opt/miniconda2/lib/python2.7/site-packages/conda/activate.py", line 551, in main assert len(argv) >= 3 AssertionError `$ -c` environment variables: CIO_TEST=<not set> CONDA_ROOT=/Users/sherbein/opt/miniconda2 PATH=/Users/sherbein/opt/miniconda2/bin:/Users/sherbein/.cargo/bin:/usr/loc al/bin:/usr/local/texlive/2016/bin/x86_64-darwin:/usr/local/sbin:/User s/sherbein/.dotfiles/bin:/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin: /Applications/VMware Fusion.app/Contents/Public:/Library/TeX/texbin:/o pt/X11/bin:/Applications/Wireshark.app/Contents/MacOS:/Users/sherbein/ .antigen/bundles/robbyrussell/oh-my- zsh/lib:/Users/sherbein/.antigen/bundles/robbyrussell/oh-my- zsh/plugins/git:/Users/sherbein/.antigen/bundles/robbyrussell/oh-my- zsh/plugins/command-not-found:/Users/sherbein/.antigen/bundles/zsh- users/zsh-syntax-highlighting PKG_CONFIG_PATH=/Users/sherbein/.local/munge/0.5.13/lib/pkgconfig REQUESTS_CA_BUNDLE=<not set> SSL_CERT_FILE=<not set> SUDO_EDITOR=emacsclient -nw active environment : None user config file : /Users/sherbein/.condarc populated config files : conda version : 4.4.7 conda-build version : not installed python version : 2.7.14.final.0 base environment : /Users/sherbein/opt/miniconda2 (writable) channel URLs : https://repo.continuum.io/pkgs/main/osx-64 https://repo.continuum.io/pkgs/main/noarch https://repo.continuum.io/pkgs/free/osx-64 https://repo.continuum.io/pkgs/free/noarch https://repo.continuum.io/pkgs/r/osx-64 https://repo.continuum.io/pkgs/r/noarch https://repo.continuum.io/pkgs/pro/osx-64 https://repo.continuum.io/pkgs/pro/noarch package cache : /Users/sherbein/opt/miniconda2/pkgs /Users/sherbein/.conda/pkgs envs directories : /Users/sherbein/opt/miniconda2/envs /Users/sherbein/.conda/envs platform : osx-64 user-agent : conda/4.4.7 requests/2.18.4 CPython/2.7.14 Darwin/17.3.0 OSX/10.13.2 UID:GID : 501:20 netrc file : None offline mode : False ``` Ah, there was an error in the command I wrote down, sorry. Can you try ``` /Users/sherbein/opt/miniconda2/bin/python -c 'import conda; conda.CondaError.__str__ = lambda self: "\n".join(map(conda.text_type, [self.message, self._kwargs])); from conda.cli.main import main; import sys; sys.argv[:]=["conda", "shell.posix", "activate", "py3"]; main()' ``` instead? ``` PS1='(py3) ' export CONDA_DEFAULT_ENV='py3' export CONDA_PREFIX='/Users/sherbein/opt/miniconda2/envs/py3' export CONDA_PROMPT_MODIFIER='(py3) ' export CONDA_PYTHON_EXE='/Users/sherbein/opt/miniconda2/bin/python' export CONDA_SHLVL='1' export PATH='/Users/sherbein/opt/miniconda2/envs/py3/bin:/Users/sherbein/opt/miniconda2/bin:/Users/sherbein/.cargo/bin:/usr/local/bin:/usr/local/texlive/2016/bin/x86_64-darwin:/usr/local/sbin:/Users/sherbein/.dotfiles/bin:/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:/Applications/VMware Fusion.app/Contents/Public:/Library/TeX/texbin:/opt/X11/bin:/Applications/Wireshark.app/Contents/MacOS:/Users/sherbein/.antigen/bundles/robbyrussell/oh-my-zsh/lib:/Users/sherbein/.antigen/bundles/robbyrussell/oh-my-zsh/plugins/git:/Users/sherbein/.antigen/bundles/robbyrussell/oh-my-zsh/plugins/command-not-found:/Users/sherbein/.antigen/bundles/zsh-users/zsh-syntax-highlighting' ``` Now that's strange.. I don't understand how your previous `source activate py3` was able to error out. The Python code that gets executed by `source activate` should be the same as with `conda shell.posix activate py3` which didn't fail somehow... Can you show ``` head /Users/sherbein/opt/miniconda2/etc/profile.d/conda.sh ``` ? ``` → head /Users/sherbein/opt/miniconda2/etc/profile.d/conda.sh _CONDA_EXE="/Users/sherbein/opt/miniconda2/bin/conda" _CONDA_ROOT="/Users/sherbein/opt/miniconda2" _conda_set_vars() { # set _CONDA_SHELL_FLAVOR if [ -n "${BASH_VERSION:+x}" ]; then _CONDA_SHELL_FLAVOR=bash elif [ -n "${ZSH_VERSION:+x}" ]; then _CONDA_SHELL_FLAVOR=zsh elif [ -n "${KSH_VERSION:+x}" ]; then _CONDA_SHELL_FLAVOR=ksh ``` And you can still reproduce the error with `source activate py3`? ``` → source activate py3 Traceback (most recent call last): File "/Users/sherbein/opt/miniconda2/lib/python2.7/site-packages/conda/gateways/logging.py", line 64, in emit msg = self.format(record) File "/Users/sherbein/opt/miniconda2/lib/python2.7/logging/__init__.py", line 734, in format return fmt.format(record) File "/Users/sherbein/opt/miniconda2/lib/python2.7/logging/__init__.py", line 465, in format record.message = record.getMessage() File "/Users/sherbein/opt/miniconda2/lib/python2.7/logging/__init__.py", line 329, in getMessage msg = msg % self.args File "/Users/sherbein/opt/miniconda2/lib/python2.7/site-packages/conda/__init__.py", line 43, in __repr__ return '%s: %s' % (self.__class__.__name__, text_type(self)) File "/Users/sherbein/opt/miniconda2/lib/python2.7/site-packages/conda/__init__.py", line 47, in __str__ return text_type(self.message % self._kwargs) ValueError: unsupported format character '{' (0x7b) at index 447 Logged from file exceptions.py, line 724 → eval $(conda shell.posix activate py3) → which python /Users/sherbein/opt/miniconda2/envs/py3/bin/python ``` What is `which activate` and `cat $(which activate)` ? ``` → which activate /Users/sherbein/opt/miniconda2/bin/activate → cat $(which activate) #!/bin/sh _CONDA_ROOT="/Users/sherbein/opt/miniconda2" #!/bin/sh _conda_set_vars() { # set _CONDA_SHELL_FLAVOR if [ -n "${BASH_VERSION:+x}" ]; then _CONDA_SHELL_FLAVOR=bash elif [ -n "${ZSH_VERSION:+x}" ]; then _CONDA_SHELL_FLAVOR=zsh elif [ -n "${KSH_VERSION:+x}" ]; then _CONDA_SHELL_FLAVOR=ksh elif [ -n "${POSH_VERSION:+x}" ]; then _CONDA_SHELL_FLAVOR=posh else # https://unix.stackexchange.com/a/120138/92065 local _q="$(ps -p$$ -o cmd="",comm="",fname="" 2>/dev/null | sed 's/^-//' | grep -oE '\w+' | head -n1)" if [ "$_q" = dash ]; then _CONDA_SHELL_FLAVOR=dash else (>&2 echo "Unrecognized shell.") return 1 fi fi # https://unix.stackexchange.com/questions/4650/determining-path-to-sourced-shell-script/ local script_dir case "$_CONDA_SHELL_FLAVOR" in bash) script_dir="$(dirname "${BASH_SOURCE[0]}")";; zsh) script_dir="$(dirname "${(%):-%x}")";; # http://stackoverflow.com/a/28336473/2127762 dash) x=$(lsof -p $$ -Fn0 | tail -1); script_dir="$(dirname "${x#n}")";; *) script_dir="$(cd "$(dirname "$_")" && echo "$PWD")";; esac if [ -z "${_CONDA_ROOT+x}" ]; then _CONDA_ROOT="$(dirname "$script_dir")" fi } _conda_script_is_sourced() { # http://stackoverflow.com/a/28776166/2127762 sourced=0 if [ -n "${ZSH_EVAL_CONTEXT:+x}" ]; then case $ZSH_EVAL_CONTEXT in *:file) sourced=1;; esac elif [ -n "${KSH_VERSION:+x}" ]; then [ "$(cd $(dirname -- $0) && pwd -P)/$(basename -- $0)" != "$(cd $(dirname -- ${.sh.file}) && pwd -P)/$(basename -- ${.sh.file})" ] && sourced=1 elif [ -n "${BASH_VERSION:+x}" ]; then [ "${BASH_SOURCE[0]}" = "$0" ] && sourced=1 else # All other shells: examine $0 for known shell binary filenames # Detects `sh` and `dash`; add additional shell filenames as needed. case ${0##*/} in sh|dash) sourced=0;; *) sourced=1;; esac fi return $sourced } if ! _conda_script_is_sourced; then ( >&2 echo "Error: activate must be sourced. Run 'source activate envname'" >&2 echo "instead of 'activate envname'." ) exit 1 fi _conda_set_vars . "$_CONDA_ROOT/etc/profile.d/conda.sh" || return $? _conda_activate "$@" ``` I have still no clue.. What does ``` (set -x ; source activate py3) ``` output? Oh. Good call. That might be it. I'm guessing this has to do with my zsh prompt (which I have been excluding from the above posts to hide my IP address/hostname, but is included here) ``` # sherbein at roaming-4-24-60.host.udel.edu in ~ [17:01:58] → (set -x ; source activate py3) +-zsh:1> source activate py3 +/Users/sherbein/opt/miniconda2/bin/activate:2> _CONDA_ROOT=/Users/sherbein/opt/miniconda2 +/Users/sherbein/opt/miniconda2/bin/activate:57> _conda_script_is_sourced +_conda_script_is_sourced:2> sourced=0 +_conda_script_is_sourced:3> [ -n x ']' +_conda_script_is_sourced:4> case toplevel:file:shfunc (*:file) +_conda_script_is_sourced:13> return 0 +/Users/sherbein/opt/miniconda2/bin/activate:65> _conda_set_vars +_conda_set_vars:2> [ -n '' ']' +_conda_set_vars:4> [ -n x ']' +_conda_set_vars:5> _CONDA_SHELL_FLAVOR=zsh +_conda_set_vars:22> local script_dir +_conda_set_vars:23> case zsh (bash) +_conda_set_vars:23> case zsh (zsh) +_conda_set_vars:25> script_dir=+_conda_set_vars:25> dirname /Users/sherbein/opt/miniconda2/bin/activate +_conda_set_vars:25> script_dir=/Users/sherbein/opt/miniconda2/bin +_conda_set_vars:30> [ -z x ']' +/Users/sherbein/opt/miniconda2/bin/activate:66> . /Users/sherbein/opt/miniconda2/etc/profile.d/conda.sh +/Users/sherbein/opt/miniconda2/etc/profile.d/conda.sh:1> _CONDA_EXE=/Users/sherbein/opt/miniconda2/bin/conda +/Users/sherbein/opt/miniconda2/etc/profile.d/conda.sh:2> _CONDA_ROOT=/Users/sherbein/opt/miniconda2 +/Users/sherbein/opt/miniconda2/etc/profile.d/conda.sh:105> _conda_set_vars +_conda_set_vars:2> [ -n '' ']' +_conda_set_vars:4> [ -n x ']' +_conda_set_vars:5> _CONDA_SHELL_FLAVOR=zsh +_conda_set_vars:15> [ -z x ']' +_conda_set_vars:29> [ -z x ']' +/Users/sherbein/opt/miniconda2/etc/profile.d/conda.sh:107> [ -z x ']' +/Users/sherbein/opt/miniconda2/bin/activate:67> _conda_activate py3 +_conda_activate:1> [ -n '' ']' +_conda_activate:8> local ask_conda +_conda_activate:9> ask_conda=+_conda_activate:9> PS1=$'\n%{\C-[[1m\C-[[34m%}#%{\C-[[00m%} %{\C-[[36m%}%n %{\C-[[37m%}at %{\C-[[32m%}roaming-4-24-60.host.udel.edu %{\C-[[37m%}in %{\C-[[1m\C-[[33m%}${PWD/#$HOME/~}%{\C-[[00m%}$(ys_hg_prompt_info)$(git_prompt_info) %{\C-[[37m%}[%*]\n%{\C-[[1m\C-[[31m%}→ %{\C-[[00m%}' /Users/sherbein/opt/miniconda2/bin/conda shell.posix activate py3 Traceback (most recent call last): File "/Users/sherbein/opt/miniconda2/lib/python2.7/site-packages/conda/gateways/logging.py", line 64, in emit msg = self.format(record) File "/Users/sherbein/opt/miniconda2/lib/python2.7/logging/__init__.py", line 734, in format return fmt.format(record) File "/Users/sherbein/opt/miniconda2/lib/python2.7/logging/__init__.py", line 465, in format record.message = record.getMessage() File "/Users/sherbein/opt/miniconda2/lib/python2.7/logging/__init__.py", line 329, in getMessage msg = msg % self.args File "/Users/sherbein/opt/miniconda2/lib/python2.7/site-packages/conda/__init__.py", line 43, in __repr__ return '%s: %s' % (self.__class__.__name__, text_type(self)) File "/Users/sherbein/opt/miniconda2/lib/python2.7/site-packages/conda/__init__.py", line 47, in __str__ return text_type(self.message % self._kwargs) ValueError: unsupported format character '{' (0x7b) at index 447 Logged from file exceptions.py, line 724 +_conda_activate:9> ask_conda='' +_conda_activate:9> return 1 ``` ``` # sherbein at roaming-4-24-60.host.udel.edu in ~ [17:04:06] → export PS1="#" # #source activate py3 (py3) # ``` Running `conda config --set changeps1 False` fixed this problem for me. Thanks @mbargull for helping me track this down! I know how frustrating remote troubleshooting can be. I appreciate your time and effort on this! You're welcome. The funny thing is that the first things I tried was a `PS1='%}'` and `PS1='→'` 🙄 . The minimal working reproducer is the concatenation `PS1='%}→` 😆. So the actual error is just ``` $ PS1=→ →source activate EncodingError: A unicode encoding or decoding error has occurred. Python 2 is the interpreter under which conda is running in your base environment. Replacing your base environment with one having Python 3 may help resolve this issue. If you still have a need for Python 2 environments, consider using 'conda create' and 'conda activate'. For example: $ conda create -n py2 python=2 $ conda activate py2 Error details: UnicodeDecodeError('ascii', '\xe2\x86\x92', 0, 1, 'ordinal not in range(128)') ``` but that gets masked by another bug that happens during error reporting: https://github.com/conda/conda/issues/6759. My general advice would be to always use Miniconda3 and use Python 2 in a separate environment, i.e., ``` conda create -n py2 python=2 ``` or if you use `anaconda` then ``` conda create -n anaconda2 python=2 anaconda ``` . That way, `conda` itself will always use Python 3 and doesn't run into those Unicode decoding errors. Many thx! @mbargull fixed! From https://github.com/conda/conda/issues/6753#issuecomment-359110810, the root-cause stack trace here is ``` Traceback (most recent call last): File "/Users/rthomas/bug/env/lib/python2.7/site-packages/conda/cli/main.py", line 111, in main return activator_main() File "/Users/rthomas/bug/env/lib/python2.7/site-packages/conda/activate.py", line 557, in main print(activator.execute(), end='') File "/Users/rthomas/bug/env/lib/python2.7/site-packages/conda/activate.py", line 149, in execute return getattr(self, self.command)() File "/Users/rthomas/bug/env/lib/python2.7/site-packages/conda/activate.py", line 135, in activate return self._finalize(self._yield_commands(self.build_activate(self.env_name_or_prefix)), File "/Users/rthomas/bug/env/lib/python2.7/site-packages/conda/activate.py", line 267, in build_activate self._update_prompt(set_vars, conda_prompt_modifier) File "/Users/rthomas/bug/env/lib/python2.7/site-packages/conda/activate.py", line 456, in _update_prompt ps1 = ps1.replace("'", "'\"'\"'") EncodingError: <unprintable EncodingError object> ```
2018-01-24T12:11:29Z
[]
[]
Traceback (most recent call last): File "/anaconda/lib/python2.7/site-packages/conda/gateways/logging.py", line 64, in emit msg = self.format(record) File "/anaconda/lib/python2.7/logging/__init__.py", line 734, in format return fmt.format(record) File "/anaconda/lib/python2.7/logging/__init__.py", line 465, in format record.message = record.getMessage() File "/anaconda/lib/python2.7/logging/__init__.py", line 329, in getMessage msg = msg % self.args File "/anaconda/lib/python2.7/site-packages/conda/__init__.py", line 43, in __repr__ return '%s: %s' % (self.__class__.__name__, text_type(self)) File "/anaconda/lib/python2.7/site-packages/conda/__init__.py", line 47, in __str__ return text_type(self.message % self._kwargs) ValueError: unsupported format character '{' (0x7b) at index 513
4,284
conda/conda
conda__conda-6847
158599161fcfcce024bf8257c7278a10fb3085e2
diff --git a/conda/common/path.py b/conda/common/path.py --- a/conda/common/path.py +++ b/conda/common/path.py @@ -8,7 +8,7 @@ import re import subprocess -from .compat import on_win, string_types +from .compat import PY2, ensure_fs_path_encoding, on_win, string_types from .. import CondaError from .._vendor.auxlib.decorators import memoize @@ -46,6 +46,8 @@ def is_path(value): def expand(path): + if on_win and PY2: + path = ensure_fs_path_encoding(path) return abspath(expanduser(expandvars(path)))
Conda(py2) Installation to a path (ascii characters only) with non-ascii characters in username Create a user with non-ascii characters in name. And try to install Miniconda2 to a path with only ascii characters ``` C:\Users\AndrésGarcía\Downloads>start /wait "" Miniconda2-4.3.30.2-Windows-x86_64.exe /InstallationType=JustMe /AddToPath=0 /RegisterPython=0 /NoRegistry=1 /S /D=C:\AG\mc2 ``` This will throw a menuinst traceback, ignore for now. See https://github.com/ContinuumIO/menuinst/issues/54 Then try to activate the base environment: ``` C:\Users\AndrésGarcía\Downloads>cd /d C:\AG\mc2 C:\AG\mc2>Scripts\activate Traceback (most recent call last): File "C:\AG\mc2\Scripts\conda-script.py", line 10, in <module> sys.exit(main()) File "C:\AG\mc2\lib\site-packages\conda\cli\main.py", line 178, in main init_loggers() File "C:\AG\mc2\lib\site-packages\conda\cli\main.py", line 84, in init_loggers from ..console import setup_verbose_handlers File "C:\AG\mc2\lib\site-packages\conda\console.py", line 9, in <module> from .base.context import context File "C:\AG\mc2\lib\site-packages\conda\base\context.py", line 771, in <module> context = Context(SEARCH_PATH, APP_NAME, None) File "C:\AG\mc2\lib\site-packages\conda\common\configuration.py", line 727, in __init__ self._set_search_path(search_path) File "C:\AG\mc2\lib\site-packages\conda\common\configuration.py", line 745, in _set_search_path self._set_raw_data(load_file_configs(search_path)) File "C:\AG\mc2\lib\site-packages\conda\common\configuration.py", line 364, in load_file_configs expanded_paths = tuple(expand(path) for path in search_path) File "C:\AG\mc2\lib\site-packages\conda\common\configuration.py", line 364, in <genexpr> expanded_paths = tuple(expand(path) for path in search_path) File "C:\AG\mc2\lib\site-packages\conda\common\path.py", line 247, in expand return abspath(expanduser(expandvars(path))) File "C:\AG\mc2\lib\ntpath.py", line 311, in expanduser return userhome + path[i:] UnicodeDecodeError: 'ascii' codec can't decode byte 0xe9 in position 13: ordinal not in range(128) C:\AG\mc2> ```
`C:\AG\mc2\lib\ntpath.py` is a part of the CPython stdlib, I don't really see what we can do about this. You just cannot have any unicode in `PATH` or in `%UserProfile%` on Python 2 AFAICT. @zooba what do you recommend? Probably best to detect Python 2 and pre-encode to `sys.getfilesystemencoding()` in `conda\common\path.py`. `expandvars` will do this automatically, but it looks like `expanduser` does not. You can preempt some of the edge-case failures by decoding back to Unicode and seeing if the path has changed (which means a username outside of the active code page), but unfortunately there's no real way to handle this properly in Python 2.x. You're sure to have plenty of other issues in this case though, so as long as you're encoding with `getfilesystemencoding` and not `ascii`, it's about the best option.
2018-02-06T17:47:23Z
[]
[]
Traceback (most recent call last): File "C:\AG\mc2\Scripts\conda-script.py", line 10, in <module> sys.exit(main()) File "C:\AG\mc2\lib\site-packages\conda\cli\main.py", line 178, in main init_loggers() File "C:\AG\mc2\lib\site-packages\conda\cli\main.py", line 84, in init_loggers from ..console import setup_verbose_handlers File "C:\AG\mc2\lib\site-packages\conda\console.py", line 9, in <module> from .base.context import context File "C:\AG\mc2\lib\site-packages\conda\base\context.py", line 771, in <module> context = Context(SEARCH_PATH, APP_NAME, None) File "C:\AG\mc2\lib\site-packages\conda\common\configuration.py", line 727, in __init__ self._set_search_path(search_path) File "C:\AG\mc2\lib\site-packages\conda\common\configuration.py", line 745, in _set_search_path self._set_raw_data(load_file_configs(search_path)) File "C:\AG\mc2\lib\site-packages\conda\common\configuration.py", line 364, in load_file_configs expanded_paths = tuple(expand(path) for path in search_path) File "C:\AG\mc2\lib\site-packages\conda\common\configuration.py", line 364, in <genexpr> expanded_paths = tuple(expand(path) for path in search_path) File "C:\AG\mc2\lib\site-packages\conda\common\path.py", line 247, in expand return abspath(expanduser(expandvars(path))) File "C:\AG\mc2\lib\ntpath.py", line 311, in expanduser return userhome + path[i:] UnicodeDecodeError: 'ascii' codec can't decode byte 0xe9 in position 13: ordinal not in range(128)
4,296
conda/conda
conda__conda-6867
31a52a8a8771aafb0582822dd2e00b3274e1ffba
diff --git a/conda/gateways/disk/update.py b/conda/gateways/disk/update.py --- a/conda/gateways/disk/update.py +++ b/conda/gateways/disk/update.py @@ -7,10 +7,11 @@ from os import rename as os_rename, utime from os.path import dirname, isdir import re +from shutil import move from . import exp_backoff_fn, mkdir_p, mkdir_p_sudo_safe from .delete import rm_rf -from .link import islink, lexists +from .link import lexists from ...common.compat import on_win from ...common.path import expand from ...exceptions import NotWritableError @@ -44,15 +45,6 @@ def update_file_in_place_as_binary(file_full_path, callback): fh.close() -def _copy_then_remove(source_path, destination_path): - from .create import copy, create_hard_link_or_copy - if islink(source_path): - copy(source_path, destination_path) - else: - create_hard_link_or_copy(source_path, destination_path) - rm_rf(source_path) - - def rename(source_path, destination_path, force=False): if lexists(destination_path) and force: rm_rf(destination_path) @@ -62,10 +54,13 @@ def rename(source_path, destination_path, force=False): os_rename(source_path, destination_path) except EnvironmentError as e: if e.errno in (EINVAL, EXDEV): - # see https://github.com/conda/conda/issues/6711 - log.trace("Could not rename do to errno [%s]. Falling back to copy/remove.", - e.errno) - _copy_then_remove(source_path, destination_path) + # https://github.com/conda/conda/issues/6811 + # https://github.com/conda/conda/issues/6711 + log.trace("Could not rename %s => %s due to errno [%s]. Falling back" + " to copy/unlink", source_path, destination_path, e.errno) + # https://github.com/moby/moby/issues/25409#issuecomment-238537855 + # shutil.move() falls back to copy+unlink + move(source_path, destination_path) else: raise else:
IsADirectoryError **I'm submitting a...** - [X] bug report ## Current Behavior When installing certain packages it will break with IsADirectoryError ``` <<HTTPS 200 OK < Accept-Ranges: bytes < Content-Disposition: attachment; filename="six-1.11.0-py35h423b573_1.tar.bz2"; filename*=UTF-8''six-1.11.0-py35h423b573_1.tar.bz2 < Content-Type: application/x-tar < Date: Mon, 29 Jan 2018 11:43:42 GMT < ETag: "576d38abfcb7ceb3b323b802b1fec665" < Last-Modified: Sat, 21 Oct 2017 18:55:29 GMT < Server: AmazonS3 < x-amz-id-2: IGiKj9lk7QKmYRYyZCVaUEM4Ax19T3slC8e1+ZVXTR3gT5Hg9UNd/8CxoS9yDWZt5PdfORpg9mo= < x-amz-request-id: 8A668C11FD79A607 < Content-Length: 21870 < Elapsed: 00:00.106578 TRACE conda.core.path_actions:execute(1238): extracting /opt/conda/pkgs/six-1.11.0-py35h423b573_1.tar.bz2 => /opt/conda/pkgs/six-1.11.0-py35h423b573_1 TRACE conda.gateways.disk.update:rename(60): renaming /opt/conda/pkgs/six-1.11.0-py35h423b573_1 => /opt/conda/pkgs/six-1.11.0-py35h423b573_1.c~ DEBUG conda.gateways.disk.create:extract_tarball(132): extracting /opt/conda/pkgs/six-1.11.0-py35h423b573_1.tar.bz2 to /opt/conda/pkgs/six-1.11.0-py35h423b573_1 TRACE conda.gateways.disk.create:write_as_json_to_file(61): writing json to file /opt/conda/pkgs/six-1.11.0-py35h423b573_1/info/repodata_record.json TRACE conda.gateways.disk.create:write_as_json_to_file(61): writing json to file /opt/conda/pkgs/six-1.11.0-py35h423b573_1/info/repodata_record.json TRACE conda.gateways.disk.delete:rm_rf(29): rm_rf /opt/conda/pkgs/six-1.11.0-py35h423b573_1.tar.bz2.c~ TRACE conda.gateways.disk.delete:rm_rf(29): rm_rf /opt/conda/pkgs/six-1.11.0-py35h423b573_1.c~ DEBUG conda.common.signals:signal_handler(56): de-registering handler for Signals.SIGIOT DEBUG conda.common.signals:signal_handler(56): de-registering handler for Signals.SIGINT DEBUG conda.common.signals:signal_handler(56): de-registering handler for Signals.SIGTERM DEBUG conda.common.signals:signal_handler(56): de-registering handler for Signals.SIGQUIT Traceback (most recent call last): File "/opt/conda/lib/python3.5/site-packages/conda/exceptions.py", line 772, in __call__ return func(*args, **kwargs) File "/opt/conda/lib/python3.5/site-packages/conda_env/cli/main.py", line 74, in do_call exit_code = getattr(module, func_name)(args, parser) File "/opt/conda/lib/python3.5/site-packages/conda_env/cli/main_update.py", line 107, in execute installer.install(prefix, specs, args, env) File "/opt/conda/lib/python3.5/site-packages/conda_env/installers/conda.py", line 38, in install pfe.execute() File "/opt/conda/lib/python3.5/site-packages/conda/core/package_cache.py", line 590, in execute raise CondaMultiError(exceptions) conda.CondaMultiError: [Errno 21] Is a directory: '/opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0' [Errno 21] Is a directory: '/opt/conda/pkgs/libedit-3.1-heed3624_0' ``` ### Steps to Reproduce - run the continuumio/miniconda3 docker image - run `conda update conda` - create a file test.yaml with the following contents: ``` name: test dependencies: - lxml==3.7.2 - anaconda::redis-py==2.10.5 - beautifulsoup4==4.5.3 - feedparser==5.2.1 - html5lib==0.999 - w3lib==1.16.0 - clinicalgraphics::libgcrypt11=1 - python=3.5 ``` - run `conda env update -f test.yaml` ## Expected Behavior It should install without problems ## Environment Information ``` active environment : None user config file : /root/.condarc populated config files : conda version : 4.4.8 conda-build version : not installed python version : 3.6.2.final.0 base environment : /opt/conda (writable) channel URLs : https://repo.continuum.io/pkgs/main/linux-64 https://repo.continuum.io/pkgs/main/noarch https://repo.continuum.io/pkgs/free/linux-64 https://repo.continuum.io/pkgs/free/noarch https://repo.continuum.io/pkgs/r/linux-64 https://repo.continuum.io/pkgs/r/noarch https://repo.continuum.io/pkgs/pro/linux-64 https://repo.continuum.io/pkgs/pro/noarch package cache : /opt/conda/pkgs /root/.conda/pkgs envs directories : /opt/conda/envs /root/.conda/envs platform : linux-64 user-agent : conda/4.4.8 requests/2.18.4 CPython/3.6.2 Linux/4.4.0-112-generic debian/8 glibc/2.19 UID:GID : 0:0 netrc file : None offline mode : False ``` </p></details> <details open><summary><code>`conda config --show-sources`</code></summary><p> ``` ``` </p></details> <details><summary><code>`conda list --show-channel-urls`</code></summary><p> <!-- between the ticks below, paste the output of 'conda list --show-channel-urls' --> ``` # packages in environment at /opt/conda: # # Name Version Build Channel asn1crypto 0.22.0 py36h265ca7c_1 defaults ca-certificates 2017.08.26 h1d4fec5_0 defaults certifi 2017.7.27.1 py36h8b7b77e_0 defaults cffi 1.10.0 py36had8d393_1 defaults chardet 3.0.4 py36h0f667ec_1 defaults conda 4.4.8 py36_0 defaults conda-env 2.6.0 h36134e3_1 defaults cryptography 2.0.3 py36ha225213_1 defaults idna 2.6 py36h82fb2a8_1 defaults libedit 3.1 heed3624_0 defaults libffi 3.2.1 h4deb6c0_3 defaults libgcc-ng 7.2.0 hcbc56d2_1 defaults libstdcxx-ng 7.2.0 h24385c6_1 defaults ncurses 6.0 h06874d7_1 defaults openssl 1.0.2l h9d1a558_3 defaults pip 9.0.1 py36h30f8307_2 defaults pycosat 0.6.3 py36h0a5515d_0 defaults pycparser 2.18 py36hf9f622e_1 defaults pyopenssl 17.2.0 py36h5cc804b_0 defaults pysocks 1.6.7 py36hd97a5b1_1 defaults python 3.6.2 h02fb82a_12 defaults readline 7.0 hac23ff0_3 defaults requests 2.18.4 py36he2e5f8d_1 defaults ruamel_yaml 0.11.14 py36ha2fb22d_2 defaults setuptools 36.5.0 py36he42e2e1_0 defaults six 1.10.0 py36hcac75e4_1 defaults sqlite 3.20.1 h6d8b0f3_1 defaults tk 8.6.7 h5979e9b_1 defaults urllib3 1.22 py36hbe7ace6_0 defaults wheel 0.29.0 py36he7f4e38_1 defaults xz 5.2.3 h2bcbf08_1 defaults yaml 0.1.7 h96e3832_1 defaults zlib 1.2.11 hfbfcf68_1 defaults ``` </p></details>
I am also seeing this with conda-build. For context, with conda-build, it happens very frequently with the cross-platform build. So, on our linux-64 docker image, when we run with CONDA_SUBDIR=linux-32 and ``linux32 conda-build``, we can get this error right away. I'm less sure about how to reproduce with the OP's report. I know that @nehaljwani has tried, and it's tough to get to happen reliably. I think that's because it depends on the state of the package cache. You must have the correct initial state (something already existing there) to get this problem to happen. This seems to be the same problem in my case, if run the command again, it seems to work. We have frozen conda to an old version as this was crashing our CI pipelines. Do you think this is related to https://github.com/conda/conda/issues/6631? That issue isn't new; it was present in conda 4.3, and probably 4.2. But something in conda 4.4 could be helping to exacerbate it. It seems to be unrelated, also this bug doesn't exist in 4.4.7. To reproduce this, exact steps: ``` # Spawn the container sudo docker run --rm -it continuumio/miniconda3:4.3.27 /bin/bash # Inside the container cat <<EOF > ~/test.yaml name: test dependencies: - anaconda::ca-certificates==2017.08.26 EOF # Update conda to 4.4.9 conda install -y conda=4.4.9 # Trigger the problem conda env update -f ~/test.yaml Solving environment: done Downloading and Extracting Packages ca-certificates 2017.08.26: ############################5 | 75% IsADirectoryError(21, 'Is a directory') ``` However, if I switch my docker daemon configuration from (overlay2/aufs): ``` { "storage-driver": "overlay2" } ``` To: ``` { "storage-driver": "devicemapper" } ``` Then the problem goes away. Logs: [log_overlay2.txt](https://github.com/conda/conda/files/1716159/log_overlay2.txt) [log_devicemapper.txt](https://github.com/conda/conda/files/1716160/log_devicemapper.txt) But, that doesn't mean that this is not a bug in Conda. Because, if I switch the update from conda 4.4.9 to 4.4.7 (as @pvanderlinden has pointed out), the problem doesn't occur, even in the case of overlay2. The difference in verbose logs is: 4.4.7: ``` TRACE conda.core.path_actions:execute(1238): extracting /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.tar.bz2 => /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0 TRACE conda.gateways.disk.update:rename(50): renaming /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0 => /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.c~ DEBUG conda.core.path_actions:execute(1251): Invalid cross-device link on rename /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0 => /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.c~ TRACE conda.gateways.disk.delete:rm_rf(29): rm_rf /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0 DEBUG conda.gateways.disk.create:extract_tarball(131): extracting /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.tar.bz2 to /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0 TRACE conda.gateways.disk.create:write_as_json_to_file(60): writing json to file /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0/info/repodata_record.json TRACE conda.gateways.disk.create:write_as_json_to_file(60): writing json to file /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0/info/repodata_record.json TRACE conda.gateways.disk.delete:rm_rf(29): rm_rf /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.tar.bz2.c~ TRACE conda.gateways.disk.delete:rm_rf(29): rm_rf /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.c~ TRACE conda.gateways.disk.delete:rm_rf(49): rm_rf failed. Not a link, file, or directory: /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.c~ ``` For conda > 4.4.7: ``` TRACE conda.core.path_actions:execute(1238): extracting /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.tar.bz2 => /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0 TRACE conda.gateways.disk.update:rename(60): renaming /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0 => /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.c~ TRACE conda.gateways.disk.update:rename(67): Could not rename do to errno [18]. Falling back to copy/remove. TRACE conda.gateways.disk.create:create_hard_link_or_copy(201): creating hard link /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0 => /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.c~ INFO conda.gateways.disk.create:create_hard_link_or_copy(204): hard link failed, so copying /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0 => /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.c~ TRACE conda.gateways.disk.create:_do_copy(251): copying /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0 => /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.c~ TRACE conda.gateways.disk.delete:rm_rf(29): rm_rf /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0 TRACE conda.core.path_actions:reverse(1206): moving /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.tar.bz2.c~ => /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.tar.bz2 TRACE conda.gateways.disk.delete:rm_rf(29): rm_rf /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.tar.bz2 TRACE conda.gateways.disk.update:rename(60): renaming /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.tar.bz2.c~ => /opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0.tar.bz2 DEBUG conda.core.package_cache:execute(586): IsADirectoryError(21, 'Is a directory') NoneType: None ``` If I revert this commit: 0b57e1ea4ff49b82df3bcd02408368cf182a3aba , then the problem mentioned in this issue goes away, but of course, that's not the right fix. Okay, seems like the [function in that commit ](https://github.com/conda/conda/blob/4.4.9/conda/gateways/disk/update.py#L47-L53)was written keeping files in mind, not directories! So, the IsADirectoryError exception is coming from https://github.com/conda/conda/blob/4.4.9/conda/gateways/disk/create.py#L259
2018-02-12T13:46:42Z
[]
[]
Traceback (most recent call last): File "/opt/conda/lib/python3.5/site-packages/conda/exceptions.py", line 772, in __call__ return func(*args, **kwargs) File "/opt/conda/lib/python3.5/site-packages/conda_env/cli/main.py", line 74, in do_call exit_code = getattr(module, func_name)(args, parser) File "/opt/conda/lib/python3.5/site-packages/conda_env/cli/main_update.py", line 107, in execute installer.install(prefix, specs, args, env) File "/opt/conda/lib/python3.5/site-packages/conda_env/installers/conda.py", line 38, in install pfe.execute() File "/opt/conda/lib/python3.5/site-packages/conda/core/package_cache.py", line 590, in execute raise CondaMultiError(exceptions) conda.CondaMultiError: [Errno 21] Is a directory: '/opt/conda/pkgs/ca-certificates-2017.08.26-h1d4fec5_0'
4,301
conda/conda
conda__conda-6909
70ef29117690193a92181f2bd4e7e0ab845158ba
diff --git a/conda/__init__.py b/conda/__init__.py --- a/conda/__init__.py +++ b/conda/__init__.py @@ -8,7 +8,7 @@ import sys from ._vendor.auxlib.packaging import get_version -from .common.compat import iteritems, text_type +from .common.compat import text_type __all__ = ( "__name__", "__version__", "__author__", "__email__", "__license__", "__summary__", "__url__", @@ -46,20 +46,20 @@ def __repr__(self): def __str__(self): try: return text_type(self.message % self._kwargs) - except TypeError: - # TypeError: not enough arguments for format string + except: debug_message = "\n".join(( "class: " + self.__class__.__name__, "message:", self.message, "kwargs:", text_type(self._kwargs), + "", )) sys.stderr.write(debug_message) raise def dump_map(self): - result = dict((k, v) for k, v in iteritems(vars(self)) if not k.startswith('_')) + result = dict((k, v) for k, v in vars(self).items() if not k.startswith('_')) result.update(exception_type=text_type(type(self)), exception_name=self.__class__.__name__, message=text_type(self), diff --git a/conda/activate.py b/conda/activate.py --- a/conda/activate.py +++ b/conda/activate.py @@ -129,7 +129,10 @@ def __init__(self, shell, arguments=None): def _finalize(self, commands, ext): commands = concatv(commands, ('',)) # add terminating newline if ext is None: - return self.command_join.join(commands) + output = self.command_join.join(commands) + if PY2: + return ensure_binary(output) + return output elif ext: with NamedTemporaryFile('w+b', suffix=ext, delete=False) as tf: # the default mode is 'w+b', and universal new lines don't work in that mode diff --git a/conda/cli/common.py b/conda/cli/common.py --- a/conda/cli/common.py +++ b/conda/cli/common.py @@ -9,6 +9,7 @@ from ..base.context import context from ..common.io import swallow_broken_pipe from ..common.path import paths_equal +from ..common.serialize import json_dump from ..models.match_spec import MatchSpec @@ -158,10 +159,7 @@ def disp_features(features): @swallow_broken_pipe def stdout_json(d): - import json - from .._vendor.auxlib.entity import EntityEncoder - json.dump(d, sys.stdout, indent=2, sort_keys=True, cls=EntityEncoder) - sys.stdout.write('\n') + print(json_dump(d)) def stdout_json_success(success=True, **kwargs): diff --git a/conda/cli/main.py b/conda/cli/main.py --- a/conda/cli/main.py +++ b/conda/cli/main.py @@ -80,28 +80,15 @@ def _main(*args): return exit_code -def _ensure_text_type(value): - # copying here from conda/common/compat.py to avoid the import - try: - return value.decode('utf-8') - except AttributeError: - # AttributeError: '<>' object has no attribute 'decode' - # In this case assume already text_type and do nothing - return value - except UnicodeDecodeError: - try: - from requests.packages.chardet import detect - except ImportError: # pragma: no cover - from pip._vendor.requests.packages.chardet import detect - encoding = detect(value).get('encoding') or 'utf-8' - return value.decode(encoding) - - def main(*args): + # conda.common.compat contains only stdlib imports + from ..common.compat import ensure_text_type # , init_std_stream_encoding + + # init_std_stream_encoding() if not args: args = sys.argv - args = tuple(_ensure_text_type(s) for s in args) + args = tuple(ensure_text_type(s) for s in args) if len(args) > 1: try: diff --git a/conda/common/compat.py b/conda/common/compat.py --- a/conda/common/compat.py +++ b/conda/common/compat.py @@ -3,6 +3,7 @@ # What is compat, and what isn't? # If a piece of code is "general" and used in multiple modules, it goes here. # If it's only used in one module, keep it in that module, preferably near the top. +# This module should contain ONLY stdlib imports. from __future__ import absolute_import, division, print_function, unicode_literals from itertools import chain @@ -49,6 +50,7 @@ from itertools import zip_longest if sys.version_info[1] >= 5: from json import JSONDecodeError + JSONDecodeError = JSONDecodeError else: JSONDecodeError = ValueError elif PY2: # pragma: py3 no cover @@ -145,6 +147,29 @@ def _clone_with_metaclass(Class): primitive_types = tuple(chain(string_types, integer_types, (float, complex, bool, NoneType))) +def _init_stream_encoding(stream): + # PY2 compat: Initialize encoding for an IO stream. + # Python 2 sets the encoding of stdout/stderr to None if not run in a + # terminal context and thus falls back to ASCII. + if getattr(stream, "encoding", True): + # avoid the imports below if they are not necessary + return stream + from codecs import getwriter + from locale import getpreferredencoding + encoding = getpreferredencoding() + try: + encoder = getwriter(encoding) + except LookupError: + encoder = getwriter("UTF-8") + base_stream = getattr(stream, "buffer", stream) + return encoder(base_stream) + + +def init_std_stream_encoding(): + sys.stdout = _init_stream_encoding(sys.stdout) + sys.stderr = _init_stream_encoding(sys.stderr) + + def ensure_binary(value): try: return value.encode('utf-8') diff --git a/conda/common/serialize.py b/conda/common/serialize.py --- a/conda/common/serialize.py +++ b/conda/common/serialize.py @@ -4,7 +4,7 @@ import json from logging import getLogger -from .compat import PY2, odict +from .compat import PY2, odict, ensure_text_type from .._vendor.auxlib.decorators import memoize from .._vendor.auxlib.entity import EntityEncoder @@ -86,5 +86,5 @@ def json_load(string): def json_dump(object): - return json.dumps(object, indent=2, sort_keys=True, - separators=(',', ': '), cls=EntityEncoder) + return ensure_text_type(json.dumps(object, indent=2, sort_keys=True, + separators=(',', ': '), cls=EntityEncoder)) diff --git a/conda/core/repodata.py b/conda/core/repodata.py --- a/conda/core/repodata.py +++ b/conda/core/repodata.py @@ -16,12 +16,12 @@ from time import time import warnings -from .. import CondaError, iteritems +from .. import CondaError from .._vendor.auxlib.ish import dals from .._vendor.auxlib.logz import stringify from ..base.constants import CONDA_HOMEPAGE_URL from ..base.context import context -from ..common.compat import (ensure_binary, ensure_text_type, ensure_unicode, text_type, +from ..common.compat import (ensure_binary, ensure_text_type, ensure_unicode, iteritems, text_type, with_metaclass) from ..common.io import ThreadLimitedThreadPoolExecutor, as_completed from ..common.url import join_url, maybe_unquote
activate script fails with conda=4.4.10 and zsh=5.3.1 with cryptic logging error <!-- Hi! This is an issue tracker for conda -- the package manager. File feature requests for conda here, as well as bug reports about something conda has messed up. If your issue is a bug report or feature request for: * a specific conda package from Anaconda ('defaults' channel): ==> file at https://github.com/ContinuumIO/anaconda-issues * a specific conda package from conda-forge: ==> file at the corresponding feedstock under https://github.com/conda-forge * repo.continuum.io access and service: ==> file at https://github.com/ContinuumIO/anaconda-issues * anaconda.org access and service: ==> file at https://anaconda.org/contact/report * commands under 'conda build': ==> file at https://github.com/conda/conda-build * commands under 'conda env': ==> please file it here! * all other conda commands that start with 'conda': ==> please file it here! --> **I'm submitting a...** - [x] bug report - [ ] feature request ## Current Behavior <!-- What actually happens? If you want to include console output, please use "Steps to Reproduce" below. --> Using ZSH 5.3.1 the activate script of `conda=4.4.10` does not work and fails with some cryptic Python logging error. It doesn't even detect an error. With 4.4.0 it at least fails at unicode-decoding `PS1` and offers to submit a report. The script works in Bash. ### Steps to Reproduce <!-- If the current behavior is a bug, please provide specific, minimal steps to independently reproduce. Include the exact conda commands that reproduce the issue and their output between the ticks below. --> ``` $ source ~/anaconda/bin/activate Traceback (most recent call last): File "/home/ondrej/anaconda/lib/python2.7/site-packages/conda/gateways/logging.py", line 64, in emit msg = self.format(record) File "/home/ondrej/anaconda/lib/python2.7/logging/__init__.py", line 734, in format return fmt.format(record) File "/home/ondrej/anaconda/lib/python2.7/logging/__init__.py", line 465, in format record.message = record.getMessage() File "/home/ondrej/anaconda/lib/python2.7/logging/__init__.py", line 329, in getMessage msg = msg % self.args File "/home/ondrej/anaconda/lib/python2.7/site-packages/conda/__init__.py", line 43, in __repr__ return '%s: %s' % (self.__class__.__name__, text_type(self)) File "/home/ondrej/anaconda/lib/python2.7/site-packages/conda/__init__.py", line 47, in __str__ return text_type(self.message % self._kwargs) ValueError: unsupported format character '{' (0x7b) at index 458 Logged from file exceptions.py, line 770 $ echo $ZSH_VERSION 5.3.1 ``` ## Expected Behavior <!-- What do you think should happen? --> The source script should work as it did with `conda<=4.3*`. My current workaround is to prepend `~/anaconda/bin` to `$PATH` manually. ## Environment Information <details open><summary><code>`conda info`</code></summary><p> <!-- between the ticks below, paste the output of 'conda info' --> ``` active environment : base active env location : /home/ondrej/anaconda shell level : 1 user config file : /home/ondrej/.condarc populated config files : /home/ondrej/.condarc conda version : 4.4.10 conda-build version : 1.19.0 python version : 2.7.13.final.0 base environment : /home/ondrej/anaconda (writable) channel URLs : https://conda.anaconda.org/anaconda-fusion/linux-64 https://conda.anaconda.org/anaconda-fusion/noarch https://repo.continuum.io/pkgs/main/linux-64 https://repo.continuum.io/pkgs/main/noarch https://repo.continuum.io/pkgs/free/linux-64 https://repo.continuum.io/pkgs/free/noarch https://repo.continuum.io/pkgs/r/linux-64 https://repo.continuum.io/pkgs/r/noarch https://repo.continuum.io/pkgs/pro/linux-64 https://repo.continuum.io/pkgs/pro/noarch package cache : /home/ondrej/anaconda/pkgs /home/ondrej/.conda/pkgs envs directories : /home/ondrej/anaconda/envs /home/ondrej/.conda/envs platform : linux-64 user-agent : conda/4.4.10 requests/2.14.2 CPython/2.7.13 Linux/4.9.0-5-amd64 debian/9 glibc/2.24 UID:GID : 1000:1000 netrc file : None offline mode : False ``` </p></details> <details open><summary><code>`conda config --show-sources`</code></summary><p> <!-- between the ticks below, paste the output of 'conda config --show-sources' --> ``` ==> /home/ondrej/.condarc <== ssl_verify: True channels: - anaconda-fusion - defaults ``` </p></details> <details><summary><code>`conda list --show-channel-urls`</code></summary><p> <!-- between the ticks below, paste the output of 'conda list --show-channel-urls' --> ``` # packages in environment at /home/ondrej/anaconda: # # Name Version Build Channel _license 1.1 py27_1 defaults _nb_ext_conf 0.3.0 py27_0 defaults abstract-rendering 0.5.1 np111py27_0 defaults accelerate 2.0.2 np19py27_p0 defaults accelerate_cudalib 2.0 0 defaults alabaster 0.7.9 py27_0 defaults anaconda custom py27h4a00acb_0 defaults anaconda-clean 1.0.0 py27_0 defaults anaconda-client 1.6.0 py27_0 defaults anaconda-navigator 1.5.0 py27_0 defaults anaconda-project 0.4.1 py27_0 defaults apipkg 1.4 <pip> argcomplete 1.0.0 py27_1 defaults astroid 1.4.9 py27_0 defaults astropy 1.3 np111py27_0 defaults babel 2.3.4 py27_0 defaults backports 1.0 py27_0 defaults backports_abc 0.5 py27_0 defaults bcolz 0.11.3 py27_0 defaults beautiful-soup 4.3.2 py27_0 <unknown> beautifulsoup4 4.5.3 py27_0 defaults binstar 0.12 2 defaults bitarray 0.8.1 py27_0 <unknown> blaze 0.10.1 py27_0 defaults blaze-core 0.9.0 py27_0 defaults blinker 1.4 py_0 conda-forge bokeh 0.12.4 py27_0 defaults boto 2.45.0 py27_0 defaults bottleneck 1.2.0 np111py27_0 defaults ca-certificates 2018.1.18 0 conda-forge cairo 1.14.8 0 defaults cdecimal 2.3 py27_2 defaults certifi 2018.1.18 py27_0 defaults cffi 1.9.1 py27_0 defaults chardet 2.3.0 <pip> chardet 2.3.0 py27_0 defaults chest 0.2.3 py27_0 defaults click 6.7 py27_0 defaults cloudpickle 0.2.2 py27_0 defaults clyent 1.2.2 py27_0 defaults colorama 0.3.7 py27_0 defaults conda 4.4.10 py27_0 defaults conda-build 1.19.0 py27_0 defaults conda-env 2.6.0 h36134e3_1 defaults conda-manager 0.3.1 py27_0 defaults configobj 5.0.6 py27_0 <unknown> configparser 3.5.0 py27_0 defaults configparser 3.3.0.post2 <pip> contextlib2 0.5.4 py27_0 defaults coverage 4.0.3 <pip> cryptography 1.7.1 py27_0 defaults cudatoolkit 7.0 1 defaults curl 7.52.1 0 defaults cycler 0.10.0 py27_0 defaults cython 0.25.2 py27_0 defaults cytoolz 0.8.2 py27_0 defaults dask 0.13.0 py27_0 defaults datashape 0.5.4 py27_0 defaults dbus 1.10.10 0 defaults decorator 4.0.11 py27_0 defaults dill 0.2.5 py27_0 defaults docutils 0.13.1 py27_0 defaults doit 0.29.0 <pip> dynd-python 0.7.2 py27_0 defaults ecdsa 0.13 <pip> entrypoints 0.2.2 py27_0 defaults enum34 1.1.6 py27_0 defaults et_xmlfile 1.0.1 py27_0 defaults execnet 1.4.1 <pip> expat 2.1.0 0 defaults fastcache 1.0.2 py27_1 defaults fftw 3.3.4 0 https://conda.binstar.org/jakirkham filelock 2.0.6 py27_0 defaults flask 0.12 py27_0 defaults flask-cors 3.0.2 py27_0 defaults flask-login 0.4.0 py27_0 defaults flask-sqlalchemy 2.3.2 py_0 conda-forge fontconfig 2.12.1 2 defaults freetype 2.5.5 2 defaults funcsigs 1.0.2 py27_0 defaults functools32 3.2.3.2 py27_0 defaults futures 3.0.5 py27_0 defaults get_terminal_size 1.0.0 py27_0 defaults gevent 1.2.1 py27_0 defaults glib 2.50.2 1 defaults greenlet 0.4.11 py27_0 defaults grin 1.2.1 py27_3 defaults gst-plugins-base 1.8.0 0 defaults gstreamer 1.8.0 0 defaults h5py 2.6.0 np111py27_2 defaults harfbuzz 0.9.39 2 defaults hdf4 4.2.11 0 defaults hdf5 1.8.17 1 defaults heapdict 1.0.0 py27_1 defaults hmmlearn 0.2.0 <pip> hmmlearn 0.3.0b np111py27_0 omnia holoviews 1.8.3 py27_0 defaults icalendar 3.10 <pip> icu 54.1 0 defaults idna 2.2 py27_0 defaults imagesize 0.7.1 py27_0 defaults ipaddress 1.0.18 py27_0 defaults ipdb 0.8.1 <pip> ipykernel 4.5.2 py27_0 defaults ipython 5.1.0 py27_0 defaults ipython-notebook 4.0.4 py27_0 defaults ipython-qtconsole 4.0.1 py27_0 defaults ipython_genutils 0.1.0 py27_0 defaults ipywidgets 5.2.2 py27_1 defaults isort 4.2.5 py27_0 defaults itsdangerous 0.24 py27_0 <unknown> javaobj-py3 0.2.3 <pip> jbig 2.1 0 defaults jdcal 1.3 py27_0 defaults jedi 0.9.0 py27_1 defaults jinja2 2.9.4 py27_0 defaults joblib 0.8.4 py27_0 defaults jpeg 9b 0 defaults jsonschema 2.5.1 py27_0 defaults jupyter 1.0.0 py27_3 defaults jupyter_client 4.4.0 py27_0 defaults jupyter_console 5.0.0 py27_0 defaults jupyter_core 4.2.1 py27_0 defaults krb5 1.13.2 0 defaults lazy-object-proxy 1.2.2 py27_0 defaults ldap3 2.3 py27hcb53f9d_0 defaults libdynd 0.7.2 0 defaults libffi 3.2.1 1 defaults libgcc 4.8.5 2 defaults libgcc-ng 7.2.0 h7cc24e2_2 defaults libgfortran 3.0.0 1 defaults libiconv 1.14 0 defaults libpng 1.6.27 0 defaults libsodium 1.0.10 0 defaults libtiff 4.0.6 3 defaults libxcb 1.12 1 defaults libxml2 2.9.4 0 defaults libxslt 1.1.29 0 defaults llvmlite 0.15.0 py27_0 defaults locket 0.2.0 py27_1 defaults lockfile 0.12.2 <pip> logilab-common 1.0.2 py27_0 defaults luigi 2.1.1 <pip> lxml 3.7.2 py27_0 defaults markupsafe 0.23 py27_2 defaults matplotlib 2.0.0 np111py27_0 defaults mistune 0.7.3 py27_0 defaults mkl 2017.0.1 0 defaults mkl-rt 11.1 p0 defaults mkl-service 1.1.2 py27_3 defaults mklfft 2.1 np19py27_p0 defaults mock 2.0.0 py27_0 defaults mpmath 0.19 py27_1 defaults multipledispatch 0.4.9 py27_0 defaults mysql 5.5.24 0 defaults mysql-python 1.2.5 py27_0 defaults nb_anacondacloud 1.2.0 py27_0 defaults nb_conda 2.0.0 py27_0 defaults nb_conda_kernels 2.0.0 py27_0 defaults nbconvert 4.2.0 py27_0 defaults nbformat 4.2.0 py27_0 defaults nbpresent 3.0.2 py27_0 defaults networkx 1.11 py27_0 defaults nltk 3.2.2 py27_0 defaults nose 1.3.7 py27_1 defaults notebook 4.3.1 py27_0 defaults numba 0.30.1 np111py27_0 defaults numbapro 0.23.1 py27_p0 defaults numbapro_cudalib 0.2 0 defaults numexpr 2.6.1 np111py27_2 defaults numpy 1.11.3 py27_0 defaults numpydoc 0.6.0 py27_0 defaults odo 0.5.0 py27_1 defaults openldap 2.4.36 1 defaults openpyxl 2.4.1 py27_0 defaults openssl 1.0.2n hb7f436b_0 defaults pandas 0.19.2 np111py27_1 defaults param 1.5.1 py27_0 defaults paramiko 1.16.0 <pip> partd 0.3.7 py27_0 defaults patchelf 0.9 0 defaults path.py 10.0 py27_0 defaults pathlib2 2.2.0 py27_0 defaults patsy 0.4.1 py27_0 defaults pbr 1.3.0 py27_0 defaults pcre 8.39 1 defaults pep8 1.7.0 py27_0 defaults pexpect 4.2.1 py27_0 defaults pickleshare 0.7.4 py27_0 defaults pillow 4.0.0 py27_0 defaults pip 9.0.1 py27_1 defaults pixman 0.34.0 0 defaults pkginfo 1.3.2 py27_0 defaults ply 3.9 py27_0 defaults progressbar 2.3 py27_0 defaults progressbar2 3.10.1 <pip> prompt_toolkit 1.0.9 py27_0 defaults psutil 5.0.1 py27_0 defaults ptyprocess 0.5.1 py27_0 defaults py 1.4.32 py27_0 defaults pyasn1 0.1.9 py27_0 defaults pycairo 1.10.0 py27_0 defaults pycosat 0.6.3 py27ha4109ae_0 defaults pycparser 2.17 py27_0 defaults pycrypto 2.6.1 py27_4 defaults pycurl 7.43.0 py27_2 defaults pyfftw 0.9.2 np19py27_1 https://conda.binstar.org/jakirkham pyflakes 1.5.0 py27_0 defaults pygments 2.1.3 py27_0 defaults pyhdf 0.9.0 np111py27_1 conda-forge pyinotify 0.9.6 <pip> pylint 1.6.4 py27_1 defaults pyopenssl 16.2.0 py27_0 defaults pyparsing 2.1.4 py27_0 defaults pyqt 5.6.0 py27_2 defaults pyserial 2.7 py27_0 defaults pytables 3.3.0 np111py27_0 defaults pytest 3.0.5 py27_0 defaults pytest-cov 2.2.1 <pip> python 2.7.13 0 defaults python-daemon 2.1.1 <pip> python-dateutil 2.6.0 py27_0 defaults python-ldap 2.4.13 py27_0 defaults python-utils 2.0.0 <pip> pytz 2016.10 py27_0 defaults pyyaml 3.12 py27_0 defaults pyzmq 16.0.2 py27_0 defaults qt 5.6.2 3 defaults qtawesome 0.4.3 py27_0 defaults qtconsole 4.2.1 py27_1 defaults qtpy 1.2.1 py27_0 defaults readline 6.2 2 defaults redis 3.2.0 0 defaults redis-py 2.10.5 py27_0 defaults requests 2.14.2 py27_0 defaults rk 0.3b1 <pip> rope 0.9.4 py27_1 <unknown> ruamel_yaml 0.11.14 py27_1 defaults ruffus 2.6.3 <pip> runipy 0.1.3 py27_0 <unknown> scandir 1.4 py27_0 defaults scikit-image 0.12.3 np111py27_1 defaults scikit-learn 0.18.1 np111py27_1 defaults scipy 0.18.1 np111py27_1 defaults seaborn 0.7.1 py27_0 defaults setuptools 38.4.0 py27_0 defaults simplegeneric 0.8.1 py27_1 defaults singledispatch 3.4.0.3 py27_0 defaults sip 4.18 py27_0 defaults six 1.10.0 py27_0 defaults snowballstemmer 1.2.1 py27_0 defaults sockjs-tornado 1.0.3 py27_0 defaults spectrum 0.1 <pip> sphinx 1.5.1 py27_0 defaults sphinx_rtd_theme 0.1.9 py27_0 defaults spyder 3.1.2 py27_0 defaults sqlalchemy 1.1.5 py27_0 defaults sqlite 3.13.0 0 defaults sqlsoup 0.9.1 <pip> ssl_match_hostname 3.4.0.2 py27_1 defaults statsmodels 0.6.1 np111py27_1 defaults subprocess32 3.2.7 py27_0 defaults sympy 1.0 py27_0 defaults system 5.8 2 <unknown> terminado 0.6 py27_0 defaults theano 0.5.0 np15py27_0 defaults tk 8.5.18 0 defaults toolz 0.8.2 py27_0 defaults tornado 4.4.2 py27_0 defaults traitlets 4.3.1 py27_0 defaults ujson 1.35 py27_0 defaults unicodecsv 0.14.1 py27_0 defaults util-linux 2.21 0 <unknown> wcwidth 0.1.7 py27_0 defaults werkzeug 0.11.15 py27_0 defaults wheel 0.29.0 py27_0 defaults widgetsnbextension 1.2.6 py27_0 defaults wrapt 1.10.8 py27_0 defaults xarray 0.9.3 py27_0 defaults xlrd 1.0.0 py27_0 defaults xlsxwriter 0.9.6 py27_0 defaults xlwt 1.2.0 py27_0 defaults xmltodict 0.10.2 <pip> xz 5.2.2 1 defaults yaml 0.1.6 0 defaults zeromq 4.1.5 0 defaults zlib 1.2.8 3 defaults ``` </p></details>
Because python2 is in the base environment, we don't have exception chaining here, and there's really no way to know the underlying root cause of this error. It's probably a unicode encoding/decoding error somewhere, being it's python2 in the base environment. So your suggestion is to give up on this and migrate to python3 in the base environment? Let's try to get to the exception without doing that. What do you get for ``` CONDA_VERBOSE=3 conda shell.posix activate ``` ? https://github.com/conda/conda/issues/6739#issuecomment-362941051 I'll submit a PR
2018-02-21T19:32:02Z
[]
[]
Traceback (most recent call last): File "/home/ondrej/anaconda/lib/python2.7/site-packages/conda/gateways/logging.py", line 64, in emit msg = self.format(record) File "/home/ondrej/anaconda/lib/python2.7/logging/__init__.py", line 734, in format return fmt.format(record) File "/home/ondrej/anaconda/lib/python2.7/logging/__init__.py", line 465, in format record.message = record.getMessage() File "/home/ondrej/anaconda/lib/python2.7/logging/__init__.py", line 329, in getMessage msg = msg % self.args File "/home/ondrej/anaconda/lib/python2.7/site-packages/conda/__init__.py", line 43, in __repr__ return '%s: %s' % (self.__class__.__name__, text_type(self)) File "/home/ondrej/anaconda/lib/python2.7/site-packages/conda/__init__.py", line 47, in __str__ return text_type(self.message % self._kwargs) ValueError: unsupported format character '{' (0x7b) at index 458
4,307
conda/conda
conda__conda-7476
d9ca6008736ff84dabfc484e115eeb5730e96e34
diff --git a/docs/scrape_help.py b/docs/scrape_help.py --- a/docs/scrape_help.py +++ b/docs/scrape_help.py @@ -50,7 +50,8 @@ def run_command(*args, **kwargs): err = b'' out, err = out.decode('utf-8'), err.decode('utf-8') if p.returncode != 0: - print("%r failed with error code %s" % (' '.join(map(quote, args[0])), p.returncode), file=sys.stderr) + print("%r failed with error code %s" % + (' '.join(map(quote, args[0])), p.returncode), file=sys.stderr) elif err: print("%r gave stderr output: %s" % (' '.join(*args), err)) @@ -128,7 +129,7 @@ def get_help(command): m = subcommands_re.match(line) if m: commands.extend(['%s %s' % (command, i) for i in - m.group(1).split(',')]) + m.group(1).split(',')]) break return commands @@ -147,7 +148,7 @@ def man_replacements(): (info['sys_rc_path'], r'\fI\,system .condarc path\/\fP'), (info['root_prefix'], r'root prefix'), - ]) + ]) return r @@ -165,7 +166,7 @@ def generate_man(command): '--version-string', conda_version, '--no-info', 'conda %s' % command, - ]) + ]) retries -= 1 if not manpage: @@ -186,11 +187,11 @@ def generate_html(command): man = Popen(['man', abspath(join(manpath, 'conda-%s.1' % command_file))], stdout=PIPE) htmlpage = check_output([ 'man2html', - '-bare', # Don't use HTML, HEAD, or BODY tags + '-bare', # Don't use HTML, HEAD, or BODY tags 'title', 'conda-%s' % command_file, - '-topm', '0', # No top margin - '-botm', '0', # No bottom margin - ], + '-topm', '0', # No top margin + '-botm', '0', # No bottom margin + ], stdin=man.stdout) with open(join(manpath, 'conda-%s.html' % command_file), 'wb') as f: @@ -233,7 +234,9 @@ def main(): 'metapackage', # 'pipbuild', 'render', - # 'sign', # let's drop this one; I've dropped support for it in 4.3.x; coming back with TUF in the near future + # let's drop this one; I've dropped support for it in 4.3.x + # coming back with TUF in the near future + # 'sign', 'skeleton', 'skeleton cpan', 'skeleton cran', @@ -267,5 +270,6 @@ def gen_command(command): else: write_rst(command, sep='build') + if __name__ == '__main__': sys.exit(main()) diff --git a/docs/source/conf.py b/docs/source/conf.py --- a/docs/source/conf.py +++ b/docs/source/conf.py @@ -17,7 +17,8 @@ # All configuration values have a default; values that are commented out # serve to show the default. -import sys, os +import sys +import os # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the @@ -27,7 +28,7 @@ # -- General configuration ----------------------------------------------------- # If your documentation needs a minimal Sphinx version, state it here. -#needs_sphinx = '1.0' +# needs_sphinx = '1.0' # Add any Sphinx extension module names here, as strings. They can be extensions # coming with Sphinx (named 'sphinx.ext.*') or your custom ones. @@ -38,7 +39,6 @@ 'sphinx.ext.ifconfig', 'sphinx.ext.inheritance_diagram', # 'sphinx.ext.napoleon', - 'numpydoc', ] # Add any paths that contain templates here, relative to this directory. @@ -48,7 +48,7 @@ source_suffix = '.rst' # The encoding of source files. -#source_encoding = 'utf-8-sig' +# source_encoding = 'utf-8-sig' # The master toctree document. master_doc = 'index' @@ -68,31 +68,31 @@ # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. -#language = None +# language = None # There are two options for replacing |today|: either, you set today to some # non-false value, then it is used: -#today = '' +# today = '' # Else, today_fmt is used as the format for a strftime call. -#today_fmt = '%B %d, %Y' +# today_fmt = '%B %d, %Y' # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. exclude_patterns = [] # The reST default role (used for this markup: `text`) to use for all documents. -#default_role = None +# default_role = None # If true, '()' will be appended to :func: etc. cross-reference text. -#add_function_parentheses = True +# add_function_parentheses = True # If true, the current module name will be prepended to all description # unit titles (such as .. function::). -#add_module_names = True +# add_module_names = True # If true, sectionauthor and moduleauthor directives will be shown in the # output. They are ignored by default. -#show_authors = False +# show_authors = False # The name of the Pygments (syntax highlighting) style to use. pygments_style = 'sphinx' @@ -112,36 +112,36 @@ html_context = { 'github_user': 'conda', - 'github_repo': 'conda', + 'github_repo': 'conda-docs', 'github_version': 'master/', 'display_github': True, 'conf_py_path': 'docs/source/', 'source_suffix': '.rst', - } +} # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the # documentation. -#html_theme_options = {} +# html_theme_options = {} # Add any paths that contain custom themes here, relative to this directory. -#html_theme_path = [] +# html_theme_path = [] # The name for this set of Sphinx documents. If None, it defaults to # "<project> v<release> documentation". -#html_title = None +# html_title = None # A shorter title for the navigation bar. Default is the same as html_title. -#html_short_title = None +# html_short_title = None # The name of an image file (relative to this directory) to place at the top # of the sidebar. -#html_logo = None +# html_logo = None # The name of an image file (within the static path) to use as favicon of the # docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32 # pixels large. -#html_favicon = None +# html_favicon = None # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, @@ -150,44 +150,44 @@ # If not '', a 'Last updated on:' timestamp is inserted at every page bottom, # using the given strftime format. -#html_last_updated_fmt = '%b %d, %Y' +# html_last_updated_fmt = '%b %d, %Y' # If true, SmartyPants will be used to convert quotes and dashes to # typographically correct entities. -#html_use_smartypants = True +# html_use_smartypants = True # Custom sidebar templates, maps document names to template names. -#html_sidebars = {} +# html_sidebars = {} # Additional templates that should be rendered to pages, maps page names to # template names. -#html_additional_pages = {} +# html_additional_pages = {} # If false, no module index is generated. -#html_domain_indices = True +# html_domain_indices = True # If false, no index is generated. -#html_use_index = True +# html_use_index = True # If true, the index is split into individual pages for each letter. -#html_split_index = False +# html_split_index = False # If true, links to the reST sources are added to the pages. -#html_show_sourcelink = True +# html_show_sourcelink = True # If true, "Created using Sphinx" is shown in the HTML footer. Default is True. -#html_show_sphinx = True +# html_show_sphinx = True # If true, "(C) Copyright ..." is shown in the HTML footer. Default is True. -#html_show_copyright = True +# html_show_copyright = True # If true, an OpenSearch description file will be output, and all pages will # contain a <link> tag referring to it. The value of this option must be the # base URL from which the finished HTML is served. -#html_use_opensearch = '' +# html_use_opensearch = '' # This is the file name suffix for HTML files (e.g. ".xhtml"). -#html_file_suffix = None +# html_file_suffix = None # Output file base name for HTML help builder. htmlhelp_basename = 'condadoc' @@ -196,42 +196,42 @@ # -- Options for LaTeX output -------------------------------------------------- latex_elements = { -# The paper size ('letterpaper' or 'a4paper'). -#'papersize': 'letterpaper', + # The paper size ('letterpaper' or 'a4paper'). + # 'papersize': 'letterpaper', -# The font size ('10pt', '11pt' or '12pt'). -#'pointsize': '10pt', + # The font size ('10pt', '11pt' or '12pt'). + # 'pointsize': '10pt', -# Additional stuff for the LaTeX preamble. -#'preamble': '', + # Additional stuff for the LaTeX preamble. + # 'preamble': '', } # Grouping the document tree into LaTeX files. List of tuples # (source start file, target name, title, author, documentclass [howto/manual]). latex_documents = [ - ('index', 'conda.tex', u'conda Documentation', - u'Anaconda, Inc.', 'manual'), + ('index', 'conda.tex', u'conda Documentation', + u'Anaconda, Inc.', 'manual'), ] # The name of an image file (relative to this directory) to place at the top of # the title page. -#latex_logo = None +# latex_logo = None # For "manual" documents, if this is true, then toplevel headings are parts, # not chapters. -#latex_use_parts = False +# latex_use_parts = False # If true, show page references after internal links. -#latex_show_pagerefs = False +# latex_show_pagerefs = False # If true, show URL addresses after external links. -#latex_show_urls = False +# latex_show_urls = False # Documents to append as an appendix to all manuals. -#latex_appendices = [] +# latex_appendices = [] # If false, no module index is generated. -#latex_domain_indices = True +# latex_domain_indices = True # -- Options for manual page output -------------------------------------------- @@ -244,7 +244,7 @@ ] # If true, show URL addresses after external links. -#man_show_urls = False +# man_show_urls = False # -- Options for Texinfo output ------------------------------------------------ @@ -253,16 +253,16 @@ # (source start file, target name, title, author, # dir menu entry, description, category) texinfo_documents = [ - ('index', 'conda', u'conda Documentation', - u'Anaconda, Inc.', 'conda', 'One line description of project.', - 'Miscellaneous'), + ('index', 'conda', u'conda Documentation', + u'Anaconda, Inc.', 'conda', 'One line description of project.', + 'Miscellaneous'), ] # Documents to append as an appendix to all manuals. -#texinfo_appendices = [] +# texinfo_appendices = [] # If false, no module index is generated. -#texinfo_domain_indices = True +# texinfo_domain_indices = True # How to display URL addresses: 'footnote', 'no', or 'inline'. -#texinfo_show_urls = 'footnote' +# texinfo_show_urls = 'footnote'
An unexpected error has occurred, please consider sending the following traceback to the conda GitHub issue tracker Hello. When trying to update anaconda recently via conda update anaconda, I received this error and was told to submit the issue here. This is the result from conda info. Traceback (most recent call last): File "/Users/danielr/anaconda/bin/conda", line 3, in <module> from conda.cli import main File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/cli/**init**.py", line 8, in <module> from conda.cli.main import main File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/cli/main.py", line 48, in <module> from conda.cli import main_create File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/cli/main_create.py", line 11, in <module> from conda.cli import common, install File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/cli/install.py", line 18, in <module> import conda.plan as plan File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/plan.py", line 22, in <module> from conda.fetch import fetch_pkg File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/fetch.py", line 24, in <module> from conda.connection import CondaSession, unparse_url, RETRIES File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/connection.py", line 22, in <module> import requests ImportError: No module named requests
Try `pip install requests` and see if you can use conda. Thanks for the suggestion. Unfortunately I got the following error. Downloading/unpacking requests Downloading requests-2.4.3-py2.py3-none-any.whl (459kB): 459kB downloaded Installing collected packages: requests Cleaning up... Exception: Traceback (most recent call last): File "/Library/Python/2.7/site-packages/pip-1.5.2-py2.7.egg/pip/basecommand.py", line 122, in main status = self.run(options, args) File "/Library/Python/2.7/site-packages/pip-1.5.2-py2.7.egg/pip/commands/install.py", line 279, in run requirement_set.install(install_options, global_options, root=options.root_path) File "/Library/Python/2.7/site-packages/pip-1.5.2-py2.7.egg/pip/req.py", line 1380, in install requirement.install(install_options, global_options, _args, *_kwargs) File "/Library/Python/2.7/site-packages/pip-1.5.2-py2.7.egg/pip/req.py", line 664, in install self.move_wheel_files(self.source_dir, root=root) File "/Library/Python/2.7/site-packages/pip-1.5.2-py2.7.egg/pip/req.py", line 894, in move_wheel_files pycompile=self.pycompile, File "/Library/Python/2.7/site-packages/pip-1.5.2-py2.7.egg/pip/wheel.py", line 202, in move_wheel_files clobber(source, lib_dir, True) File "/Library/Python/2.7/site-packages/pip-1.5.2-py2.7.egg/pip/wheel.py", line 189, in clobber os.makedirs(destsubdir) File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/os.py", line 157, in makedirs mkdir(name, mode) OSError: [Errno 13] Permission denied: '/Library/Python/2.7/site-packages/requests' Storing debug log for failure in /Users/danielr/Library/Logs/pip.log Can you show me the output of `conda info -a`? Hi @asmeurer. Thanks for your interest in my issue. The result of `conda info -a` is below. It seems the problem is with the missing requests package. But I am unable to install it as you see above. I'm confused as to what to do about this. Seems like a catch 22. Traceback (most recent call last): File "/Users/danielr/anaconda/bin/conda", line 3, in <module> from conda.cli import main File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/cli/**init**.py", line 8, in <module> from conda.cli.main import main File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/cli/main.py", line 48, in <module> from conda.cli import main_create File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/cli/main_create.py", line 11, in <module> from conda.cli import common, install File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/cli/install.py", line 18, in <module> import conda.plan as plan File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/plan.py", line 22, in <module> from conda.fetch import fetch_pkg File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/fetch.py", line 24, in <module> from conda.connection import CondaSession, unparse_url, RETRIES File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/connection.py", line 22, in <module> import requests ImportError: No module named requests Just to let you know that I uninstalled anaconda and reinstalled it and now it seems to be working... Thanks for your help.
2018-06-30T16:05:22Z
[]
[]
Traceback (most recent call last): File "/Users/danielr/anaconda/bin/conda", line 3, in <module> from conda.cli import main File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/cli/**init**.py", line 8, in <module> from conda.cli.main import main File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/cli/main.py", line 48, in <module> from conda.cli import main_create File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/cli/main_create.py", line 11, in <module> from conda.cli import common, install File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/cli/install.py", line 18, in <module> import conda.plan as plan File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/plan.py", line 22, in <module> from conda.fetch import fetch_pkg File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/fetch.py", line 24, in <module> from conda.connection import CondaSession, unparse_url, RETRIES File "/Users/danielr/anaconda/lib/python2.7/site-packages/conda/connection.py", line 22, in <module> import requests ImportError: No module named requests
4,364
conda/conda
conda__conda-7919
1471f043eed980d62f46944e223f0add6a9a790b
diff --git a/conda/base/context.py b/conda/base/context.py --- a/conda/base/context.py +++ b/conda/base/context.py @@ -229,7 +229,7 @@ class Context(Configuration): anaconda_upload = PrimitiveParameter(None, aliases=('binstar_upload',), element_type=(bool, NoneType)) _croot = PrimitiveParameter('', aliases=('croot',)) - _conda_build = MapParameter(string_types, aliases=('conda-build',)) + _conda_build = MapParameter(string_types, aliases=('conda-build', 'conda_build')) def __init__(self, search_path=None, argparse_args=None): if search_path is None:
config appears to not pick up conda_build/matrix_base_dir in 4.6.0b1 In my .condarc: ``` conda_build: matrix_base_dir: ~/code/automated-build/c3i_configurations/anaconda_public ``` This is for c3i. Somehow this isn't getting loaded properly: ``` c3i one-off py36_win python-3.6-feedstock -p "win*" Traceback (most recent call last): File "/Users/msarahan/miniconda3/bin/c3i", line 11, in <module> load_entry_point('conda-concourse-ci', 'console_scripts', 'c3i')() File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/cli.py", line 240, in main execute.submit_one_off(pass_throughs=pass_throughs, **args.__dict__) File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/execute.py", line 802, in submit_one_off config_root_dir = os.path.expanduser(config_root_dir) File "/Users/msarahan/miniconda3/lib/python3.6/posixpath.py", line 235, in expanduser path = os.fspath(path) TypeError: expected str, bytes or os.PathLike object, not NoneType ``` The code that loads this setting is at https://github.com/conda/conda-concourse-ci/blob/master/conda_concourse_ci/cli.py#L43
I isolated this to https://github.com/conda/conda/commit/1418be0fc7368ddbbfdae5994168b98f01b0c190#diff-b48dce471b4b8aed241c3b09306aa176 That commit causes the conda_build key to not be loaded at all. I am investigating.
2018-10-28T14:26:40Z
[]
[]
Traceback (most recent call last): File "/Users/msarahan/miniconda3/bin/c3i", line 11, in <module> load_entry_point('conda-concourse-ci', 'console_scripts', 'c3i')() File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/cli.py", line 240, in main execute.submit_one_off(pass_throughs=pass_throughs, **args.__dict__) File "/Users/msarahan/code/conda-concourse-ci/conda_concourse_ci/execute.py", line 802, in submit_one_off config_root_dir = os.path.expanduser(config_root_dir) File "/Users/msarahan/miniconda3/lib/python3.6/posixpath.py", line 235, in expanduser path = os.fspath(path) TypeError: expected str, bytes or os.PathLike object, not NoneType
4,393
conda/conda
conda__conda-8184
8eeb1bba428cceb56c6dae120c51e35b3332b854
diff --git a/conda/common/pkg_formats/python.py b/conda/common/pkg_formats/python.py --- a/conda/common/pkg_formats/python.py +++ b/conda/common/pkg_formats/python.py @@ -113,7 +113,8 @@ def _check_files(self): for fname in self.MANDATORY_FILES: if self._metadata_dir_full_path: fpath = join(self._metadata_dir_full_path, fname) - assert isfile(fpath) + if not isfile(fpath): + raise OSError(ENOENT, strerror(ENOENT), fpath) def _check_path_data(self, path, checksum, size): """Normalizes record data content and format.""" diff --git a/conda/core/prefix_data.py b/conda/core/prefix_data.py --- a/conda/core/prefix_data.py +++ b/conda/core/prefix_data.py @@ -256,7 +256,8 @@ def _load_site_packages(self): for af in non_conda_anchor_files: try: python_record = read_python_record(self.prefix_path, af, python_pkg_record.version) - except EnvironmentError: + except EnvironmentError as e: + log.info("Python record ignored for anchor path '%s'\n due to %s", af, e) continue except ValidationError: import sys
conda pip interop AssertionError in 4.6.1 This is the error with the most reports for 4.6.1 (over 100 in the last 24 hours). I've got a PR in the works to resolve this one. ``` error: AssertionError() command: D:\Users\rahul\Anaconda3\Scripts\conda-script.py list user_agent: conda/4.6.1 requests/2.18.4 CPython/3.6.5 Windows/10 Windows/10.0.17763 _messageid: -9223372036617369010 _messagetime: 1548792058544 / 2019-01-29 14:00:58 Traceback (most recent call last): File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\exceptions.py", line 1001, in __call__ return func(*args, **kwargs) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\cli\main.py", line 84, in _main exit_code = do_call(args, p) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\cli\conda_argparse.py", line 81, in do_call exit_code = getattr(module, func_name)(args, parser) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\cli\main_list.py", line 142, in execute show_channel_urls=context.show_channel_urls) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\cli\main_list.py", line 80, in print_packages show_channel_urls=show_channel_urls) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\cli\main_list.py", line 45, in list_packages installed = sorted(PrefixData(prefix, pip_interop_enabled=True).iter_records(), File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\core\prefix_data.py", line 116, in iter_records return itervalues(self._prefix_records) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\core\prefix_data.py", line 145, in _prefix_records return self.__prefix_records or self.load() or self.__prefix_records File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\core\prefix_data.py", line 69, in load self._load_site_packages() File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\core\prefix_data.py", line 258, in _load_site_packages python_record = read_python_record(self.prefix_path, af, python_pkg_record.version) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\gateways\disk\read.py", line 245, in read_python_record pydist = PythonDistribution.init(prefix_path, anchor_file, python_version) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\common\pkg_formats\python.py", line 78, in init return PythonInstalledDistribution(prefix_path, anchor_file, python_version) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\common\pkg_formats\python.py", line 382, in __init__ super(PythonInstalledDistribution, self).__init__(anchor_full_path, python_version) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\common\pkg_formats\python.py", line 106, in __init__ self._check_files() File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\common\pkg_formats\python.py", line 116, in _check_files assert isfile(fpath) AssertionError ```
2019-01-30T14:49:40Z
[]
[]
Traceback (most recent call last): File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\exceptions.py", line 1001, in __call__ return func(*args, **kwargs) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\cli\main.py", line 84, in _main exit_code = do_call(args, p) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\cli\conda_argparse.py", line 81, in do_call exit_code = getattr(module, func_name)(args, parser) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\cli\main_list.py", line 142, in execute show_channel_urls=context.show_channel_urls) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\cli\main_list.py", line 80, in print_packages show_channel_urls=show_channel_urls) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\cli\main_list.py", line 45, in list_packages installed = sorted(PrefixData(prefix, pip_interop_enabled=True).iter_records(), File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\core\prefix_data.py", line 116, in iter_records return itervalues(self._prefix_records) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\core\prefix_data.py", line 145, in _prefix_records return self.__prefix_records or self.load() or self.__prefix_records File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\core\prefix_data.py", line 69, in load self._load_site_packages() File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\core\prefix_data.py", line 258, in _load_site_packages python_record = read_python_record(self.prefix_path, af, python_pkg_record.version) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\gateways\disk\read.py", line 245, in read_python_record pydist = PythonDistribution.init(prefix_path, anchor_file, python_version) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\common\pkg_formats\python.py", line 78, in init return PythonInstalledDistribution(prefix_path, anchor_file, python_version) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\common\pkg_formats\python.py", line 382, in __init__ super(PythonInstalledDistribution, self).__init__(anchor_full_path, python_version) File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\common\pkg_formats\python.py", line 106, in __init__ self._check_files() File "D:\Users\rahul\Anaconda3\lib\site-packages\conda\common\pkg_formats\python.py", line 116, in _check_files assert isfile(fpath) AssertionError
4,409
conda/conda
conda__conda-8248
e419d60b4398c4b1c2c8f871792942c7962274f3
diff --git a/conda/_vendor/auxlib/type_coercion.py b/conda/_vendor/auxlib/type_coercion.py --- a/conda/_vendor/auxlib/type_coercion.py +++ b/conda/_vendor/auxlib/type_coercion.py @@ -222,8 +222,11 @@ def typify(value, type_hint=None): if isinstance(type_hint, type) and issubclass(type_hint, Enum): try: return type_hint(value) - except ValueError: - return type_hint[value] + except ValueError as e: + try: + return type_hint[value] + except KeyError: + raise TypeCoercionError(value, text_type(e)) type_hint = set(type_hint) if not (type_hint - NUMBER_TYPES_SET): return numberify(value) diff --git a/conda/base/constants.py b/conda/base/constants.py --- a/conda/base/constants.py +++ b/conda/base/constants.py @@ -199,6 +199,15 @@ def __str__(self): return self.value +class SatSolverChoice(Enum): + PYCOSAT = 'pycosat' + PYCRYPTOSAT = 'pycryptosat' + PYSAT = 'pysat' + + def __str__(self): + return self.value + + # Magic files for permissions determination PACKAGE_CACHE_MAGIC_FILE = 'urls.txt' PREFIX_MAGIC_FILE = join('conda-meta', 'history') diff --git a/conda/base/context.py b/conda/base/context.py --- a/conda/base/context.py +++ b/conda/base/context.py @@ -14,7 +14,7 @@ DEFAULT_AGGRESSIVE_UPDATE_PACKAGES, DEFAULT_CHANNELS, DEFAULT_CHANNEL_ALIAS, DEFAULT_CUSTOM_CHANNELS, DepsModifier, ERROR_UPLOAD_URL, PLATFORM_DIRECTORIES, PREFIX_MAGIC_FILE, PathConflict, - ROOT_ENV_NAME, SEARCH_PATH, SafetyChecks, UpdateModifier) + ROOT_ENV_NAME, SEARCH_PATH, SafetyChecks, SatSolverChoice, UpdateModifier) from .. import __version__ as CONDA_VERSION from .._vendor.appdirs import user_data_dir from .._vendor.auxlib.decorators import memoize, memoizedproperty @@ -25,8 +25,8 @@ from ..common.compat import NoneType, iteritems, itervalues, odict, on_win, string_types from ..common.configuration import (Configuration, ConfigurationLoadError, MapParameter, PrimitiveParameter, SequenceParameter, ValidationError) -from ..common.path import expand, paths_equal from ..common.os.linux import linux_get_libc_version +from ..common.path import expand, paths_equal from ..common.url import has_scheme, path_to_url, split_scheme_auth_token try: @@ -217,7 +217,7 @@ class Context(Configuration): # ###################################################### deps_modifier = PrimitiveParameter(DepsModifier.NOT_SET) update_modifier = PrimitiveParameter(UpdateModifier.UPDATE_SPECS) - sat_solver = PrimitiveParameter(None, element_type=string_types + (NoneType,)) + sat_solver = PrimitiveParameter(SatSolverChoice.PYCOSAT) solver_ignore_timestamps = PrimitiveParameter(False) # no_deps = PrimitiveParameter(NULL, element_type=(type(NULL), bool)) # CLI-only @@ -675,7 +675,6 @@ def category_map(self): 'allow_non_channel_urls', )), ('Basic Conda Configuration', ( # TODO: Is there a better category name here? - 'env_prompt', 'envs_dirs', 'pkgs_dirs', )), @@ -722,6 +721,7 @@ def category_map(self): 'always_yes', 'auto_activate_base', 'changeps1', + 'env_prompt', 'json', 'notify_outdated_conda', 'quiet', diff --git a/conda/cli/main_config.py b/conda/cli/main_config.py --- a/conda/cli/main_config.py +++ b/conda/cli/main_config.py @@ -15,7 +15,7 @@ from .._vendor.auxlib.entity import EntityEncoder from .._vendor.toolz import concat, groupby from ..base.constants import (ChannelPriority, DepsModifier, PathConflict, SafetyChecks, - UpdateModifier) + UpdateModifier, SatSolverChoice) from ..base.context import context, sys_rc_path, user_rc_path from ..common.compat import (Mapping, Sequence, isiterable, iteritems, itervalues, string_types, text_type) @@ -357,6 +357,7 @@ def enum_representer(dumper, data): yaml.representer.RoundTripRepresenter.add_representer(DepsModifier, enum_representer) yaml.representer.RoundTripRepresenter.add_representer(UpdateModifier, enum_representer) yaml.representer.RoundTripRepresenter.add_representer(ChannelPriority, enum_representer) + yaml.representer.RoundTripRepresenter.add_representer(SatSolverChoice, enum_representer) try: with open(rc_path, 'w') as rc: diff --git a/conda/common/logic.py b/conda/common/logic.py --- a/conda/common/logic.py +++ b/conda/common/logic.py @@ -34,6 +34,7 @@ from logging import DEBUG, getLogger from .compat import iteritems, odict +from ..base.constants import SatSolverChoice log = getLogger(__name__) @@ -264,33 +265,32 @@ def process_solution(self, sat_solution): return solution -def get_sat_solver_cls(name=None): +def get_sat_solver_cls(sat_solver_choice=SatSolverChoice.PYCOSAT): solvers = odict([ - ('pycosat', PycoSatSolver), - ('pycryptosat', CryptoMiniSatSolver), - ('pysat', PySatSolver), + (SatSolverChoice.PYCOSAT, PycoSatSolver), + (SatSolverChoice.PYCRYPTOSAT, CryptoMiniSatSolver), + (SatSolverChoice.PYSAT, PySatSolver), ]) - if name is not None: - try: - cls = solvers[name] - except KeyError: - raise ValueError('Unknown SAT solver interface: "{}".'.format(name)) - try: - cls().run(0) - except Exception: - raise RuntimeError('Could not run SAT solver through interface "{}".'.format(name)) - else: - log.debug('Using SAT solver interface "{}".'.format(name)) - return cls - for name, cls in solvers.items(): + cls = solvers[sat_solver_choice] + try: + cls().run(0) + except Exception as e: + log.warning("Could not run SAT solver through interface '%s'.", sat_solver_choice) + log.debug("SAT interface error due to: %s", e, exc_info=True) + else: + log.debug("Using SAT solver interface '%s'.", sat_solver_choice) + return cls + for solver_choice, cls in solvers.items(): try: cls().run(0) - except Exception: - log.warn('Could not run SAT solver through interface "{}".'.format(name)) + except Exception as e: + log.debug("Attempted SAT interface '%s' but unavailable due to: %s", + sat_solver_choice, e) else: - log.debug('Using SAT solver interface "{}".'.format(name)) + log.debug("Falling back to SAT solver interface '%s'.", sat_solver_choice) return cls - raise RuntimeError('Could not run any SAT solver.') + from ..exceptions import CondaDependencyError + raise CondaDependencyError("Cannot run solver. No functioning SAT implementations available.") # Code that uses special cases (generates no clauses) is in ADTs/FEnv.h in
RuntimeError: Could not run any SAT solver. ``` error: RuntimeError('Could not run any SAT solver.') command: C:\Tools\Progs\python\anaconda3\scripts\conda-script.py install -p C:\Tools\Progs\python\anaconda3 -y anaconda-client==1.7.2 user_agent: conda/4.6.1 requests/2.21.0 CPython/3.7.1 Windows/10 Windows/10.0.15063 _messageid: -9223372036617369659 _messagetime: 1548779675918 / 2019-01-29 10:34:35 Traceback (most recent call last): File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\exceptions.py", line 1001, in __call__ return func(*args, **kwargs) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\cli\main.py", line 84, in _main exit_code = do_call(args, p) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\cli\conda_argparse.py", line 81, in do_call exit_code = getattr(module, func_name)(args, parser) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\cli\main_install.py", line 21, in execute install(args, parser, 'install') File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\cli\install.py", line 250, in install force_reinstall=context.force_reinstall or context.force, File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\core\solve.py", line 107, in solve_for_transaction force_remove, force_reinstall) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\core\solve.py", line 145, in solve_for_diff force_remove) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\core\solve.py", line 240, in solve_final_state ssc = self._find_inconsistent_packages(ssc) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\core\solve.py", line 342, in _find_inconsistent_packages _, inconsistent_precs = ssc.r.bad_installed(ssc.solution_precs, ()) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\resolve.py", line 896, in bad_installed C = r2.gen_clauses() File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\common\io.py", line 85, in decorated return f(*args, **kwds) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\resolve.py", line 668, in gen_clauses C = Clauses(sat_solver_cls=get_sat_solver_cls(context.sat_solver)) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\_vendor\auxlib\decorators.py", line 56, in _memoized_func result = func(*args, **kwargs) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\common\logic.py", line 293, in get_sat_solver_cls raise RuntimeError('Could not run any SAT solver.') RuntimeError: Could not run any SAT solver. ``` This shouldn't be a `RuntimeError` that results in an ugly stack trace, but rather some type of `CondaError` that says they've probably managed to uninstall `pycosat`.
This one is related. We just need to improve the error handling here. ``` error: RuntimeError('Could not run SAT solver through interface "pycryptosat".') command: /home/brentp/miniconda3/bin/conda install -y conda-build user_agent: conda/4.6.2 requests/2.21.0 CPython/3.7.1 Linux/4.15.0-43-generic ubuntu/16.04.5 glibc/2.23 _messageid: -9223372036616970284 _messagetime: 1548885913575 / 2019-01-30 16:05:13 Traceback (most recent call last): File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/common/logic.py", line 279, in get_sat_solver_cls cls().run(0) File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/common/logic.py", line 175, in run solver = self.setup(m, **run_kwargs) File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/common/logic.py", line 223, in setup from pycryptosat import Solver ModuleNotFoundError: No module named 'pycryptosat' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/exceptions.py", line 1001, in __call__ return func(*args, **kwargs) File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/cli/main.py", line 84, in _main exit_code = do_call(args, p) File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/cli/conda_argparse.py", line 82, in do_call exit_code = getattr(module, func_name)(args, parser) File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/cli/main_install.py", line 20, in execute install(args, parser, 'install') File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/cli/install.py", line 253, in install force_reinstall=context.force_reinstall or context.force, File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/core/solve.py", line 107, in solve_for_transaction force_remove, force_reinstall) File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/core/solve.py", line 145, in solve_for_diff force_remove) File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/core/solve.py", line 240, in solve_final_state ssc = self._find_inconsistent_packages(ssc) File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/core/solve.py", line 342, in _find_inconsistent_packages _, inconsistent_precs = ssc.r.bad_installed(ssc.solution_precs, ()) File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/resolve.py", line 896, in bad_installed C = r2.gen_clauses() File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/common/io.py", line 85, in decorated return f(*args, **kwds) File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/resolve.py", line 668, in gen_clauses C = Clauses(sat_solver_cls=get_sat_solver_cls(context.sat_solver)) File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/_vendor/auxlib/decorators.py", line 59, in _memoized_func result = func(*args, **kwargs) File "/home/brentp/miniconda3/lib/python3.7/site-packages/conda/common/logic.py", line 281, in get_sat_solver_cls raise RuntimeError('Could not run SAT solver through interface "{}".'.format(name)) RuntimeError: Could not run SAT solver through interface "pycryptosat". ``` I got this error today and was befuddled enough to just uninstall and reinstall anaconda. (also some cursing...) My hunch was that activate wasn't working correctly when doing an env create with an environment.yml file because I also ended up with a mixture of python 3.6 and 3.7 in the base environment (which was on 3.6). When I went to try and fix that base environment, I think this SAT silver error was the next error I got.
2019-02-11T02:13:42Z
[]
[]
Traceback (most recent call last): File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\exceptions.py", line 1001, in __call__ return func(*args, **kwargs) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\cli\main.py", line 84, in _main exit_code = do_call(args, p) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\cli\conda_argparse.py", line 81, in do_call exit_code = getattr(module, func_name)(args, parser) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\cli\main_install.py", line 21, in execute install(args, parser, 'install') File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\cli\install.py", line 250, in install force_reinstall=context.force_reinstall or context.force, File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\core\solve.py", line 107, in solve_for_transaction force_remove, force_reinstall) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\core\solve.py", line 145, in solve_for_diff force_remove) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\core\solve.py", line 240, in solve_final_state ssc = self._find_inconsistent_packages(ssc) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\core\solve.py", line 342, in _find_inconsistent_packages _, inconsistent_precs = ssc.r.bad_installed(ssc.solution_precs, ()) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\resolve.py", line 896, in bad_installed C = r2.gen_clauses() File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\common\io.py", line 85, in decorated return f(*args, **kwds) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\resolve.py", line 668, in gen_clauses C = Clauses(sat_solver_cls=get_sat_solver_cls(context.sat_solver)) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\_vendor\auxlib\decorators.py", line 56, in _memoized_func result = func(*args, **kwargs) File "C:\Tools\Progs\python\anaconda3\lib\site-packages\conda\common\logic.py", line 293, in get_sat_solver_cls raise RuntimeError('Could not run any SAT solver.') RuntimeError: Could not run any SAT solver.
4,415
conda/conda
conda__conda-8328
662df07042b2b0f5096334c4594f4351a5685978
diff --git a/conda/cli/install.py b/conda/cli/install.py --- a/conda/cli/install.py +++ b/conda/cli/install.py @@ -201,7 +201,11 @@ def install(args, parser, command='install'): specs = [] if args.file: for fpath in args.file: - specs.extend(common.specs_from_url(fpath, json=context.json)) + try: + specs.extend(common.specs_from_url(fpath, json=context.json)) + except UnicodeError: + raise CondaError("Error reading file, file should be a text file containing" + " packages \nconda create --help for details") if '@EXPLICIT' in specs: explicit(specs, prefix, verbose=not context.quiet, index_args=index_args) return
warn on incorrect use of `conda install --file` argument ``` error: UnicodeDecodeError('charmap', b'BZh91AY&SY+\xeah\xb3\x02<M\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xe2\xc30.\xddt\x84r|s)\xf7\xbdel\x1b\xee\x95\xef\xbb\xbe\xb9\xf7\xcd\xf7\xce\xef\xa1\x93\xe8\xfa\xb7\xdf\x0c\xfb\xef\xb7\xaez`\x07@\x03)\x04\x00\x00\x0fW\xbd\xe0\xf1\xd3@\xf6\x18\x07;s\xdfp\xe8\xaa_6\xe0\x1f{\xbc\x06\xf1c\xa0\x01@(\xf7\x86\x9e\xeb\xb8\xd36\xae6\xd9m\xa1\xaa\xa1}\x9d\x8d\xca+\xa0\x18\xaa6\x07@\xcb\xe3v\xfaE\xefM=\x05HSZ(g\xd8\x00\xf46{\xef\x80\x00Wz^\xdb\xe0;/{\x9b\xb9\xcf\x838{\xcf\xa0<\xde\xbeoW\xa6\xb4ty\xef\xb6\xaf\xb7\xc7\xde\xe1\xdc\xf7Z\x1e>\xect\x0e\xaf@\x00k\xe3\xef\xbe\xef\x07\xbe\xde\xf8\xbd\xb0\x00\x00\xf6\xcd\x10\xbd\xa5-\xeb\xdcu\xed\xef9\xd7\x80\x03=\xe9m\xdd\xd7\xdb\xed\x14\x1d:\x9b\x0fn\xf9\xb5}\xef\xb5\xef\x8d\xefy\xd9J/\xb6\x93\xd6\x9aV\x8dh4\xb6\x01\xf7U\x03Y\x18@\xa2\xb5\x8a\x87\xdc}\xde\xf9\xb6\xef\xbe\xd5\x01\xa0\xb3kT[\xb9\xd1\xdb\xbc\xf1\xf3\x0e\x90\x14\x00\x15\xdb\xefw\xddgZ\x01\xa9\x00|1\xde\xcf\x90\xfa\x03\x00\x00>\x8e8\xd0)\x88\x04Z\xfba]\xb0\x92\xa1\xe1\x0e\xfb\xe3\xc3!\xb2\x1b\x1c\xc0\x03\xdd\xef\xbet\x06\x82G\x9fw\xb9\xd5\x00\xad\xa8^\x9d\xf3\x9e\x015\xf4\xd0\x17\xb9{\x92|\xbe\xb7\x03\x8f\xa0\x1e\xea^\xd6\x05qe\xb5\xa5\xa98\x1c\x00\xd8\xb2\xedu\xe3\x85h;\xef{\xdf}\x8fG{\xdc\xeb\xb9\x81\xb1\xf5\xba\xef}\xc6^\xd5\x93k%\xa9g\xd1\xeb\xaa\xdf\x07\x9b\x8f\xaeq\x9d\xf7:\xaa\x16E\xdd\xf7\x9b3\x16\xc0\xdb\x06\xf6eD\x06\x9a\x1bc\xd7\x9d\xe7{6\xbeGo{\xdd:\xe8q\x97\x9a\xf5\xf2w\xd9\xf7s\xde\xde\xbe\xef\xbc\xf6\xb6s;\xb4\x9e\xdes\xe0\x03\xd1\xd7_\x1e\x9f@c\xeb\xbb];\xb7\xde\xf5\xd8\x95ml\rbh\xa5\xedTP\x02\xb3\xday\xbbt\xad\xdd\xdd\xa4i\x8f\xbe\xbe\xdb\xef\xbe\xd5|\x1an\xa6\xecrV\xd7Z\xf2\xf1\xd0T\x85\xe0;\xd3\xde\xd8\xe9\xdd\xb3\xafx\xc5b\xfb\x10]\x87\xbd\'\xb8(oc\\\xf3\x1d\xd8Nfks\xde\xe0\tU\x0e\xf6\x0b\xb0V\x9d\xec<u\x9e\x81\xad(2\x0e\xb9\x03\x93\xb6\xaf^\x05\x17A\xdb\x00[{\xc7s\xc9\xbd_-\xdcs\xc77>\x01\xdf9\xee\xaf\x14\xa5B\x82\x95\xf7\x1f=\xb7\xcc}w\xb7\x94\xfa\xd9\xcd\x8e\x9eC\xc3^\xb4\x1e\xcb\xbe\xd3\xd7\xd06\xc1\xad\xedJ7\xde\xe0\x00\x1e\xf8\xc3\xeb\xdb\x01R\x15B\xda\x10\xddz\xb1\xbb}}\xf7ot\xeb\xe8\x0f^\xb5\xd7\xbe\xfbe\xf7=\x8e=\xb7h\xf8\xfb{Qx\x84\t6\xcf\x9b\xd7\xbbL\xcb\xee(\r<@\x06\xde\xe0\x1e\x81\xe4\xf4\x05\x00\x00\x02\xd9\xe8\x00\x00\x00\x00\x06\xdfx\x02\xa8\xf2/G\xbe\xc0\x01\xd3M\x00\xfb\xb0\x18\xe5F<\xfb}\xf6\x01\xa0\x0f@\x1dk\xd9\x9c\xb3}\xb7n\xec\xde\xb7\x9e\xcfy\xe3\xb0\x07\x9a\xdb\x8e\xdc\xbb\x8e\xf5\xb6[)T\x1e\xdbom.\xee\xeb7Z\xfb\xdcv\xb6l\x0bywJ\x9f{\x9d\xb7\xb5\xd8\t\xb6\xe7jq\xb8v\xcd\xd6\xef\xae\xd1\x0e\xec\xf7\xdaN\xf7\xdfww\xbb\xb1\xf3\xa3{\xab\xeam\x8aP\xf44\x9d\x06\xa7l=\xf7\xdd{\xef\x8a\xf9\xef\x9f=\xef\x1ez.\xf6\xed\xb2z\xf8X\xdfgk\xe1\xef\x80\xfb\xd9\xb7\xc6\xf9\xf7\xbd\xa7{\xac\xdch\xf7\xde\xeep\xfb\x1b\xde\xed\xd7\xbdw\xbb\xb9;\xef>\xd3\xb7\xb1\xf7\xc5\xbb\xe9\xbboy^\xf5V\xd4\thP\x1a\xa0\n\x00P\xb3\rn\x1e\xd3\x1f-\xcd\xf7\xbb\xd5|6\xa7\xb6\xd9\x97\x04\xda\xaa\xb6\xd6m\x1b}\xf6<\xaf}\x8e\xd7t\xc3d)\xa3\xad\x14(h\xe9\xca \xd8V6\xac\x02\xdbZ\x0b\xb2\xdd\xcb\xad\xbb\xbb]>\xf9\xddvb}\x9a\xb3\x00\x14\xf5NB\x95\x13\xa5\x86\xfb1Vvn\xdc\xa6u\xba\xe5\xa6\xb2\xd5\x94\xd0\x00\xd6q\x8aQJ\xdd\x9c\x07r\xc7wjZa\xd3w\xc7\x0e\xae\xcd\xb6Jj\xf4\xf7\x03^%.|\xefV/\xbb\xe8\x9dxz\x1f{f\xf1\x1f\x00>\xe3\xb9f\x1bc\x96\xad\xb6\xc7N\xec\xeb\xb5\x0f\xb7;oNq\x1fRn\x06\x9a\x1a\xdb\x8e\x1c\xea\xde\x8ak.\xed\xdbu\x95\xa4\xbd\xcdS\x96\xd8\xbdqu\x07\x9e`\x07C\xeeg\xd7\xa3l&\xc1\xd3\xb9\x83v+\xa3\xa1\xb6\xf9wh\x80\x86\xad\xb5Zl\x1a\x16\xc5\xa3\x8f,\xf3\xc3,\xad\xb7\xa6:y\xde\x8e\x9d4\x00\x07M\x07\xa1Ww\x1c\xcb\x1ba\x85k\x00\x1a\t\xbe\xdb<\xf2=+\xa0\x0e{\xba\xf5\xbd\xde\xb1Mh\x00\x00\n\x1b\xa2\xeek\xb3n\xeeu\xdd0\xeb\x1dw\xb6\xf5\xbc\xde\xcfv\xf4\xefz\xdc\xa6\xebrs\xb5\xedx\xd5\xa0\xfb\x9e\xec\x93\xdb\xef\x99\xf2y\xf0\x03\xed\x80\xd6\xd6\x1d\xb0:\xa4v\x19\xd8+\x15;g|\xf8\x8e\xe8\x1f}\xaf[\x9e\xeb\x89\xed\xe5\xefs\xbe=\x80\xdeWpVj\xde\xe7O_;\xee\x80c\xbbN\x15w\xde\xf5\xd7\xa5\xf5\xd7\xc7\xdfZdto)\xdb\x80\x00\x03M(\x02\xf6=\xca+z\x08\xeb\xb0=\xe7q\xe4fu\x17}\xefJ\xab\xbd\xbe\xdf[O\xb6\xbb\xaf\xb7\xaf\xbd\xae\xd8S\x80\xd0\x03\x94\x8d[/]\xe7\x9a\xb6=\x9b\xb5n\xef8>\xee\xcd\xceA\xf7\x8e\xf7\x93\x0f\xb7\xbep\xf9\xbe\xfb\xce\xfb]{\xe6[\xd7P\xfah\xfam\x81\xad\x1b`hk\xa7\xce\xba\xf9wv\xb7\xbbw\xcf=\xbd\xdb\xaf\x91\r\xef\xb7\xaf\xb5\xa6\xde\xc7\xbd\xcf]Y\xbb\xed\xdd\xeb\x9f7R{b\xf5\x9d\xb3\xba\xee\x1e>\xfb\xe9|\xeb\x8e\xbe\xed\xd3\xdf}\xe9\xf16\xcf\xa6\xee\xed\xbb\xbe\xec\xfb\x9d\xe7\xdb;9{\xb7\xaf}>\xfb\x81\xf2\x1e\x95G\xabl\xf5}\xd5\xdbg\x1d;\xbd\xef\x81\xdb\xd9\xee\xee{F\x9d\xb3E\xed\xde|\x9f}\xf5\xda\x8e\xfb\xbaAdLm\x8e\xe3\xdf^\xfb\xe6\xfa"T\x8a\x97mT_[\xeb#u\xdfk{\xe5\xab\xeb5\xf3\x14-}\xa7\r\xf3}\xf6}6SX\x0b\xebg\xbe\xea\x9d\xee+\x17\xa7\xbeg\xbe\xe73\xc4\xb7\xd7{\xc7\xda{\xde\xee\x8b\xab\xd2n>\x9e\xec[\x1bL\xed_{\xdd\xb3\xd7*\x17\xcfw+=s^\xec\x0e> \xa2\x12\t\x01K|\xfb\xd5\xb6RQ(\x0c\xe6;\xbe\xfb\xa3\xbd\xdd\xef\x9f\x1b\xdfx>\xbc\xdf{\xb7x5_k\xcd\xc6\x85\x8eis\x93!\x8erp\x00\xf4\x05\xf7\xbe\xe7"\r\xeb\xc3\xe0\x06\x8f@|;\x1e\xfb\xc2\x14\x03\xdd\xb8\xf9w\x03P\xa1\r\x11\x04\x00L\x80\x00\x00\x00\x00\x00F\x04\xc2d\r\x00\x00\x00\x00\x00\x00\x00\x04\xc0\x00\x00\x04\xc10\x00\x04a\x00\x00\x00\x1a\x00\t\xb4hA\xa0\x80@\x08\x00&\x80F@\x00\t\xa6\x9a\x00&\x00\x00\x98\x9944\rL\xd4i\x84\xd3\x02d\x06\x9a4\xd3\x056\x8c\x99Ldd\xd0d\xd2\x8f\x11\x84\xf4\x10\xc9O\xc24\xd4\xf4\xc1L\xd0\x13hI\xe2OH%2\x11\x04\x08\x10\t\xa0\x08\xd3 \x024\xd3\t\x84\xc9\xa6\x93\x02z\r4\rS\xc4\xd4\xf2e=\x13\xc6\x80\xa3\xcaz\x98\x9e\xa6d\x994\x9e&\x05=\x90dh\x13$\xf2G\xa4\xc6Dd\xf4\xc8\x9a1\xa9\xb4M=LL\x9e\x9a\x086\x88\xd3A&\x92$ @\x04\x01\x1a4\x98\x13\x04\xc8b\ny5=\xa2\x9f\x90\xd3@L\x04\xc1\r\x1a\'\xa9\xb4\x9bS\xd0\xc1T\xf7\xa4=\x06\x91\xb5)\xfbEM\xe6\x89\xea\xa7\xe2)\xb4\xd9\'\xa1O\xc5O\xf5\x18\xa6\xd0\xd4\xf6\xa9\xe2\x9a\x99\x8fR6\x88\x8d\xe5Oh\xa62M\x19=5\x02$\x88A\r\x04d\x04\xc2\r0\x86\x91\x80&\x13L#\nyO)\xe8\xd3M\x18\x99\x06\x934\x99S=0\x06\x83*y0\x89\xe0\x9aeO\x19\x02j\x9e\x9bBg\xa6\x93\x19I\xe4\x9e\xa9\xec\xa9\xa7\xe4\xd3\xd2\x08\xd3S=S\xf2\xa3\xf4M\xa8\xc1#\xcd\x14\xde\xa4\x11$\x81\x02\x00\x9a\x00L OA\x88\x02hmL\x0112z#M2h\x1a\r\r\x0cS\xd0\x057\xa8\xd4\xa7\xfa\x01\x88\xa4\xf6\x9bM4\x8d\x90e3\x13Jbz\x9f\xa8\xc4&\xcax\x94\xdbH\x9eS\xda&\xa6\xf5F\xf52\x9bS\xf4I\xa6\xd4\x07\xa9\xa7\xae\xebg\xaey\xb4\x0b\xfdi\xee4ouqm\xe8i\xedp\xe1|\xe9\xd2\xf1\xcb\xccV\x92\x97F\xe2\x1e\xb5\xacY\x00nVF\x91hV;\x90`B\xf1\xc5\xd7\xba3(\xe5\xc5.\xb2\x1e%\xf7\xf0,p3\xfa,{N\xccV\xc4y\xb7S\x81\xa0A]B\xf2\xaf\x88\xac\xde\xaa\x05\x12a\x84u\xe7+h\xc2\xce\'%S\xb5\x84A\xc93hC\x88L.\x9c({V7\xdf\xb4@\x01\xe2\xf8c\xd8k\xf78\xf5\xe5\xdf\xcd\x96x_\xb7\x16\xbd\x80\\/\nW^\xff|\xc6\xc2v\xc2h\xa9\xb9\xb3\x07\xd9J\xc8;\x05X\x7f\xcdH\xdd\x13!\x0f6A%\xe2\x9dL\xac\xccmge9\x86\xd6\xadf\x8b.\x0f\x90`\xa5F\x8cE\x10\x14\x86\x8c\xb1c\n\xf2\xff\xfa\xb5mY\x9a\xb6\x01:rAA(\xf0a\xa6\x83\xcfm\xc8B\\\x18\xde\x83\x06\x80\xe2a\x17\xc6\xcf-\'\x889\x88\xa2H\'\x04&\xaaqX\x9cS Q\xf0$\x00\xfcD\x0f\xc7@k\x8d\x052\x94\x1f\xc7\xb7\xc5s\x7fv\xf3\xf8\xff\xba\xf2\xf29\x1a\xfc\x97\xef\xfd\xa7\xb5\xa2\xe9Qw\xb2\xbc?\xf5\xf3\x83A\x0c\x17\xf1\x1c\xe58\xad\x89\x90\x15\x18\xaa\xc2\xb5\x00`\xc0\xe8\xbf)1!\x1d\xbf\x99\xf5\xef~{\x0f\x89\xa2$\xef9\'\xa2\xcb\x9e\xf2\xca\xfc\xa4V?\xff\x1f\xd5\xffK\xff\xbb\x7f\xc5!\x8a\xb2C\x08dI4\xc6\x8a?\xf37\xf5\xcc\xaa\xb1\xd9~\x0fi6&S@\xc3\x92\xba\xc2k\x11\x00D\t\x8a\xc1A\xf2\xb9\x93\x08L#\xd4\xa0\xd5n7\x93\x0c_\xf8\x9f\x9b\xff\xb0\xab5\xfa\xec\xcf\xdcq\xf8\'p\xe2\r\xe66,\xce\x19]8\x8b9\x82\xce\x0c\x03\x06I\xb0\x9f\xb3\xfe\xda\xad\xe4\xb4a\xbe\x88cc\x92zG!\x95\x83\x06K\xa2Zn3r0\x02\x01\xd6O\xfd\x11\x16\xa9t\xc0qo\x90\xc82m\x8c\x8a\x9b\xcc\xf9\xd0?m\xde\xf5\xf8\xbe\xa0\xc0U\x93\xc1\xc8G \x19\x97\xeeq-\x1e\xa49A\xfa@tK\xaa\x1d\x90\x8f\x8c\x99\x12n\x87\xb4#%.bZJH\x9f\x03\xe0\x11\xa1\xf5\xf8\x8cX\x80\nG\x07u\x86\x94\xe9\x88\x85"\x0b0\x15d\xcc\x18pv\x11w\xfa\xad\xb8\x04\xa8Yy\xd4\xa2\xe8\xf2\x86M\xb2\xda?;`\xda\xb2\x8bu\x08\xc9\x9b\x83\r\xe0\xa2\x9e\xaeA\xf8-A--))C&\x04\x98\x93\x12aI\x17\x87d&\xa4\x82d\x80\x85\x06\xe4\x80{\x11P*\xca\xd0E\x80W%\xd9\xda\xa4\xa3\xa0m\xf9\xd8\xebl|x\x8e\xda7\x9ah\xcc\xe2\xe1\x84(\x84\x83\x83\xd2W\x85\xfd<\x01\x9a\x9d<\\\xd9\xee 8\xad\xbd{\x8b\x9f>n.\xbe~.~-:)\xa3\xc0\xdf\xe8o\xea\x1a\x878\xb4\xb3F\xf0\xec\xf4P\x07\x04s\xc1\x0eFa5\xbe\xa8\xe2\xef\xc1G\x8ex\x8f\xb2^>\xa6n\x0eB\x18\xb8\x85x\x00o4\x8b\x87\xbd\xca\x00c\xc81\xf5\x8f\xc5\x18\x88\xc3\x8bTr\xc7\x8bC\xca\xae+ \x12\xaa\xabP\xca\x86\x82((\xa2\x07Qa\xde\xe6\x98\xf60\xe6u!\xe2W\x04\xe4^\xf7%D\xd5\xedd\x0c:\xe1\xc3:\x99\xdf\x19\xdc\xc3\xd2\x0fM\x07\xce\xbe\x7f\x86\xe6/\xcb\xfa=d\xec9\xbeOC2\x8b\x07S\x05#k\xb7\xe9n;\xa3\xd1\x80\xcd\xa3\x08\xb3\x1f]a\xd4@\x0e\x0c\xdau\xf4U\xcf\x86\n\xb4\x81\xac,\xb0\x0b\x05\x83Q\t\xf2\x10\xd9\x1b\x17\xd4\xc1\x983\xa41\x03!dg\xce\xcc\xf5:\xa5\xba\xf5\xcdp\xa5\x88B3\x02o\xf24\x9b\xfe\x92i\xda\x9c\xac\xd7\x12\x1c\x92\x07#\x91\xea\xb9V\xf4\x0e\xc5\x80+\xeb\x88\xaf0\x8a\xb5T\x10\xfa\x94\x1d\xf0v\xfe\xc6?\x95;\xff\x06\xfb\xadE\xd7\xce\xde[\xa1T\xe6=\xd6qZ\x06i\xbe\xfb\x9a?\xcaj1\xad8\x95\x90kACp\x81\x87\xb9\xa1\xbb*\x88jn\r\xbb\x0c\xb6U\xd0k\x99\xcb\xe7s*b\x97<\xd6\x06\x05[|B\x8b\xe2qX\xc3]C]\x8b\xa6\xd9\x87\xcc\xc9\xa0>\x92\x0c\x10\x06dFu\x08}\x9d\x91\xa7R`a\xff\xd1*\xa3(\xb6\x13\n&\xaa\x8e\x8d\x84\xc0B\xe5N\x00\xbb\xcf\x15,\xf5-MO\x90\xee\xc4v}>\xc4v=\x7f_n+\xe8/X\xeb\x88PE\xd9\xb2\xf7\x8du\xce\x16\x16\x81\xdaZv\xdb`\xecrVD\x053\x01^;u6\xdcx\xed\x18\xee\xb5;Kq\xdc\x06a|\xcb\xb16\x1eC,\xbb(\x06\xa2=\xb4\x0f;\x08\x81\x10\x01\x02\x9e\x93\x94\x9e\x9f\x9cVrqi\xf9\xc6"jyT\xe7\x12\xbc\xf2\x98\xe9\xf7\x95\x18Q\xd1Q\xd0\xb1\rQF\xbdvq\x7ff:\x9d\xbb_T/1\xd9\xef\x1b\xa3\xd76\x0b\x16\x86\xac\xcd\xf6\x19\x9b\x1ag%\x8c\xd0rL\xfe\xba\xb8\xb9\x90p6\x9aje\x14\xe0\xac\xe0+\xae\xaa\xa6\x18\xeac$%\x07\xd9\xa2\x0b< t\x8a\xddSg\x9a\x1a\xc0\x88uD@\xe4\xc2$X\xf6\xee\xaf\x0f\xc3\xae\xe7\x04\xe2\xd5\x13#^\xf6\xba2\x04"\xb3\x0f\xe11\xb5\x0f\xb4\xd6\xe29\xe8\x8c6\xa7\xe8\xd97\xbaH\xa45\xf7\x8d\xb8\xb8\x95\xdfqr-\xc9\xc3\xa9\xf5\xb9\x8e\xfe[\rb\x02Y\x81\x9e\xa5(\x82\x1a*\x91\xfd\xde\x8f/\xde\x9a\xd9\xe1\x11\xfc"\xad\x9f,\x84\x1c\x8a$b\xc4g\xb6\xd45\xf1\xdc\xb2\xf8Q\xf7P_:\x01?\xe9\xc6\xcb\xaf*\x00\xf8\xe2$G\xa9\x8a\xde\x022\xf4\x88\x88B\nZ\x1e\xce\x00\x8d\xa0\x82\xb2B\xc4!\xac20\xa9P$\x86x\xa9\x88\xb3E\xa6\xbd\x9b\x9f\x8dd\x80\x98[\xf3f[\xfb\x99\xbeg+k\xd0\xc1\xe5\xd66z\x8dT\xe7\x9e\xc2k\x08\x93a\x1f_6\x14C\xa4\x81\xdd\xf4=\x7f@;\xffdQ\xdeX\xae\xfe\xbb\xc2\xc6n\xbb\xd9[\xbd\xcf\xe6X*s5\xf6\x14Q\xdd\x90\t\xff\xea\x84\xc0\xeb\x8a(\x90U\x8cE\x88\xa8\x8cT\x15F#\x15"\x83\xe8\x14\xf8\x12[\x0f\x9e5,\xcf\x90L\xc9\xed\xa6\xa9\x9f\x0e\xcc~F\x8b\xc3y\xfaM\xea\xa7\xa3N\xd1\xf8\x0c\x00:\x8e\x14`\x83\x07\x9aE\x11\n5aZN\xce\xcc1\x044\x85\n\r\x8d\x88\xd0\xa2\x08\x94,:\x19\x1b\x98u\xb8\x98\xcdfv^\xd7\rm\n\t\xbd\xe1\xf7\xd6\xa8\x00\'\x85$4.\x90\xcd\x9e\xa4\x1b@\xa2\xc5\t$\x86\xd0\xb7G\x13\x81d\xe1\xa1\xb2^%\xd4r\xc3x\xc2\xbeK\x19\xf4\xfa\xca\x1c4\x14A\x00~\xc4Dx\x07[\xd3\xd8\x0c\x0c\xf1\xd6\xac\xabQR\xc9\xaf\xaa\xd5\xf0\x9f)\xc0\xb9.vk/\xe1cZ\xdeN\xad*\x8a\xbd"\xad\xd5\xbe\x0e\xcdq\xde\xfc_\xaa\xe9\x11\xe9\xa8\x0b\x88\xdfSMH\xea\x9a/\x84\x1f\x11\x99\x9a\xb7\xcd\xf2\xe7\xb3D\xa6]\xda\xa8\x93\x0c\xa8\x1a\x04>\x8f\xb3\xf5\x9d\xb4\xc3\xbd\xf0f\xdf\xc6\xcf}\xdf\xf5h\x03\x807\xd7?\xd8ox0\x1d\xa6\xfd\xaap\xd7{\xb5\xa4\xfb\xdbb\xe3$}\xa2\x0c\ny\xe9\x94\xde\xa6\x04T)\xf2\xa3\x1e\xc6\xfc\xb9\x1e\x8c\xb8\'\xa0\xc8\x0fR\x1aI\xa2\x02\x89\xd05k\xe7#\xe7\xf9\xdf\xc1\xdd\xb0\x88@8\r\xb7\x8f`\xd2\xfb\x16\x7f\xb3\xe0\xfaj\xa4v,x\xba\x9a\xfeW\rj\x88|\x87\x1fE\xca\xce\xee\xc3\xf0\x00C\x11B\xc3\x07\xf7\xfb\x13\x1b\xdb\xa3\x00\xba\xf1\xd1\x90\xc1\x83!\xed\xee\xe7\x06\x17\x87S\x0c\xee\x7f\xc6\xea^\xb7Fk\xddU\x94\x84\x84#\x92MY\x0e7]\xe6O\x8d\xf5\x1bc\xddY\xf4:\xfb\xb9\xf6\xdf\xeb\xff\x86\xa3\xfd\xd5^Z$P\x98\xc4~\x15\x0c\x95\x1fP\xa3\np&\xb3\xe6\xe5\xd7\x07\xfdZ/\xd8\x18g\xff4\xcc\x0bKm\x1f\xb3\x84.>9\tu?\xb4\x84\xe0kKiN\x8b\xfd\xf9pU\'\xac\xcc\x12\xbeO\xff\xa1\x14G\xf6\xfc\x1e\x0f/\xb1\xb3\xe2!\xba\xa2@\xdf\x03\x95=\x18\x14\xd4\xf0p\xado\x8b\x82A\xd9!\xbc\xe2\x80$q\xb7\xd58\xb5wg"0\xec\'Y\xee\xf6\xb9i\xc62\x84\xc2\x06k\\\xa8R\x11\xb3\xee\xe5Q<v\xea\rW\x9f\x96\xe3\xc2\xb3,\xbb]\xa5\xd9\xa2\xaf\xd87\x9e_5fL\x9d\xa7|a\x95L\x9aE\xc5\xee\x82\x93\xd8\x13\xd0\xc8x\xb6i\x02N\x15~\x88\x90\xe8\xb0\xd5\xcd\xf8?\x10\x0c\x8ds\xbd\xf8p\x9c\x81w\x9f?a>7\x85\xb6\xe9s\x10\x86\x9dz\x1b\xe1t\xb4a\x81\x04z\xe1\xd5\xe3\xfdF\x8c\xec\xc6X$\r\xb4\xd3\xe3\xcc\xc1\xde\xb8\x83\xfeuu(\xc3\xaf\xf0}K\xb6\xa9x+\x01M\xcb\xf5\xc6\xdd\xca\xc7Ls\xf3\x13N\xf0\xea\x19\x9830"\x19\x92\xb8\x97\x9c$\xf2\x1d1f\x0c\xc0#\x039\x96QZ\x95\xd1\x15\'/"\xa7%\xc4b\x8ab;:\'\xd3\x7f\xb5`\x0cG\xac8\xcd\x06y\xdcM\xae%&\xff\x87NL+\x9f\xf8\x9e\x8e\x8e\xc3\x01p\x8fpGQ\xa6?\x8a\x08\xacZc\x13\xe9\xbc\x7f8\xfc\x9f{\x9b\xf1[\xd7\xdeg\xe9d!Y\xee\xbd1?\xba\x16\x87\xf5Y\xff\x81\x81\xfc\xb5O\xc3\xf0^S\xf4^y\xd3vY\xe8\xf9_\xab\xa5!\xfeB\xaa\xa4_\x94lR\x1f}\xf5\xcd\x1a<\xaf\xae\x88`\xcc\x94T\xc3\x94\xa5\xd59\xb4\x8d\xe0a\xdf}\xf9\xbfMWX\x90\xd1\x16\x84!",%\x19\xa0es\xbd"\x1f8\x80\xa0\xbcN\x84\x90\x95\xa0\xe6E\xf9\xe6\xa8|\x1erD9\x8b\xa3\x04\x1a_\xa6\xda\x96\xff\xaa\xcb\x97\xfe~\xa1\xf1\xcd\xfcD\xf6Up\x18$/\xa1c\x17\xf4q\x10\x10@\xa3v76\x8f\xcf\xb0Nl\xaf\x05dY{\x8b\x97\x17\x88\xd1\x83M\x9e\x1e\x1cTn\xa3F\xcc\xe6L*\\\x93#\x8aK\xb6]S\x94\x10H7\xc7\xd8\xf4\xb9\xaa\x7f\xcc\xe9 7a\x84>\x11\x137KT\x8a&\x918\x81{rD\xe7p\xc1\xf1\xe9A\x11beu\xb0\x8b}\x8c\x83\x85\x86\xce\xebP\xbb\x91\x0c=\xa4\x00-\x81\xe4\x82+\x1c4\x84\xdd\xe9\xac\xdf\xe0\x8c\x83\xfd\xe0G\xc4[\xc9\xd3\xd5\xa4\xb7A\xb6\x05\xd8\x92\xc3ov\xd1\x10P\xd6\x80$hhTe\xc9\xf5\xee\xcf\xc1\xf9\xfb\xa6\x10V\xf4S\xbf\x897\xe3\xaa\xc2_\x9b\xbb\xa2\x84\xad\x0fSK\x9b\x9f*}\x910\x81\x8fX\x89w\xc6\xbf\xe3{z\xb7w\xdbX\x8fu\xb3\xd6F~\xb9\xd6\\]\xfd\x92\xc2\xef\x13\xe5\x83\x1f\xaf?\'\x176Gm#\x1e/\xdc\\O\xa6\xc6\xea\n\xe3\x17{$\x94S\xea\xf5\x89\xf5\xf7\xe3^_\xb3]\xae\xc4\xa7\xf4\xc2\xf4\xdd\x93\xd0\xf4\xa23@T!\rlU\xbd)\x86\xbd\xcc\xd8\x15Z\xe4E\x0f\x15\xed%\xd92Ae\x0b\xd0\x17\x1cm\xb7\xb1\xbeb\x18O\xdf\xa5\x12A\xbc6\x0566\x0c\\\xda+;`\xa4\x19\x0b\x95-\xee\xce-\x8b2\x06Cb@\x8b\x06\xca\xfa3@\x1c(\x98h\x9c\xda\xa82\xfd(\x18\x98\xc7\xda\xfb\xaa.\x9el+\xd8\xcc/{C7W^\xff\xe1\xcb\xd8\xe7v\xdc\x15\xdfn\x04\x0e\x04\xe5uK\xdb\x8b\x83\xf5\xe4\xedT\xd9\xa5\xe1\x06d<.\xdc\x023\xbe,\xab*\xf7\x8e\x01\x05\xc8\xc8\x80\xfd\x0c\x89\x1f)\x85\x1f\x98\x94v\xee\x8d\x95\xc1U\xee\xfdK\xe2\x84\xcf\xa8w\xb7X\xd7\xe5\x95 \x0ei\xfd&\x81Q\x9f~U``\x07\xeeb\xbc\xd5HnF\xe8\xc8\x8c\x1duY\x14\x16C\xcf.xP\x89D\x90\xd5\x01\xbc\xdd\xe5\x18\xaeA\x93&\x00q\xbd\xf0\xda\xd3M`\x9f\x99\xe1\xfe\x0f\xbf\xec\xf9\xf5\xce\xe7\xe5\xecx\xd9\x08x\x18W\x81v:\xdf\x91\xee\xf3\xa2\x97\xd7\xd8\x158@\'\x82\x00\xd8l\xb0k\xb0 \xcd\x00_nXW\xae\x12\xa4\xb8\x02D JPa\x9fj\xc8\x07\xb8\x1a\x03J\x17\x0e\x1f\xf1\xcdH\x0b\xfc\x8atx\x9dx\x1d#\x91\xd9\xec\x81\xf2\xe2\xa9\xad\xb7N\x94\xb1\r\xaa\'\xb5\x1a\xb8\xb5\xed\xa6\x05\xb5\xfa#a\xd7#Al\xec\xb3\xa7\xa3{g\xc8\xf1\xb8\x1a\xb5\x10\xd5D\xd5ia/vX\xb0\x19pl\xc5\x8f\x9c\xef\xd91\xcdl\xd6\x83t,\xf9\xc6f]\x19\x9a\x92\x89\x81\xa8\xc2\x071\x13N\x92\tu"\x98\xf8\xaa\x92\x0c\xe8\xf1\x86\xed\xc1.\x1aH\x92g*\x84\x95U\x14LQN\x81\x89\x0e\xa6D\xa6\xdc\xd5\xcd\xa6\xaaiL\xb2^\xf1U\xf5KH\xc6:\xa1:m`R\xe7Y\x18\xa6\xab\x17\xd6\xb1\xb6\x0c\xcb\x1b\x9cU\xc1E\xef\x87V\x91z7U\\\xdd\x8b\xe0~\x1a\xa1\xe4 =\x1a>\x8a\xee\x00\x05!B\xb0H\x00\xd9P\xba\x97\x05\x0f\xd9\x12\xe2\xbcCN\xcc\xe0\xcb\x1b_ \xe2\xf9\'\xaf\xde\xcet9\x9e&l\xb83;;\xc03/\x1d\xc4X\x06\x7fU\x8b\xefa4\xe1\x12\xd0\xd1o\xbc\xfb\x970\x80\xa8\xa0K\xbeV\x16r\x9aZ\x03\xc3\xd3\t#\xc6\xe1\xdf\x04dS<\xc6\x85\xc1\xd1YB\xc0\x81\x98\xf1C\x145\x1c"\xc2\x18\x9f\x1bI\x90t\t\xeaH\x81\xaa\xa1\t\xbb\xe3v\x1dnpD\xe7D2\x93\x88\xf9y\xa8\xe7c"u\x14Q!\xc8\xbe\xfb\\\xe9\x9d6\xd3\x91\xa9f\x05 \x18\xc7\xa0z\x9aE4n\\\xf9\xa0\xafH\x90\x80\x82\xc6\x89,U\xc4\x90\x81+.\x00\x92V\x9d\xbcv\xadL?\x07\rc\x14\xcbx\x9a\xe3Qr\xd8\xd1\xc3v\xf5\xb9\x9c\x94\x99\x07*\xa2\x81Pz(u*%n\xd2\x8adb\x02\xc3\x9eRY-P6\x13\x1c\x14C0\xa2\xacU@3d\x02\x86\x01\x994GsI\x19:p\x9a\xe7\xdd\xa1\xbe\xb2\r\xf2&\xa0\xdbk\xfa!\x1d;\x89\x15\x99\x17L\x85\xd2\xedn \xc4\x0b\x98\x17;\x041m\xb3\xea]m\xc7\xe8\x1fE&h\xd6\xaa\xaaE\xe7\xa0\xde}\xf1\xf0:_R\xee\xa6\xea\xf9\x1eF:\x03y\x9a\xc6\xfc\xd9\xdb\xd6\xb9[^E\x17/\n\x9bp\xad\xcb\xf93\x1f3\x8c\x1e\x9b\xc8\xc2\xd0XL\x08.\\\xb4p\x06d`KI\x90\x96\xb14\x05\x14U\xb6t\x05\x865\xcc\xf4\xdb\x96\x11s\xa2m\x13\xbes\x14\xd4\x94m\xd8\x06\xa0\xf0\x08\xe6%\x16\xfbF\x82\xbd-\x9f\xec\xdd>\xa4\x80)\x94 l\x8c\x05\xce\xce\xd4N\xda\x13L.|\xb7\xfb\x89\x9aG\x0f\x03@\xee\xe0\xa5\xac=t\xf1\x1b\xb1\xbc\x0b\xbd/\x1bj\xf0\x82\x02\xee\n\xe1\xb4\xbb\xa7b\xee|{\xa4\x1c^\xe3\xb5?\xd6^\xf4\xfd\x1dq\x83\xc9`\xc0.\x0c \xaa\x89j\x11\'\x0fZ\xcf\x12\xecn\x8c\x0c\xc8\x1f\xbaA\xfb\x90\xd4\x00T\xd3\x95n\x8b\x88)\x91w\xf4\xaf\x9eg\xadi;\xd2\xfc\xf4\x15\x9f\x08\xfdfd\x1a9\xa9=*\xc1\xe4DV\x0c<\xc0\x1c\xd8 2V\x9e\xbdY\x13\x96\xcd\x10\x91\xeeb8\x8a\x0c\xa6\xa2\xa5 7\xe1\xf8+9\xd25\x82!\xce\x1e&/\x1b\x1b\x88\t>\xa1t\xc9e\xe4G\xbd\xe8f\xcb\xc8oW0\xf2\xc6\x100f-{\xd0\x80\xf5m\xf6z\xdbg\xd3\xe5\xfbq\x8f\xdc\xf3\x17\xf4G\x97F\xff\xbf\t\xc28\xf4\xb2\x8cy\x86@\xd6\x10\xc0`\xbc\xd8\x17\xb9\x19\x9f\x0en\xb8\x98|L\xc0k\x81C!f\xb7\xb8\xa3\xe0-[\x1a\x82\x84\x17\x8c\xceq\xac\x9c~\x9d\x01H\x06\x14j\xc3\xd8\xed\x02)\xca\xd0q\x00*\x037=B\xae\xc9\xa1\xbbn*\xfd\xea\x87\xc0\xb0|\xbb\xa8\x9f4\xf3\x1d\xac\xac\xc7\x06\x97\x90\xa0\x80\'\x927\xab\xe9$&0\x81\xa1A\x97Fv\x80WWS\xc4\x01\x80f\xd2$\x05q\xefH\xbe\xbd6\xd9\xf6\x9a\xb9\xe6\x81\x0f\xa58\xa0\xeb\xa0A\x84\xb7$\x88Ts\xf4\xb8\x1df0\xf4\xcdixt\xaa\xa5\xf5K\xef\x11\xd8gc\xaeOsj\x0cBq\x12f\xf5\xd3\xd2@\x1c\xf6@\xacn\x99\n\x95\xa7\xb4\x95\x9dT\x8b\x08\x81vr!\xaeb\x99\xa9\xd7gi;&\xd8?\x7fYq\xffd\x0fvg\xa8\xec\xba\x81\x92\x17\x03\x19\xf0\xab\x85K\x8d]e\xf4v|R|\xe6\xc9<\xb8\xa95"\x8cST\x94bx\xca\xc9\x95\xfd\xa8\xe2\xb2n\x08\xa3 \x05\xd9g\xdc\xd0\xffc\xd7\xe6d\x0f.kL\x80\x19N\xf1\xe8\xcc;8\x93\xa1\xb4q\xfc@\xd2\x02\xcf\x0c\x10\x99h\xf30\xf3\xd8E\x87\xb1\xe6\xeav.\x87\xbe\x80-\xd7\xddn\xc6\x94\xef\x87a\xce\xda\xa2_\x90<\x1d\xe1\xc5\x00\xf9hg\xb3\x010\xd4!\xb5T\xa6\xcca8\x89O\x04\x12\x9e\xd7\xa5\xa4\x9d\xb7\xc8\xb7\xac\xbd\xd6w|Vppq\x82\xb98\xf3\xe6\xd9\x16\x02=\x1fGY\x9bm!\xae\xe8N9\xc1C\xc6s\x02\x8bZg\xfa\xf5{\'e~\xb2\xd3@"\xe1\x80<M\xd5\x03\x1ep\xa2&\x06W\x93\x9a\x80\x9edK\xda\xef\x06\x07\xb1\x1a\x07\xcf\xd9=\x9c\x9a\xa9\xee\xa5wg\xf9\x9c\x8e\xb3\x86\xc1\x81\xd2P\x9b\xf1aspa`\xa5\xc8\x12\x08\x12\\\xd1\xeel\xfcZ\xaao\x0b\x1c\xc5h\xe6\x04;\x99oG\xe1\xd7\x07\xfc\xd3\xa1\xcdH\t]\xa8[\xae\x1b\x96?#\xd4p\xde\xdd\xc7\xe3;\xda\x81\xbc\xa9\xf2{\x06\x9b\x1c7P\x14r\xb4\x89\xcc\xb9G^\r\xdc\xc3\x98v\xe4\x18\xf0d\xc5\x03c\xd1\r\xcc\xeen\x19\xb5\xa39\xbajo0\xb6\x06\xd7o~\x1b\x83\x12\xa6\xa21\xea9\xe9\x93\xe2\xfa\xa35\xf5\x17\xf6\xd7=\\\xdd\x14k\xf48\xa5\xbbS\xc1\tV\xd0!6\xfe\xae\x17s\xbb\xf5M\xb3\x99\xf4Fs$\xf0w\x1472\x14\xe76e\'@\xda\x88N*,t&,}\xe6\xd4k\x8b`\x0f\x18s\xa3\x80S\x8e\x01\x85\x0c\x01\xbf\xa9\xddN\xc4\x14\x91;\xc5^\xd9\x89u\x82\xb5BP\xb8G[\x06\x88\xb4!3\xc6P\xef\xa5\x17"\xa5\xf4\x1c\xac\xadDJ\x14%v\x9b\xa1\t\xcfj\x8a\xbb\xb0:\xee\xf0\xcb\\\xcb\xa09\x08\x07}\x9d\x1c\x0b\xe0\xa7Uh\x87a\x87\x8e\x1e\xa2~\xcd\x14\xef\xd3V\x7f\x7f\xd9`@U\xcdP\x84\x92L\xb0a\x17\xa3\x9a\xa6E\xf7N\xbeeK\xa4;\xa6\x06\xb2HI\xc6\xc1\xd4C\x06+2[|\x9b\xd4\xfd@gpeR\xbbqd\xd6"\xb0D\x83\x85i\x0fF\xee\xc6,\x90\xd8@\xb4[K\xb7\xc4\xad+3>\x05\x10\x852\x0eC\xa8|\x9b\xdc\x84\xbe\xf1\xe0\x9f\xb1>>,\x14\xbbp\xb8\xeb\x83\xa3q\x9b\xf8\x17O\xd3\xcd\xd19u\xaf3;{\xdar3\\\xc7n|=\xaf*}=u\xd4n\r\xda8[\xce\xea\x18}7 \xb3\xd2n;{=\xcf\xa7C\xb8\xc3\xad\xe2}\x99`%\x17u\xfb{=\xf6\x94\xcd\x86:D\xee\xcb>\x93\xf3\xb9\x16H\x1b(\xd6\xce>\xeb(\x9b\xe6\xeb\xe3r\xf8\xfc=\xb2\xc1\xae}\xba\xe0\x1a\xde)\xc91\n\x822\xe4\x1c\xb6^\xbe\xecf\xf8p\xc7[\xcf\xe6\xee4t\x06\xa1\xda!F\xee\xd0s%|\xb6\xf0HG\xa8\xbe\x99\x9e,z\x8f\xc0\x83\xd9\xcf\xeb\xf9\xd3\xb1\x9a\xdb>\xd3\x07\x1e\xac\x00q(\x81\xb3\xba\xe2\xb1\x7f\x84\xc5\xc6\xa0\x1f\xf0\xa0\x0f\xb3\xaf\xee\xdetHB\xce\xa8\xa3\xcc\x80\xc8`\xe5d\xe6#\x18\xc8O7\xb18\x92\xd78`\x14\xb47\x97\x91\x8d\xae\xb7WZ\xa4\xa3\xa7t\x89dX\x99\xd8\x01=\x059_eA\x12\xb8\x89\xca\xe2\x18\x87\x85@\xba\xc3\xc9\x0c\xa5D\xab\xcb\xfd|\x13\x83K9W\xa0\xd0\xb6\xf2\x16s\x85\x94\x8a\x91\xde?I@\t\xc0\x15l\xa4\x95\xaf\xb3a\x00ec(c\x91\x16\xf1\x86\x83\x11\xe1\x00\xc1\xacW\xd9J\x07\xb0\xed\xf9\xed\x88ns\x9ezDQ\x87\x91\xf0\xad_Gna7B\xde\xc2j\xbeE\xbcls\x15\\\xbe\x11M\'d\xbc\xc9\x00\xd6~\x12\xe4sswJ\x88HxP\x8a\x0bld%\xaey\n,\xbc\x9c\x16\x12\x92\xd5\\\xa4\x9b,m\xcc(\xa2r!\xf2\x01\x00$\xea\x12.!k\x81\t\xe8eT\x9a\x12\xf9\x16\xf5/Q\xcf\x92,T\xe6D\xad\xc1\xa9\x9f\xafk\x85\x87i\'\x17/\x87E;\xb3\x82\xcc\xb8}\x8f\xba\xa1\xde\xd9\x11\xf8<\xe3\nI\x9f=r\x90\xcf9\x8cP\xadF\x01o\xe9Q\x87\x11X\xa2\x8d:}\x054v\x02\x81\xbd\x0c\xa36!\xb4\x14\xd3\x86@J\t\xc4\xb8~`\x82I\x92\xb4\xd5\x07\xc3\xcc\x04\xde\xd3_t\x90\xadE\xbdu\x8b\xdfE\xf8y\xf2:f\xd1:n*q\'#\x95\xcd\xa1\x86J\x16#\x1e\xca\x04c\t\xe6\x03\tR\x8c5\x9a\xf4\xd4\xa3\x17\xde\xc9\xbd\xc8\x836\xab@\xe3\x99H[\xca\xaan\x11\xe7\x16\xad\x08\xf3\xf4\xc1\x1e\x07\xc9U\x950A\xb8t\x17\x1b\xfc\x8d\xf8=\x08\xcd\xa0N2R\x13{\x0f\xa1\x8c\x82}\x16\xa7\xe7o\xf3@\xd1M\x97H{\x83\x92z\xb6%\\q\xcf4{\x1cS\xc8[\xe5g.\x16\xeawS\xfd\x1c\x8b\xaa\x01\xadeO\x96\xbaR\\)\xf2\xd7\xb3\xb6Y2e\x9a)9\x82F]\x11\xc4f\xaen\xd4\x95\xa9)\x15\\[IP\xb2A\xda\xdf\xe8g\x87D\xc0\xe3\x1cY\xcc\x15H\xf6\x8e\x81A{\\\xeb\xcd!&\xb4\x82\x02\xf4`\xad\xe3:\xb2p\x87\x91\xbef\xb5\xd7\xa8\xb9<\xe2\x93\xb55P\xf3\xd9\xb4\xb1\xa3\n\xe3\xf3T\xb0\xbc\x84\xab7u\xb5\xf6\xd4\xb8\xde\xa7eOBv\x12\x831p\xb2C\x85\xd3j\x15TI+\xa9H\x1b\xba\xb7(U\xb1y\xcd\x91\x9b~\xb8G\xc8F\xd8?p\xf5\x9bh\x1e\x9a\x8ax\xe4\xaeXaa\xb2\xa3\xb7\xba\xd6\x8bv\x82\x00lFS!{\x86\x83\x94\xcdd\xb1j5L`6d\x81S\xaa\n2\x99\x99\x11\xcf/3\x15\xf7?U\xb3\x03\xa8g\x95Cm\x1f!9\x93qZ#}\t\x0c]}>*\xf1\x8f~\xb5]LG\x8d1\xf4\x1a#j\xb0@b\xce\xf2\x95\x11A},||\xc5\x06\xcbs\xbb\x82y`(Y\xf2\xe1\x04\x99n\xb6\xb3\x01\x8b\x11k\nK:\xda\xcb`!\xef\xa7x`\x10\x1c\xd4_\x05uUa\x85C\x1c\xce\x95\xec\xe2\x15\xe9@\x08\x88\x0fN.\xe6\xb8\xf5\x07\xb7R!\xe4C\xe7\xacQ\xb4\x9a\x08\x84\xf5T\x8a\xb9\x10\x91"\xf3\x88\xb9y\xf2\xe5\x90\xe6\x01\xd2\xf1\xa1^pc\x07\xa5\xec9\x16\xec.\xf7\x92\x9e\x06\xcbW\xa2\xb3h\xfa7z\xc0\xdd\xab\x8c&\x9a\x0b\x13\x97\xb9TBdq\x9d2\xf6K\xbbb h\xea^\xed\xf1\x97!=\xed\x8b\xf9\xf2\x0c[t\x16$\x11\xde\x16\xc4\xfb44\x11{\xd0CsJ.7[\xa6\x97\xdd\x04\xd6g\xbe\xd3Kk\x1b\xfc\xca\xfd\x9f\x88\xbd\xe0]\x0b\x1d\xf6\xa3\x04\xda\xb0M<4P>\xdba\n\xfe\xf3\x9b\xf7\xdf\xe4\x1f\xc8:\xc3Ix~\x01\x89[\xdc\xc3XV]\xb8\x82PG0\xcb\xdc\xacww\xfc`\xe28\xdc\xdb\xeeh\x04*\xfdq\xd6\x1da\x8c=\xa0Dy\x08\x0e\xa7\xb0 c\xcb\x91N\x17}\xaa\x8e\xbbp\xf1\xe6\x93\xee\xa8\x90\x13.\xae\xdf\xc8\xbb3\xc6\x08\xfd?\r,1\xbfO8\xaf\xcdIa\xf7\t\x96/\xe2f\x03\x0cR\x18\xd8\xbfq\xb1\xca\xc6\x03T\x98\xc7,@(\xca\xbe`\xdb\x10R\xee\xfa\xc4\xbb\xbf5\xfd\xf0\x97~\x1f\x00a\x91\x84\xe9\xac\xd8y\xfd?u\xe6\xee\x0fy\xb5\xbe3\x0fh\x7fp\xd3\xa4\x0e\x86\x14a\xee\xa9\xaf{A\t\xf7\xb8\xf4\x96\x97@;\t\xf9\xbc\xaeOK\xfe\x7f\xe1\xf6\x19\xf3\x1e\xe7\x99J\x16\xc9\x1d\xee\x9b\xe6\xde\xc9\xa40\xe9\xf5\xf88\xa9\x82\x04\xf0l\x81\x89\xa8\x01\xc9S\x17\x9b1\xae\xfe\xbf\x08\xa3\xccG\xf0\x17\xee\x17\xe0\xa2\xf1\xfa\x9f\xe5\xeb\xed\xbc\xec\x90@\xf9\xc7xl\x0b\xe6\x07^\x142\xe9\xc5\xad\xda\x9c\x1d\xd8@\x86\x1e^\x0f\xd4\xd9\xfd\x9d\x81A\xe5\x06\x08m\x8b\x05\xa5h\xa0^\x19\x01\x07\x07\x86\x86,^\xa2\xa6\xdc\xe7\x1e\xca@D\x18\xf4\xef\xa0~\xaf:\x02 4\xea\xd5\xf8\x9f}\xa6\x80X\xce\xf6o\xd4\x8d\x14[\xf5\xd0\x12\x98\xaa(f\x08\xdd\xb4\x04\x8c\xec\xb8\x7f\xed\xee\xf5\x9e\x04@:\xb2\x0c\xb7p\x06\xb9\xd1qA\xe4\x8f\xd8\xce\xe6\x01vq\x9ch\x9f\x0b\x04\xcf\x01\xb0\x80k\x80\xd0\x8c0"\\\x03(MV\x9a)\x8d\x16[\x0b5?\xe0\x19\xf31Q\xfan\x84\xa8\x89\xb9B(5Rj\xf0k\x01\xc3\x97\xc7\x90\x1e\x0f3\x1d#~\x0f\xf1A\x926t\x16\xd8\x11\xc8\x12\x02#O\xd2\xf8\x87U\xc0!\xb5\xeb\x01\xb2\xc9\xa7\xa7\xd9\xb8\xec\xf2\xfe_\x8fQ)\x06P\xb2\xe6\x80!\x84\xbc\xbd"\x91\xe0C\x81\x19\x0e\x19\xc9\x14L\x92\xec\x1a\x83XsiB\'\xe1\xc2)TA4p\xc5\xc4\x91!\xd3EVJ\x13\xc6\x9b\xd8\n5\x81\x00\xc8\x00,\xd5!G\xc2\x1b\xd1F\x00\xc3\xbc\xaaZ\xcc\x91\xf1\xb3\xdd3\xf4\xc33|\x8f\xa1\xcf\x14\xfe\xf7m\xc2\xdez\xfa\xe1p\x97\x89\xc0\x13\x95\xc5\xb5\x96', 142, 143, 'character maps to <undefined>') command: C:\ProgramData\Anaconda3\Scripts\conda install --file https://repo.continuum.io/pkgs/free/win-32/pyyaml-3.11-py27_3.tar.bz2 user_agent: conda/4.4.7 requests/2.18.4 CPython/3.6.3 Windows/7 Windows/6.1.7601 _messageid: -9223372036837525397 _messagetime: 1515518595057 / 2018-01-09 11:23:15 Traceback (most recent call last): File "C:\ProgramData\Anaconda3\lib\site-packages\conda\exceptions.py", line 743, in __call__ return func(*args, **kwargs) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\cli\main.py", line 78, in _main exit_code = do_call(args, p) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\cli\conda_argparse.py", line 76, in do_call exit_code = getattr(module, func_name)(args, parser) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\cli\main_install.py", line 11, in execute install(args, parser, 'install') File "C:\ProgramData\Anaconda3\lib\site-packages\conda\cli\install.py", line 182, in install specs.extend(common.specs_from_url(fpath, json=context.json)) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\cli\common.py", line 126, in specs_from_url for line in open(path): File "C:\ProgramData\Anaconda3\lib\encodings\cp1252.py", line 23, in decode return codecs.charmap_decode(input,self.errors,decoding_table)[0] UnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 142: character maps to <undefined> ```
The `--file` parameter on `conda install` isn't meant for something like `conda install --file package.tar.bz2` or for `environment.yml` files. Basically, if we have a problem parsing the contents of the file around https://github.com/conda/conda/blob/4.4.11/conda/cli/install.py#L182, we should raise an appropriate `CondaError` with a helpful message. similar to #6697
2019-02-23T18:28:46Z
[]
[]
Traceback (most recent call last): File "C:\ProgramData\Anaconda3\lib\site-packages\conda\exceptions.py", line 743, in __call__ return func(*args, **kwargs) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\cli\main.py", line 78, in _main exit_code = do_call(args, p) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\cli\conda_argparse.py", line 76, in do_call exit_code = getattr(module, func_name)(args, parser) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\cli\main_install.py", line 11, in execute install(args, parser, 'install') File "C:\ProgramData\Anaconda3\lib\site-packages\conda\cli\install.py", line 182, in install specs.extend(common.specs_from_url(fpath, json=context.json)) File "C:\ProgramData\Anaconda3\lib\site-packages\conda\cli\common.py", line 126, in specs_from_url for line in open(path): File "C:\ProgramData\Anaconda3\lib\encodings\cp1252.py", line 23, in decode return codecs.charmap_decode(input,self.errors,decoding_table)[0] UnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 142: character maps to <undefined>
4,421
conda/conda
conda__conda-8917
296a67dd74c1c6c78058c1b94d6aeec33691fe3b
diff --git a/conda/resolve.py b/conda/resolve.py --- a/conda/resolve.py +++ b/conda/resolve.py @@ -113,7 +113,8 @@ def __init__(self, index, processed=False, channels=()): self._pool_cache = {} self._strict_channel_cache = {} - self._system_precs = {_ for _ in index if _.package_type == PackageType.VIRTUAL_SYSTEM} + self._system_precs = {_ for _ in index if ( + hasattr(_, 'package_type') and _.package_type == PackageType.VIRTUAL_SYSTEM)} # sorting these in reverse order is effectively prioritizing # contstraint behavior from newer packages. It is applying broadening
Errors with conda 4.7.6 on conda-forge builds: AttributeError: 'Dist' object has no attribute 'package_type' It seems that the conda 4.7.6 or conda-build 3.17.8 upgrade is causing errors on all recent conda-forge builds: https://dev.azure.com/conda-forge/feedstock-builds/_build/results?buildId=49363&view=logs&jobId=10bda42a-d420-5c28-a43f-bc4e66405e5c&taskId=bef775d0-a1a2-55a3-9721-17147029baf0&lineStart=256&lineEnd=274&colStart=1&colEnd=62 ``` + xargs -n1 cat + xargs -r /usr/bin/sudo -n yum install -y + find /home/conda/conda-recipes -mindepth 2 -maxdepth 2 -type f -name yum_requirements.txt + grep -v -e '^#' -e '^$' + test 1 == 1 + python /home/conda/.ci_support/build_all.py /home/conda/conda-recipes Building ipymidicontrols with conda-forge/label/main Traceback (most recent call last): File "/home/conda/.ci_support/build_all.py", line 131, in <module> build_all(args.recipes_dir, args.arch) File "/home/conda/.ci_support/build_all.py", line 68, in build_all build_folders(recipes_dir, new_comp_folders, arch, channel_urls) File "/home/conda/.ci_support/build_all.py", line 94, in build_folders conda_resolve = conda_build.conda_interface.Resolve(index) File "/opt/conda/lib/python3.7/site-packages/conda/resolve.py", line 116, in __init__ self._system_precs = {_ for _ in index if _.package_type == PackageType.VIRTUAL_SYSTEM} File "/opt/conda/lib/python3.7/site-packages/conda/resolve.py", line 116, in <setcomp> self._system_precs = {_ for _ in index if _.package_type == PackageType.VIRTUAL_SYSTEM} AttributeError: 'Dist' object has no attribute 'package_type' ``` Some examples of builds having this error, all having conda 4.7.6, conda-build 3.18.7: * https://dev.azure.com/conda-forge/feedstock-builds/_build/results?buildId=49404&view=logs&jobId=10bda42a-d420-5c28-a43f-bc4e66405e5c&taskId=bef775d0-a1a2-55a3-9721-17147029baf0&lineStart=183&lineEnd=185&colStart=1&colEnd=1 * https://dev.azure.com/conda-forge/feedstock-builds/_build/results?buildId=49419&view=logs&jobId=10bda42a-d420-5c28-a43f-bc4e66405e5c&taskId=bef775d0-a1a2-55a3-9721-17147029baf0&lineStart=183&lineEnd=185&colStart=1&colEnd=1 * https://dev.azure.com/conda-forge/feedstock-builds/_build/results?buildId=49420&view=logs&jobId=10bda42a-d420-5c28-a43f-bc4e66405e5c&taskId=bef775d0-a1a2-55a3-9721-17147029baf0&lineStart=183&lineEnd=185&colStart=1&colEnd=1 Examples of successful builds just before these, all having conda 4.7.5, conda-build 3.17.8: * https://dev.azure.com/conda-forge/feedstock-builds/_build/results?buildId=49197&view=logs&jobId=240f1fee-52bc-5498-a14a-8361bde76ba0&taskId=2fb5b0a7-737c-5f5c-8f3f-f6db6174bacf&lineStart=99&lineEnd=101&colStart=1&colEnd=1 * https://dev.azure.com/conda-forge/feedstock-builds/_build/results?buildId=49327&view=logs&jobId=10bda42a-d420-5c28-a43f-bc4e66405e5c&taskId=bef775d0-a1a2-55a3-9721-17147029baf0&lineStart=183&lineEnd=185&colStart=1&colEnd=1 * https://dev.azure.com/conda-forge/feedstock-builds/_build/results?buildId=49127&view=logs&jobId=10bda42a-d420-5c28-a43f-bc4e66405e5c&taskId=bef775d0-a1a2-55a3-9721-17147029baf0&lineStart=183&lineEnd=185&colStart=1&colEnd=1
2019-07-12T14:20:14Z
[]
[]
Traceback (most recent call last): File "/home/conda/.ci_support/build_all.py", line 131, in <module> build_all(args.recipes_dir, args.arch) File "/home/conda/.ci_support/build_all.py", line 68, in build_all build_folders(recipes_dir, new_comp_folders, arch, channel_urls) File "/home/conda/.ci_support/build_all.py", line 94, in build_folders conda_resolve = conda_build.conda_interface.Resolve(index) File "/opt/conda/lib/python3.7/site-packages/conda/resolve.py", line 116, in __init__ self._system_precs = {_ for _ in index if _.package_type == PackageType.VIRTUAL_SYSTEM} File "/opt/conda/lib/python3.7/site-packages/conda/resolve.py", line 116, in <setcomp> self._system_precs = {_ for _ in index if _.package_type == PackageType.VIRTUAL_SYSTEM} AttributeError: 'Dist' object has no attribute 'package_type'
4,443
conda/conda
conda__conda-8999
a4f8a776d78e47d3665e0adcf175139d718d1a5a
diff --git a/conda/cli/install.py b/conda/cli/install.py --- a/conda/cli/install.py +++ b/conda/cli/install.py @@ -229,6 +229,8 @@ def install(args, parser, command='install'): elif REPODATA_FN not in repodata_fns: repodata_fns.append(REPODATA_FN) + args_set_update_modifier = hasattr(args, "update_modifier") and args.update_modifier != NULL + for repodata_fn in repodata_fns: try: update_modifier = context.update_modifier @@ -281,9 +283,12 @@ def install(args, parser, command='install'): if not hasattr(args, 'update_modifier'): if repodata_fn == repodata_fns[-1]: raise e - elif args.update_modifier == NULL: + elif not args_set_update_modifier or args.update_modifier not in ( + UpdateModifier.FREEZE_INSTALLED, + UpdateModifier.UPDATE_SPECS): try: - if not args.json: + if not args.json and (not args_set_update_modifier or + args.update_modifier == UpdateModifier.FREEZE_INSTALLED): print("Initial quick solve with frozen env failed. " "Unfreezing env and trying again.") unlink_link_transaction = solver.solve_for_transaction( diff --git a/conda/core/solve.py b/conda/core/solve.py --- a/conda/core/solve.py +++ b/conda/core/solve.py @@ -195,8 +195,10 @@ def solve_final_state(self, update_modifier=NULL, deps_modifier=NULL, prune=NULL the solved state of the environment. """ + update_modifier_set = True if update_modifier is NULL: update_modifier = context.update_modifier + update_modifier_set = False else: update_modifier = UpdateModifier(text_type(update_modifier).lower()) if deps_modifier is NULL: @@ -248,8 +250,14 @@ def solve_final_state(self, update_modifier=NULL, deps_modifier=NULL, prune=NULL context.json): ssc = self._collect_all_metadata(ssc) - fail_message = ("failed\n" if self._repodata_fn == REPODATA_FN else "failed with %s, " - "will retry with next repodata source.\n" % self._repodata_fn) + if not update_modifier_set or update_modifier != UpdateModifier.UPDATE_SPECS: + fail_message = "failed with initial frozen solve. Retrying with flexible solve.\n" + elif self._repodata_fn != REPODATA_FN: + fail_message = ("failed with repodata from %s, will retry with next repodata" + " source.\n" % self._repodata_fn) + else: + fail_message = "failed\n" + with Spinner("Solving environment", not context.verbosity and not context.quiet, context.json, fail_message=fail_message): ssc = self._remove_specs(ssc)
Solving environment: failed with current_repodata.json, will retry with next repodata source. UnsatisfiableError: <!-- Hi! Read this; it's important. This is an issue tracker for conda -- the package manager. File feature requests for conda here, as well as bug reports about something conda has messed up. If your issue is a bug report for: * a specific conda package from Anaconda ('defaults' channel): ==> file at https://github.com/ContinuumIO/anaconda-issues * a specific conda package from conda-forge: ==> file at the corresponding feedstock under https://github.com/conda-forge * repo.anaconda.com access and service: ==> file at https://github.com/ContinuumIO/anaconda-issues * anaconda.org access and service: ==> file at https://anaconda.org/contact/report * commands under 'conda build': ==> file at https://github.com/conda/conda-build * commands under 'conda env': ==> please file it here! * all other conda commands that start with 'conda': ==> please file it here! If you continue on, **please include all requested information below.** If a maintainer determines the information is required to understand your issue, and if it is not provided, your issue may be closed automatically. --> ## Current Behavior <!-- What actually happens? If you want to include console output, please use "Steps to Reproduce" below. --> Cannot install or update packages when conda-forge is added ### Steps to Reproduce <!-- If the current behaviour is a bug, please provide specific, minimal steps to independently reproduce. Include the exact conda commands that reproduce the issue and their output between the ticks below. --> 1. Install Anaconda3-2019.07-Windows-x86_64.exe (https://repo.continuum.io/archive/Anaconda3-2019.07-Windows-x86_64.exe) 2. Enable conda-forge ``` conda config --add channels conda-forge conda config --set channel_priority strict ``` Here's what happens: ``` blim09@DESKTOP-J3L8E29:C:\Windows\system32 $ conda update --all Collecting package metadata (current_repodata.json): done Solving environment: failed with current_repodata.json, will retry with next repodata source. UnsatisfiableError: blim09@DESKTOP-J3L8E29:C:\Windows\system32 $ ``` ``` blim09@DESKTOP-J3L8E29:C:\Windows\system32 $ conda update --all --verbose Collecting package metadata (current_repodata.json): ...working... done Solving environment: ...working... failed with current_repodata.json, will retry with next repodata source. Traceback (most recent call last): File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\exceptions.py", line 1062, in __call__ return func(*args, **kwargs) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\cli\main.py", line 84, in _main exit_code = do_call(args, p) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\cli\conda_argparse.py", line 82, in do_call exit_code = getattr(module, func_name)(args, parser) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\cli\main_update.py", line 20, in execute install(args, parser, 'update') File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\cli\install.py", line 307, in install raise e File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\cli\install.py", line 256, in install force_reinstall=context.force_reinstall or context.force, File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\core\solve.py", line 112, in solve_for_transaction force_remove, force_reinstall) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\core\solve.py", line 150, in solve_for_diff force_remove) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\core\solve.py", line 262, in solve_final_state ssc = self._run_sat(ssc) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\common\io.py", line 88, in decorated return f(*args, **kwds) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\core\solve.py", line 780, in _run_sat repodata_fn=self._repodata_fn) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\common\io.py", line 88, in decorated return f(*args, **kwds) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\resolve.py", line 1219, in solve raise UnsatisfiableError({}) conda.exceptions.UnsatisfiableError blim09@DESKTOP-J3L8E29:C:\Windows\system32 $ ``` Everything works correctly when conda-forge is removed ## Expected Behavior <!-- What do you think should happen? --> Packages should install or update normally ## Environment Information <details open><summary><code>`conda info`</code></summary><p> <!-- between the ticks below, paste the output of 'conda info' --> ``` active environment: None user config file : C:\Users\blim09\.condarc populated config files : C:\Users\blim09\.condarc conda version : 4.7.10 conda-build version : 3.18.9 python version : 3.7.3.final.0 virtual packages : base environment : C:\Users\blim09\Anaconda3 (writable) channel URLs : https://conda.anaconda.org/conda-forge/win-64 https://conda.anaconda.org/conda-forge/noarch https://repo.anaconda.com/pkgs/main/win-64 https://repo.anaconda.com/pkgs/main/noarch https://repo.anaconda.com/pkgs/r/win-64 https://repo.anaconda.com/pkgs/r/noarch https://repo.anaconda.com/pkgs/msys2/win-64 https://repo.anaconda.com/pkgs/msys2/noarch package cache : C:\Users\blim09\Anaconda3\pkgs C:\Users\blim09\.conda\pkgs C:\Users\blim09\AppData\Local\conda\conda\pkgs envs directories : C:\Users\blim09\Anaconda3\envs C:\Users\blim09\.conda\envs C:\Users\blim09\AppData\Local\conda\conda\envs platform : win-64 user-agent : conda/4.7.10 requests/2.22.0 CPython/3.7.3 Windows/10 Windows/10.0.18362 administrator : True netrc file : None offline mode : False ``` </p></details> <details open><summary><code>`conda config --show-sources`</code></summary><p> <!-- between the ticks below, paste the output of 'conda config --show-sources' --> ``` ==> C:\Users\blim09\.condarc <== channel_priority: strict channels: - conda-forge - defaults ``` </p></details> <details><summary><code>`conda list --show-channel-urls`</code></summary><p> <!-- between the ticks below, paste the output of 'conda list --show-channel-urls' --> ``` # packages in environment at C:\Users\blim09\Anaconda3: # # Name Version Build Channel _ipyw_jlab_nb_ext_conf 0.1.0 py37_0 defaults alabaster 0.7.12 py37_0 defaults anaconda 2019.07 py37_0 defaults anaconda-client 1.7.2 py37_0 defaults anaconda-navigator 1.9.7 py37_0 defaults anaconda-project 0.8.3 py_0 defaults asn1crypto 0.24.0 py37_0 defaults astroid 2.2.5 py37_0 defaults astropy 3.2.1 py37he774522_0 defaults atomicwrites 1.3.0 py37_1 defaults attrs 19.1.0 py37_1 defaults babel 2.7.0 py_0 defaults backcall 0.1.0 py37_0 defaults backports 1.0 py_2 defaults backports.functools_lru_cache 1.5 py_2 defaults backports.os 0.1.1 py37_0 defaults backports.shutil_get_terminal_size 1.0.0 py37_2 defaults backports.tempfile 1.0 py_1 defaults backports.weakref 1.0.post1 py_1 defaults beautifulsoup4 4.7.1 py37_1 defaults bitarray 0.9.3 py37he774522_0 defaults bkcharts 0.2 py37_0 defaults blas 1.0 mkl defaults bleach 3.1.0 py37_0 defaults blosc 1.16.3 h7bd577a_0 defaults bokeh 1.2.0 py37_0 defaults boto 2.49.0 py37_0 defaults bottleneck 1.2.1 py37h452e1ab_1 defaults bzip2 1.0.8 he774522_0 defaults ca-certificates 2019.5.15 0 defaults certifi 2019.6.16 py37_0 defaults cffi 1.12.3 py37h7a1dbc1_0 defaults chardet 3.0.4 py37_1 defaults click 7.0 py37_0 defaults cloudpickle 1.2.1 py_0 defaults clyent 1.2.2 py37_1 defaults colorama 0.4.1 py37_0 defaults comtypes 1.1.7 py37_0 defaults conda 4.7.10 py37_0 defaults conda-build 3.18.9 py37_0 defaults conda-env 2.6.0 1 defaults conda-package-handling 1.3.11 py37_0 defaults conda-verify 3.4.2 py_1 defaults console_shortcut 0.1.1 3 defaults contextlib2 0.5.5 py37_0 defaults cryptography 2.7 py37h7a1dbc1_0 defaults curl 7.65.2 h2a8f88b_0 defaults cycler 0.10.0 py37_0 defaults cython 0.29.12 py37ha925a31_0 defaults cytoolz 0.10.0 py37he774522_0 defaults dask 2.1.0 py_0 defaults dask-core 2.1.0 py_0 defaults decorator 4.4.0 py37_1 defaults defusedxml 0.6.0 py_0 defaults distributed 2.1.0 py_0 defaults docutils 0.14 py37_0 defaults entrypoints 0.3 py37_0 defaults et_xmlfile 1.0.1 py37_0 defaults fastcache 1.1.0 py37he774522_0 defaults filelock 3.0.12 py_0 defaults flask 1.1.1 py_0 defaults freetype 2.9.1 ha9979f8_1 defaults future 0.17.1 py37_0 defaults get_terminal_size 1.0.0 h38e98db_0 defaults gevent 1.4.0 py37he774522_0 defaults glob2 0.7 py_0 defaults greenlet 0.4.15 py37hfa6e2cd_0 defaults h5py 2.9.0 py37h5e291fa_0 defaults hdf5 1.10.4 h7ebc959_0 defaults heapdict 1.0.0 py37_2 defaults html5lib 1.0.1 py37_0 defaults icc_rt 2019.0.0 h0cc432a_1 defaults icu 58.2 ha66f8fd_1 defaults idna 2.8 py37_0 defaults imageio 2.5.0 py37_0 defaults imagesize 1.1.0 py37_0 defaults importlib_metadata 0.17 py37_1 defaults intel-openmp 2019.4 245 defaults ipykernel 5.1.1 py37h39e3cac_0 defaults ipython 7.6.1 py37h39e3cac_0 defaults ipython_genutils 0.2.0 py37_0 defaults ipywidgets 7.5.0 py_0 defaults isort 4.3.21 py37_0 defaults itsdangerous 1.1.0 py37_0 defaults jdcal 1.4.1 py_0 defaults jedi 0.13.3 py37_0 defaults jinja2 2.10.1 py37_0 defaults joblib 0.13.2 py37_0 defaults jpeg 9b hb83a4c4_2 defaults json5 0.8.4 py_0 defaults jsonschema 3.0.1 py37_0 defaults jupyter 1.0.0 py37_7 defaults jupyter_client 5.3.1 py_0 defaults jupyter_console 6.0.0 py37_0 defaults jupyter_core 4.5.0 py_0 defaults jupyterlab 1.0.2 py37hf63ae98_0 defaults jupyterlab_server 1.0.0 py_0 defaults keyring 18.0.0 py37_0 defaults kiwisolver 1.1.0 py37ha925a31_0 defaults krb5 1.16.1 hc04afaa_7 defaults lazy-object-proxy 1.4.1 py37he774522_0 defaults libarchive 3.3.3 h0643e63_5 defaults libcurl 7.65.2 h2a8f88b_0 defaults libiconv 1.15 h1df5818_7 defaults liblief 0.9.0 ha925a31_2 defaults libpng 1.6.37 h2a8f88b_0 defaults libsodium 1.0.16 h9d3ae62_0 defaults libssh2 1.8.2 h7a1dbc1_0 defaults libtiff 4.0.10 hb898794_2 defaults libxml2 2.9.9 h464c3ec_0 defaults libxslt 1.1.33 h579f668_0 defaults llvmlite 0.29.0 py37ha925a31_0 defaults locket 0.2.0 py37_1 defaults lxml 4.3.4 py37h1350720_0 defaults lz4-c 1.8.1.2 h2fa13f4_0 defaults lzo 2.10 h6df0209_2 defaults m2w64-gcc-libgfortran 5.3.0 6 defaults m2w64-gcc-libs 5.3.0 7 defaults m2w64-gcc-libs-core 5.3.0 7 defaults m2w64-gmp 6.1.0 2 defaults m2w64-libwinpthread-git 5.0.0.4634.697f757 2 defaults markupsafe 1.1.1 py37he774522_0 defaults matplotlib 3.1.0 py37hc8f65d3_0 defaults mccabe 0.6.1 py37_1 defaults menuinst 1.4.16 py37he774522_0 defaults mistune 0.8.4 py37he774522_0 defaults mkl 2019.4 245 defaults mkl-service 2.0.2 py37he774522_0 defaults mkl_fft 1.0.12 py37h14836fe_0 defaults mkl_random 1.0.2 py37h343c172_0 defaults mock 3.0.5 py37_0 defaults more-itertools 7.0.0 py37_0 defaults mpmath 1.1.0 py37_0 defaults msgpack-python 0.6.1 py37h74a9793_1 defaults msys2-conda-epoch 20160418 1 defaults multipledispatch 0.6.0 py37_0 defaults navigator-updater 0.2.1 py37_0 defaults nbconvert 5.5.0 py_0 defaults nbformat 4.4.0 py37_0 defaults networkx 2.3 py_0 defaults nltk 3.4.4 py37_0 defaults nose 1.3.7 py37_2 defaults notebook 6.0.0 py37_0 defaults numba 0.44.1 py37hf9181ef_0 defaults numexpr 2.6.9 py37hdce8814_0 defaults numpy 1.16.4 py37h19fb1c0_0 defaults numpy-base 1.16.4 py37hc3f5095_0 defaults numpydoc 0.9.1 py_0 defaults olefile 0.46 py37_0 defaults openpyxl 2.6.2 py_0 defaults openssl 1.1.1c he774522_1 defaults packaging 19.0 py37_0 defaults pandas 0.24.2 py37ha925a31_0 defaults pandoc 2.2.3.2 0 defaults pandocfilters 1.4.2 py37_1 defaults parso 0.5.0 py_0 defaults partd 1.0.0 py_0 defaults path.py 12.0.1 py_0 defaults pathlib2 2.3.4 py37_0 defaults patsy 0.5.1 py37_0 defaults pep8 1.7.1 py37_0 defaults pickleshare 0.7.5 py37_0 defaults pillow 6.1.0 py37hdc69c19_0 defaults pip 19.1.1 py37_0 defaults pkginfo 1.5.0.1 py37_0 defaults pluggy 0.12.0 py_0 defaults ply 3.11 py37_0 defaults powershell_shortcut 0.0.1 2 defaults prometheus_client 0.7.1 py_0 defaults prompt_toolkit 2.0.9 py37_0 defaults psutil 5.6.3 py37he774522_0 defaults py 1.8.0 py37_0 defaults py-lief 0.9.0 py37ha925a31_2 defaults pycodestyle 2.5.0 py37_0 defaults pycosat 0.6.3 py37hfa6e2cd_0 defaults pycparser 2.19 py37_0 defaults pycrypto 2.6.1 py37hfa6e2cd_9 defaults pycurl 7.43.0.3 py37h7a1dbc1_0 defaults pyflakes 2.1.1 py37_0 defaults pygments 2.4.2 py_0 defaults pylint 2.3.1 py37_0 defaults pyodbc 4.0.26 py37ha925a31_0 defaults pyopenssl 19.0.0 py37_0 defaults pyparsing 2.4.0 py_0 defaults pyqt 5.9.2 py37h6538335_2 defaults pyreadline 2.1 py37_1 defaults pyrsistent 0.14.11 py37he774522_0 defaults pysocks 1.7.0 py37_0 defaults pytables 3.5.2 py37h1da0976_1 defaults pytest 5.0.1 py37_0 defaults pytest-arraydiff 0.3 py37h39e3cac_0 defaults pytest-astropy 0.5.0 py37_0 defaults pytest-doctestplus 0.3.0 py37_0 defaults pytest-openfiles 0.3.2 py37_0 defaults pytest-remotedata 0.3.1 py37_0 defaults python 3.7.3 h8c8aaf0_1 defaults python-dateutil 2.8.0 py37_0 defaults python-libarchive-c 2.8 py37_11 defaults pytz 2019.1 py_0 defaults pywavelets 1.0.3 py37h8c2d366_1 defaults pywin32 223 py37hfa6e2cd_1 defaults pywinpty 0.5.5 py37_1000 defaults pyyaml 5.1.1 py37he774522_0 defaults pyzmq 18.0.0 py37ha925a31_0 defaults qt 5.9.7 vc14h73c81de_0 defaults qtawesome 0.5.7 py37_1 defaults qtconsole 4.5.1 py_0 defaults qtpy 1.8.0 py_0 defaults requests 2.22.0 py37_0 defaults rope 0.14.0 py_0 defaults ruamel_yaml 0.15.46 py37hfa6e2cd_0 defaults scikit-image 0.15.0 py37ha925a31_0 defaults scikit-learn 0.21.2 py37h6288b17_0 defaults scipy 1.2.1 py37h29ff71c_0 defaults seaborn 0.9.0 py37_0 defaults send2trash 1.5.0 py37_0 defaults setuptools 41.0.1 py37_0 defaults simplegeneric 0.8.1 py37_2 defaults singledispatch 3.4.0.3 py37_0 defaults sip 4.19.8 py37h6538335_0 defaults six 1.12.0 py37_0 defaults snappy 1.1.7 h777316e_3 defaults snowballstemmer 1.9.0 py_0 defaults sortedcollections 1.1.2 py37_0 defaults sortedcontainers 2.1.0 py37_0 defaults soupsieve 1.8 py37_0 defaults sphinx 2.1.2 py_0 defaults sphinxcontrib 1.0 py37_1 defaults sphinxcontrib-applehelp 1.0.1 py_0 defaults sphinxcontrib-devhelp 1.0.1 py_0 defaults sphinxcontrib-htmlhelp 1.0.2 py_0 defaults sphinxcontrib-jsmath 1.0.1 py_0 defaults sphinxcontrib-qthelp 1.0.2 py_0 defaults sphinxcontrib-serializinghtml 1.1.3 py_0 defaults sphinxcontrib-websupport 1.1.2 py_0 defaults spyder 3.3.6 py37_0 defaults spyder-kernels 0.5.1 py37_0 defaults sqlalchemy 1.3.5 py37he774522_0 defaults sqlite 3.29.0 he774522_0 defaults statsmodels 0.10.0 py37h8c2d366_0 defaults sympy 1.4 py37_0 defaults tblib 1.4.0 py_0 defaults terminado 0.8.2 py37_0 defaults testpath 0.4.2 py37_0 defaults tk 8.6.8 hfa6e2cd_0 defaults toolz 0.10.0 py_0 defaults tornado 6.0.3 py37he774522_0 defaults tqdm 4.32.1 py_0 defaults traitlets 4.3.2 py37_0 defaults unicodecsv 0.14.1 py37_0 defaults urllib3 1.24.2 py37_0 defaults vc 14.1 h0510ff6_4 defaults vs2015_runtime 14.15.26706 h3a45250_4 defaults wcwidth 0.1.7 py37_0 defaults webencodings 0.5.1 py37_1 defaults werkzeug 0.15.4 py_0 defaults wheel 0.33.4 py37_0 defaults widgetsnbextension 3.5.0 py37_0 defaults win_inet_pton 1.1.0 py37_0 defaults win_unicode_console 0.5 py37_0 defaults wincertstore 0.2 py37_0 defaults winpty 0.4.3 4 defaults wrapt 1.11.2 py37he774522_0 defaults xlrd 1.2.0 py37_0 defaults xlsxwriter 1.1.8 py_0 defaults xlwings 0.15.8 py37_0 defaults xlwt 1.3.0 py37_0 defaults xz 5.2.4 h2fa13f4_4 defaults yaml 0.1.7 hc54c509_2 defaults zeromq 4.3.1 h33f27b4_3 defaults zict 1.0.0 py_0 defaults zipp 0.5.1 py_0 defaults zlib 1.2.11 h62dcd97_3 defaults zstd 1.3.7 h508b16e_0 defaults ``` </p></details>
2019-07-25T19:47:35Z
[]
[]
Traceback (most recent call last): File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\exceptions.py", line 1062, in __call__ return func(*args, **kwargs) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\cli\main.py", line 84, in _main exit_code = do_call(args, p) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\cli\conda_argparse.py", line 82, in do_call exit_code = getattr(module, func_name)(args, parser) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\cli\main_update.py", line 20, in execute install(args, parser, 'update') File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\cli\install.py", line 307, in install raise e File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\cli\install.py", line 256, in install force_reinstall=context.force_reinstall or context.force, File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\core\solve.py", line 112, in solve_for_transaction force_remove, force_reinstall) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\core\solve.py", line 150, in solve_for_diff force_remove) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\core\solve.py", line 262, in solve_final_state ssc = self._run_sat(ssc) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\common\io.py", line 88, in decorated return f(*args, **kwds) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\core\solve.py", line 780, in _run_sat repodata_fn=self._repodata_fn) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\common\io.py", line 88, in decorated return f(*args, **kwds) File "C:\Users\blim09\Anaconda3\lib\site-packages\conda\resolve.py", line 1219, in solve raise UnsatisfiableError({}) conda.exceptions.UnsatisfiableError
4,448
dagster-io/dagster
dagster-io__dagster-14785
96058ec8434f97130ddac1fcab36102e6e33627f
diff --git a/python_modules/dagster/setup.py b/python_modules/dagster/setup.py --- a/python_modules/dagster/setup.py +++ b/python_modules/dagster/setup.py @@ -106,7 +106,7 @@ def get_version() -> str: "docstring-parser", "universal_pathlib", # https://github.com/pydantic/pydantic/issues/5821 - "pydantic != 1.10.7", + "pydantic != 1.10.7,<2.0.0", ], extras_require={ "docker": ["docker"],
Support Pydantic V2 and/or peg it to V1 for now ### What's the use case? Pydantic is beta testing a V2 version, which replaces the core of the library with Rust and overhauls the API. I'd like to use it as a dependency in a pre-release project to validate my own data, however, Dagster now bundles Pydantic V1 and hasn't yet capped the version. This is what I get when I try to upgrade: ``` Traceback (most recent call last): File "/Users/admin/Desktop/hello-dagster/venv/bin/dagster", line 5, in <module> from dagster.cli import main File "/Users/admin/Desktop/hello-dagster/venv/lib/python3.10/site-packages/dagster/__init__.py", line 100, in <module> from dagster._config.pythonic_config import ( File "/Users/admin/Desktop/hello-dagster/venv/lib/python3.10/site-packages/dagster/_config/pythonic_config/__init__.py", line 22, in <module> from pydantic import ConstrainedFloat, ConstrainedInt, ConstrainedStr File "/Users/admin/Desktop/hello-dagster/venv/lib/python3.10/site-packages/pydantic/_migration.py", line 279, in wrapper raise PydanticImportError(f'`{import_path}` has been removed in V2.') pydantic.errors.PydanticImportError: `pydantic:ConstrainedFloat` has been removed in V2. For further information visit https://errors.pydantic.dev/2.0b2/u/import-error ``` ### Ideas of implementation I think as a pre-release this won't cause problems for anyone except people like me who want to use the new version, but a) Dagster should probably change requirements.text to reflect `pydantic<2` and b) maybe support it when it's no longer a pre-release. The V2 does include some V1 shims, so it shouldn't be too bad, but some parts of the API did change in breaking ways. ### Additional information _No response_ ### Message from the maintainers Impacted by this issue? Give it a 👍! We factor engagement into prioritization.
@benpankow - thoughts on this one?
2023-06-13T23:29:25Z
[]
[]
Traceback (most recent call last): File "/Users/admin/Desktop/hello-dagster/venv/bin/dagster", line 5, in <module> from dagster.cli import main File "/Users/admin/Desktop/hello-dagster/venv/lib/python3.10/site-packages/dagster/__init__.py", line 100, in <module> from dagster._config.pythonic_config import ( File "/Users/admin/Desktop/hello-dagster/venv/lib/python3.10/site-packages/dagster/_config/pythonic_config/__init__.py", line 22, in <module> from pydantic import ConstrainedFloat, ConstrainedInt, ConstrainedStr File "/Users/admin/Desktop/hello-dagster/venv/lib/python3.10/site-packages/pydantic/_migration.py", line 279, in wrapper raise PydanticImportError(f'`{import_path}` has been removed in V2.') pydantic.errors.PydanticImportError: `pydantic:ConstrainedFloat` has been removed in V2.
4,518
dagster-io/dagster
dagster-io__dagster-2852
9719aa0b6fef7abf29aefd78e1d738596031f181
diff --git a/python_modules/libraries/dagster-ge/setup.py b/python_modules/libraries/dagster-ge/setup.py --- a/python_modules/libraries/dagster-ge/setup.py +++ b/python_modules/libraries/dagster-ge/setup.py @@ -25,6 +25,6 @@ def get_version(): "Operating System :: OS Independent", ], packages=find_packages(exclude=["test"]), - install_requires=["dagster", "dagster-pandas", "pandas", "great_expectations"], + install_requires=["dagster", "dagster-pandas", "pandas", "great_expectations >=0.11.9"], zip_safe=False, )
dagster-ge great_expectations pin It looks like dagster-ge should have a bottom pin of `0.11.9`: ``` import: 'dagster_ge' Traceback (most recent call last): File "/home/conda/feedstock_root/build_artifacts/dagster_1598270757893/test_tmp/run_test.py", line 2, in <module> import dagster_ge File "/home/conda/feedstock_root/build_artifacts/dagster_1598270757893/_test_env_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehol/lib/python3.7/site-packages/dagster_ge/__init__.py", line 1, in <module> from .factory import ge_validation_solid_factory File "/home/conda/feedstock_root/build_artifacts/dagster_1598270757893/_test_env_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehol/lib/python3.7/site-packages/dagster_ge/factory.py", line 6, in <module> from great_expectations.render.page_renderer_util import ( ModuleNotFoundError: No module named 'great_expectations.render.page_renderer_util' ``` This file is present in 0.11.9, but not in 0.11.8: - https://github.com/great-expectations/great_expectations/tree/0.11.8/great_expectations/render - https://github.com/great-expectations/great_expectations/tree/0.11.9/great_expectations/render
2020-08-24T12:43:34Z
[]
[]
Traceback (most recent call last): File "/home/conda/feedstock_root/build_artifacts/dagster_1598270757893/test_tmp/run_test.py", line 2, in <module> import dagster_ge File "/home/conda/feedstock_root/build_artifacts/dagster_1598270757893/_test_env_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehol/lib/python3.7/site-packages/dagster_ge/__init__.py", line 1, in <module> from .factory import ge_validation_solid_factory File "/home/conda/feedstock_root/build_artifacts/dagster_1598270757893/_test_env_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehol/lib/python3.7/site-packages/dagster_ge/factory.py", line 6, in <module> from great_expectations.render.page_renderer_util import ( ModuleNotFoundError: No module named 'great_expectations.render.page_renderer_util'
4,539
dagster-io/dagster
dagster-io__dagster-7788
cba914464ee73a9e9161e87cbf4762ee9a029315
diff --git a/python_modules/dagster/dagster/core/decorator_utils.py b/python_modules/dagster/dagster/core/decorator_utils.py --- a/python_modules/dagster/dagster/core/decorator_utils.py +++ b/python_modules/dagster/dagster/core/decorator_utils.py @@ -62,7 +62,7 @@ def param_is_var_keyword(param: funcsigs.Parameter) -> bool: def format_docstring_for_description(fn: Callable) -> Optional[str]: if fn.__doc__ is not None: docstring = fn.__doc__ - if docstring[0].isspace(): + if len(docstring) > 0 and docstring[0].isspace(): return textwrap.dedent(docstring).strip() else: first_newline_pos = docstring.find("\n")
Empty docstring causes crash ## Summary <!-- A brief description of the issue and what you expect to happen instead --> Empty docstring causes a crash. I was doing the cereal tutorial and was typing the code (vs cut/paste). Out of habit I added an empty doc string. Running the job `dagster job execute -f cereal.py` throws an error. ## Reproduction <!-- A minimal example that exhibits the behavior --> Add an empty docstring. E.g. ```python @op def download_cereals(): '''''' response = requests.get("https://docs.dagster.io/assets/cereal.csv") lines = response.text.split("\n") return [row for row in csv.DictReader(lines)] ``` ```bash ❯ dagster job execute -f cereal.py Using temporary directory /Users/jasonwirth/code/dagster_examples/tmpyz8xbih8 for storage. This will be removed when ``dagster job execute`` exits. To persist information across sessions, set the environment variable DAGSTER_HOME to a directory to use. 0it [00:00, ?it/s] 0it [00:00, ?it/s] Traceback (most recent call last): File "/Users/jasonwirth/miniconda3/envs/dagster/bin/dagster", line 8, in <module> sys.exit(main()) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/cli/__init__.py", line 50, in main cli(auto_envvar_prefix=ENV_PREFIX) # pylint:disable=E1123 File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/click/core.py", line 1130, in __call__ return self.main(*args, **kwargs) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/click/core.py", line 1055, in main rv = self.invoke(ctx) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/click/core.py", line 1657, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/click/core.py", line 1657, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/click/core.py", line 1404, in invoke return ctx.invoke(self.callback, **ctx.params) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/click/core.py", line 760, in invoke return __callback(*args, **kwargs) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/cli/job.py", line 131, in job_execute_command execute_execute_command(instance, kwargs, True) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/core/telemetry.py", line 110, in wrap result = f(*args, **kwargs) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/cli/pipeline.py", line 390, in execute_execute_command pipeline_origin = get_pipeline_or_job_python_origin_from_kwargs(kwargs, using_job_op_graph_apis) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/cli/workspace/cli_target.py", line 445, in get_pipeline_or_job_python_origin_from_kwargs repository_origin = get_repository_python_origin_from_kwargs(kwargs) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/cli/workspace/cli_target.py", line 578, in get_repository_python_origin_from_kwargs code_pointer_dict = _get_code_pointer_dict_from_kwargs(kwargs) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/cli/workspace/cli_target.py", line 500, in _get_code_pointer_dict_from_kwargs for loadable_target in get_loadable_targets( File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/grpc/utils.py", line 33, in get_loadable_targets else loadable_targets_from_python_file(python_file, working_directory) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/core/workspace/autodiscovery.py", line 26, in loadable_targets_from_python_file loaded_module = load_python_file(python_file, working_directory) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/core/code_pointer.py", line 86, in load_python_file return import_module_from_path(module_name, python_file) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/seven/__init__.py", line 50, in import_module_from_path spec.loader.exec_module(module) File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/Users/jasonwirth/code/dagster_examples/cereal.py", line 9, in <module> def download_cereals(): File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/core/definitions/decorators/op_decorator.py", line 299, in op return _Op()(name) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/core/definitions/decorators/op_decorator.py", line 130, in __call__ description=self.description or format_docstring_for_description(fn), File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/core/decorator_utils.py", line 65, in format_docstring_for_description if docstring[0].isspace(): IndexError: string index out of range ``` ## Dagit UI/UX Issue Screenshots <!-- (Optional) --> ## Additional Info about Your Environment <!-- (Optional) --> ❯ dagster --version dagster, version 0.14.14 --- #### Message from the maintainers: Impacted by this bug? Give it a 👍. We factor engagement into prioritization.
2022-05-08T15:31:14Z
[]
[]
Traceback (most recent call last): File "/Users/jasonwirth/miniconda3/envs/dagster/bin/dagster", line 8, in <module> sys.exit(main()) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/cli/__init__.py", line 50, in main cli(auto_envvar_prefix=ENV_PREFIX) # pylint:disable=E1123 File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/click/core.py", line 1130, in __call__ return self.main(*args, **kwargs) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/click/core.py", line 1055, in main rv = self.invoke(ctx) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/click/core.py", line 1657, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/click/core.py", line 1657, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/click/core.py", line 1404, in invoke return ctx.invoke(self.callback, **ctx.params) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/click/core.py", line 760, in invoke return __callback(*args, **kwargs) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/cli/job.py", line 131, in job_execute_command execute_execute_command(instance, kwargs, True) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/core/telemetry.py", line 110, in wrap result = f(*args, **kwargs) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/cli/pipeline.py", line 390, in execute_execute_command pipeline_origin = get_pipeline_or_job_python_origin_from_kwargs(kwargs, using_job_op_graph_apis) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/cli/workspace/cli_target.py", line 445, in get_pipeline_or_job_python_origin_from_kwargs repository_origin = get_repository_python_origin_from_kwargs(kwargs) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/cli/workspace/cli_target.py", line 578, in get_repository_python_origin_from_kwargs code_pointer_dict = _get_code_pointer_dict_from_kwargs(kwargs) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/cli/workspace/cli_target.py", line 500, in _get_code_pointer_dict_from_kwargs for loadable_target in get_loadable_targets( File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/grpc/utils.py", line 33, in get_loadable_targets else loadable_targets_from_python_file(python_file, working_directory) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/core/workspace/autodiscovery.py", line 26, in loadable_targets_from_python_file loaded_module = load_python_file(python_file, working_directory) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/core/code_pointer.py", line 86, in load_python_file return import_module_from_path(module_name, python_file) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/seven/__init__.py", line 50, in import_module_from_path spec.loader.exec_module(module) File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/Users/jasonwirth/code/dagster_examples/cereal.py", line 9, in <module> def download_cereals(): File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/core/definitions/decorators/op_decorator.py", line 299, in op return _Op()(name) File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/core/definitions/decorators/op_decorator.py", line 130, in __call__ description=self.description or format_docstring_for_description(fn), File "/Users/jasonwirth/miniconda3/envs/dagster/lib/python3.10/site-packages/dagster/core/decorator_utils.py", line 65, in format_docstring_for_description if docstring[0].isspace(): IndexError: string index out of range
4,582
docker/compose
docker__compose-10
b4c905dc83fb44d6fb5ac0b4a0e8d5c63205c28b
diff --git a/fig/cli/log_printer.py b/fig/cli/log_printer.py --- a/fig/cli/log_printer.py +++ b/fig/cli/log_printer.py @@ -31,28 +31,18 @@ def _make_log_generators(self): def _make_log_generator(self, container, color_fn): prefix = color_fn(container.name + " | ") - websocket = self._attach(container) - return (prefix + line for line in split_buffer(read_websocket(websocket), '\n')) + for line in split_buffer(self._attach(container), '\n'): + yield prefix + line def _attach(self, container): params = { - 'stdin': False, 'stdout': True, 'stderr': True, - 'logs': False, 'stream': True, } params.update(self.attach_params) params = dict((name, 1 if value else 0) for (name, value) in list(params.items())) - return container.attach_socket(params=params, ws=True) - -def read_websocket(websocket): - while True: - data = websocket.recv() - if data: - yield data - else: - break + return container.attach(**params) def split_buffer(reader, separator): """ diff --git a/fig/container.py b/fig/container.py --- a/fig/container.py +++ b/fig/container.py @@ -122,6 +122,9 @@ def links(self): links.append(bits[2]) return links + def attach(self, *args, **kwargs): + return self.client.attach(self.id, *args, **kwargs) + def attach_socket(self, **kwargs): return self.client.attach_socket(self.id, **kwargs)
Getting ValueError: scheme unix is invalid Getting this error when running docker v0.7.2. Here's the corresponding traceback: ``` Traceback (most recent call last): File "/usr/local/bin/fig", line 9, in <module> load_entry_point('fig==0.0.2', 'console_scripts', 'fig')() File "/usr/local/lib/python2.7/dist-packages/fig/cli/main.py", line 37, in main command.sys_dispatch() File "/usr/local/lib/python2.7/dist-packages/fig/cli/docopt_command.py", line 19, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/local/lib/python2.7/dist-packages/fig/cli/docopt_command.py", line 22, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/local/lib/python2.7/dist-packages/fig/cli/docopt_command.py", line 25, in perform_command handler(command_options) File "/usr/local/lib/python2.7/dist-packages/fig/cli/main.py", line 226, in up log_printer = LogPrinter(containers) File "/usr/local/lib/python2.7/dist-packages/fig/cli/log_printer.py", line 13, in __init__ self.generators = self._make_log_generators() File "/usr/local/lib/python2.7/dist-packages/fig/cli/log_printer.py", line 26, in _make_log_generators generators.append(self._make_log_generator(container, color_fn)) File "/usr/local/lib/python2.7/dist-packages/fig/cli/log_printer.py", line 32, in _make_log_generator websocket = self._attach(container) File "/usr/local/lib/python2.7/dist-packages/fig/cli/log_printer.py", line 45, in _attach return container.attach_socket(params=params, ws=True) File "/usr/local/lib/python2.7/dist-packages/fig/container.py", line 131, in attach_socket return self.client.attach_socket(self.id, **kwargs) File "/usr/local/lib/python2.7/dist-packages/docker/client.py", line 251, in attach_socket return self._attach_websocket(container, params) File "/usr/local/lib/python2.7/dist-packages/docker/client.py", line 217, in _attach_websocket return self._create_websocket_connection(full_url) File "/usr/local/lib/python2.7/dist-packages/docker/client.py", line 220, in _create_websocket_connection return websocket.create_connection(url) File "/usr/local/lib/python2.7/dist-packages/websocket.py", line 178, in create_connection websock.connect(url, **options) File "/usr/local/lib/python2.7/dist-packages/websocket.py", line 419, in connect hostname, port, resource, is_secure = _parse_url(url) File "/usr/local/lib/python2.7/dist-packages/websocket.py", line 141, in _parse_url raise ValueError("scheme %s is invalid" % scheme) ValueError: scheme unix is invalid ```
Seems like it's for an old version of docker and expecting a tcp socket, while newer ones mount on /var/run/docker.sock. @aronasorman Shouldn't be the case. I get this on 0.7.4. Unless you mean that fig expects older versions? I am having the same issue with Docker 0.7.5
2014-01-14T12:51:14Z
[]
[]
Traceback (most recent call last): File "/usr/local/bin/fig", line 9, in <module> load_entry_point('fig==0.0.2', 'console_scripts', 'fig')() File "/usr/local/lib/python2.7/dist-packages/fig/cli/main.py", line 37, in main command.sys_dispatch() File "/usr/local/lib/python2.7/dist-packages/fig/cli/docopt_command.py", line 19, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/local/lib/python2.7/dist-packages/fig/cli/docopt_command.py", line 22, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/local/lib/python2.7/dist-packages/fig/cli/docopt_command.py", line 25, in perform_command handler(command_options) File "/usr/local/lib/python2.7/dist-packages/fig/cli/main.py", line 226, in up log_printer = LogPrinter(containers) File "/usr/local/lib/python2.7/dist-packages/fig/cli/log_printer.py", line 13, in __init__ self.generators = self._make_log_generators() File "/usr/local/lib/python2.7/dist-packages/fig/cli/log_printer.py", line 26, in _make_log_generators generators.append(self._make_log_generator(container, color_fn)) File "/usr/local/lib/python2.7/dist-packages/fig/cli/log_printer.py", line 32, in _make_log_generator websocket = self._attach(container) File "/usr/local/lib/python2.7/dist-packages/fig/cli/log_printer.py", line 45, in _attach return container.attach_socket(params=params, ws=True) File "/usr/local/lib/python2.7/dist-packages/fig/container.py", line 131, in attach_socket return self.client.attach_socket(self.id, **kwargs) File "/usr/local/lib/python2.7/dist-packages/docker/client.py", line 251, in attach_socket return self._attach_websocket(container, params) File "/usr/local/lib/python2.7/dist-packages/docker/client.py", line 217, in _attach_websocket return self._create_websocket_connection(full_url) File "/usr/local/lib/python2.7/dist-packages/docker/client.py", line 220, in _create_websocket_connection return websocket.create_connection(url) File "/usr/local/lib/python2.7/dist-packages/websocket.py", line 178, in create_connection websock.connect(url, **options) File "/usr/local/lib/python2.7/dist-packages/websocket.py", line 419, in connect hostname, port, resource, is_secure = _parse_url(url) File "/usr/local/lib/python2.7/dist-packages/websocket.py", line 141, in _parse_url raise ValueError("scheme %s is invalid" % scheme) ValueError: scheme unix is invalid
4,618
docker/compose
docker__compose-1026
66555aa69b1ebfde8ba16183ea66f6c7e4b491e7
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -27,7 +27,7 @@ def find_version(*file_paths): install_requires = [ 'docopt >= 0.6.1, < 0.7', 'PyYAML >= 3.10, < 4', - 'requests >= 2.2.1, < 3', + 'requests >= 2.2.1, < 2.5.0', 'texttable >= 0.8.1, < 0.9', 'websocket-client >= 0.11.0, < 1.0', 'docker-py >= 0.6.0, < 0.8',
'requests' version conflict docker-compose install the latest `requests` version (2.5.1) but docker-py expects `requests` version lower than 2.5.0. #### docker-compose(fig) https://github.com/docker/fig/blob/master/setup.py ``` install_requires = [ ... 'requests >= 2.2.1, < 3', 'docker-py >= 0.6.0, < 0.8', ... ] ``` #### docker-py https://github.com/docker/docker-py/blob/master/setup.py ``` install_requires = [ ... 'requests >= 2.2.1, < 2.5.0', ... ] ``` Here is the error message: ``` Traceback (most recent call last): File "/usr/local/bin/docker-compose", line 5, in <module> from pkg_resources import load_entry_point File "/usr/local/lib/python2.7/dist-packages/pkg_resources/__init__.py", line 3018, in <module> working_set = WorkingSet._build_master() File "/usr/local/lib/python2.7/dist-packages/pkg_resources/__init__.py", line 614, in _build_master return cls._build_from_requirements(__requires__) File "/usr/local/lib/python2.7/dist-packages/pkg_resources/__init__.py", line 627, in _build_from_requirements dists = ws.resolve(reqs, Environment()) File "/usr/local/lib/python2.7/dist-packages/pkg_resources/__init__.py", line 810, in resolve raise VersionConflict(dist, req).with_context(dependent_req) pkg_resources.ContextualVersionConflict: (requests 2.5.1 (/usr/local/lib/python2.7/dist-packages), Requirement.parse('requests<2.5.0,>=2.2.1'), set(['docker-py'])) ```
To check conflict ``` $ pip install pip-conflict-checker $ pipconflictchecker -------------------------------------------------- Conflicts Detected -------------------------------------------------- - requests(2.5.1) docker-py(<2.5.0,>=2.2.1) ``` I get the same error. Are there any workarounds until a permanent fix is landed? For a temporary workaround, you can clone docker-compose and update `requests` version in setup.py: ``` $ git clone git@github.com:docker/fig.git $ cd fig ``` replace the string `'requests >= 2.2.1, < 3',` to `'requests >= 2.2.1, < 2.5.0',` in setup.py and then run: `$ pip install .` Thanks for the tip @blackrosezy. Uninstalling requests and installing «valid» version also worked: ``` bash pip uninstall requests pip install requests==2.4.3 ``` docker-py should be updated to support newer versions, is there a corresponding ticket for that? docker/docker-py#469 seems to address it.
2015-02-25T17:37:00Z
[]
[]
Traceback (most recent call last): File "/usr/local/bin/docker-compose", line 5, in <module> from pkg_resources import load_entry_point File "/usr/local/lib/python2.7/dist-packages/pkg_resources/__init__.py", line 3018, in <module> working_set = WorkingSet._build_master() File "/usr/local/lib/python2.7/dist-packages/pkg_resources/__init__.py", line 614, in _build_master return cls._build_from_requirements(__requires__) File "/usr/local/lib/python2.7/dist-packages/pkg_resources/__init__.py", line 627, in _build_from_requirements dists = ws.resolve(reqs, Environment()) File "/usr/local/lib/python2.7/dist-packages/pkg_resources/__init__.py", line 810, in resolve raise VersionConflict(dist, req).with_context(dependent_req) pkg_resources.ContextualVersionConflict: (requests 2.5.1 (/usr/local/lib/python2.7/dist-packages), Requirement.parse('requests<2.5.0,>=2.2.1'), set(['docker-py']))
4,620
docker/compose
docker__compose-1159
03535a61588e205dfe687707996b1e0d35d065cd
diff --git a/compose/__init__.py b/compose/__init__.py --- a/compose/__init__.py +++ b/compose/__init__.py @@ -1,4 +1,4 @@ from __future__ import unicode_literals from .service import Service # noqa:flake8 -__version__ = '1.1.0' +__version__ = '1.2.0' diff --git a/compose/cli/command.py b/compose/cli/command.py --- a/compose/cli/command.py +++ b/compose/cli/command.py @@ -4,9 +4,9 @@ import logging import os import re -import yaml import six +from .. import config from ..project import Project from ..service import ConfigError from .docopt_command import DocoptCommand @@ -25,7 +25,7 @@ class Command(DocoptCommand): def dispatch(self, *args, **kwargs): try: super(Command, self).dispatch(*args, **kwargs) - except SSLError, e: + except SSLError as e: raise errors.UserError('SSL error: %s' % e) except ConnectionError: if call_silently(['which', 'docker']) != 0: @@ -69,18 +69,11 @@ def get_client(self, verbose=False): return verbose_proxy.VerboseProxy('docker', client) return client - def get_config(self, config_path): - try: - with open(config_path, 'r') as fh: - return yaml.safe_load(fh) - except IOError as e: - raise errors.UserError(six.text_type(e)) - def get_project(self, config_path, project_name=None, verbose=False): try: - return Project.from_config( + return Project.from_dicts( self.get_project_name(config_path, project_name), - self.get_config(config_path), + config.load(config_path), self.get_client(verbose=verbose)) except ConfigError as e: raise errors.UserError(six.text_type(e)) diff --git a/compose/cli/docker_client.py b/compose/cli/docker_client.py --- a/compose/cli/docker_client.py +++ b/compose/cli/docker_client.py @@ -32,4 +32,4 @@ def docker_client(): ) timeout = int(os.environ.get('DOCKER_CLIENT_TIMEOUT', 60)) - return Client(base_url=base_url, tls=tls_config, version='1.14', timeout=timeout) + return Client(base_url=base_url, tls=tls_config, version='1.15', timeout=timeout) diff --git a/compose/cli/log_printer.py b/compose/cli/log_printer.py --- a/compose/cli/log_printer.py +++ b/compose/cli/log_printer.py @@ -46,7 +46,7 @@ def no_color(text): if monochrome: color_fn = no_color else: - color_fn = color_fns.next() + color_fn = next(color_fns) generators.append(self._make_log_generator(container, color_fn)) return generators diff --git a/compose/cli/main.py b/compose/cli/main.py --- a/compose/cli/main.py +++ b/compose/cli/main.py @@ -1,26 +1,26 @@ from __future__ import print_function from __future__ import unicode_literals +from inspect import getdoc +from operator import attrgetter import logging -import sys import re import signal -from operator import attrgetter +import sys -from inspect import getdoc +from docker.errors import APIError import dockerpty from .. import __version__ from ..project import NoSuchService, ConfigurationError from ..service import BuildError, CannotBeScaledError +from ..config import parse_environment from .command import Command +from .docopt_command import NoSuchCommand +from .errors import UserError from .formatter import Formatter from .log_printer import LogPrinter from .utils import yesno -from docker.errors import APIError -from .errors import UserError -from .docopt_command import NoSuchCommand - log = logging.getLogger(__name__) @@ -238,8 +238,8 @@ def rm(self, project, options): Usage: rm [options] [SERVICE...] Options: - --force Don't ask to confirm removal - -v Remove volumes associated with containers + -f, --force Don't ask to confirm removal + -v Remove volumes associated with containers """ all_containers = project.containers(service_names=options['SERVICE'], stopped=True) stopped_containers = [c for c in all_containers if not c.is_running] @@ -276,6 +276,7 @@ def run(self, project, options): new container name. --entrypoint CMD Override the entrypoint of the image. -e KEY=VAL Set an environment variable (can be used multiple times) + -u, --user="" Run as specified username or uid --no-deps Don't start linked services. --rm Remove container after run. Ignored in detached mode. --service-ports Run command with the service's ports enabled and mapped @@ -293,7 +294,7 @@ def run(self, project, options): if len(deps) > 0: project.up( service_names=deps, - start_links=True, + start_deps=True, recreate=False, insecure_registry=insecure_registry, detach=options['-d'] @@ -316,28 +317,31 @@ def run(self, project, options): } if options['-e']: - for option in options['-e']: - if 'environment' not in service.options: - service.options['environment'] = {} - k, v = option.split('=', 1) - service.options['environment'][k] = v + # Merge environment from config with -e command line + container_options['environment'] = dict( + parse_environment(service.options.get('environment')), + **parse_environment(options['-e'])) if options['--entrypoint']: container_options['entrypoint'] = options.get('--entrypoint') + + if options['--user']: + container_options['user'] = options.get('--user') + + if not options['--service-ports']: + container_options['ports'] = [] + container = service.create_container( one_off=True, insecure_registry=insecure_registry, **container_options ) - service_ports = None - if options['--service-ports']: - service_ports = service.options['ports'] if options['-d']: - service.start_container(container, ports=service_ports, one_off=True) + service.start_container(container) print(container.name) else: - service.start_container(container, ports=service_ports, one_off=True) + service.start_container(container) dockerpty.start(project.client, container.id, interactive=not options['-T']) exit_code = container.wait() if options['--rm']: @@ -389,17 +393,29 @@ def stop(self, project, options): They can be started again with `docker-compose start`. - Usage: stop [SERVICE...] + Usage: stop [options] [SERVICE...] + + Options: + -t, --timeout TIMEOUT Specify a shutdown timeout in seconds. + (default: 10) """ - project.stop(service_names=options['SERVICE']) + timeout = options.get('--timeout') + params = {} if timeout is None else {'timeout': int(timeout)} + project.stop(service_names=options['SERVICE'], **params) def restart(self, project, options): """ Restart running containers. - Usage: restart [SERVICE...] + Usage: restart [options] [SERVICE...] + + Options: + -t, --timeout TIMEOUT Specify a shutdown timeout in seconds. + (default: 10) """ - project.restart(service_names=options['SERVICE']) + timeout = options.get('--timeout') + params = {} if timeout is None else {'timeout': int(timeout)} + project.restart(service_names=options['SERVICE'], **params) def up(self, project, options): """ @@ -418,30 +434,33 @@ def up(self, project, options): Usage: up [options] [SERVICE...] Options: - --allow-insecure-ssl Allow insecure connections to the docker - registry - -d Detached mode: Run containers in the background, - print new container names. - --no-color Produce monochrome output. - --no-deps Don't start linked services. - --no-recreate If containers already exist, don't recreate them. - --no-build Don't build an image, even if it's missing + --allow-insecure-ssl Allow insecure connections to the docker + registry + -d Detached mode: Run containers in the background, + print new container names. + --no-color Produce monochrome output. + --no-deps Don't start linked services. + --no-recreate If containers already exist, don't recreate them. + --no-build Don't build an image, even if it's missing + -t, --timeout TIMEOUT When attached, use this timeout in seconds + for the shutdown. (default: 10) + """ insecure_registry = options['--allow-insecure-ssl'] detached = options['-d'] monochrome = options['--no-color'] - start_links = not options['--no-deps'] + start_deps = not options['--no-deps'] recreate = not options['--no-recreate'] service_names = options['SERVICE'] project.up( service_names=service_names, - start_links=start_links, + start_deps=start_deps, recreate=recreate, insecure_registry=insecure_registry, - detach=options['-d'], + detach=detached, do_build=not options['--no-build'], ) @@ -460,7 +479,9 @@ def handler(signal, frame): signal.signal(signal.SIGINT, handler) print("Gracefully stopping... (press Ctrl+C again to force)") - project.stop(service_names=service_names) + timeout = options.get('--timeout') + params = {} if timeout is None else {'timeout': int(timeout)} + project.stop(service_names=service_names, **params) def list_containers(containers): diff --git a/compose/config.py b/compose/config.py new file mode 100644 --- /dev/null +++ b/compose/config.py @@ -0,0 +1,432 @@ +import os +import yaml +import six + + +DOCKER_CONFIG_KEYS = [ + 'cap_add', + 'cap_drop', + 'cpu_shares', + 'command', + 'detach', + 'dns', + 'dns_search', + 'domainname', + 'entrypoint', + 'env_file', + 'environment', + 'hostname', + 'image', + 'links', + 'mem_limit', + 'net', + 'ports', + 'privileged', + 'restart', + 'stdin_open', + 'tty', + 'user', + 'volumes', + 'volumes_from', + 'working_dir', +] + +ALLOWED_KEYS = DOCKER_CONFIG_KEYS + [ + 'build', + 'expose', + 'external_links', + 'name', +] + +DOCKER_CONFIG_HINTS = { + 'cpu_share' : 'cpu_shares', + 'link' : 'links', + 'port' : 'ports', + 'privilege' : 'privileged', + 'priviliged': 'privileged', + 'privilige' : 'privileged', + 'volume' : 'volumes', + 'workdir' : 'working_dir', +} + + +def load(filename): + working_dir = os.path.dirname(filename) + return from_dictionary(load_yaml(filename), working_dir=working_dir, filename=filename) + + +def from_dictionary(dictionary, working_dir=None, filename=None): + service_dicts = [] + + for service_name, service_dict in list(dictionary.items()): + if not isinstance(service_dict, dict): + raise ConfigurationError('Service "%s" doesn\'t have any configuration options. All top level keys in your docker-compose.yml must map to a dictionary of configuration options.' % service_name) + loader = ServiceLoader(working_dir=working_dir, filename=filename) + service_dict = loader.make_service_dict(service_name, service_dict) + service_dicts.append(service_dict) + + return service_dicts + + +def make_service_dict(name, service_dict, working_dir=None): + return ServiceLoader(working_dir=working_dir).make_service_dict(name, service_dict) + + +class ServiceLoader(object): + def __init__(self, working_dir, filename=None, already_seen=None): + self.working_dir = working_dir + self.filename = filename + self.already_seen = already_seen or [] + + def make_service_dict(self, name, service_dict): + if self.signature(name) in self.already_seen: + raise CircularReference(self.already_seen) + + service_dict = service_dict.copy() + service_dict['name'] = name + service_dict = resolve_environment(service_dict, working_dir=self.working_dir) + service_dict = self.resolve_extends(service_dict) + return process_container_options(service_dict, working_dir=self.working_dir) + + def resolve_extends(self, service_dict): + if 'extends' not in service_dict: + return service_dict + + extends_options = process_extends_options(service_dict['name'], service_dict['extends']) + + if self.working_dir is None: + raise Exception("No working_dir passed to ServiceLoader()") + + other_config_path = expand_path(self.working_dir, extends_options['file']) + other_working_dir = os.path.dirname(other_config_path) + other_already_seen = self.already_seen + [self.signature(service_dict['name'])] + other_loader = ServiceLoader( + working_dir=other_working_dir, + filename=other_config_path, + already_seen=other_already_seen, + ) + + other_config = load_yaml(other_config_path) + other_service_dict = other_config[extends_options['service']] + other_service_dict = other_loader.make_service_dict( + service_dict['name'], + other_service_dict, + ) + validate_extended_service_dict( + other_service_dict, + filename=other_config_path, + service=extends_options['service'], + ) + + return merge_service_dicts(other_service_dict, service_dict) + + def signature(self, name): + return (self.filename, name) + + +def process_extends_options(service_name, extends_options): + error_prefix = "Invalid 'extends' configuration for %s:" % service_name + + if not isinstance(extends_options, dict): + raise ConfigurationError("%s must be a dictionary" % error_prefix) + + if 'service' not in extends_options: + raise ConfigurationError( + "%s you need to specify a service, e.g. 'service: web'" % error_prefix + ) + + for k, _ in extends_options.items(): + if k not in ['file', 'service']: + raise ConfigurationError( + "%s unsupported configuration option '%s'" % (error_prefix, k) + ) + + return extends_options + + +def validate_extended_service_dict(service_dict, filename, service): + error_prefix = "Cannot extend service '%s' in %s:" % (service, filename) + + if 'links' in service_dict: + raise ConfigurationError("%s services with 'links' cannot be extended" % error_prefix) + + if 'volumes_from' in service_dict: + raise ConfigurationError("%s services with 'volumes_from' cannot be extended" % error_prefix) + + if 'net' in service_dict: + if get_service_name_from_net(service_dict['net']) is not None: + raise ConfigurationError("%s services with 'net: container' cannot be extended" % error_prefix) + + +def process_container_options(service_dict, working_dir=None): + for k in service_dict: + if k not in ALLOWED_KEYS: + msg = "Unsupported config option for %s service: '%s'" % (service_dict['name'], k) + if k in DOCKER_CONFIG_HINTS: + msg += " (did you mean '%s'?)" % DOCKER_CONFIG_HINTS[k] + raise ConfigurationError(msg) + + service_dict = service_dict.copy() + + if 'volumes' in service_dict: + service_dict['volumes'] = resolve_host_paths(service_dict['volumes'], working_dir=working_dir) + + if 'build' in service_dict: + service_dict['build'] = resolve_build_path(service_dict['build'], working_dir=working_dir) + + return service_dict + + +def merge_service_dicts(base, override): + d = base.copy() + + if 'environment' in base or 'environment' in override: + d['environment'] = merge_environment( + base.get('environment'), + override.get('environment'), + ) + + if 'volumes' in base or 'volumes' in override: + d['volumes'] = merge_volumes( + base.get('volumes'), + override.get('volumes'), + ) + + if 'image' in override and 'build' in d: + del d['build'] + + if 'build' in override and 'image' in d: + del d['image'] + + list_keys = ['ports', 'expose', 'external_links'] + + for key in list_keys: + if key in base or key in override: + d[key] = base.get(key, []) + override.get(key, []) + + list_or_string_keys = ['dns', 'dns_search'] + + for key in list_or_string_keys: + if key in base or key in override: + d[key] = to_list(base.get(key)) + to_list(override.get(key)) + + already_merged_keys = ['environment', 'volumes'] + list_keys + list_or_string_keys + + for k in set(ALLOWED_KEYS) - set(already_merged_keys): + if k in override: + d[k] = override[k] + + return d + + +def merge_environment(base, override): + env = parse_environment(base) + env.update(parse_environment(override)) + return env + + +def parse_links(links): + return dict(parse_link(l) for l in links) + + +def parse_link(link): + if ':' in link: + source, alias = link.split(':', 1) + return (alias, source) + else: + return (link, link) + + +def get_env_files(options, working_dir=None): + if 'env_file' not in options: + return {} + + if working_dir is None: + raise Exception("No working_dir passed to get_env_files()") + + env_files = options.get('env_file', []) + if not isinstance(env_files, list): + env_files = [env_files] + + return [expand_path(working_dir, path) for path in env_files] + + +def resolve_environment(service_dict, working_dir=None): + service_dict = service_dict.copy() + + if 'environment' not in service_dict and 'env_file' not in service_dict: + return service_dict + + env = {} + + if 'env_file' in service_dict: + for f in get_env_files(service_dict, working_dir=working_dir): + env.update(env_vars_from_file(f)) + del service_dict['env_file'] + + env.update(parse_environment(service_dict.get('environment'))) + env = dict(resolve_env_var(k, v) for k, v in six.iteritems(env)) + + service_dict['environment'] = env + return service_dict + + +def parse_environment(environment): + if not environment: + return {} + + if isinstance(environment, list): + return dict(split_env(e) for e in environment) + + if isinstance(environment, dict): + return environment + + raise ConfigurationError( + "environment \"%s\" must be a list or mapping," % + environment + ) + + +def split_env(env): + if '=' in env: + return env.split('=', 1) + else: + return env, None + + +def resolve_env_var(key, val): + if val is not None: + return key, val + elif key in os.environ: + return key, os.environ[key] + else: + return key, '' + + +def env_vars_from_file(filename): + """ + Read in a line delimited file of environment variables. + """ + if not os.path.exists(filename): + raise ConfigurationError("Couldn't find env file: %s" % filename) + env = {} + for line in open(filename, 'r'): + line = line.strip() + if line and not line.startswith('#'): + k, v = split_env(line) + env[k] = v + return env + + +def resolve_host_paths(volumes, working_dir=None): + if working_dir is None: + raise Exception("No working_dir passed to resolve_host_paths()") + + return [resolve_host_path(v, working_dir) for v in volumes] + + +def resolve_host_path(volume, working_dir): + container_path, host_path = split_volume(volume) + if host_path is not None: + host_path = os.path.expanduser(host_path) + host_path = os.path.expandvars(host_path) + return "%s:%s" % (expand_path(working_dir, host_path), container_path) + else: + return container_path + + +def resolve_build_path(build_path, working_dir=None): + if working_dir is None: + raise Exception("No working_dir passed to resolve_build_path") + + _path = expand_path(working_dir, build_path) + if not os.path.exists(_path) or not os.access(_path, os.R_OK): + raise ConfigurationError("build path %s either does not exist or is not accessible." % _path) + else: + return _path + + +def merge_volumes(base, override): + d = dict_from_volumes(base) + d.update(dict_from_volumes(override)) + return volumes_from_dict(d) + + +def dict_from_volumes(volumes): + if volumes: + return dict(split_volume(v) for v in volumes) + else: + return {} + + +def volumes_from_dict(d): + return [join_volume(v) for v in d.items()] + + +def split_volume(string): + if ':' in string: + (host, container) = string.split(':', 1) + return (container, host) + else: + return (string, None) + + +def join_volume(pair): + (container, host) = pair + if host is None: + return container + else: + return ":".join((host, container)) + + +def expand_path(working_dir, path): + return os.path.abspath(os.path.join(working_dir, path)) + + +def to_list(value): + if value is None: + return [] + elif isinstance(value, six.string_types): + return [value] + else: + return value + + +def get_service_name_from_net(net_config): + if not net_config: + return + + if not net_config.startswith('container:'): + return + + _, net_name = net_config.split(':', 1) + return net_name + + +def load_yaml(filename): + try: + with open(filename, 'r') as fh: + return yaml.safe_load(fh) + except IOError as e: + raise ConfigurationError(six.text_type(e)) + + +class ConfigurationError(Exception): + def __init__(self, msg): + self.msg = msg + + def __str__(self): + return self.msg + + +class CircularReference(ConfigurationError): + def __init__(self, trail): + self.trail = trail + + @property + def msg(self): + lines = [ + "{} in {}".format(service_name, filename) + for (filename, service_name) in self.trail + ] + return "Circular reference:\n {}".format("\n extends ".join(lines)) diff --git a/compose/container.py b/compose/container.py --- a/compose/container.py +++ b/compose/container.py @@ -2,6 +2,7 @@ from __future__ import absolute_import import six +from functools import reduce class Container(object): diff --git a/compose/project.py b/compose/project.py --- a/compose/project.py +++ b/compose/project.py @@ -2,6 +2,8 @@ from __future__ import absolute_import import logging +from functools import reduce +from .config import get_service_name_from_net, ConfigurationError from .service import Service from .container import Container from docker.errors import APIError @@ -18,6 +20,15 @@ def sort_service_dicts(services): def get_service_names(links): return [link.split(':')[0] for link in links] + def get_service_dependents(service_dict, services): + name = service_dict['name'] + return [ + service for service in services + if (name in get_service_names(service.get('links', [])) or + name in service.get('volumes_from', []) or + name == get_service_name_from_net(service.get('net'))) + ] + def visit(n): if n['name'] in temporary_marked: if n['name'] in get_service_names(n.get('links', [])): @@ -28,8 +39,7 @@ def visit(n): raise DependencyError('Circular import between %s' % ' and '.join(temporary_marked)) if n in unmarked: temporary_marked.add(n['name']) - dependents = [m for m in services if (n['name'] in get_service_names(m.get('links', []))) or (n['name'] in m.get('volumes_from', []))] - for m in dependents: + for m in get_service_dependents(n, services): visit(m) temporary_marked.remove(n['name']) unmarked.remove(n) @@ -59,20 +69,12 @@ def from_dicts(cls, name, service_dicts, client): for service_dict in sort_service_dicts(service_dicts): links = project.get_links(service_dict) volumes_from = project.get_volumes_from(service_dict) + net = project.get_net(service_dict) - project.services.append(Service(client=client, project=name, links=links, volumes_from=volumes_from, **service_dict)) + project.services.append(Service(client=client, project=name, links=links, net=net, + volumes_from=volumes_from, **service_dict)) return project - @classmethod - def from_config(cls, name, config, client): - dicts = [] - for service_name, service in list(config.items()): - if not isinstance(service, dict): - raise ConfigurationError('Service "%s" doesn\'t have any configuration options. All top level keys in your docker-compose.yml must map to a dictionary of configuration options.' % service_name) - service['name'] = service_name - dicts.append(service) - return cls.from_dicts(name, dicts, client) - def get_service(self, name): """ Retrieve a service by name. Raises NoSuchService @@ -84,31 +86,31 @@ def get_service(self, name): raise NoSuchService(name) - def get_services(self, service_names=None, include_links=False): + def get_services(self, service_names=None, include_deps=False): """ Returns a list of this project's services filtered by the provided list of names, or all services if service_names is None or []. - If include_links is specified, returns a list including the links for + If include_deps is specified, returns a list including the dependencies for service_names, in order of dependency. Preserves the original order of self.services where possible, - reordering as needed to resolve links. + reordering as needed to resolve dependencies. Raises NoSuchService if any of the named services do not exist. """ if service_names is None or len(service_names) == 0: return self.get_services( service_names=[s.name for s in self.services], - include_links=include_links + include_deps=include_deps ) else: unsorted = [self.get_service(name) for name in service_names] services = [s for s in self.services if s in unsorted] - if include_links: - services = reduce(self._inject_links, services, []) + if include_deps: + services = reduce(self._inject_deps, services, []) uniques = [] [uniques.append(s) for s in services if s not in uniques] @@ -145,6 +147,28 @@ def get_volumes_from(self, service_dict): del service_dict['volumes_from'] return volumes_from + def get_net(self, service_dict): + if 'net' in service_dict: + net_name = get_service_name_from_net(service_dict.get('net')) + + if net_name: + try: + net = self.get_service(net_name) + except NoSuchService: + try: + net = Container.from_id(self.client, net_name) + except APIError: + raise ConfigurationError('Serivce "%s" is trying to use the network of "%s", which is not the name of a service or container.' % (service_dict['name'], net_name)) + else: + net = service_dict['net'] + + del service_dict['net'] + + else: + net = 'bridge' + + return net + def start(self, service_names=None, **options): for service in self.get_services(service_names): service.start(**options) @@ -170,13 +194,13 @@ def build(self, service_names=None, no_cache=False): def up(self, service_names=None, - start_links=True, + start_deps=True, recreate=True, insecure_registry=False, detach=False, do_build=True): running_containers = [] - for service in self.get_services(service_names, include_links=start_links): + for service in self.get_services(service_names, include_deps=start_deps): if recreate: for (_, container) in service.recreate_containers( insecure_registry=insecure_registry, @@ -193,7 +217,7 @@ def up(self, return running_containers def pull(self, service_names=None, insecure_registry=False): - for service in self.get_services(service_names, include_links=True): + for service in self.get_services(service_names, include_deps=True): service.pull(insecure_registry=insecure_registry) def remove_stopped(self, service_names=None, **options): @@ -206,19 +230,22 @@ def containers(self, service_names=None, stopped=False, one_off=False): for service in self.get_services(service_names) if service.has_container(container, one_off=one_off)] - def _inject_links(self, acc, service): - linked_names = service.get_linked_names() + def _inject_deps(self, acc, service): + net_name = service.get_net_name() + dep_names = (service.get_linked_names() + + service.get_volumes_from_names() + + ([net_name] if net_name else [])) - if len(linked_names) > 0: - linked_services = self.get_services( - service_names=linked_names, - include_links=True + if len(dep_names) > 0: + dep_services = self.get_services( + service_names=list(set(dep_names)), + include_deps=True ) else: - linked_services = [] + dep_services = [] - linked_services.append(service) - return acc + linked_services + dep_services.append(service) + return acc + dep_services class NoSuchService(Exception): @@ -230,13 +257,5 @@ def __str__(self): return self.msg -class ConfigurationError(Exception): - def __init__(self, msg): - self.msg = msg - - def __str__(self): - return self.msg - - class DependencyError(ConfigurationError): pass diff --git a/compose/service.py b/compose/service.py --- a/compose/service.py +++ b/compose/service.py @@ -3,55 +3,20 @@ from collections import namedtuple import logging import re -import os from operator import attrgetter import sys +import six from docker.errors import APIError +from docker.utils import create_host_config +from .config import DOCKER_CONFIG_KEYS from .container import Container, get_container_name from .progress_stream import stream_output, StreamOutputError log = logging.getLogger(__name__) -DOCKER_CONFIG_KEYS = [ - 'cap_add', - 'cap_drop', - 'cpu_shares', - 'command', - 'detach', - 'dns', - 'dns_search', - 'domainname', - 'entrypoint', - 'env_file', - 'environment', - 'hostname', - 'image', - 'mem_limit', - 'net', - 'ports', - 'privileged', - 'restart', - 'stdin_open', - 'tty', - 'user', - 'volumes', - 'volumes_from', - 'working_dir', -] -DOCKER_CONFIG_HINTS = { - 'cpu_share' : 'cpu_shares', - 'link' : 'links', - 'port' : 'ports', - 'privilege' : 'privileged', - 'priviliged': 'privileged', - 'privilige' : 'privileged', - 'volume' : 'volumes', - 'workdir' : 'working_dir', -} - DOCKER_START_KEYS = [ 'cap_add', 'cap_drop', @@ -87,7 +52,7 @@ class ConfigError(ValueError): class Service(object): - def __init__(self, name, client=None, project='default', links=None, external_links=None, volumes_from=None, **options): + def __init__(self, name, client=None, project='default', links=None, external_links=None, volumes_from=None, net=None, **options): if not re.match('^%s+$' % VALID_NAME_CHARS, name): raise ConfigError('Invalid service name "%s" - only %s are allowed' % (name, VALID_NAME_CHARS)) if not re.match('^%s+$' % VALID_NAME_CHARS, project): @@ -95,26 +60,13 @@ def __init__(self, name, client=None, project='default', links=None, external_li if 'image' in options and 'build' in options: raise ConfigError('Service %s has both an image and build path specified. A service can either be built to image or use an existing image, not both.' % name) - for filename in get_env_files(options): - if not os.path.exists(filename): - raise ConfigError("Couldn't find env file for service %s: %s" % (name, filename)) - - supported_options = DOCKER_CONFIG_KEYS + ['build', 'expose', - 'external_links'] - - for k in options: - if k not in supported_options: - msg = "Unsupported config option for %s service: '%s'" % (name, k) - if k in DOCKER_CONFIG_HINTS: - msg += " (did you mean '%s'?)" % DOCKER_CONFIG_HINTS[k] - raise ConfigError(msg) - self.name = name self.client = client self.project = project self.links = links or [] self.external_links = external_links or [] self.volumes_from = volumes_from or [] + self.net = net or None self.options = options def containers(self, stopped=False, one_off=False): @@ -217,6 +169,7 @@ def create_container(self, one_off=False, insecure_registry=False, do_build=True, + intermediate_container=None, **override_options): """ Create a container for this service. If the image doesn't exist, attempt to pull @@ -224,7 +177,9 @@ def create_container(self, """ container_options = self._get_container_create_options( override_options, - one_off=one_off) + one_off=one_off, + intermediate_container=intermediate_container, + ) if (do_build and self.can_be_built() and @@ -289,57 +244,33 @@ def recreate_container(self, container, **override_options): entrypoint=['/bin/echo'], command=[], detach=True, + host_config=create_host_config(volumes_from=[container.id]), ) - intermediate_container.start(volumes_from=container.id) + intermediate_container.start() intermediate_container.wait() container.remove() options = dict(override_options) - new_container = self.create_container(do_build=False, **options) - self.start_container(new_container, intermediate_container=intermediate_container) + new_container = self.create_container( + do_build=False, + intermediate_container=intermediate_container, + **options + ) + self.start_container(new_container) intermediate_container.remove() return (intermediate_container, new_container) - def start_container_if_stopped(self, container, **options): + def start_container_if_stopped(self, container): if container.is_running: return container else: log.info("Starting %s..." % container.name) - return self.start_container(container, **options) - - def start_container(self, container, intermediate_container=None, **override_options): - options = dict(self.options, **override_options) - port_bindings = build_port_bindings(options.get('ports') or []) - - volume_bindings = dict( - build_volume_binding(parse_volume_spec(volume)) - for volume in options.get('volumes') or [] - if ':' in volume) - - privileged = options.get('privileged', False) - net = options.get('net', 'bridge') - dns = options.get('dns', None) - dns_search = options.get('dns_search', None) - cap_add = options.get('cap_add', None) - cap_drop = options.get('cap_drop', None) - - restart = parse_restart_spec(options.get('restart', None)) + return self.start_container(container) - container.start( - links=self._get_links(link_to_self=options.get('one_off', False)), - port_bindings=port_bindings, - binds=volume_bindings, - volumes_from=self._get_volumes_from(intermediate_container), - privileged=privileged, - network_mode=net, - dns=dns, - dns_search=dns_search, - restart_policy=restart, - cap_add=cap_add, - cap_drop=cap_drop, - ) + def start_container(self, container): + container.start() return container def start_or_create_containers( @@ -363,6 +294,15 @@ def start_or_create_containers( def get_linked_names(self): return [s.name for (s, _) in self.links] + def get_volumes_from_names(self): + return [s.name for s in self.volumes_from if isinstance(s, Service)] + + def get_net_name(self): + if isinstance(self.net, Service): + return self.net.name + else: + return + def _next_container_name(self, all_containers, one_off=False): bits = [self.project, self.name] if one_off: @@ -398,7 +338,6 @@ def _get_volumes_from(self, intermediate_container=None): for volume_source in self.volumes_from: if isinstance(volume_source, Service): containers = volume_source.containers(stopped=True) - if not containers: volumes_from.append(volume_source.create_container().id) else: @@ -412,7 +351,26 @@ def _get_volumes_from(self, intermediate_container=None): return volumes_from - def _get_container_create_options(self, override_options, one_off=False): + def _get_net(self): + if not self.net: + return "bridge" + + if isinstance(self.net, Service): + containers = self.net.containers() + if len(containers) > 0: + net = 'container:' + containers[0].id + else: + log.warning("Warning: Service %s is trying to use reuse the network stack " + "of another service that is not running." % (self.net.name)) + net = None + elif isinstance(self.net, Container): + net = 'container:' + self.net.id + else: + net = self.net + + return net + + def _get_container_create_options(self, override_options, one_off=False, intermediate_container=None): container_options = dict( (k, self.options[k]) for k in DOCKER_CONFIG_KEYS if k in self.options) @@ -450,8 +408,6 @@ def _get_container_create_options(self, override_options, one_off=False): (parse_volume_spec(v).internal, {}) for v in container_options['volumes']) - container_options['environment'] = merge_environment(container_options) - if self.can_be_built(): container_options['image'] = self.full_name else: @@ -461,8 +417,47 @@ def _get_container_create_options(self, override_options, one_off=False): for key in DOCKER_START_KEYS: container_options.pop(key, None) + container_options['host_config'] = self._get_container_host_config(override_options, one_off=one_off, intermediate_container=intermediate_container) + return container_options + def _get_container_host_config(self, override_options, one_off=False, intermediate_container=None): + options = dict(self.options, **override_options) + port_bindings = build_port_bindings(options.get('ports') or []) + + volume_bindings = dict( + build_volume_binding(parse_volume_spec(volume)) + for volume in options.get('volumes') or [] + if ':' in volume) + + privileged = options.get('privileged', False) + cap_add = options.get('cap_add', None) + cap_drop = options.get('cap_drop', None) + + dns = options.get('dns', None) + if isinstance(dns, six.string_types): + dns = [dns] + + dns_search = options.get('dns_search', None) + if isinstance(dns_search, six.string_types): + dns_search = [dns_search] + + restart = parse_restart_spec(options.get('restart', None)) + + return create_host_config( + links=self._get_links(link_to_self=one_off), + port_bindings=port_bindings, + binds=volume_bindings, + volumes_from=self._get_volumes_from(intermediate_container), + privileged=privileged, + network_mode=self._get_net(), + dns=dns, + dns_search=dns_search, + restart_policy=restart, + cap_add=cap_add, + cap_drop=cap_drop, + ) + def _get_image_name(self, image): repo, tag = parse_repository_tag(image) if tag == "": @@ -482,7 +477,7 @@ def build(self, no_cache=False): try: all_events = stream_output(build_output, sys.stdout) - except StreamOutputError, e: + except StreamOutputError as e: raise BuildError(self, unicode(e)) image_id = None @@ -590,8 +585,7 @@ def parse_repository_tag(s): def build_volume_binding(volume_spec): internal = {'bind': volume_spec.internal, 'ro': volume_spec.mode == 'ro'} - external = os.path.expanduser(volume_spec.external) - return os.path.abspath(os.path.expandvars(external)), internal + return volume_spec.external, internal def build_port_bindings(ports): @@ -620,54 +614,3 @@ def split_port(port): external_ip, external_port, internal_port = parts return internal_port, (external_ip, external_port or None) - - -def get_env_files(options): - env_files = options.get('env_file', []) - if not isinstance(env_files, list): - env_files = [env_files] - return env_files - - -def merge_environment(options): - env = {} - - for f in get_env_files(options): - env.update(env_vars_from_file(f)) - - if 'environment' in options: - if isinstance(options['environment'], list): - env.update(dict(split_env(e) for e in options['environment'])) - else: - env.update(options['environment']) - - return dict(resolve_env(k, v) for k, v in env.iteritems()) - - -def split_env(env): - if '=' in env: - return env.split('=', 1) - else: - return env, None - - -def resolve_env(key, val): - if val is not None: - return key, val - elif key in os.environ: - return key, os.environ[key] - else: - return key, '' - - -def env_vars_from_file(filename): - """ - Read in a line delimited file of environment variables. - """ - env = {} - for line in open(filename, 'r'): - line = line.strip() - if line and not line.startswith('#'): - k, v = split_env(line) - env[k] = v - return env diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -27,10 +27,10 @@ def find_version(*file_paths): install_requires = [ 'docopt >= 0.6.1, < 0.7', 'PyYAML >= 3.10, < 4', - 'requests >= 2.2.1, < 2.5.0', + 'requests >= 2.2.1, < 2.6', 'texttable >= 0.8.1, < 0.9', 'websocket-client >= 0.11.0, < 1.0', - 'docker-py >= 0.6.0, < 0.8', + 'docker-py >= 1.0.0, < 1.2', 'dockerpty >= 0.3.2, < 0.4', 'six >= 1.3.0, < 2', ]
Setting environment variable on CLI raises exception when service has environment definition It looks like I can't set an environment variable in the fig.yml and on the console at the same time. fig version: `1.0.1` python version: `2.7 (homebrew)` OS: OSX `10.10.2` _fig.yml:_ ``` bliep: image: ubuntu:14.04 environment: - SECRET_KEY ``` _full output:_ ``` $ fig run -e BLA=bla bliep Traceback (most recent call last): File "/usr/local/bin/fig", line 9, in <module> load_entry_point('fig==1.0.1', 'console_scripts', 'fig')() File "/usr/local/lib/python2.7/site-packages/fig/cli/main.py", line 31, in main command.sys_dispatch() File "/usr/local/lib/python2.7/site-packages/fig/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/local/lib/python2.7/site-packages/fig/cli/command.py", line 28, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/local/lib/python2.7/site-packages/fig/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/local/lib/python2.7/site-packages/fig/cli/command.py", line 56, in perform_command handler(project, command_options) File "/usr/local/lib/python2.7/site-packages/fig/cli/main.py", line 312, in run service.options['environment'][k] = v TypeError: list indices must be integers, not unicode ```
hmm, I thought this was fixed. Maybe it's in 1.1 or still in a PR somewhere. I believe the workaround for now is to use the mapping form in the fig.yml ``` yaml image: ubuntu:14.04 environment: SECRET_KEY: ``` sorry, nope: ``` $ cat fig.yml bliep: image: ubuntu:14.04 environment: - BLOP: $ fig run -e BLOP=bla bliep Traceback (most recent call last): File "/usr/local/bin/fig", line 9, in <module> load_entry_point('fig==1.0.1', 'console_scripts', 'fig')() File "/usr/local/lib/python2.7/site-packages/fig/cli/main.py", line 31, in main command.sys_dispatch() File "/usr/local/lib/python2.7/site-packages/fig/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/local/lib/python2.7/site-packages/fig/cli/command.py", line 28, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/local/lib/python2.7/site-packages/fig/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/local/lib/python2.7/site-packages/fig/cli/command.py", line 56, in perform_command handler(project, command_options) File "/usr/local/lib/python2.7/site-packages/fig/cli/main.py", line 312, in run service.options['environment'][k] = v TypeError: list indices must be integers, not unicode ``` You did not make the change I suggested. You still have a list in `environment`. You need to remove the dash. sorry, you are right and yes, that workaround works.
2015-03-23T17:56:16Z
[]
[]
Traceback (most recent call last): File "/usr/local/bin/fig", line 9, in <module> load_entry_point('fig==1.0.1', 'console_scripts', 'fig')() File "/usr/local/lib/python2.7/site-packages/fig/cli/main.py", line 31, in main command.sys_dispatch() File "/usr/local/lib/python2.7/site-packages/fig/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/local/lib/python2.7/site-packages/fig/cli/command.py", line 28, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/local/lib/python2.7/site-packages/fig/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/local/lib/python2.7/site-packages/fig/cli/command.py", line 56, in perform_command handler(project, command_options) File "/usr/local/lib/python2.7/site-packages/fig/cli/main.py", line 312, in run service.options['environment'][k] = v TypeError: list indices must be integers, not unicode
4,621
docker/compose
docker__compose-128
e6016bd5b4a24953ef8e4c8549e9018cd560d76f
diff --git a/fig/cli/main.py b/fig/cli/main.py --- a/fig/cli/main.py +++ b/fig/cli/main.py @@ -8,7 +8,7 @@ from inspect import getdoc from .. import __version__ -from ..project import NoSuchService, DependencyError +from ..project import NoSuchService, ConfigurationError from ..service import CannotBeScaledError from .command import Command from .formatter import Formatter @@ -40,7 +40,7 @@ def main(): except KeyboardInterrupt: log.error("\nAborting.") sys.exit(1) - except (UserError, NoSuchService, DependencyError) as e: + except (UserError, NoSuchService, ConfigurationError) as e: log.error(e.msg) sys.exit(1) except NoSuchCommand as e: diff --git a/fig/project.py b/fig/project.py --- a/fig/project.py +++ b/fig/project.py @@ -67,6 +67,8 @@ def from_dicts(cls, name, service_dicts, client): def from_config(cls, name, config, client): dicts = [] for service_name, service in list(config.items()): + if not isinstance(service, dict): + raise ConfigurationError('Service "%s" doesn\'t have any configuration options. All top level keys in your fig.yml must map to a dictionary of configuration options.') service['name'] = service_name dicts.append(service) return cls.from_dicts(name, dicts, client) @@ -156,9 +158,13 @@ def __str__(self): return self.msg -class DependencyError(Exception): +class ConfigurationError(Exception): def __init__(self, msg): self.msg = msg def __str__(self): return self.msg + +class DependencyError(ConfigurationError): + pass +
OSX - 'str' object does not support item assignment System specs: - Mac OSX 10.8.5 - Python 2.7.2 (built-in) - Fig 0.2.2 My fig.yml: ``` image: stackbrew/ubuntu ports: - 3306:3306 - 6379:6379 links: - mysql - redis mysql: image: orchardup/mysql redis: image: orchardup/redis ``` When running `fig up`, I get the following error: ``` Traceback (most recent call last): File "/usr/local/bin/fig", line 9, in <module> load_entry_point('fig==0.2.2', 'console_scripts', 'fig')() File "/Library/Python/2.7/site-packages/fig/cli/main.py", line 39, in main command.sys_dispatch() File "/Library/Python/2.7/site-packages/fig/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/Library/Python/2.7/site-packages/fig/cli/command.py", line 26, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/Library/Python/2.7/site-packages/fig/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/Library/Python/2.7/site-packages/fig/cli/docopt_command.py", line 27, in perform_command handler(command_options) File "/Library/Python/2.7/site-packages/fig/cli/main.py", line 287, in up (old, new) = self.project.recreate_containers(service_names=options['SERVICE']) File "/Library/Python/2.7/site-packages/fig/cli/utils.py", line 22, in get x = self._property_cache[f] = f(self) File "/Library/Python/2.7/site-packages/fig/cli/command.py", line 75, in project return Project.from_config(self.project_name, config, self.client) File "/Library/Python/2.7/site-packages/fig/project.py", line 64, in from_config service['name'] = service_name TypeError: 'str' object does not support item assignment ``` I tried removing everything from the fig.yml shy of the image entry, and the same error occurs regardless. Any help would be greatly appreciated.
Everything in your `fig.yml` needs to be inside a service. So if, for example, your main service is a web app, it should look like this: ``` web: image: stackbrew/ubuntu ports: - 3306:3306 - 6379:6379 links: - mysql - redis mysql: image: orchardup/mysql redis: image: orchardup/redis ``` We should throw a better error message here though, so I'm going to leave this open. There I go, not reading the documentation carefully enough. Thank you for a polite response and for taking the effort to improve the error messages. No worries – thanks for helping debug! Is there any part of the documentation that wasn't clear enough that we could improve?
2014-03-03T16:22:20Z
[]
[]
Traceback (most recent call last): File "/usr/local/bin/fig", line 9, in <module> load_entry_point('fig==0.2.2', 'console_scripts', 'fig')() File "/Library/Python/2.7/site-packages/fig/cli/main.py", line 39, in main command.sys_dispatch() File "/Library/Python/2.7/site-packages/fig/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/Library/Python/2.7/site-packages/fig/cli/command.py", line 26, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/Library/Python/2.7/site-packages/fig/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/Library/Python/2.7/site-packages/fig/cli/docopt_command.py", line 27, in perform_command handler(command_options) File "/Library/Python/2.7/site-packages/fig/cli/main.py", line 287, in up (old, new) = self.project.recreate_containers(service_names=options['SERVICE']) File "/Library/Python/2.7/site-packages/fig/cli/utils.py", line 22, in get x = self._property_cache[f] = f(self) File "/Library/Python/2.7/site-packages/fig/cli/command.py", line 75, in project return Project.from_config(self.project_name, config, self.client) File "/Library/Python/2.7/site-packages/fig/project.py", line 64, in from_config service['name'] = service_name TypeError: 'str' object does not support item assignment
4,627
docker/compose
docker__compose-1374
7e0ab0714fc6d023f2f1c19e8b6cb58cf6eef869
diff --git a/compose/service.py b/compose/service.py --- a/compose/service.py +++ b/compose/service.py @@ -476,6 +476,11 @@ def build(self, no_cache=False): except StreamOutputError as e: raise BuildError(self, unicode(e)) + # Ensure the HTTP connection is not reused for another + # streaming command, as the Docker daemon can sometimes + # complain about it + self.client.close() + image_id = None for event in all_events:
APIError: 500 Server Error on first docker-compose up / build cc @aanand I noticed this when testing the most recent RC (4) Steps to reproduce: 1. Clone https://github.com/nathanleclaire/hubfwd 2. Run `docker-compose up` in the repo Expected behavior: It doesn't blow up Actual behavior: It blows up On the first `up` and the first `up` only, so build smells suspicious. ``` Creating hubfwd_app_1... Building app... Step 0 : FROM golang:1.4.2 ---> 121a93c90463 Step 1 : RUN go get -u github.com/codegangsta/negroni ---> Running in 5e2161a172f9 ---> 623f1c94741b Removing intermediate container 5e2161a172f9 Step 2 : RUN go get -u github.com/gorilla/mux ---> Running in c74924a6c8fd ---> 7923dd360f79 Removing intermediate container c74924a6c8fd Step 3 : RUN go get -u github.com/Sirupsen/logrus ---> Running in 93443d6cf298 ---> 3ae5e3801312 Removing intermediate container 93443d6cf298 Step 4 : RUN mkdir -p /go/src/github.com/nathanleclaire/hubfwd ---> Running in 8deddcbb0f1d ---> 6586dfbe5b2e Removing intermediate container 8deddcbb0f1d Step 5 : WORKDIR /go/src/github.com/nathanleclaire/hubfwd ---> Running in bb42cbdf1032 ---> 0d824f6e8519 Removing intermediate container bb42cbdf1032 Step 6 : COPY . /go/src/github.com/nathanleclaire/hubfwd ---> ad6983d66cf7 Removing intermediate container e16e62829fb7 Step 7 : CMD go build ---> Running in 550e4ab79b39 ---> 15ebeafc0600 Removing intermediate container 550e4ab79b39 Successfully built 15ebeafc0600 Attaching to hubfwd_app_1 Exception in thread Thread-1: Traceback (most recent call last): File "/compose/build/docker-compose/out00-PYZ.pyz/threading", line 810, in __bootstrap_inner File "/compose/build/docker-compose/out00-PYZ.pyz/threading", line 763, in run File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.multiplexer", line 41, in _enqueue_output File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.log_printer", line 59, in _make_log_generator File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.utils", line 77, in split_buffer File "/compose/build/docker-compose/out00-PYZ.pyz/docker.client", line 199, in _multiplexed_response_stream_helper File "/compose/build/docker-compose/out00-PYZ.pyz/docker.client", line 143, in _get_raw_response_socket File "/compose/build/docker-compose/out00-PYZ.pyz/docker.client", line 95, in _raise_for_status APIError: 500 Server Error: Internal Server Error ("http: Hijack is incompatible with use of CloseNotifier") ``` I'm on OSX using a VM created by `docker-machine` in VirtualBox (boot2docker).
Bump on this - I got it with https://github.com/nathanleclaire/laraveldocker as well: (rc4) ``` Attaching to laraveldocker_db_1, laraveldocker_redis_1, laraveldocker_beanstalkd_1, laraveldocker_web_1 Exception in thread Thread-1: Traceback (most recent call last): File "/compose/build/docker-compose/out00-PYZ.pyz/threading", line 810, in __bootstrap_inner File "/compose/build/docker-compose/out00-PYZ.pyz/threading", line 763, in run File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.multiplexer", line 41, in _enqueue_output File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.log_printer", line 59, in _make_log_generator File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.utils", line 77, in split_buffer File "/compose/build/docker-compose/out00-PYZ.pyz/docker.client", line 199, in _multiplexed_response_stream_helper File "/compose/build/docker-compose/out00-PYZ.pyz/docker.client", line 143, in _get_raw_response_socket File "/compose/build/docker-compose/out00-PYZ.pyz/docker.client", line 95, in _raise_for_status APIError: 500 Server Error: Internal Server Error ("http: Hijack is incompatible with use of CloseNotifier") ``` Also getting this after `docker-compose up`: Versions: - docker-compose 1.2.0 (installed by `sudo pip install -U https://github.com/docker/compose/archive/1.2.0.zip`) - Docker version 1.6.0, build 4749651 - Boot2Docker-cli version: v1.6.0 and via virtualbox - OSX Yosemite ``` Exception in thread Thread-1: Traceback (most recent call last): File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/threading.py", line 810, in __bootstrap_inner self.run() File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/threading.py", line 763, in run self.__target(*self.__args, **self.__kwargs) File "/Library/Python/2.7/site-packages/compose/cli/multiplexer.py", line 41, in _enqueue_output for item in generator: File "/Library/Python/2.7/site-packages/compose/cli/log_printer.py", line 59, in _make_log_generator for line in line_generator: File "/Library/Python/2.7/site-packages/compose/cli/utils.py", line 77, in split_buffer for data in reader: File "/Library/Python/2.7/site-packages/docker/client.py", line 225, in _multiplexed_response_stream_helper socket = self._get_raw_response_socket(response) File "/Library/Python/2.7/site-packages/docker/client.py", line 167, in _get_raw_response_socket self._raise_for_status(response) File "/Library/Python/2.7/site-packages/docker/client.py", line 119, in _raise_for_status raise errors.APIError(e, response, explanation=explanation) APIError: 500 Server Error: Internal Server Error ("http: Hijack is incompatible with use of CloseNotifier") Exception in thread Thread-5: Traceback (most recent call last): File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/threading.py", line 810, in __bootstrap_inner self.run() File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/threading.py", line 763, in run self.__target(*self.__args, **self.__kwargs) File "/Library/Python/2.7/site-packages/compose/cli/multiplexer.py", line 41, in _enqueue_output for item in generator: File "/Library/Python/2.7/site-packages/compose/cli/log_printer.py", line 59, in _make_log_generator for line in line_generator: File "/Library/Python/2.7/site-packages/compose/cli/utils.py", line 77, in split_buffer for data in reader: File "/Library/Python/2.7/site-packages/docker/client.py", line 225, in _multiplexed_response_stream_helper socket = self._get_raw_response_socket(response) File "/Library/Python/2.7/site-packages/docker/client.py", line 167, in _get_raw_response_socket self._raise_for_status(response) File "/Library/Python/2.7/site-packages/docker/client.py", line 119, in _raise_for_status raise errors.APIError(e, response, explanation=explanation) APIError: 500 Server Error: Internal Server Error ("http: Hijack is incompatible with use of CloseNotifier") ``` :+1: also experiencing this issue. :+1: Have you figured out the solution already? :+1: For anyone reading, it seems that if I do `docker-compose build` manually and then do a `docker-compose up` everything works fine. That should hopefully serve as a workaround until this gets addressed. I can reproduce. @nathanleclaire's workaround (`docker-compose build` first) does indeed fix it for me. Looking at the `docker inspect` output of a container in both the working and failing case, I can't find anything obvious that's wrong. Very strange. Me too, and running `docker-compose build` and then `docker-compose up` like @nathanleclaire said fixes it Yay, this workaround worked for me too. Experiencing the same issue here, same versions +1, same issue/version, thanks for the workaround! +1, same problem for me Somehow your `up` supports http://golang.org/pkg/net/http/#CloseNotifier and `build` not :) I think problem is in keep-alive, so in one connection you `build` and `attach`. To fix this you need to create new connection, not sure how can we fix it on docker side. https://go-review.googlesource.com/#/c/3821/ +1 @LK4D4: nice find. I wonder if there's a way to force docker-py/requests to create a new connection for attaching. @shin-? It looks like `docker.Client` extends http://docs.python-requests.org/en/latest/api/#request-sessions, so I think we could call `client.close()` and that would force it to re-establish a new connection.
2015-04-29T16:56:45Z
[]
[]
Traceback (most recent call last): File "/compose/build/docker-compose/out00-PYZ.pyz/threading", line 810, in __bootstrap_inner File "/compose/build/docker-compose/out00-PYZ.pyz/threading", line 763, in run File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.multiplexer", line 41, in _enqueue_output File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.log_printer", line 59, in _make_log_generator File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.utils", line 77, in split_buffer File "/compose/build/docker-compose/out00-PYZ.pyz/docker.client", line 199, in _multiplexed_response_stream_helper File "/compose/build/docker-compose/out00-PYZ.pyz/docker.client", line 143, in _get_raw_response_socket File "/compose/build/docker-compose/out00-PYZ.pyz/docker.client", line 95, in _raise_for_status APIError: 500 Server Error: Internal Server Error ("http: Hijack is incompatible with use of CloseNotifier")
4,630
docker/compose
docker__compose-1376
878d90febf822dbf9c38fc23d75fbda908a9cb0f
diff --git a/compose/service.py b/compose/service.py --- a/compose/service.py +++ b/compose/service.py @@ -462,8 +462,10 @@ def _get_container_host_config(self, override_options, one_off=False, intermedia def build(self, no_cache=False): log.info('Building %s...' % self.name) + path = six.binary_type(self.options['build']) + build_output = self.client.build( - self.options['build'], + path=path, tag=self.full_name, stream=True, rm=True,
UnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position 5: ordinal not in range(128) Ubuntu 14.04 x64 ``` shell $ docker --version Docker version 1.6.0, build 4749651 $ docker-compose --version docker-compose 1.2.0 ``` Getting error on `$ sudo docker-compose build` ``` shell $ sudo docker-compose build Building web... Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 31, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 27, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 59, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 119, in build File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 191, in build File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 475, in build File "/code/build/docker-compose/out00-PYZ.pyz/docker.client", line 300, in build File "/code/build/docker-compose/out00-PYZ.pyz/docker.utils.utils", line 73, in tar File "/code/build/docker-compose/out00-PYZ.pyz/os", line 294, in walk File "/code/build/docker-compose/out00-PYZ.pyz/os", line 294, in walk File "/code/build/docker-compose/out00-PYZ.pyz/os", line 294, in walk File "/code/build/docker-compose/out00-PYZ.pyz/os", line 294, in walk File "/code/build/docker-compose/out00-PYZ.pyz/os", line 294, in walk File "/code/build/docker-compose/out00-PYZ.pyz/os", line 284, in walk File "/code/build/docker-compose/out00-PYZ.pyz/posixpath", line 80, in join UnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position 5: ordinal not in range(128) ``` docker-compose.yml ``` yml web: build: . command: gulp serve ports: - 3001:3001 volumes: - .:/app - /bower_components:./app/bower_components volumes_from: - bower_components ``` Dockerfile ``` shell FROM ubuntu:14.04 ENV NVM_VERSION 0.24.1 ENV NODE_VERSION 0.12.0 ENV NPM_VERSION 2.5.1 RUN wget -qO- https://raw.githubusercontent.com/creationix/nvm/v$NVM_VERSION/install.sh | bash \ && source ~/.bashrc \ && nvm install $NODE_VERSION \ && nvm alias default $NODE_VERSION \ && npm install -g npm@"$NPM_VERSION" gulp bower \ WORKDIR /app CMD tail -f /dev/null ```
NOTE: with`docker-comose -v 1.1.0-rc2` everything works well I can't reproduce, but from the stack trace it looks like something's going wrong when trying to tar your build directory - specifically, when trying to build a path. Do you have a file somewhere with an odd name? I have the same probleme than @itsNikolay. Ubuntu 14.04 x64 ``` $ docker --version Docker version 1.6.0, build 4749651 $ docker-compose --version docker-compose 1.2.0 ``` ``` $ docker-compose build db uses an image, skipping Building web... Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 31, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 27, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 59, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 119, in build File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 191, in build File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 475, in build File "/code/build/docker-compose/out00-PYZ.pyz/docker.client", line 300, in build File "/code/build/docker-compose/out00-PYZ.pyz/docker.utils.utils", line 73, in tar File "/code/build/docker-compose/out00-PYZ.pyz/os", line 294, in walk File "/code/build/docker-compose/out00-PYZ.pyz/os", line 294, in walk File "/code/build/docker-compose/out00-PYZ.pyz/os", line 294, in walk File "/code/build/docker-compose/out00-PYZ.pyz/os", line 284, in walk File "/code/build/docker-compose/out00-PYZ.pyz/posixpath", line 80, in join UnicodeDecodeError: 'ascii' codec can't decode byte 0xc3 in position 8: ordinal not in range(128) ``` Works with docker-compose 1.1.0. It's would be great to have a better error message to locate a possible file with odd name Having the same issue with docker-compose 1.1.0: $ docker --version Docker version 1.3.1, build 4e9bbfa $ docker-compose --version docker-compose 1.1.0 Even after running docker-compose rm && docker-compose up -d it stalls on `'ascii' codec can't decode byte 0xc3 in position 13: ordinal not in range(128)` Full trace: ``` $ docker-compose up -d Recreating adoveoweb_redisphpsessionhandler_1... Recreating adoveoweb_mailcatcher_1... Creating adoveoweb_phpfiles_1... Traceback (most recent call last): File "<string>", line 3, in <module> File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 31, in main File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 27, in dispatch File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 59, in perform_command File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 445, in up File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.project", line 184, in up File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.service", line 259, in recreate_containers File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.service", line 227, in create_container File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.service", line 453, in _get_container_create_options File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.service", line 636, in merge_environment File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.service", line 670, in env_vars_from_file UnicodeDecodeError: 'ascii' codec can't decode byte 0xc3 in position 13: ordinal not in range(128) ``` UPDATE: Upgraded docker-compose to 1.2.0 and the problem went away UPDATE: Upgraded docker to 1.6.0 and the problem has not yet resurfaced - everything works afaik I can reproduce this bug with a minimal project: ``` $ touch donées.txt $ cat > docker-compose.yml << eof data: build: . eof $ cat > Dockerfile <<eof FROM scratch COPY donées.txt . eof ``` after I have the bug ``` $ docker-compose build Building data... Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 31, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 27, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 59, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 119, in build File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 191, in build File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 475, in build File "/code/build/docker-compose/out00-PYZ.pyz/docker.client", line 300, in build File "/code/build/docker-compose/out00-PYZ.pyz/docker.utils.utils", line 73, in tar File "/code/build/docker-compose/out00-PYZ.pyz/os", line 284, in walk File "/code/build/docker-compose/out00-PYZ.pyz/posixpath", line 80, in join UnicodeDecodeError: 'ascii' codec can't decode byte 0xc3 in position 4: ordinal not in range(128) ``` If I rename the file `donées.txt` as `data.txt` it's work. Looks like a codepage / conversion thing; http://stackoverflow.com/questions/24475393/unicodedecodeerror-ascii-codec-cant-decode-byte-0xc3-in-position-23-ordinal I can't even run `docker build` when there's a file with that name (Mac OS X 10.10.3). When I run @oalbiez's commands: ``` $ docker build Sending build context to Docker daemon 4.608 kB Sending build context to Docker daemon Step 0 : FROM scratch ---> Step 1 : COPY donées.txt . INFO[0000] donées.txt: no such file or directory ``` `docker-compose build` gives me the same error, unsurprisingly. That output looks like utf-8 treated as non-UTF8 yes. I have the same problem with compose 1.2 but not 1.1, Ubuntu 15.04, Docker 1.6.0. Even a compose yml with one service to build with a Dockerfile with one line did not work but "docker build ." has no problem. Getting conflicting reports about what versions of Compose it does and doesn't work with. @itsNikolay, @motin, @madaray: 1. Could you try @oalbiez's minimal failing case with both Compose 1.1 and 1.2? 2. If you can't reproduce the error with that, can you reduce your failing case to a minimal one? I recommend passing `--no-cache` to both `docker build` and `docker-compose build`, just to be sure we don't get any false negatives. I tracked this down to our "apple-­touch-­icon-­precomposed.png". After copying apple-­touch-­icon.png it works. Not sure why this doesn't happen with 1.1.0. UPDATE: The name was corrupted somehow: apple-\302\255touch-\302\255icon-\302\255precomposed.png @aanand, I am having the same issue with 1.2.0, but not 1.1.0 (Ubuntu 14.04 with kernel version 3.16.0-30-generic). The traceback information are exactly the same as @itsNikolay 's. ``` Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 31, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 27, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 59, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 464, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 208, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 214, in recreate_containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 187, in create_container File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 475, in build File "/code/build/docker-compose/out00-PYZ.pyz/docker.client", line 300, in build File "/code/build/docker-compose/out00-PYZ.pyz/docker.utils.utils", line 73, in tar File "/code/build/docker-compose/out00-PYZ.pyz/os", line 294, in walk File "/code/build/docker-compose/out00-PYZ.pyz/os", line 294, in walk File "/code/build/docker-compose/out00-PYZ.pyz/os", line 284, in walk File "/code/build/docker-compose/out00-PYZ.pyz/posixpath", line 80, in join UnicodeDecodeError: 'ascii' codec can't decode byte 0xe5 in position 39: ordinal not in range(128) ``` My Dockerfile and docker-compose.yml are pretty simple. I am trying to run a simple node application talk with redis database to keep the session. ``` FROM node:0.10-onbuild MAINTAINER YIRAN Mao <yiranmao@gmail.com> RUN mkdir -p /usr/src/app WORKDIR /usr/src/app EXPOSE 3000 EXPOSE 3001 ``` ``` redis: image: redis:2.8.19 command: redis-server --appendonly yes volumes: - /data/:/data/ ports: - "6379:6379" keystone: build: . command: node app.js ports: - "5000:3000" - "5001:3001" volumes: - ./:/usr/src/app ``` @yeelan0319 I think you're leaving out important details here. Could you paste your `package.json`, for a start? @aanand Glad to hear that may help. My application is just a keystone sample application which you can get it easily at [keystonejs.com](keystonejs.com). And here is the package.json: ``` { "name": "keystonejs", "version": "0.0.0", "private": true, "dependencies": { "async": "~0.9.0", "cloudinary": "^1.0.12", "connect-redis": "^2.2.0", "dotenv": "~0.4.0", "express-handlebars": "~1.1.0", "handlebars": "~2.0.0", "i18n": "^0.5.0", "keystone": "~0.3.0", "moment": "~2.8.1", "node-sass": "~1.0.1", "node-sass-middleware": "^0.4.0", "underscore": "~1.7.0" }, "devDependencies": { "gulp": "~3.7.0", "gulp-jshint": "~1.9.0", "jshint-stylish": "~0.1.3", "gulp-watch": "~0.6.5" }, "engines": { "node": ">=0.10.22", "npm": ">=1.3.14" }, "scripts": { "start": "node keystone.js" }, "main": "keystone.js" } ```
2015-04-30T10:58:59Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 31, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 27, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 59, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 119, in build File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 191, in build File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 475, in build File "/code/build/docker-compose/out00-PYZ.pyz/docker.client", line 300, in build File "/code/build/docker-compose/out00-PYZ.pyz/docker.utils.utils", line 73, in tar File "/code/build/docker-compose/out00-PYZ.pyz/os", line 294, in walk File "/code/build/docker-compose/out00-PYZ.pyz/os", line 294, in walk File "/code/build/docker-compose/out00-PYZ.pyz/os", line 294, in walk File "/code/build/docker-compose/out00-PYZ.pyz/os", line 294, in walk File "/code/build/docker-compose/out00-PYZ.pyz/os", line 294, in walk File "/code/build/docker-compose/out00-PYZ.pyz/os", line 284, in walk File "/code/build/docker-compose/out00-PYZ.pyz/posixpath", line 80, in join UnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position 5: ordinal not in range(128)
4,631
docker/compose
docker__compose-140
71533791dd31d0e4b511d07ee547497b97cc990d
diff --git a/fig/container.py b/fig/container.py --- a/fig/container.py +++ b/fig/container.py @@ -86,7 +86,10 @@ def human_readable_state(self): @property def human_readable_command(self): self.inspect_if_not_inspected() - return ' '.join(self.dictionary['Config']['Cmd']) + if self.dictionary['Config']['Cmd']: + return ' '.join(self.dictionary['Config']['Cmd']) + else: + return '' @property def environment(self):
Container List Fails I have a basic fig file with: ``` redis: image: dockerfile/redis ports: - "6379:6379" ``` and starting fig with `fig up -d` succeeds, however trying to list processes running throws the following error: ``` vagrant@docker:~$ fig ps Traceback (most recent call last): File "/usr/local/bin/fig", line 9, in <module> load_entry_point('fig==0.3.0', 'console_scripts', 'fig')() File "/usr/local/lib/python2.7/dist-packages/fig/cli/main.py", line 39, in main command.sys_dispatch() File "/usr/local/lib/python2.7/dist-packages/fig/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/local/lib/python2.7/dist-packages/fig/cli/command.py", line 30, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/local/lib/python2.7/dist-packages/fig/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/local/lib/python2.7/dist-packages/fig/cli/command.py", line 47, in perform_command return super(Command, self).perform_command(options, *args, **kwargs) File "/usr/local/lib/python2.7/dist-packages/fig/cli/docopt_command.py", line 27, in perform_command handler(command_options) File "/usr/local/lib/python2.7/dist-packages/fig/cli/main.py", line 158, in ps command = container.human_readable_command File "/usr/local/lib/python2.7/dist-packages/fig/container.py", line 89, in human_readable_command return ' '.join(self.dictionary['Config']['Cmd']) TypeError ```
Ahah – this is a Docker 0.8.1 thing, where `['Config']['Cmd']` might return `None`. I'll get this fixed for you.
2014-03-04T11:36:30Z
[]
[]
Traceback (most recent call last): File "/usr/local/bin/fig", line 9, in <module> load_entry_point('fig==0.3.0', 'console_scripts', 'fig')()
4,634
docker/compose
docker__compose-1642
db7e5124f312422b75f9ef537fc16987c729ef7d
diff --git a/compose/project.py b/compose/project.py --- a/compose/project.py +++ b/compose/project.py @@ -226,6 +226,9 @@ def up(self, services = self.get_services(service_names, include_deps=start_deps) + for service in services: + service.remove_duplicate_containers() + plans = self._get_convergence_plans( services, allow_recreate=allow_recreate, diff --git a/compose/service.py b/compose/service.py --- a/compose/service.py +++ b/compose/service.py @@ -378,6 +378,26 @@ def start_container(self, container): container.start() return container + def remove_duplicate_containers(self, timeout=DEFAULT_TIMEOUT): + for c in self.duplicate_containers(): + log.info('Removing %s...' % c.name) + c.stop(timeout=timeout) + c.remove() + + def duplicate_containers(self): + containers = sorted( + self.containers(stopped=True), + key=lambda c: c.get('Created'), + ) + + numbers = set() + + for c in containers: + if c.number in numbers: + yield c + else: + numbers.add(c.number) + def config_hash(self): return json_hash(self.config_dict())
Duplicate container is left over when recreate fails This is related to #1045. After hitting the issue mentioned in #1045, subsequent `docker-compose up` fails with: ``` $ sudo docker-compose up -d Recreating foo_bar_1... Traceback (most recent call last): ... requests.exceptions.ReadTimeout: UnixHTTPConnectionPool(host='localhost', port=None): Read timed out. (read timeout=60) $ sudo DOCKER_CLIENT_TIMEOUT=300 docker-compose up Recreating foo_bar_1... Recreating 9c1b31a841_foo_bar_1... Conflict. The name "foo_bar_1" is already in use by container 4af37fe03b76. You have to delete (or rename) that container to be able to reuse that name. ```
I also see a lot of these. I'm not sure if those are caused by #1045 Ubuntu 15.04; Docker 1.7 Looks related to #1349 - Compose has somehow gotten into a situation where there are two containers with the suffix `_1`, one of which has been renamed but then failed to recreate (perhaps because there was an HTTP timeout). I think after the failure, `9c1b31a841_foo_bar_1` still has the correct labels, so it's still considered an instance of the service. So there are two instances with the id of 1. compose would need to only recreate once per instance id. In many ways this is related to #1516. OK yeah, looks like this happens if we successfully create the new container but fail to start it. I can reproduce by sticking a `sys.exit()` in `recreate_container`: ``` python new_container = self.create_container(...) sys.exit(1) self.start_container(new_container) ``` This results in a renamed old container, plus an unstarted new container, both with `container-number=1`. > In many ways this is related to #1516. Random container numbers would solve the immediate error, yeah, but we'd still be left with two containers where we only wanted one. > I also see a lot of these. > > Ubuntu 15.04; Docker 1.7 Same combination, same problem. I also doubt it's #1045 I could get back to work by coming back to Docker 1.4.1 and docker-compose 1.2.0. This bug seems to show up if there is an exception (e.g. with #1045) or if you Ctrl+C in the middle of a `docker-compose up`. Essentially, docker-compose fails to clean up the temporary containers. Which version can I downgrade to so that I can get around this issue? @iMerica Run `docker-compose rm -f` to clear everything out, then downgrade to 1.2.0.
2015-07-03T09:47:08Z
[]
[]
Traceback (most recent call last): ... requests.exceptions.ReadTimeout: UnixHTTPConnectionPool(host='localhost', port=None): Read timed out. (read timeout=60)
4,651
docker/compose
docker__compose-1658
e9da790f760c8998412f1a44e99df47a71aa1719
diff --git a/compose/service.py b/compose/service.py --- a/compose/service.py +++ b/compose/service.py @@ -65,6 +65,10 @@ def __init__(self, service): self.service = service +class NoSuchImageError(Exception): + pass + + VolumeSpec = namedtuple('VolumeSpec', 'external internal mode') @@ -225,8 +229,11 @@ def ensure_image_exists(self, do_build=True, insecure_registry=False): - if self.image(): + try: + self.image() return + except NoSuchImageError: + pass if self.can_be_built(): if do_build: @@ -241,7 +248,7 @@ def image(self): return self.client.inspect_image(self.image_name) except APIError as e: if e.response.status_code == 404 and e.explanation and 'No such image' in str(e.explanation): - return None + raise NoSuchImageError("Image '{}' not found".format(self.image_name)) else: raise @@ -275,7 +282,17 @@ def convergence_plan(self, return ConvergencePlan('recreate', containers) def _containers_have_diverged(self, containers): - config_hash = self.config_hash() + config_hash = None + + try: + config_hash = self.config_hash() + except NoSuchImageError as e: + log.debug( + 'Service %s has diverged: %s', + self.name, six.text_type(e), + ) + return True + has_diverged = False for c in containers:
x-smart-recreate crashes when modified image name/tag Just got this error with docker-compose 1.3.0rc2 : ``` $ docker-compose up --x-smart-recreate Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 31, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 34, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 66, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 469, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 221, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 261, in _get_convergence_plans File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 285, in convergence_plan File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 299, in _containers_have_diverged File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 415, in config_hash File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 420, in config_dict TypeError: 'NoneType' object has no attribute '__getitem__' ``` Resolved after `docker-compose pull`
Can you provide more information please? What sequence of commands did you type? What does your YAML look like? Does the problem only occur if you use `--x-smart-recreate`? Simple case (tested with 1.4.0dev) : ``` $ mkdir demo $ cd demo $ cat <<EOF > docker-compose.yml example: image: debian:latest command: echo "hello world" EOF $ docker-compose up -d --x-smart-recreate Pulling example (debian:latest)... latest: Pulling from debian 64e5325c0d9d: Pull complete bf84c1d84a8f: Already exists debian:latest: The image you are pulling has been verified. Important: image verification is a tech preview feature and should not be relied on to provide security. Digest: sha256:2613dd69166e1bcc0a3e4b1f7cfe30d3dfde7762aea0e2f467632bda681d9765 Status: Downloaded newer image for debian:latest Creating demo_example_1... ``` => OK ``` $ cat <<EOF > docker-compose.yml example: image: debian:8.1 command: echo "hello world" EOF $ docker-compose up -d --x-smart-recreate Traceback (most recent call last): File "/usr/local/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.4.0.dev0', 'console_scripts', 'docker-compose')() File "/Users/joseph/git-workplace/contrib/compose/compose/cli/main.py", line 33, in main command.sys_dispatch() File "/Users/joseph/git-workplace/contrib/compose/compose/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/Users/joseph/git-workplace/contrib/compose/compose/cli/command.py", line 34, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/Users/joseph/git-workplace/contrib/compose/compose/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/Users/joseph/git-workplace/contrib/compose/compose/cli/command.py", line 66, in perform_command handler(project, command_options) File "/Users/joseph/git-workplace/contrib/compose/compose/cli/main.py", line 465, in up timeout=timeout File "/Users/joseph/git-workplace/contrib/compose/compose/project.py", line 232, in up smart_recreate=smart_recreate, File "/Users/joseph/git-workplace/contrib/compose/compose/project.py", line 273, in _get_convergence_plans smart_recreate=smart_recreate, File "/Users/joseph/git-workplace/contrib/compose/compose/service.py", line 264, in convergence_plan if smart_recreate and not self._containers_have_diverged(containers): File "/Users/joseph/git-workplace/contrib/compose/compose/service.py", line 278, in _containers_have_diverged config_hash = self.config_hash() File "/Users/joseph/git-workplace/contrib/compose/compose/service.py", line 382, in config_hash return json_hash(self.config_dict()) File "/Users/joseph/git-workplace/contrib/compose/compose/service.py", line 387, in config_dict 'image_id': self.image()['Id'], TypeError: 'NoneType' object has no attribute '__getitem__' ``` => FAILED with a new image tag ``` $ docker pull debian:8.1 8.1: Pulling from debian 64e5325c0d9d: Already exists bf84c1d84a8f: Already exists debian:8.1: The image you are pulling has been verified. Important: image verification is a tech preview feature and should not be relied on to provide security. Digest: sha256:44241b371e469ff74cd6b401231bfef80832108b5ee0e6bb7a2ac26a70326dd5 Status: Downloaded newer image for debian:8.1 $ docker-compose up -d --x-smart-recreate Recreating demo_example_1... ``` => OK after pulling with Docker cli ``` $ cat <<EOF > docker-compose.yml example: image: alpine:3.1 command: echo "hello world" EOF $ docker-compose up -d --x-smart-recreate Traceback (most recent call last): File "/usr/local/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.4.0.dev0', 'console_scripts', 'docker-compose')() File "/Users/joseph/git-workplace/contrib/compose/compose/cli/main.py", line 33, in main command.sys_dispatch() File "/Users/joseph/git-workplace/contrib/compose/compose/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/Users/joseph/git-workplace/contrib/compose/compose/cli/command.py", line 34, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/Users/joseph/git-workplace/contrib/compose/compose/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/Users/joseph/git-workplace/contrib/compose/compose/cli/command.py", line 66, in perform_command handler(project, command_options) File "/Users/joseph/git-workplace/contrib/compose/compose/cli/main.py", line 465, in up timeout=timeout File "/Users/joseph/git-workplace/contrib/compose/compose/project.py", line 232, in up smart_recreate=smart_recreate, File "/Users/joseph/git-workplace/contrib/compose/compose/project.py", line 273, in _get_convergence_plans smart_recreate=smart_recreate, File "/Users/joseph/git-workplace/contrib/compose/compose/service.py", line 264, in convergence_plan if smart_recreate and not self._containers_have_diverged(containers): File "/Users/joseph/git-workplace/contrib/compose/compose/service.py", line 278, in _containers_have_diverged config_hash = self.config_hash() File "/Users/joseph/git-workplace/contrib/compose/compose/service.py", line 382, in config_hash return json_hash(self.config_dict()) File "/Users/joseph/git-workplace/contrib/compose/compose/service.py", line 387, in config_dict 'image_id': self.image()['Id'], TypeError: 'NoneType' object has no attribute '__getitem__' ``` => FAILED with a new image name ``` $ docker pull alpine:3.1 3.1: Pulling from alpine 878b6301beda: Already exists alpine:3.1: The image you are pulling has been verified. Important: image verification is a tech preview feature and should not be relied on to provide security. Digest: sha256:5083b4c32bc726ff6f074e5513615ec0befcc9a69224c3be9261513cabe2cf6e Status: Downloaded newer image for alpine:3.1 $ docker-compose up -d --x-smart-recreate Recreating demo_example_1... ``` => OK after pulling with Docker cli The bug appears when changing the tag or the name of the image.
2015-07-06T14:34:04Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 31, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 34, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 66, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 469, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 221, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 261, in _get_convergence_plans File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 285, in convergence_plan File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 299, in _containers_have_diverged File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 415, in config_hash File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 420, in config_dict TypeError: 'NoneType' object has no attribute '__getitem__'
4,656
docker/compose
docker__compose-166
a3374ac51d7a83d09ac83980264ffdd6e68d6e7e
diff --git a/fig/cli/log_printer.py b/fig/cli/log_printer.py --- a/fig/cli/log_printer.py +++ b/fig/cli/log_printer.py @@ -18,7 +18,7 @@ def __init__(self, containers, attach_params=None): def run(self): mux = Multiplexer(self.generators) for line in mux.loop(): - sys.stdout.write(line) + sys.stdout.write(line.encode(sys.__stdout__.encoding or 'utf8')) def _make_log_generators(self): color_fns = cycle(colors.rainbow()) diff --git a/fig/cli/socketclient.py b/fig/cli/socketclient.py --- a/fig/cli/socketclient.py +++ b/fig/cli/socketclient.py @@ -81,7 +81,7 @@ def recv(self, socket, stream): chunk = socket.recv(4096) if chunk: - stream.write(chunk) + stream.write(chunk.encode(stream.encoding or 'utf8')) stream.flush() else: break diff --git a/fig/service.py b/fig/service.py --- a/fig/service.py +++ b/fig/service.py @@ -295,7 +295,7 @@ def build(self): match = re.search(r'Successfully built ([0-9a-f]+)', line) if match: image_id = match.group(1) - sys.stdout.write(line) + sys.stdout.write(line.encode(sys.__stdout__.encoding or 'utf8')) if image_id is None: raise BuildError()
UnicodeEncodeError: 'ascii' codec can't encode character u'\ufffd' in position 66: ordinal not in range(128) I sporadically get the following error while building a container with PHP 5: ``` db uses an image, skipping Building web... Step 1 : FROM doos/base ---> e12d62b19d01 Step 2 : RUN add-apt-repository ppa:ondrej/php5 ---> Running in f79de09aa4e8 Executing: gpg --ignore-time-conflict --no-options --no-default-keyring --secret-keyring /tmp/tmp.cCNHEbGwpL --trustdb-name /etc/apt/trustdb.gpg --keyring /etc/apt/trusted.gpg --primary-keyring /etc/apt/trusted.gpg --keyserver hkp://keyserver.ubuntu.com:80/ --recv 14AA40EC0831756756D7F66C4F4EA0AAE5267A6C gpg: requesting key E5267A6C from hkp server keyserver.ubuntu.com gpg: Traceback (most recent call last): File "/usr/local/bin/fig", line 9, in <module> load_entry_point('fig==0.2.2', 'console_scripts', 'fig')() File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/fig/cli/main.py", line 39, in main command.sys_dispatch() File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/fig/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/fig/cli/command.py", line 26, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/fig/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/fig/cli/docopt_command.py", line 27, in perform_command handler(command_options) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/fig/cli/main.py", line 103, in build self.project.build(service_names=options['SERVICE']) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/fig/project.py", line 128, in build service.build(**options) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/fig/service.py", line 284, in build sys.stdout.write(line) UnicodeEncodeError: 'ascii' codec can't encode character u'\ufffd' in position 66: ordinal not in range(128) ``` Here are my Dockerfiles and Fig file: http://pastebin.com/raw.php?i=ghJ32AcD I'm on Mavericks 10.9.2 using VirtualBox 4.3.6, Vagrant 1.4.3, docker-osx & docker 0.8.0, and fig 0.2.2. Any idea what might cause this?
2014-03-25T13:43:42Z
[]
[]
Traceback (most recent call last): File "/usr/local/bin/fig", line 9, in <module> load_entry_point('fig==0.2.2', 'console_scripts', 'fig')() File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/fig/cli/main.py", line 39, in main command.sys_dispatch() File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/fig/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/fig/cli/command.py", line 26, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/fig/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/fig/cli/docopt_command.py", line 27, in perform_command handler(command_options) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/fig/cli/main.py", line 103, in build self.project.build(service_names=options['SERVICE']) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/fig/project.py", line 128, in build service.build(**options) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/fig/service.py", line 284, in build sys.stdout.write(line) UnicodeEncodeError: 'ascii' codec can't encode character u'\ufffd' in position 66: ordinal not in range(128)
4,657
docker/compose
docker__compose-1705
5a462305552de3272d9b54f25644f1e1de39f174
diff --git a/compose/legacy.py b/compose/legacy.py --- a/compose/legacy.py +++ b/compose/legacy.py @@ -149,7 +149,7 @@ def _get_legacy_containers_iter( for service in services: for container in containers: - if LABEL_VERSION in container['Labels']: + if LABEL_VERSION in (container.get('Labels') or {}): continue name = get_container_name(container)
_get_legacy_containers_iter() TypeError: argument of type 'NoneType' is not iterable I just upgraded from 1.1.0 to 1.3.2, and after dealing with the /tmp directory being mounted as noexec (issue #1339), I ran into another issue that I couldn't find in the backlog; ``` $ docker-compose up -d Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 32, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 34, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 66, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 471, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 230, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 398, in remove_duplicate_containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 405, in duplicate_containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 112, in containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.legacy", line 56, in check_for_legacy_containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.legacy", line 138, in get_legacy_containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.legacy", line 152, in _get_legacy_containers_iter TypeError: argument of type 'NoneType' is not iterable ``` Downgrading to 1.3.1 seems to alleviate this behavior.
Which docker version are you using? In docker 1.6 "no labels" is an empty mapping, but this errors seems like it's returning `null` instead of the empty mapping. My docker version: ``` Docker version 1.6.2, build 7c8fca2 ``` I suspect this pull is related: #1643 Can confirm the same behaviour - `Docker version 1.6.2, build 7c8fca2`. Upgraded from docker-compose 1.1.0 to 1.3.2 via `pip install -U docker-compose` and I get the following result from `docker-compose migrate-to-labels`: ``` Traceback (most recent call last): File "/usr/local/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.3.2', 'console_scripts', 'docker-compose')() File "/usr/local/lib/python2.7/dist-packages/compose/cli/main.py", line 32, in main command.sys_dispatch() File "/usr/local/lib/python2.7/dist-packages/compose/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/local/lib/python2.7/dist-packages/compose/cli/command.py", line 34, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/local/lib/python2.7/dist-packages/compose/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/local/lib/python2.7/dist-packages/compose/cli/command.py", line 66, in perform_command handler(project, command_options) File "/usr/local/lib/python2.7/dist-packages/compose/cli/main.py", line 515, in migrate_to_labels legacy.migrate_project_to_labels(project) File "/usr/local/lib/python2.7/dist-packages/compose/legacy.py", line 121, in migrate_project_to_labels one_off=False, File "/usr/local/lib/python2.7/dist-packages/compose/legacy.py", line 138, in get_legacy_containers one_off=one_off, File "/usr/local/lib/python2.7/dist-packages/compose/legacy.py", line 152, in _get_legacy_containers_iter if LABEL_VERSION in container['Labels']: TypeError: argument of type 'NoneType' is not iterable ``` Suggest simply prefixing the erroneous conditional on the last frame of that stacktrace (compose/legacy.py:152) with `container['Labels'] and`. If I get a chance later today I'll submit a PR myself, but I'm happy for someone to beat me to the punch! Seeing this with `Docker version 1.7.0, build 0baf609`, so it's not just a docker-1.6 issue. Can also confirm this same behavior with `Docker version 1.7.0, build 0baf609` and `Docker version 1.6.1, build 97cd073` with `docker-compose version: 1.3.2`. After reverting back to `docker-compose version: 1.2.0`, I no longer see the issue. I think the easy fix is to change `container['Labels']` to `container.get('Labels', {})`, but I'm still not sure how this is happening, since it seems like the default empty value should always be the empty mapping `{}` not None. Maybe this behaviour differs based on the version of docker that was used to create the container? Do you happen to know if the old containers were created with a version of docker < 1.6 ? A paste of the `docker inspect <container name>` for one of the old containers would be really helpful for debugging this. I'm seeing this and I think had a set of container that I upgraded to labels with the `docker-compose migrate-to-labels` command.
2015-07-15T16:14:53Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 32, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 34, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 66, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 471, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 230, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 398, in remove_duplicate_containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 405, in duplicate_containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 112, in containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.legacy", line 56, in check_for_legacy_containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.legacy", line 138, in get_legacy_containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.legacy", line 152, in _get_legacy_containers_iter TypeError: argument of type 'NoneType' is not iterable
4,660
docker/compose
docker__compose-1835
6cb8e512f211696d6c1ba5482606d791eaf302a3
diff --git a/compose/container.py b/compose/container.py --- a/compose/container.py +++ b/compose/container.py @@ -22,10 +22,14 @@ def from_ps(cls, client, dictionary, **kwargs): """ Construct a container object from the output of GET /containers/json. """ + name = get_container_name(dictionary) + if name is None: + return None + new_dictionary = { 'Id': dictionary['Id'], 'Image': dictionary['Image'], - 'Name': '/' + get_container_name(dictionary), + 'Name': '/' + name, } return cls(client, new_dictionary, **kwargs) diff --git a/compose/project.py b/compose/project.py --- a/compose/project.py +++ b/compose/project.py @@ -310,11 +310,11 @@ def containers(self, service_names=None, stopped=False, one_off=False): else: service_names = self.service_names - containers = [ + containers = filter(None, [ Container.from_ps(self.client, container) for container in self.client.containers( all=stopped, - filters={'label': self.labels(one_off=one_off)})] + filters={'label': self.labels(one_off=one_off)})]) def matches_service_names(container): return container.labels.get(LABEL_SERVICE) in service_names diff --git a/compose/service.py b/compose/service.py --- a/compose/service.py +++ b/compose/service.py @@ -101,11 +101,11 @@ def __init__(self, name, client=None, project='default', links=None, external_li self.options = options def containers(self, stopped=False, one_off=False): - containers = [ + containers = filter(None, [ Container.from_ps(self.client, container) for container in self.client.containers( all=stopped, - filters={'label': self.labels(one_off=one_off)})] + filters={'label': self.labels(one_off=one_off)})]) if not containers: check_for_legacy_containers( @@ -494,12 +494,13 @@ def get_container_name(self, number, one_off=False): # TODO: this would benefit from github.com/docker/docker/pull/11943 # to remove the need to inspect every container def _next_container_number(self, one_off=False): - numbers = [ - Container.from_ps(self.client, container).number + containers = filter(None, [ + Container.from_ps(self.client, container) for container in self.client.containers( all=True, filters={'label': self.labels(one_off=one_off)}) - ] + ]) + numbers = [c.number for c in containers] return 1 if not numbers else max(numbers) + 1 def _get_links(self, link_to_self):
TypeError: coercing to Unicode: need string or buffer, NoneType found The first time I do `docker-compose up -d` everything works The second time I get the following error ``` Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 31, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 34, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 66, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 460, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 221, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 261, in _get_convergence_plans File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 278, in convergence_plan File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 101, in containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.container", line 28, in from_ps TypeError: coercing to Unicode: need string or buffer, NoneType found ``` ``` docker-compose --version docker-compose version: 1.3.1 CPython version: 2.7.9 OpenSSL version: OpenSSL 1.0.1e 11 Feb 2013 ``` ``` docker info Containers: 8 Images: 98 Storage Driver: aufs Root Dir: /var/lib/docker/aufs Backing Filesystem: extfs Dirs: 114 Dirperm1 Supported: false Execution Driver: native-0.2 Logging Driver: json-file Kernel Version: 3.13.0-53-generic Operating System: Ubuntu 14.04.2 LTS CPUs: 1 Total Memory: 992.5 MiB Name: ip-10-0-11-99 ID: 3RM6:3ZR5:FK7P:N6TM:CE35:2VFA:BUJ7:UDBR:IUDR:OSB7:7BP2:TEHS WARNING: No swap limit support ```
@sstarcher try downgrading to docker-compose v1.2.0 I've been having a similar issue with a php-fpm container. The root cause is the "affinity:container" ENV variable that is being added when docker-compose **recreates** containers. See https://github.com/docker/compose/issues/1598 for details. +1 @sstarcher @devinrsmith I can't reproduce. Could you paste your `docker-compose.yml`? Was it previously working with Compose 1.2 or is this your first time running it? @aanand I'll get you an example tomorrow. It happened on several of our servers with different docker compose files. It was previously working on 1.2. Broke on 1.3. Downgraded to 1.2, back and working. We see the same error from docker-compose 1.3.1 and docker 1.7.0 using AUFS. It's working for me, but exactly the same setup isn't working for my Greek colleague who gets the same error as in the issue. Downgrading to docker-compose 1.2.0 is one solution, but is there a better solution? I have the same issue. However, when I use `docker-compose migrate-to-labels`, the containers are created and started correctly. When I removed some old (dead and exited) containers, `docker-compose up -d` is working again. [Identical Issue here too](https://gist.github.com/bb1daba5966d2958fdc1) I _might_ try to fix this tomorrow and submit a pull request (_unless someone is already on it!_) Same issue here, I was able to fix it by removing some dead containers, everything is going well again. b368272 at least tries to fix this; I'm not sure how `get_container_name()` should change for a "Dead" container though. docker-compose 1.3.1 docker 1.7.0 docker-compose.yml ``` { "image1": { "image": "ubuntu", "volumes": [ "/:/logstashforwarder:ro" ], "command": "sleep 10000" }, "image2": { "image": "ubuntu", "command": "sleep 10000", "volumes": [ "/:/hostfs:ro" ] } } ``` Run the following twice docker-compose up -d ``` Recreating opt_image2_1... Cannot destroy container 982a0d62eab102811a5a697a3df2de46fad5c7acd34a9bb5bea44e98d47fb1ee: Driver aufs failed to remove root filesystem 982a0d62eab102811a5a697a3df2de46fad5c7acd34a9bb5bea44e98d47fb1ee: rename /var/lib/docker/aufs/mnt/982a0d62eab102811a5a697a3df2de46fad5c7acd34a9bb5bea44e98d47fb1ee /var/lib/docker/aufs/mnt/982a0d62eab102811a5a697a3df2de46fad5c7acd34a9bb5bea44e98d47fb1ee-removing: device or resource busy ``` Afterwards if you try docker-compose rm you will get ``` docker-compose rm Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 31, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 34, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 66, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 247, in rm File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 281, in containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.container", line 28, in from_ps TypeError: coercing to Unicode: need string or buffer, NoneType found ``` FWIW, confirmed with 1.3.3, rolling back to 1.2.0 indeed fixes the issue. @prologic Can you paste the JSON of a "Dead" container? I can't reproduce one locally. I'm also not sure what we should do if we can't determine a container's name. I get this error (1.3.3) when doing `compose run --rm run bash`. I have a service called `run`. Changing the service name works (e.g. `runa`). This had been working (I think 1.3.0 was the last version I had - it was 1.3.something) up until now that I have 1.3.3. I can confirm this: ``` docker-compose up -d Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 32, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 34, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 66, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 471, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 230, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 398, in remove_duplicate_containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 405, in duplicate_containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 112, in containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.legacy", line 56, in check_for_legacy_containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.legacy", line 138, in get_legacy_containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.legacy", line 152, in _get_legacy_containers_iter TypeError: argument of type 'NoneType' is not iterable ``` Docker Compose ``` docker-compose --version docker-compose version: 1.3.2 CPython version: 2.7.9 OpenSSL version: OpenSSL 1.0.1e 11 Feb 2013 ``` Docker ``` docker version Client version: 1.7.1 Client API version: 1.19 Go version (client): go1.4.2 Git commit (client): 786b29d OS/Arch (client): linux/amd64 Server version: 1.7.1 Server API version: 1.19 Go version (server): go1.4.2 Git commit (server): 786b29d OS/Arch (server): linux/amd64 ``` docker-compose.yml ``` wildfly: build: ./sp ports: - "8080:8080" - "9990:9990" - "39331:8787" links: - mysql:db volumes: - ./volumes/remote_deploy:/opt/jboss/wildfly/standalone/deployments - ./volumes/sp_properties:/opt/jboss/wildfly/standalone/sp_properties mysql: image: mysql environment: MYSQL_USER: mysql MYSQL_PASSWORD: mysql MYSQL_DATABASE: test ``` A rollback to 1.2.0 fixed the issue. I think same issue here: ``` # docker-compose logs Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 32, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 34, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 66, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 158, in logs File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 301, in containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.container", line 28, in from_ps TypeError: coercing to Unicode: need string or buffer, NoneType found # docker-compose --version docker-compose version: 1.3.3 CPython version: 2.7.9 OpenSSL version: OpenSSL 1.0.1e 11 Feb 2013 ``` when `docker-compose up -d <service_name>` in progress (linked service starting) - same issue ``` $ docker-compose ps Traceback (most recent call last): File "/usr/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.3.3', 'console_scripts', 'docker-compose')() File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 32, in main command.sys_dispatch() File "/usr/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 34, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 66, in perform_command handler(project, command_options) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 195, in ps project.containers(service_names=options['SERVICE'], stopped=True) + File "/usr/lib/python2.7/site-packages/compose/project.py", line 301, in containers filters={'label': self.labels(one_off=one_off)})] File "/usr/lib/python2.7/site-packages/compose/container.py", line 28, in from_ps 'Name': '/' + get_container_name(dictionary), TypeError: coercing to Unicode: need string or buffer, NoneType found ``` Versions ``` $ sudo docker version Client version: 1.7.1 Client API version: 1.19 Go version (client): go1.4.2 Git commit (client): 786b29d OS/Arch (client): linux/amd64 Server version: 1.7.1 Server API version: 1.19 Go version (server): go1.4.2 Git commit (server): 786b29d OS/Arch (server): linux/amd64 sudo docker-compose --version docker-compose version: 1.3.3 CPython version: 2.7.10 OpenSSL version: OpenSSL 1.0.2d 9 Jul 2015 ``` we are receiving the same error with latest pip release fnigi]# docker-compose -v dockdocker-compose version: 1.4.0rc2 fnigi]# docker -v Docker version 1.6.0, build 4749651 the same with version 1.7.1 and 1.3.3 od docker-compose the same configuration work on an upgraded python env on Python 2.6.6 (r266:84292, Jan 22 2014, 09:42:36) seem broken Anyone experiencing this issue: please paste the output of `docker ps -a` immediately after you get the error. in compose/container.py add line `print('get_container_name - container:', container) # debug` in get_container_name function: ``` def get_container_name(container): print('get_container_name - container:', container) # debug if not container.get('Name') and not container.get('Names'): return None # inspect if 'Name' in container: return container['Name'] # ps shortest_name = min(container['Names'], key=lambda n: len(n.split('/'))) return shortest_name.split('/')[-1] ``` When ``` $ sudo docker-compose up -d <service-name> ``` Then Immediately call ``` $ sudo docker-compose ps ``` And result is ``` .... (u'get_container_name - container:', {u'Status': u'Removal In Progress', u'Created': 1438267340, u'Image': u'ib_frontend', u'Labels': {u'com.docker.compose.service': u'frontend', u'com.docker.compose.config-hash': u'e527b462027b756c37e4c631b21f143dac9f1d39c986ee23b4d0e9acef2f62b1', u'com.docker.compose.project': u'ib', u'com.docker.compose.version': u'1.3.3', u'com.docker.compose.oneoff': u'False', u'com.docker.compose.container-number': u'1'}, u'Ports': [], u'Command': u'gulp --debug', u'Names': None, u'Id': u'2b8755775154015d668ef4212d6470f07ad3b83dd01c24e57bd6c2997de69d28'}) Traceback (most recent call last): File "/usr/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.3.3', 'console_scripts', 'docker-compose')() File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 32, in main command.sys_dispatch() File "/usr/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 34, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 66, in perform_command handler(project, command_options) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 195, in ps project.containers(service_names=options['SERVICE'], stopped=True) + File "/usr/lib/python2.7/site-packages/compose/project.py", line 301, in containers filters={'label': self.labels(one_off=one_off)})] File "/usr/lib/python2.7/site-packages/compose/container.py", line 28, in from_ps 'Name': '/' + get_container_name(dictionary), TypeError: coercing to Unicode: need string or buffer, NoneType found ``` On this dictionary get_container_name return None and we have TypeError ``` {u'Command': u'gulp --debug', u'Created': 1438267340, u'Id': u'2b8755775154015d668ef4212d6470f07ad3b83dd01c24e57bd6c2997de69d28', u'Image': u'ib_frontend', u'Labels': {u'com.docker.compose.config-hash': u'e527b462027b756c37e4c631b21f143dac9f1d39c986ee23b4d0e9acef2f62b1', u'com.docker.compose.container-number': u'1', u'com.docker.compose.oneoff': u'False', u'com.docker.compose.project': u'ib', u'com.docker.compose.service': u'frontend', u'com.docker.compose.version': u'1.3.3'}, u'Names': None, u'Ports': [], u'Status': u'Removal In Progress'} ``` Names are None After <service-name> started `docker-complse ps` works without any errors. When immediately call of `sudo docker ps` ``` $ sudo docker ps .... waiting about 1-2 seconds and then produced output ... CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 4628aaee3ed7 ib_backend "/var/webapps/IB/con About a minute ago Up About a minute 8000/tcp ib_backend_run_1 dc80f79e28da memcached "memcached -m 64" 26 minutes ago Up 26 minutes 11211/tcp ib_memcached_1 546f431ea39d elasticsearch "/var/webapps/IB/con 26 minutes ago Up 26 minutes 9000/tcp, 9300/tcp, 0.0.0.0:9200->9200/tcp ib_elasticsearch_1 b4a5d058d1c0 ib_redis "/usr/bin/redis-serv 26 minutes ago Up 26 minutes 6379/tcp ib_redis_1 27e3541b7778 ib_postgres "/var/webapps/IB/con 26 minutes ago Up 26 minutes 5432/tcp ``` Sorry for my poor English. and another shot of `docker-compose ps` and `docker ps` when `docker-compose up -d` in progress ``` $ sudo docker-compose ps ... (u'get_container_name - container:', {u'Status': u'Removal In Progress', u'Created': 1438280708, u'Image': u'elasticsearch', u'Labels': {u'com.docker.compose.service': u'elasticsearch', u'com.docker.compose.config-hash': u'e83fac7baf70ed5d259c9c73b36224d7df63fc0c979750388ad1205350a5503f', u'com.docker.compose.project': u'ib', u'com.docker.compose.version': u'1.3.3', u'com.docker.compose.oneoff': u'False', u'com.docker.compose.container-number': u'1'}, u'Ports': [], u'Command': u'/var/webapps/IB/containers/elasticsearch_entrypoint.sh elasticsearch -Des.config=/var/webapps/IB/containers/elasticsearch.yml', u'Names': None, u'Id': u'546f431ea39ddff8c3c3c8b798c17a93348013d4c5a17cf46696087533349ff6'}) Traceback (most recent call last): File "/usr/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.3.3', 'console_scripts', 'docker-compose')() File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 32, in main command.sys_dispatch() File "/usr/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 34, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 66, in perform_command handler(project, command_options) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 195, in ps project.containers(service_names=options['SERVICE'], stopped=True) + File "/usr/lib/python2.7/site-packages/compose/project.py", line 301, in containers filters={'label': self.labels(one_off=one_off)})] File "/usr/lib/python2.7/site-packages/compose/container.py", line 28, in from_ps 'Name': '/' + get_container_name(dictionary), TypeError: coercing to Unicode: need string or buffer, NoneType found $ sudo docker ps CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES e8a2ea48e3b0 elasticsearch "/var/webapps/IB/con 4 seconds ago Up 1 seconds 9000/tcp, 9300/tcp, 0.0.0.0:9200->9200/tcp ib_elasticsearch_1 d07381a7f001 ib_redis "/usr/bin/redis-serv 19 seconds ago Up 14 seconds 6379/tcp ib_redis_1 e07808b3640e ib_postgres "/var/webapps/IB/con 30 seconds ago Up 27 seconds 5432/tcp ib_postgres_1 dc80f79e28da memcached "memcached -m 64" 33 minutes ago Up 33 minutes 11211/tcp ib_memcached_1 ``` Sorry, I need the output of `docker ps -a` - updated the comment. In your first paste, it looks like you've got a container without a name - the one running the `ib_postgres` image. It looks like the `/containers/json` API endpoint can sometimes return a container JSON blob without a name. That definitely sounds like a Docker bug to me, so let's figure out how to reproduce it. It could be related to renaming containers. Anyone who's experiencing this, please: 1. Clear everything out with `docker-compose kill && docker-compose rm -f` 2. Use `docker-compose --verbose up -d` to start/recreate your containers 3. When you encounter the error, paste the output of the last `docker-compose --verbose up -d` command you ran 4. Also paste the output of `docker ps -a` > 1. Clear everything out with docker-compose kill && docker-compose rm -f @aanand, this issue correspond to container state `Removal in Progress` and when `docker-compose (up|rm|kill)` will be recreate/remove some containers for linking it to another. this patch worked for me: ``` diff diff --git a/compose/cli/docker_client.py b/compose/cli/docker_client.py index 244bcbe..56cf4f8 100644 --- a/compose/cli/docker_client.py +++ b/compose/cli/docker_client.py @@ -4,6 +4,15 @@ import ssl import os +# issue-1593 compose not interesting in 'Removal In Progress' containers +class ComposeDockerClient(Client): + def containers(self, *args, **kwargs): + return list(filter( + lambda i: i['Status'] != 'Removal In Progress', + super(ComposeDockerClient, self).containers(*args, **kwargs), + )) + + def docker_client(): """ Returns a docker-py client configured using environment variables @@ -34,4 +43,6 @@ def docker_client(): ) timeout = int(os.environ.get('DOCKER_CLIENT_TIMEOUT', 60)) - return Client(base_url=base_url, tls=tls_config, version=api_version, timeout=timeout) + return ComposeDockerClient( + base_url=base_url, tls=tls_config, version=api_version, timeout=timeout + ) ``` Sorry for my poor English. PS. may be docker-py must have new filter params to exclude `Removal in Progress` containers ``` Compose version 1.3.2 Docker base_url: http+docker://localunixsocket Docker version: KernelVersion=2.6.32-504.16.2.el6.x86_64, Arch=amd64, ApiVersion=1.18, Version=1.6.2, GitCommit=7c8fca2, Os=linux, GoVersion=go1.4.2 docker containers <- (all=True, filters={u'label': [u'com.docker.compose.project=dockerregistry', u'com.docker.compose.service=DockerRegistry', u'com.docker.compose.oneoff=False']}) docker containers -> (list with 0 items) docker containers <- (all=True) docker containers -> (list with 0 items) docker containers <- (all=True, filters={u'label': [u'com.docker.compose.project=dockerregistry', u'com.docker.compose.service=DockerRegistryUI', u'com.docker.compose.oneoff=False']}) docker containers -> (list with 0 items) docker containers <- (all=True) docker containers -> (list with 0 items) docker containers <- (all=True, filters={u'label': [u'com.docker.compose.project=dockerregistry', u'com.docker.compose.service=DockerRegistry', u'com.docker.compose.oneoff=False']}) docker containers -> (list with 0 items) docker containers <- (all=True) docker containers -> (list with 0 items) docker containers <- (all=True, filters={u'label': [u'com.docker.compose.project=dockerregistry', u'com.docker.compose.service=DockerRegistryUI', u'com.docker.compose.oneoff=False']}) docker containers -> (list with 0 items) docker containers <- (all=True) docker containers -> (list with 0 items) docker inspect_image <- ('docker-registry.mitre.org:8443/autobuild/docker_registry:9') docker inspect_image -> {u'Architecture': u'amd64', u'Author': u'Venkat Natarajan <venkat@mitre.org>', u'Comment': u'', u'Config': {u'AttachStderr': False, u'AttachStdin': False, u'AttachStdout': False, u'Cmd': None, u'CpuShares': 0, u'Cpuset': u'', u'Domainname': u'', ... docker containers <- (all=True, filters={u'label': [u'com.docker.compose.project=dockerregistry', u'com.docker.compose.service=DockerRegistry', u'com.docker.compose.oneoff=False']}) docker containers -> (list with 0 items) docker inspect_image <- ('docker-registry.mitre.org:8443/autobuild/docker_registry:9') docker inspect_image -> {u'Architecture': u'amd64', u'Author': u'Venkat Natarajan <venkat@mitre.org>', u'Comment': u'', u'Config': {u'AttachStderr': False, u'AttachStdin': False, u'AttachStdout': False, u'Cmd': None, u'CpuShares': 0, u'Cpuset': u'', u'Domainname': u'', ... Creating dockerregistry_DockerRegistry_1... docker create_container <- (name=u'dockerregistry_DockerRegistry_1', image='docker-registry.mitre.org:8443/autobuild/docker_registry:9', labels={u'com.docker.compose.service': u'DockerRegistry', 'com.docker.compose.config-hash': 'f5d00447e880c7fa0ccc47dde09aad90fb7136e944c2769897980ea46a1af4ed', u'com.docker.compose.project': u'dockerregistry', 'com.docker.compose.version': u'1.3.2', 'mitre.release': '@@RELEASE@@', u'com.docker.compose.oneoff': u'False', 'com.docker.compose.container-number': '1', 'mitre.environment': '@@ENVIRONMENT@@'}, host_config={'Links': [], 'PortBindings': {'443/tcp': [{'HostPort': '8443', 'HostIp': ''}], '5000/tcp': [{'HostPort': '5000', 'HostIp': u'127.0.0.1'}]}, 'Binds': [u'/data/docker_registry/config:/data/docker_registry/config:ro', u'/etc/pki/tls/private/r10a-venkat-docker.key:/etc/ssl/private/ssl-cert-snakeoil.key:rw', u'/data/docker_registry/includes/authnz_ldap.conf:/etc/apache2/conf-available/authnz_ldap.conf:ro', u'/data/docker_registry:/docker_registry:rw', u'/logs/docker_registry:/var/log/apache2:rw', u'/etc/pki/tls/certs/r10a-venkat-docker.crt:/etc/ssl/certs/ssl-cert-snakeoil.pem:rw'], 'RestartPolicy': {u'MaximumRetryCount': 0, u'Name': u'always'}, 'ExtraHosts': [], 'LogConfig': {'Type': u'json-file', 'Config': {}}, 'VolumesFrom': []}, environment={'SETTINGS_FLAVOR': 'dev', 'SEARCH_BACKEND': 'sqlalchemy', 'SQLALCHEMY_INDEX_DATABASE': 'sqlite:////tmp/docker-registry.db', 'DOCKER_REGISTRY_CONFIG': '/data/docker_registry/config/config.yml'}, volumes={u'/data/docker_registry/config': {}, u'/etc/ssl/private/ssl-cert-snakeoil.key': {}, u'/etc/apache2/conf-available/authnz_ldap.conf': {}, u'/docker_registry': {}, u'/var/log/apache2': {}, u'/etc/ssl/certs/ssl-cert-snakeoil.pem': {}}, detach=True, ports=[u'443', u'5000']) docker create_container -> {u'Id': u'd95374a1ead2b98f089c4d5e0e3a6a42fb0702ef267f2edde0b380a81c592c42', u'Warnings': None} docker inspect_container <- (u'd95374a1ead2b98f089c4d5e0e3a6a42fb0702ef267f2edde0b380a81c592c42') docker inspect_container -> {u'AppArmorProfile': u'', u'Args': [u'-c', u'/etc/supervisord.conf'], u'Config': {u'AttachStderr': False, u'AttachStdin': False, u'AttachStdout': False, u'Cmd': None, u'CpuShares': 0, u'Cpuset': u'', u'Domainname': u'', u'Entrypoint': [u'/usr/bin/supervisord', ... docker start <- (u'd95374a1ead2b98f089c4d5e0e3a6a42fb0702ef267f2edde0b380a81c592c42') docker start -> None docker inspect_image <- ('docker-registry.mitre.org:8443/autobuild/docker_registry_ui:9') docker inspect_image -> {u'Architecture': u'amd64', u'Author': u'Venkatesan Natarajan <venkat@mitre.org>', u'Comment': u'', u'Config': {u'AttachStderr': False, u'AttachStdin': False, u'AttachStdout': False, u'Cmd': [u'/bin/sh', u'-c', u'$START_SCRIPT'], u'CpuShares': 0, u'Cpuset': u'', u'Domainname': u'', ... docker containers <- (all=True, filters={u'label': [u'com.docker.compose.project=dockerregistry', u'com.docker.compose.service=DockerRegistryUI', u'com.docker.compose.oneoff=False']}) docker containers -> (list with 0 items) docker inspect_image <- ('docker-registry.mitre.org:8443/autobuild/docker_registry_ui:9') docker inspect_image -> {u'Architecture': u'amd64', u'Author': u'Venkatesan Natarajan <venkat@mitre.org>', u'Comment': u'', u'Config': {u'AttachStderr': False, u'AttachStdin': False, u'AttachStdout': False, u'Cmd': [u'/bin/sh', u'-c', u'$START_SCRIPT'], u'CpuShares': 0, u'Cpuset': u'', u'Domainname': u'', ... docker containers <- (all=False, filters={u'label': [u'com.docker.compose.project=dockerregistry', u'com.docker.compose.service=DockerRegistry', u'com.docker.compose.oneoff=False']}) docker containers -> (list with 1 items) docker inspect_container <- (u'd95374a1ead2b98f089c4d5e0e3a6a42fb0702ef267f2edde0b380a81c592c42') docker inspect_container -> {u'AppArmorProfile': u'', u'Args': [u'-c', u'/etc/supervisord.conf'], u'Config': {u'AttachStderr': False, u'AttachStdin': False, u'AttachStdout': False, u'Cmd': None, u'CpuShares': 0, u'Cpuset': u'', u'Domainname': u'', u'Entrypoint': [u'/usr/bin/supervisord', ... Creating dockerregistry_DockerRegistryUI_1... docker create_container <- (name=u'dockerregistry_DockerRegistryUI_1', image='docker-registry.mitre.org:8443/autobuild/docker_registry_ui:9', labels={u'com.docker.compose.service': u'DockerRegistryUI', 'com.docker.compose.config-hash': '8949aef8e893f2bd04b031dd96801466d7dc22ed2923920769e7201c2c939ca8', u'com.docker.compose.project': u'dockerregistry', 'com.docker.compose.version': u'1.3.2', 'mitre.release': '@@RELEASE@@', u'com.docker.compose.oneoff': u'False', 'com.docker.compose.container-number': '1', 'mitre.environment': '@@ENVIRONMENT@@'}, host_config={'Links': ['dockerregistry_DockerRegistry_1:DockerRegistry_1', 'dockerregistry_DockerRegistry_1:docker-registry.mitre.org', 'dockerregistry_DockerRegistry_1:dockerregistry_DockerRegistry_1'], 'PortBindings': {'443/tcp': [{'HostPort': '444', 'HostIp': ''}]}, 'Binds': [u'/etc/pki/tls/private/r10a-venkat-docker.key:/etc/apache2/docker.key:ro', u'/data/docker_registry_ui/includes/authnz_ldap.conf:/etc/apache2/conf-enabled/authnz_ldap.conf:rw', u'/logs/docker_registry_ui:/var/log/apache2:rw', u'/etc/pki/tls/certs/r10a-venkat-docker.crt:/etc/apache2/docker.crt:ro'], 'RestartPolicy': {u'MaximumRetryCount': 0, u'Name': u'always'}, 'ExtraHosts': [], 'LogConfig': {'Type': u'json-file', 'Config': {}}, 'VolumesFrom': []}, environment={'ENV_DOCKER_REGISTRY_PORT': '8443', 'ENV_DOCKER_REGISTRY_HOST': 'docker-registry.mitre.org', 'ENV_USE_SSL': 'yes'}, volumes={u'/etc/apache2/docker.key': {}, u'/etc/apache2/conf-enabled/authnz_ldap.conf': {}, u'/var/log/apache2': {}, u'/etc/apache2/docker.crt': {}}, detach=True, ports=[u'443']) docker create_container -> {u'Id': u'03373d7eaba13df0c686aa3097ca1602c492d2a0800a5ea97330024464b47c1e', u'Warnings': None} docker inspect_container <- (u'03373d7eaba13df0c686aa3097ca1602c492d2a0800a5ea97330024464b47c1e') docker inspect_container -> {u'AppArmorProfile': u'', u'Args': [u'-c', u'$START_SCRIPT'], u'Config': {u'AttachStderr': False, u'AttachStdin': False, u'AttachStdout': False, u'Cmd': [u'/bin/sh', u'-c', u'$START_SCRIPT'], u'CpuShares': 0, u'Cpuset': u'', u'Domainname': u'', u'Entrypoint': None, ... docker start <- (u'03373d7eaba13df0c686aa3097ca1602c492d2a0800a5ea97330024464b47c1e') docker start -> None docker containers <- (all=False, filters={u'label': [u'com.docker.compose.project=dockerregistry', u'com.docker.compose.service=DockerRegistry', u'com.docker.compose.oneoff=False']}) docker containers -> (list with 1 items) docker containers <- (all=False, filters={u'label': [u'com.docker.compose.project=dockerregistry', u'com.docker.compose.service=DockerRegistryUI', u'com.docker.compose.oneoff=False']}) docker containers -> (list with 1 items) ----------------------------------------------------------- Compose version 1.3.2 Docker base_url: http+docker://localunixsocket Docker version: KernelVersion=2.6.32-504.16.2.el6.x86_64, Arch=amd64, ApiVersion=1.18, Version=1.6.2, GitCommit=7c8fca2, Os=linux, GoVersion =go1.4.2 docker containers <- (all=True, filters={u'label': [u'com.docker.compose.project=dockerregistry', u'com.docker.compose.service=DockerRegistr y', u'com.docker.compose.oneoff=False']}) docker containers -> (list with 1 items) Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 32, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 34, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 66, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 471, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 230, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 398, in remove_duplicate_containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 405, in duplicate_containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 106, in containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.container", line 28, in from_ps TypeError: coercing to Unicode: need string or buffer, NoneType found ``` ``` [venkat@r10a-venkat-docker bin]$ docker ps -a CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES ef9075284e6c docker-registry.mitre.org:8443/autobuild/mitre_open_docker_registry_ui:18 "/bin/sh -c $START_S 3 minutes ago Dead 03373d7eaba1 docker-registry.mitre.org:8443/autobuild/docker_registry_ui:9 "/bin/sh -c $START_S 3 minutes ago Exited (137) 2 minutes ago dockerregistry_DockerRegistryUI_1 08ab0eaf1b69 docker-registry.mitre.org:8443/autobuild/mitre_open_docker_registry_ui:18 "/bin/sh -c $START_S 22 minutes ago Dead loving_meitner 9dcfb52a9a1a docker-registry.mitre.org:8443/autobuild/mitre_open_docker_registry_ui:18 "/bin/sh -c $START_S 34 minutes ago Dead angry_fermi 471f00dccbe2 docker-registry.mitre.org:8443/autobuild/docker_registry_ui:9 "/bin/sh -c $START_S 35 minutes ago Dead loving_hodgkin fe121f0792a4 docker-registry.mitre.org:8443/autobuild/docker_registry:9 "/usr/bin/supervisor 35 minutes ago Dead 8d713fdd23ea docker-registry.mitre.org:8443/autobuild/mitre_open_docker_registry_ui:18 "/bin/sh -c $START_S 2 weeks ago Dead angry_darwin f2f298d71e6f docker-registry.mitre.org:8443/autobuild/docker_registry_ui:9 "/bin/sh -c $START_S 2 weeks ago Dead adoring_davinci ``` I've implemented a fix here: https://github.com/aanand/fig/compare/bump-1.4.0...aanand:fix-name-none Please try it out if you've been experiencing this issue: http://cl.ly/0G3c3y0y1k19/docker-compose-Darwin-x86_64 http://cl.ly/0r3J020O2s2C/docker-compose-Linux-x86_64 Works for me call `ps` without fix ``` $ docker-compose ps Traceback (most recent call last): File "/usr/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.3.3', 'console_scripts', 'docker-compose')() File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 32, in main command.sys_dispatch() File "/usr/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 34, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 66, in perform_command handler(project, command_options) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 195, in ps project.containers(service_names=options['SERVICE'], stopped=True) + File "/usr/lib/python2.7/site-packages/compose/project.py", line 301, in containers filters={'label': self.labels(one_off=one_off)})] File "/usr/lib/python2.7/site-packages/compose/container.py", line 28, in from_ps 'Name': '/' + get_container_name(dictionary), TypeError: coercing to Unicode: need string or buffer, NoneType found ``` and call `ps` with fix ``` $ /tmp/docker-compose-Linux-x86_64 ps Name Command State Ports ------------------------------------------------------------------------------------------------------------ 64ffb1d005_ib_redis_1 /usr/bin/redis-server Exit 0 ib_backend_1 /var/webapps/IB/containers ... Up 0.0.0.0:8000->8000/tcp ib_celery_1 /var/webapps/IB/containers ... Up ib_elasticsearch_1 /var/webapps/IB/containers ... Up 9000/tcp, 0.0.0.0:9200->9200/tcp, 9300/tcp ib_frontend_1 gulp --debug Up 0.0.0.0:5000->5000/tcp ib_memcached_1 memcached -m 64 Up 11211/tcp ib_postgres_1 /var/webapps/IB/containers ... Up 5432/tcp ib_redis_1 /usr/bin/redis-server Up 6379/tcp ib_ws_1 /var/webapps/IB/containers ... Up 0.0.0.0:8080->8080/tcp ``` When do double call `up -d` without fix ``` $ docker-compose up -d Traceback (most recent call last): File "/usr/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.3.3', 'console_scripts', 'docker-compose')() File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 32, in main command.sys_dispatch() File "/usr/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 34, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 66, in perform_command handler(project, command_options) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 471, in up timeout=timeout File "/usr/lib/python2.7/site-packages/compose/project.py", line 230, in up service.remove_duplicate_containers() File "/usr/lib/python2.7/site-packages/compose/service.py", line 398, in remove_duplicate_containers for c in self.duplicate_containers(): File "/usr/lib/python2.7/site-packages/compose/service.py", line 405, in duplicate_containers self.containers(stopped=True), File "/usr/lib/python2.7/site-packages/compose/service.py", line 106, in containers filters={'label': self.labels(one_off=one_off)})] File "/usr/lib/python2.7/site-packages/compose/container.py", line 28, in from_ps 'Name': '/' + get_container_name(dictionary), TypeError: coercing to Unicode: need string or buffer, NoneType found ``` When do double call `up -d` with fix ``` $ /tmp/docker-compose-Linux-x86_64 up -d Removing ib_elasticsearch_1... Removing ib_memcached_1... Cannot destroy container 7cce897e11ad63816286bcf70cdf742f28d006e28d5b6831fd49922a5c690db7: Driver devicemapper failed to remove init filesystem 7cce897e11ad63816286bcf70cdf742f28d006e28d5b6831fd49922a5c690db7-init: Device is Busy ``` Where `double call` is call `docker-compose up -d` at second time util first `up -d` is not done even with the fix, when running a `docker-compose run` having the service being `up` already, returns a weird message: ``` Could not get container for ``` I also noticed that with the patch, my containers no longer carry the name that `docker-compose` lists on the logs when `up`ing it. So when I do `docker-compose ps` and `docker ps` I see the container but the name doesn't show on the second command. May be something with docker itself? I'm running the latest `1.4.1` there @robertoandrade Do you mean 1.7.1? We don't support 1.4.x any more.
2015-08-10T15:28:55Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 31, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 34, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 66, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 460, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 221, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.project", line 261, in _get_convergence_plans File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 278, in convergence_plan File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 101, in containers File "/code/build/docker-compose/out00-PYZ.pyz/compose.container", line 28, in from_ps TypeError: coercing to Unicode: need string or buffer, NoneType found
4,671
docker/compose
docker__compose-1859
7d5e26bd37f740944ca2640b23c144ec85dd856e
diff --git a/compose/config/config.py b/compose/config/config.py --- a/compose/config/config.py +++ b/compose/config/config.py @@ -186,8 +186,16 @@ def resolve_extends(self, service_dict): already_seen=other_already_seen, ) + base_service = extends_options['service'] other_config = load_yaml(other_config_path) - other_service_dict = other_config[extends_options['service']] + + if base_service not in other_config: + msg = ( + "Cannot extend service '%s' in %s: Service not found" + ) % (base_service, other_config_path) + raise ConfigurationError(msg) + + other_service_dict = other_config[base_service] other_loader.detect_cycle(extends_options['service']) other_service_dict = other_loader.make_service_dict( service_dict['name'],
Stack trace when trying to extend a nonexistent service ``` yaml web: image: busybox extends: service: foo ``` ``` $ docker-compose up -d Traceback (most recent call last): File "/Users/aanand/.virtualenvs/docker-compose/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.4.0.dev0', 'console_scripts', 'docker-compose')() File "/Users/aanand/work/docker/compose/compose/cli/main.py", line 39, in main command.sys_dispatch() File "/Users/aanand/work/docker/compose/compose/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/Users/aanand/work/docker/compose/compose/cli/command.py", line 27, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/Users/aanand/work/docker/compose/compose/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/Users/aanand/work/docker/compose/compose/cli/command.py", line 57, in perform_command verbose=options.get('--verbose')) File "/Users/aanand/work/docker/compose/compose/cli/command.py", line 78, in get_project config.load(config_details), File "/Users/aanand/work/docker/compose/compose/config/config.py", line 132, in load service_dict = loader.make_service_dict(service_name, service_dict) File "/Users/aanand/work/docker/compose/compose/config/config.py", line 156, in make_service_dict service_dict = self.resolve_extends(service_dict) File "/Users/aanand/work/docker/compose/compose/config/config.py", line 183, in resolve_extends other_service_dict = other_config[extends_options['service']] KeyError: 'foo' ```
2015-08-12T21:36:21Z
[]
[]
Traceback (most recent call last): File "/Users/aanand/.virtualenvs/docker-compose/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.4.0.dev0', 'console_scripts', 'docker-compose')() File "/Users/aanand/work/docker/compose/compose/cli/main.py", line 39, in main command.sys_dispatch() File "/Users/aanand/work/docker/compose/compose/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/Users/aanand/work/docker/compose/compose/cli/command.py", line 27, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/Users/aanand/work/docker/compose/compose/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/Users/aanand/work/docker/compose/compose/cli/command.py", line 57, in perform_command verbose=options.get('--verbose')) File "/Users/aanand/work/docker/compose/compose/cli/command.py", line 78, in get_project config.load(config_details), File "/Users/aanand/work/docker/compose/compose/config/config.py", line 132, in load service_dict = loader.make_service_dict(service_name, service_dict) File "/Users/aanand/work/docker/compose/compose/config/config.py", line 156, in make_service_dict service_dict = self.resolve_extends(service_dict) File "/Users/aanand/work/docker/compose/compose/config/config.py", line 183, in resolve_extends other_service_dict = other_config[extends_options['service']] KeyError: 'foo'
4,672
docker/compose
docker__compose-2018
e9871b084e65001aa3d260fea18263018491568d
diff --git a/compose/utils.py b/compose/utils.py --- a/compose/utils.py +++ b/compose/utils.py @@ -21,7 +21,6 @@ def parallel_execute(objects, obj_callable, msg_index, msg): """ stream = get_output_stream(sys.stdout) lines = [] - errors = {} for obj in objects: write_out_msg(stream, lines, msg_index(obj), msg) @@ -29,16 +28,17 @@ def parallel_execute(objects, obj_callable, msg_index, msg): q = Queue() def inner_execute_function(an_callable, parameter, msg_index): + error = None try: result = an_callable(parameter) except APIError as e: - errors[msg_index] = e.explanation + error = e.explanation result = "error" except Exception as e: - errors[msg_index] = e + error = e result = 'unexpected_exception' - q.put((msg_index, result)) + q.put((msg_index, result, error)) for an_object in objects: t = Thread( @@ -49,15 +49,17 @@ def inner_execute_function(an_callable, parameter, msg_index): t.start() done = 0 + errors = {} total_to_execute = len(objects) while done < total_to_execute: try: - msg_index, result = q.get(timeout=1) + msg_index, result, error = q.get(timeout=1) if result == 'unexpected_exception': - raise errors[msg_index] + errors[msg_index] = result, error if result == 'error': + errors[msg_index] = result, error write_out_msg(stream, lines, msg_index, msg, status='error') else: write_out_msg(stream, lines, msg_index, msg) @@ -65,10 +67,14 @@ def inner_execute_function(an_callable, parameter, msg_index): except Empty: pass - if errors: - stream.write("\n") - for error in errors: - stream.write("ERROR: for {} {} \n".format(error, errors[error])) + if not errors: + return + + stream.write("\n") + for msg_index, (result, error) in errors.items(): + stream.write("ERROR: for {} {} \n".format(msg_index, error)) + if result == 'unexpected_exception': + raise error def get_output_stream(stream):
test_scale_with_api_returns_unexpected_exception fails frequently on jenkins `tests/integration/service_test.py ServiceTest.test_scale_with_api_returns_unexpected_exception` aka `Test that when scaling if the API returns an error, that is not of type` fails in many of our jenkins run - https://jenkins.dockerproject.org/job/Compose-PRs/1213/console with error: ``` Traceback (most recent call last): File "/code/tests/integration/testcases.py", line 22, in tearDown self.client.remove_container(c['Id']) File "/code/.tox/py27/lib/python2.7/site-packages/docker/utils/decorators.py", line 20, in wrapped return f(self, resource_id, *args, **kwargs) File "/code/.tox/py27/lib/python2.7/site-packages/docker/client.py", line 696, in remove_container self._raise_for_status(res) File "/code/.tox/py27/lib/python2.7/site-packages/docker/clientbase.py", line 103, in _raise_for_status raise errors.NotFound(e, response, explanation=explanation) NotFound: 404 Client Error: Not Found ("Cannot destroy container faeacc68496793312980e36e20be6c7ae2a04a9b5f98c12f815a2235963ae834: Driver overlay failed to remove root filesystem faeacc68496793312980e36e20be6c7ae2a04a9b5f98c12f815a2235963ae834: stat /var/lib/docker/overlay/faeacc68496793312980e36e20be6c7ae2a04a9b5f98c12f815a2235963ae834: no such file or directory") ``` This is a docker engine bug, but something about this specific test triggers it frequently. Looking at the test, I can't really tell what it might be. The failure is coming from `tearDown()`. We could wrap this cleanup in a try/except, which would have the added benefit of completing the rest of the cleanup.
Ok, I spent some time looking into this today, and I've made some progress. If I add a `time.sleep(0.1)` [here](https://github.com/docker/compose/blob/1.4.0/tests/integration/testcases.py#L21) between the `kill` and the `rm` it doesn't fail. So it likely has to do with container state. If I inspect the container before the `rm` it's in state created, but the directory for the container does not exist on disk. I'll continue trying to track down the error in the engine, but for now I think our options are either to try/except and skip the remove, or attempt removing the container a couple times with a short sleep between attempts. Ok, so this isn't a daemon bug after-all! It's a problem with how we handle the threads in `parallel_execute()`. When we encounter an error, we're not joining the threads, so some can still be running after the first one hits an exception. In some cases, the `mock.patch()` is removed before the thread runs `create()` so we actually create a container, and when we attempt to remove it, the daemon is still in the process of creating it. There's probably a missing lock in the daemon somewhere (preventing commands from trying to inspect or remove a half-created container), but we can still fix this on the compose side by waiting for the threads to complete before exiting the function.
2015-09-09T18:44:15Z
[]
[]
Traceback (most recent call last): File "/code/tests/integration/testcases.py", line 22, in tearDown self.client.remove_container(c['Id']) File "/code/.tox/py27/lib/python2.7/site-packages/docker/utils/decorators.py", line 20, in wrapped return f(self, resource_id, *args, **kwargs) File "/code/.tox/py27/lib/python2.7/site-packages/docker/client.py", line 696, in remove_container self._raise_for_status(res) File "/code/.tox/py27/lib/python2.7/site-packages/docker/clientbase.py", line 103, in _raise_for_status raise errors.NotFound(e, response, explanation=explanation) NotFound: 404 Client Error: Not Found ("Cannot destroy container faeacc68496793312980e36e20be6c7ae2a04a9b5f98c12f815a2235963ae834: Driver overlay failed to remove root filesystem faeacc68496793312980e36e20be6c7ae2a04a9b5f98c12f815a2235963ae834: stat /var/lib/docker/overlay/faeacc68496793312980e36e20be6c7ae2a04a9b5f98c12f815a2235963ae834: no such file or directory")
4,686
docker/compose
docker__compose-2055
f49fc1867c2d4ea80a1de33cc9531834efc96d03
diff --git a/compose/cli/multiplexer.py b/compose/cli/multiplexer.py --- a/compose/cli/multiplexer.py +++ b/compose/cli/multiplexer.py @@ -2,6 +2,8 @@ from threading import Thread +from six.moves import _thread as thread + try: from Queue import Queue, Empty except ImportError: @@ -38,6 +40,9 @@ def loop(self): yield item except Empty: pass + # See https://github.com/docker/compose/issues/189 + except thread.error: + raise KeyboardInterrupt() def _init_readers(self): for iterator in self.iterators:
Ctrl-C on `docker-compose up` raises thread.error I get this intermittently (sometimes it does shutdown gracefully): ``` ^CGracefully stopping... (press Ctrl+C again to force) Stopping dailylead_db_1... Stopping dailylead_queue_1... Traceback (most recent call last): File "<string>", line 3, in <module> File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.main", line 39, in main File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 21, in sys_dispatch File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.command", line 30, in dispatch File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 24, in dispatch File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.command", line 47, in perform_command File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 27, in perform_command File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.main", line 316, in up File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.log_printer", line 20, in run File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.multiplexer", line 20, in loop File "/Users/ben/fig/build/fig/out00-PYZ.pyz/Queue", line 182, in get thread.error: release unlocked lock ```
me too I've seen this too - perhaps KeyboardInterrupt messes with Queue somehow. Not sure if there's a clean solution - perhaps we should just catch `thread.error` in this loop: https://github.com/orchardup/fig/blob/master/fig/cli/multiplexer.py#L24-L3 On Mon, Jul 21, 2014 at 6:24 AM, Patrick Arminio notifications@github.com wrote: > me too > > — > Reply to this email directly or view it on GitHub > https://github.com/orchardup/fig/issues/189#issuecomment-49604372. So the combination of threads and signals (SIGINT in this case) doesn't play well together (IIRC, signals are dispatched to one thread and it's non-deterministic which thread that will be). I'm sure this can be fixed, though I may just change how the multiplexer works to use `select()` instead of threads and blocking IO. That _should_ just make this problem just disappear, but we'll see. When you have multiple streams you want to read from, the idiomatic approach is to use `select()` in any case. I should have time to do that this weekend, as this is biting us too. Still seeing this on 0.5.2, is it fixed in mainline? not yet. happens to me too. I'm new to fig. Anyone know how you end up clearing the process it leaves behind. It leaves behind my nodejs process running and with a state of `Exit -1`? Edit: As soon as I asked, I figured it out. `fig rm`. Sorry about that. +1, happening to me as well. Using the passenger image. +1 while using `spotify/cassandra` Same here with 1.0.1 using custom image and mysql. I just hit this. It seems hard to reproduce. I ran it ten times and I didn't hit the "release unlocked lock" bug, but I did hit this: ``` Traceback (most recent call last): File "/code/build/fig/out00-PYZ.pyz/threading", line 552, in __bootstrap_inner File "/code/build/fig/out00-PYZ.pyz/threading", line 505, in run File "/code/build/fig/out00-PYZ.pyz/fig.cli.multiplexer", line 41, in _enqueue_output File "/code/build/fig/out00-PYZ.pyz/fig.cli.log_printer", line 56, in _make_log_generator File "/code/build/fig/out00-PYZ.pyz/fig.cli.utils", line 77, in split_buffer File "/code/build/fig/out00-PYZ.pyz/docker.client", line 332, in _multiplexed_socket_stream_helper File "/code/build/fig/out00-PYZ.pyz/docker.client", line 319, in recvall error: [Errno 4] Interrupted system call ``` In my setup I'm using fig to just run one service. Same here with **1.1.0-rc2**. ``` Traceback (most recent call last): File "<string>", line 3, in <module> File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 31, in main File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 27, in dispatch File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 59, in perform_command File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 455, in up File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.cli.log_printer", line 22, in run File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/compose.cli.multiplexer", line 25, in loop File "/Users/ben/fig/build/docker-compose/out00-PYZ.pyz/Queue", line 182, in get thread.error: release unlocked lock ``` Yep same here = only seems to happen when I introduced a queuing server (rabbitmq) Definitely seeing it here; for me, it's when I ctrl-c to stop a multi-container app that includes couchdb, and it happens on the stop of the couchdb container. Any news on this? `Multiplexer` still uses threading, so this still happens. If anyone wants to take a crack at refactoring it to use IO streams + `select()` instead of generators + threading, go for it. ``` ^CGracefully stopping... (press Ctrl+C again to force) Stopping blog_phpmyadmin_1... Stopping blog_web_1... Stopping blog_mysql_1... Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 32, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 34, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 66, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 481, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.log_printer", line 22, in run File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.multiplexer", line 25, in loop File "/code/build/docker-compose/out00-PYZ.pyz/Queue", line 182, in get thread.error: release unlocked lock ``` ``` $ docker-compose -v docker-compose version: 1.3.2 CPython version: 2.7.9 OpenSSL version: OpenSSL 1.0.1e 11 Feb 2013 $ docker -v Docker version 1.7.1, build 786b29d $ docker info Containers: 3 Images: 117 Storage Driver: aufs Root Dir: /var/lib/docker/aufs Backing Filesystem: extfs Dirs: 190 Dirperm1 Supported: false Execution Driver: native-0.2 Logging Driver: json-file Kernel Version: 3.13.0-24-generic Operating System: Ubuntu 14.04.2 LTS CPUs: 4 Total Memory: 14.69 GiB Name: desktop ID: MQKY:YIVZ:XCTH:DRAP:QYGC:SLX5:Z3HU:EZSW:E4T5:GYLP:4WMR:RSMI WARNING: No swap limit support ``` +1 :+1: ``` ^CGracefully stopping... (press Ctrl+C again to force) Stopping nginx_registry_1... Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 32, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 34, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 66, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 481, in up File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.log_printer", line 22, in run File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.multiplexer", line 25, in loop File "/code/build/docker-compose/out00-PYZ.pyz/Queue", line 182, in get thread.error: release unlocked lock ``` It also happen to me. My docker version like following. ``` $ docker-compose -v docker-compose version: 1.3.3 CPython version: 2.7.9 OpenSSL version: OpenSSL 1.0.1e 11 Feb 2013 ``` ``` $ docker -v Docker version 1.8.1, build d12ea79 ``` ``` $ docker info Containers: 78 Images: 360 Storage Driver: devicemapper Pool Name: docker-253:0-134331561-pool Pool Blocksize: 65.54 kB Backing Filesystem: xfs Data file: /dev/loop0 Metadata file: /dev/loop1 Data Space Used: 8.428 GB Data Space Total: 107.4 GB Data Space Available: 38.98 GB Metadata Space Used: 19.23 MB Metadata Space Total: 2.147 GB Metadata Space Available: 2.128 GB Udev Sync Supported: true Deferred Removal Enabled: false Data loop file: /var/lib/docker/devicemapper/devicemapper/data Metadata loop file: /var/lib/docker/devicemapper/devicemapper/metadata Library Version: 1.02.93-RHEL7 (2015-01-28) Execution Driver: native-0.2 Logging Driver: json-file Kernel Version: 3.10.0-229.11.1.el7.x86_64 Operating System: CentOS Linux 7 (Core) CPUs: 4 Total Memory: 7.585 GiB Name: nexhub06.ndeploy ID: CKJK:WFUB:JXXS:OG7K:U7ST:GUNU:HR3M:2J2A:LWTL:MVCW:PBWP:FXY3 ``` Seeing it here too. Is system info helpful or superfluous? I've been using compose from `pip install` and never once saw this issue. I just recently tried the binary and I get it almost every time I ctrl-c the logs. I think it might have something to do with the version of python we're building against, or pyinstaller itself.
2015-09-16T00:57:48Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.main", line 39, in main File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 21, in sys_dispatch File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.command", line 30, in dispatch File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 24, in dispatch File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.command", line 47, in perform_command File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 27, in perform_command File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.main", line 316, in up File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.log_printer", line 20, in run File "/Users/ben/fig/build/fig/out00-PYZ.pyz/fig.cli.multiplexer", line 20, in loop File "/Users/ben/fig/build/fig/out00-PYZ.pyz/Queue", line 182, in get thread.error: release unlocked lock
4,694
docker/compose
docker__compose-2132
13d5efcd8b646eaf1d2d1f4733737e5b28c0df4c
diff --git a/compose/cli/log_printer.py b/compose/cli/log_printer.py --- a/compose/cli/log_printer.py +++ b/compose/cli/log_printer.py @@ -6,8 +6,8 @@ from . import colors from .multiplexer import Multiplexer -from .utils import split_buffer from compose import utils +from compose.utils import split_buffer class LogPrinter(object): @@ -75,7 +75,7 @@ def build_no_log_generator(container, prefix, color_func): def build_log_generator(container, prefix, color_func): # Attach to container before log printer starts running stream = container.attach(stdout=True, stderr=True, stream=True, logs=True) - line_generator = split_buffer(stream, u'\n') + line_generator = split_buffer(stream) for line in line_generator: yield prefix + line diff --git a/compose/cli/utils.py b/compose/cli/utils.py --- a/compose/cli/utils.py +++ b/compose/cli/utils.py @@ -7,7 +7,6 @@ import ssl import subprocess -import six from docker import version as docker_py_version from six.moves import input @@ -36,31 +35,6 @@ def yesno(prompt, default=None): return None -def split_buffer(reader, separator): - """ - Given a generator which yields strings and a separator string, - joins all input, splits on the separator and yields each chunk. - - Unlike string.split(), each chunk includes the trailing - separator, except for the last one if none was found on the end - of the input. - """ - buffered = six.text_type('') - separator = six.text_type(separator) - - for data in reader: - buffered += data.decode('utf-8') - while True: - index = buffered.find(separator) - if index == -1: - break - yield buffered[:index + 1] - buffered = buffered[index + 1:] - - if len(buffered) > 0: - yield buffered - - def call_silently(*args, **kwargs): """ Like subprocess.call(), but redirects stdout and stderr to /dev/null. diff --git a/compose/container.py b/compose/container.py --- a/compose/container.py +++ b/compose/container.py @@ -212,9 +212,6 @@ def links(self): def attach(self, *args, **kwargs): return self.client.attach(self.id, *args, **kwargs) - def attach_socket(self, **kwargs): - return self.client.attach_socket(self.id, **kwargs) - def __repr__(self): return '<Container: %s (%s)>' % (self.name, self.id[:6]) diff --git a/compose/progress_stream.py b/compose/progress_stream.py --- a/compose/progress_stream.py +++ b/compose/progress_stream.py @@ -1,7 +1,3 @@ -import json - -import six - from compose import utils @@ -16,10 +12,7 @@ def stream_output(output, stream): lines = {} diff = 0 - for chunk in output: - if six.PY3: - chunk = chunk.decode('utf-8') - event = json.loads(chunk) + for event in utils.json_stream(output): all_events.append(event) if 'progress' in event or 'progressDetail' in event: diff --git a/compose/service.py b/compose/service.py --- a/compose/service.py +++ b/compose/service.py @@ -34,6 +34,7 @@ from .utils import json_hash from .utils import parallel_execute + log = logging.getLogger(__name__) diff --git a/compose/utils.py b/compose/utils.py --- a/compose/utils.py +++ b/compose/utils.py @@ -1,6 +1,7 @@ import codecs import hashlib import json +import json.decoder import logging import sys from threading import Thread @@ -13,6 +14,8 @@ log = logging.getLogger(__name__) +json_decoder = json.JSONDecoder() + def parallel_execute(objects, obj_callable, msg_index, msg): """ @@ -83,6 +86,71 @@ def get_output_stream(stream): return codecs.getwriter('utf-8')(stream) +def stream_as_text(stream): + """Given a stream of bytes or text, if any of the items in the stream + are bytes convert them to text. + + This function can be removed once docker-py returns text streams instead + of byte streams. + """ + for data in stream: + if not isinstance(data, six.text_type): + data = data.decode('utf-8') + yield data + + +def line_splitter(buffer, separator=u'\n'): + index = buffer.find(six.text_type(separator)) + if index == -1: + return None, None + return buffer[:index + 1], buffer[index + 1:] + + +def split_buffer(stream, splitter=None, decoder=lambda a: a): + """Given a generator which yields strings and a splitter function, + joins all input, splits on the separator and yields each chunk. + + Unlike string.split(), each chunk includes the trailing + separator, except for the last one if none was found on the end + of the input. + """ + splitter = splitter or line_splitter + buffered = six.text_type('') + + for data in stream_as_text(stream): + buffered += data + while True: + item, rest = splitter(buffered) + if not item: + break + + buffered = rest + yield item + + if buffered: + yield decoder(buffered) + + +def json_splitter(buffer): + """Attempt to parse a json object from a buffer. If there is at least one + object, return it and the rest of the buffer, otherwise return None. + """ + try: + obj, index = json_decoder.raw_decode(buffer) + rest = buffer[json.decoder.WHITESPACE.match(buffer, index).end():] + return obj, rest + except ValueError: + return None, None + + +def json_stream(stream): + """Given a stream of text, return a stream of json objects. + This handles streams which are inconsistently buffered (some entries may + be newline delimited, and others are not). + """ + return split_buffer(stream_as_text(stream), json_splitter, json_decoder.decode) + + def write_out_msg(stream, lines, msg_index, msg, status="done"): """ Using special ANSI code characters we can write out the msg over the top of
docker-compose crashes on build - json data from daemon is not parsed correctly When trying to build my images using docker-compose, it does not work. Using `docker build` does work. This is the command I run. ``` $ docker-compose --verbose build sonar Compose version 1.1.0 Docker base_url: http://192.168.59.103:2375 Docker version: KernelVersion=3.18.5-tinycore64, Arch=amd64, ApiVersion=1.17, Version=1.5.0, GitCommit=a8a31ef, Os=linux, GoVersion=go1.4.1 Building sonar... docker build <- ('sonar', rm=True, tag=u'sonardocker_sonar', nocache=False, stream=True) docker build -> <generator object _stream_helper at 0x10534b190> Step 0 : FROM xxx.xxx:5000/base-java ---> 82af2109b0f0 Step 1 : MAINTAINER Tim Soethout Traceback (most recent call last): File "/usr/local/Cellar/fig/1.1.0/libexec/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.1.0', 'console_scripts', 'docker-compose')() File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/main.py", line 31, in main command.sys_dispatch() File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/command.py", line 27, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/command.py", line 59, in perform_command handler(project, command_options) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/main.py", line 119, in build project.build(service_names=options['SERVICE'], no_cache=no_cache) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/project.py", line 167, in build service.build(no_cache) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/service.py", line 484, in build all_events = stream_output(build_output, sys.stdout) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/progress_stream.py", line 18, in stream_output event = json.loads(chunk) File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/json/__init__.py", line 338, in loads return _default_decoder.decode(s) File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/json/decoder.py", line 368, in decode raise ValueError(errmsg("Extra data", s, end, len(s))) ValueError: Extra data: line 2 column 1 - line 4 column 1 (char 39 - 162) ``` This last line suggest something in the parsing of the daemon response is not working as expected. On the docker daemon side (boot2docker) I have the following logs. It seems that the build works as expected, only docker-compose can't handle the json it gets back from the daemon. ``` time="2015-03-04T19:31:36Z" level="debug" msg="Calling GET /version" time="2015-03-04T19:31:36Z" level="info" msg="GET /v1.14/version" time="2015-03-04T19:31:36Z" level="info" msg="+job version()" time="2015-03-04T19:31:36Z" level="info" msg="-job version() = OK (0)" time="2015-03-04T19:31:36Z" level="debug" msg="Calling POST /build" time="2015-03-04T19:31:36Z" level="info" msg="POST /v1.14/build?pull=True&nocache=False&q=False&t=sonardocker_sonar&forcerm=False&rm=True" time="2015-03-04T19:31:36Z" level="info" msg="+job build()" time="2015-03-04T19:31:36Z" level="debug" msg="[BUILDER] Use cached version" time="2015-03-04T19:31:36Z" level="debug" msg="[BUILDER] Command to be executed: [/bin/sh -c apt-get update && apt-get install procps -y]" time="2015-03-04T19:31:36Z" level="debug" msg="[BUILDER] Use cached version" time="2015-03-04T19:31:36Z" level="debug" msg="[BUILDER] Command to be executed: [/bin/sh -c apt-get update && apt-get install -y --force-yes sonar]" time="2015-03-04T19:31:36Z" level="debug" msg="[BUILDER] Use cached version" time="2015-03-04T19:31:36Z" level="debug" msg="[BUILDER] Use cached version" time="2015-03-04T19:31:36Z" level="debug" msg="[BUILDER] Command to be executed: [/bin/sh -c chmod 755 /app/init]" time="2015-03-04T19:31:36Z" level="debug" msg="[BUILDER] Use cached version" time="2015-03-04T19:31:36Z" level="debug" msg="[BUILDER] Use cached version" time="2015-03-04T19:31:36Z" level="debug" msg="[BUILDER] Use cached version" time="2015-03-04T19:31:36Z" level="debug" msg="[BUILDER] Use cached version" time="2015-03-04T19:31:37Z" level="debug" msg="[BUILDER] Use cached version" time="2015-03-04T19:31:37Z" level="info" msg="-job build() = OK (0)" ``` What I notice is that docker-compose uses the `v1.14` version of the api, while the "normal" `docker build` uses the `v1.17` api: ``` time="2015-03-04T19:42:51Z" level="debug" msg="Calling POST /build" time="2015-03-04T19:42:51Z" level="info" msg="POST /v1.17/build?dockerfile=Dockerfile&rm=1&t=sonar" time="2015-03-04T19:42:51Z" level="info" msg="+job build()" time="2015-03-04T19:42:51Z" level="debug" msg="[BUILDER] Use cached version" time="2015-03-04T19:42:51Z" level="debug" msg="[BUILDER] Command to be executed: [/bin/sh -c apt-get update && apt-get install procps -y]" time="2015-03-04T19:42:51Z" level="debug" msg="[BUILDER] Use cached version" time="2015-03-04T19:42:51Z" level="debug" msg="[BUILDER] Command to be executed: [/bin/sh -c apt-get update && apt-get install -y --force-yes sonar]" time="2015-03-04T19:42:51Z" level="debug" msg="[BUILDER] Use cached version" time="2015-03-04T19:42:51Z" level="debug" msg="[BUILDER] Cache miss" time="2015-03-04T19:42:51Z" level="debug" msg="CopyFileWithTar(/mnt/sda1/var/lib/docker/tmp/docker-build999836939/assets/init, /mnt/sda1/var/lib/docker/aufs/mnt/46a584cbc5150d0a3286716ad43f18817572ef84a990138185d5285d1ef163ee/app/init)" time="2015-03-04T19:42:51Z" level="debug" msg="Skipping excluded path: .wh..wh.aufs" time="2015-03-04T19:42:51Z" level="debug" msg="Skipping excluded path: .wh..wh.orph" time="2015-03-04T19:42:51Z" level="debug" msg="Skipping excluded path: .wh..wh.plnk" time="2015-03-04T19:42:51Z" level="debug" msg="[BUILDER] Command to be executed: [/bin/sh -c chmod 755 /app/init]" time="2015-03-04T19:42:51Z" level="debug" msg="[BUILDER] Cache miss" time="2015-03-04T19:42:51Z" level="info" msg="+job allocate_interface(6814090adc83ba62ebe24f13ad889344216ec1434066a45980b52809ad8ff4c0)" time="2015-03-04T19:42:51Z" level="info" msg="-job allocate_interface(6814090adc83ba62ebe24f13ad889344216ec1434066a45980b52809ad8ff4c0) = OK (0)" time="2015-03-04T19:42:51Z" level="info" msg="+job log(start, 6814090adc83ba62ebe24f13ad889344216ec1434066a45980b52809ad8ff4c0, cfd71b715943)" time="2015-03-04T19:42:51Z" level="info" msg="-job log(start, 6814090adc83ba62ebe24f13ad889344216ec1434066a45980b52809ad8ff4c0, cfd71b715943) = OK (0)" time="2015-03-04T19:42:51Z" level="info" msg="+job logs(6814090adc83ba62ebe24f13ad889344216ec1434066a45980b52809ad8ff4c0)" time="2015-03-04T19:42:52Z" level="info" msg="+job log(die, 6814090adc83ba62ebe24f13ad889344216ec1434066a45980b52809ad8ff4c0, cfd71b715943)" time="2015-03-04T19:42:52Z" level="info" msg="-job log(die, 6814090adc83ba62ebe24f13ad889344216ec1434066a45980b52809ad8ff4c0, cfd71b715943) = OK (0)" time="2015-03-04T19:42:52Z" level="info" msg="+job release_interface(6814090adc83ba62ebe24f13ad889344216ec1434066a45980b52809ad8ff4c0)" time="2015-03-04T19:42:52Z" level="info" msg="-job release_interface(6814090adc83ba62ebe24f13ad889344216ec1434066a45980b52809ad8ff4c0) = OK (0)" time="2015-03-04T19:42:52Z" level="info" msg="-job logs(6814090adc83ba62ebe24f13ad889344216ec1434066a45980b52809ad8ff4c0) = OK (0)" time="2015-03-04T19:42:52Z" level="debug" msg="Skipping excluded path: .wh..wh.aufs" time="2015-03-04T19:42:52Z" level="debug" msg="Skipping excluded path: .wh..wh.orph" time="2015-03-04T19:42:52Z" level="debug" msg="Skipping excluded path: .wh..wh.plnk" time="2015-03-04T19:42:52Z" level="debug" msg="[BUILDER] Cache miss" time="2015-03-04T19:42:52Z" level="debug" msg="Creating dest directory: /mnt/sda1/var/lib/docker/vfs/dir/5599661b259884f341f5506b4ede76ca2fe62e0b8974597f25adbf62fa3c2c07" time="2015-03-04T19:42:52Z" level="debug" msg="Calling TarUntar(/mnt/sda1/var/lib/docker/aufs/mnt/a5557c694ad9e0f870eb2f08c25111daae94611fda8366cbc40ac86bfc559433/opt/sonar/extensions, /mnt/sda1/var/lib/docker/vfs/dir/5599661b259884f341f5506b4ede76ca2fe62e0b8974597f25adbf62fa3c2c07)" time="2015-03-04T19:42:52Z" level="debug" msg="TarUntar(/mnt/sda1/var/lib/docker/aufs/mnt/a5557c694ad9e0f870eb2f08c25111daae94611fda8366cbc40ac86bfc559433/opt/sonar/extensions /mnt/sda1/var/lib/docker/vfs/dir/5599661b259884f341f5506b4ede76ca2fe62e0b8974597f25adbf62fa3c2c07)" time="2015-03-04T19:42:52Z" level="debug" msg="Skipping excluded path: .wh..wh.aufs" time="2015-03-04T19:42:52Z" level="debug" msg="Skipping excluded path: .wh..wh.orph" time="2015-03-04T19:42:52Z" level="debug" msg="Skipping excluded path: .wh..wh.plnk" time="2015-03-04T19:42:52Z" level="debug" msg="[BUILDER] Cache miss" time="2015-03-04T19:42:52Z" level="debug" msg="Creating dest directory: /mnt/sda1/var/lib/docker/vfs/dir/8c569d91bc95aac5c308b2a3f9a23431a225ef4a7bb9b00641c61f89967b8a0d" time="2015-03-04T19:42:52Z" level="debug" msg="Calling TarUntar(/mnt/sda1/var/lib/docker/aufs/mnt/a9b31d20747c1cac990def5433cf8c829e9ecf920cbe250498fcbd776a5bf705/opt/sonar/extensions, /mnt/sda1/var/lib/docker/vfs/dir/8c569d91bc95aac5c308b2a3f9a23431a225ef4a7bb9b00641c61f89967b8a0d)" time="2015-03-04T19:42:52Z" level="debug" msg="TarUntar(/mnt/sda1/var/lib/docker/aufs/mnt/a9b31d20747c1cac990def5433cf8c829e9ecf920cbe250498fcbd776a5bf705/opt/sonar/extensions /mnt/sda1/var/lib/docker/vfs/dir/8c569d91bc95aac5c308b2a3f9a23431a225ef4a7bb9b00641c61f89967b8a0d)" time="2015-03-04T19:42:52Z" level="debug" msg="Skipping excluded path: .wh..wh.aufs" time="2015-03-04T19:42:52Z" level="debug" msg="Skipping excluded path: .wh..wh.orph" time="2015-03-04T19:42:52Z" level="debug" msg="Skipping excluded path: .wh..wh.plnk" time="2015-03-04T19:42:52Z" level="debug" msg="[BUILDER] Cache miss" time="2015-03-04T19:42:52Z" level="debug" msg="Creating dest directory: /mnt/sda1/var/lib/docker/vfs/dir/b72ddfc443cb6b8632337bdd60a5b4ea62c33d274e2b540e846c3e13a1a0b7b5" time="2015-03-04T19:42:52Z" level="debug" msg="Calling TarUntar(/mnt/sda1/var/lib/docker/aufs/mnt/5da66e951afbee05bc863a68db8c7b3c788a97ed2e31b849b5d0b8356c8d243a/opt/sonar/extensions, /mnt/sda1/var/lib/docker/vfs/dir/b72ddfc443cb6b8632337bdd60a5b4ea62c33d274e2b540e846c3e13a1a0b7b5)" time="2015-03-04T19:42:52Z" level="debug" msg="TarUntar(/mnt/sda1/var/lib/docker/aufs/mnt/5da66e951afbee05bc863a68db8c7b3c788a97ed2e31b849b5d0b8356c8d243a/opt/sonar/extensions /mnt/sda1/var/lib/docker/vfs/dir/b72ddfc443cb6b8632337bdd60a5b4ea62c33d274e2b540e846c3e13a1a0b7b5)" time="2015-03-04T19:42:52Z" level="debug" msg="Skipping excluded path: .wh..wh.aufs" time="2015-03-04T19:42:52Z" level="debug" msg="Skipping excluded path: .wh..wh.orph" time="2015-03-04T19:42:52Z" level="debug" msg="Skipping excluded path: .wh..wh.plnk" time="2015-03-04T19:42:52Z" level="debug" msg="[BUILDER] Cache miss" time="2015-03-04T19:42:52Z" level="debug" msg="Creating dest directory: /mnt/sda1/var/lib/docker/vfs/dir/6288cf8325d559a91f4c75093c459ce4e53b2fe28dcae315a43c7835f98dfe47" time="2015-03-04T19:42:52Z" level="debug" msg="Calling TarUntar(/mnt/sda1/var/lib/docker/aufs/mnt/7bc83dd46e21b0a1da9cb77acc36470d0941df6c709d87f03bd9d3811bf36c00/opt/sonar/extensions, /mnt/sda1/var/lib/docker/vfs/dir/6288cf8325d559a91f4c75093c459ce4e53b2fe28dcae315a43c7835f98dfe47)" time="2015-03-04T19:42:52Z" level="debug" msg="TarUntar(/mnt/sda1/var/lib/docker/aufs/mnt/7bc83dd46e21b0a1da9cb77acc36470d0941df6c709d87f03bd9d3811bf36c00/opt/sonar/extensions /mnt/sda1/var/lib/docker/vfs/dir/6288cf8325d559a91f4c75093c459ce4e53b2fe28dcae315a43c7835f98dfe47)" time="2015-03-04T19:42:53Z" level="debug" msg="Skipping excluded path: .wh..wh.aufs" time="2015-03-04T19:42:53Z" level="debug" msg="Skipping excluded path: .wh..wh.orph" time="2015-03-04T19:42:53Z" level="debug" msg="Skipping excluded path: .wh..wh.plnk" time="2015-03-04T19:42:53Z" level="info" msg="-job build() = OK (0)" ```
i have the same issue I've the same error too ``` File "/usr/local/Cellar/fig/1.1.0/libexec/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.1.0', 'console_scripts', 'docker-compose')() File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/main.py", line 31, in main command.sys_dispatch() File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/command.py", line 27, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/command.py", line 59, in perform_command handler(project, command_options) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/main.py", line 445, in up do_build=not options['--no-build'], File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/project.py", line 184, in up do_build=do_build): File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/service.py", line 259, in recreate_containers **override_options) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/service.py", line 244, in create_container stream_output(output, sys.stdout) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/progress_stream.py", line 37, in stream_output print_output_event(event, stream, is_terminal) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/progress_stream.py", line 50, in print_output_event raise StreamOutputError(event['errorDetail']['message']) compose.progress_stream.StreamOutputError: Get https://index.docker.io/v1/repositories/library/postgres/images: x509: certificate has expired or is not yet valid ``` I have the same issue as @steve0hh +1 Are you still able to reproduce this? @steve0hh that is a different issue and should have been fixed a while ago @dnephin yeah. got it! thanks!
2015-10-02T23:52:52Z
[]
[]
Traceback (most recent call last): File "/usr/local/Cellar/fig/1.1.0/libexec/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.1.0', 'console_scripts', 'docker-compose')() File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/main.py", line 31, in main command.sys_dispatch() File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/command.py", line 27, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/command.py", line 59, in perform_command handler(project, command_options) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/cli/main.py", line 119, in build project.build(service_names=options['SERVICE'], no_cache=no_cache) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/project.py", line 167, in build service.build(no_cache) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/service.py", line 484, in build all_events = stream_output(build_output, sys.stdout) File "/usr/local/Cellar/fig/1.1.0/libexec/lib/python2.7/site-packages/compose/progress_stream.py", line 18, in stream_output event = json.loads(chunk) File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/json/__init__.py", line 338, in loads return _default_decoder.decode(s) File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/json/decoder.py", line 368, in decode raise ValueError(errmsg("Extra data", s, end, len(s))) ValueError: Extra data: line 2 column 1 - line 4 column 1 (char 39 - 162)
4,700
docker/compose
docker__compose-2210
cd0b63879b177f39d87345381384bc56b6b782ef
diff --git a/compose/config/config.py b/compose/config/config.py --- a/compose/config/config.py +++ b/compose/config/config.py @@ -14,7 +14,6 @@ from .validation import validate_against_service_schema from .validation import validate_extended_service_exists from .validation import validate_extends_file_path -from .validation import validate_service_names from .validation import validate_top_level_object @@ -165,16 +164,6 @@ def find_candidates_in_parent_dirs(filenames, path): return (candidates, path) -@validate_top_level_object -@validate_service_names -def pre_process_config(config): - """ - Pre validation checks and processing of the config file to interpolate env - vars returning a config dict ready to be tested against the schema. - """ - return interpolate_environment_variables(config) - - def load(config_details): """Load the configuration from a working directory and a list of configuration files. Files are loaded in order, and merged on top @@ -194,7 +183,7 @@ def build_service(filename, service_name, service_dict): return service_dict def load_file(filename, config): - processed_config = pre_process_config(config) + processed_config = interpolate_environment_variables(config) validate_against_fields_schema(processed_config) return [ build_service(filename, name, service_config) @@ -209,7 +198,10 @@ def merge_services(base, override): } config_file = config_details.config_files[0] + validate_top_level_object(config_file.config) for next_file in config_details.config_files[1:]: + validate_top_level_object(next_file.config) + config_file = ConfigFile( config_file.filename, merge_services(config_file.config, next_file.config)) @@ -283,9 +275,9 @@ def validate_and_construct_extends(self): ) self.extended_service_name = extends['service'] - full_extended_config = pre_process_config( - load_yaml(self.extended_config_path) - ) + config = load_yaml(self.extended_config_path) + validate_top_level_object(config) + full_extended_config = interpolate_environment_variables(config) validate_extended_service_exists( self.extended_service_name, diff --git a/compose/config/validation.py b/compose/config/validation.py --- a/compose/config/validation.py +++ b/compose/config/validation.py @@ -2,7 +2,6 @@ import logging import os import sys -from functools import wraps import six from docker.utils.ports import split_port @@ -65,27 +64,21 @@ def format_boolean_in_environment(instance): return True -def validate_service_names(func): - @wraps(func) - def func_wrapper(config): - for service_name in config.keys(): - if type(service_name) is int: - raise ConfigurationError( - "Service name: {} needs to be a string, eg '{}'".format(service_name, service_name) - ) - return func(config) - return func_wrapper +def validate_service_names(config): + for service_name in config.keys(): + if not isinstance(service_name, six.string_types): + raise ConfigurationError( + "Service name: {} needs to be a string, eg '{}'".format( + service_name, + service_name)) -def validate_top_level_object(func): - @wraps(func) - def func_wrapper(config): - if not isinstance(config, dict): - raise ConfigurationError( - "Top level object needs to be a dictionary. Check your .yml file that you have defined a service at the top level." - ) - return func(config) - return func_wrapper +def validate_top_level_object(config): + if not isinstance(config, dict): + raise ConfigurationError( + "Top level object needs to be a dictionary. Check your .yml file " + "that you have defined a service at the top level.") + validate_service_names(config) def validate_extends_file_path(service_name, extends_options, filename):
Stack trace when an override file is empty If I create an empty `docker-compose.override.yml` in a project, `docker-compose up` throws the following error: ``` Traceback (most recent call last): File "<string>", line 3, in <module> File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 53, in main File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 23, in sys_dispatch File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 26, in dispatch File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 163, in perform_command File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 54, in project_from_options File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 90, in get_project File "/compose/build/docker-compose/out00-PYZ.pyz/compose.config.config", line 215, in load File "/compose/build/docker-compose/out00-PYZ.pyz/compose.config.config", line 205, in merge_services TypeError: 'NoneType' object is not iterable ``` We should instead show the same `Top level object needs to be a dictionary` error that you get if your `docker-compose.yml` file is empty.
2015-10-16T19:49:57Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 53, in main File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 23, in sys_dispatch File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 26, in dispatch File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 163, in perform_command File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 54, in project_from_options File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 90, in get_project File "/compose/build/docker-compose/out00-PYZ.pyz/compose.config.config", line 215, in load File "/compose/build/docker-compose/out00-PYZ.pyz/compose.config.config", line 205, in merge_services TypeError: 'NoneType' object is not iterable
4,703
docker/compose
docker__compose-2239
6f78271b826c59d7f58d03e68cca13438436bdbf
diff --git a/compose/config/config.py b/compose/config/config.py --- a/compose/config/config.py +++ b/compose/config/config.py @@ -1,3 +1,4 @@ +import codecs import logging import os import sys @@ -455,6 +456,8 @@ def parse_environment(environment): def split_env(env): + if isinstance(env, six.binary_type): + env = env.decode('utf-8') if '=' in env: return env.split('=', 1) else: @@ -477,7 +480,7 @@ def env_vars_from_file(filename): if not os.path.exists(filename): raise ConfigurationError("Couldn't find env file: %s" % filename) env = {} - for line in open(filename, 'r'): + for line in codecs.open(filename, 'r', 'utf-8'): line = line.strip() if line and not line.startswith('#'): k, v = split_env(line)
UnicodeDecodeError - for environment variable with non-ascii character value ``` $ docker-compose --version docker-compose version: 1.3.2 CPython version: 2.7.9 OpenSSL version: OpenSSL 1.0.1e 11 Feb 2013 $ cat docker-compose.yml code: image: busybox $ docker-compose run -e FOO=bär code ls Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 32, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 34, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 66, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 342, in run File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 225, in create_container File "/code/build/docker-compose/out00-PYZ.pyz/compose.container", line 38, in create File "/code/build/docker-compose/out00-PYZ.pyz/docker.client", line 237, in create_container File "/code/build/docker-compose/out00-PYZ.pyz/docker.utils.utils", line 528, in create_container_config UnicodeDecodeError: 'ascii' codec can't decode byte 0xc3 in position 1: ordinal not in range(128) ```
This happens also when the environment variables are loaded from a file with `env_file: ./somefile` and inside the file some environment variables have values containing non-ASCII characters. ``` $ docker-compose version docker-compose version: 1.4.2 docker-py version: 1.3.1 CPython version: 2.7.9 OpenSSL version: OpenSSL 1.0.1j 15 Oct 2014 ``` Error output: ``` Traceback (most recent call last): File "<string>", line 3, in <module> File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 39, in main File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 27, in dispatch File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 59, in perform_command File "/compose/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 313, in run File "/compose/build/docker-compose/out00-PYZ.pyz/compose.project", line 274, in up File "/compose/build/docker-compose/out00-PYZ.pyz/compose.service", line 377, in execute_convergence_plan File "/compose/build/docker-compose/out00-PYZ.pyz/compose.service", line 284, in create_container File "/compose/build/docker-compose/out00-PYZ.pyz/compose.container", line 42, in create File "/compose/build/docker-compose/out00-PYZ.pyz/docker.client", line 243, in create_container File "/compose/build/docker-compose/out00-PYZ.pyz/docker.utils.utils", line 534, in create_container_config UnicodeDecodeError: 'ascii' codec can't decode byte 0xc3 in position 3: ordinal not in range(128) ``` Both of these cases are fixed in master and the 1.5.0 release. You can try it out with the release candidate: https://github.com/docker/compose/releases/tag/1.5.0rc1 Oops, I was wrong. They're only fixed on python3.
2015-10-22T16:13:37Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 32, in main File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 21, in sys_dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 34, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.docopt_command", line 24, in dispatch File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.command", line 66, in perform_command File "/code/build/docker-compose/out00-PYZ.pyz/compose.cli.main", line 342, in run File "/code/build/docker-compose/out00-PYZ.pyz/compose.service", line 225, in create_container File "/code/build/docker-compose/out00-PYZ.pyz/compose.container", line 38, in create File "/code/build/docker-compose/out00-PYZ.pyz/docker.client", line 237, in create_container File "/code/build/docker-compose/out00-PYZ.pyz/docker.utils.utils", line 528, in create_container_config UnicodeDecodeError: 'ascii' codec can't decode byte 0xc3 in position 1: ordinal not in range(128)
4,705
docker/compose
docker__compose-2326
67dc90ec0e71282bbf3ec27f30d2b1232406b76d
diff --git a/compose/config/config.py b/compose/config/config.py --- a/compose/config/config.py +++ b/compose/config/config.py @@ -457,7 +457,7 @@ def parse_environment(environment): def split_env(env): if isinstance(env, six.binary_type): - env = env.decode('utf-8') + env = env.decode('utf-8', 'replace') if '=' in env: return env.split('=', 1) else: diff --git a/compose/utils.py b/compose/utils.py --- a/compose/utils.py +++ b/compose/utils.py @@ -95,7 +95,7 @@ def stream_as_text(stream): """ for data in stream: if not isinstance(data, six.text_type): - data = data.decode('utf-8') + data = data.decode('utf-8', 'replace') yield data
UnicodeDecodeError exception on non-unicode character output (1.5.0) #### Description of problem: Containers outputing non-unicode characters throws a fatal exception. #### How reproducible: Since version 1.5.0 (final version and both RC) #### Steps to Reproduce: - Create the following `docker-compose.yml` file: ``` web: image: ubuntu command: "/code/test.sh" volumes: - .:/code ``` - Create a `test.sh` file in the same folder (**Important**: file should include some ISO8859-1 encoded characters or from other non-unicode charset). For example download this one https://www.dropbox.com/s/clsuzzv5g9zc2qj/test.sh?dl=0 ``` #!/bin/sh echo ae��u ``` - Run `docker-compose up` #### Actual Results: ``` Creating composer_web_1 Attaching to composer_web_1 Gracefully stopping... (press Ctrl+C again to force) Stopping composer_web_1 ... done Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/compose/cli/main.py", line 54, in main File "/code/compose/cli/docopt_command.py", line 23, in sys_dispatch File "/code/compose/cli/docopt_command.py", line 26, in dispatch File "/code/compose/cli/main.py", line 170, in perform_command File "/code/compose/cli/main.py", line 588, in up File "/code/compose/cli/main.py", line 658, in attach_to_logs File "/code/compose/cli/log_printer.py", line 27, in run File "/code/compose/cli/multiplexer.py", line 35, in loop UnicodeDecodeError: 'utf8' codec can't decode byte 0xed in position 2: invalid continuation byte docker-compose returned -1 ``` #### Expected Results: ``` Starting composer_web_1 Attaching to composer_web_1 web_1 | ae��u composer_web_1 exited with code 0 Gracefully stopping... (press Ctrl+C again to force) ``` #### Additional info: - Containers are running ok with `docker run`, it's a docker-compose issue - On 1.5.0 RC1 and 1.5.0 RC2 it only fails after running a `sleep` command inside the script, on final version it fails on every non-unicode output #### `uname -a`: ``` Linux smarconi 3.19.0-31-generic #36-Ubuntu SMP Wed Oct 7 15:04:02 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux ``` #### `docker-compose version`: ``` docker-compose version: 1.5.0 docker-py version: 1.5.0 CPython version: 2.7.9 OpenSSL version: OpenSSL 1.0.1e 11 Feb 2013 ``` #### `docker version`: ``` Client version: 1.7.1 Client API version: 1.19 Go version (client): go1.4.2 Git commit (client): 786b29d OS/Arch (client): linux/amd64 Server version: 1.7.1 Server API version: 1.19 Go version (server): go1.4.2 Git commit (server): 786b29d OS/Arch (server): linux/amd64 ``` #### `docker info`: ``` Containers: 0 Images: 415 Storage Driver: aufs Root Dir: /var/lib/docker/aufs Backing Filesystem: extfs Dirs: 419 Dirperm1 Supported: true Execution Driver: native-0.2 Logging Driver: json-file Kernel Version: 3.19.0-31-generic Operating System: Ubuntu 15.04 CPUs: 8 Total Memory: 7.516 GiB Name: smarconi ID: JWSO:PS6C:FOC4:NSTQ:DA2N:FVTD:ZJM3:JJCN:CDOW:EORB:YSJW:GYCJ WARNING: No swap limit support ```
I guess we can use `replace` or `ignore` in the `decode()`.
2015-11-04T20:57:48Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/compose/cli/main.py", line 54, in main File "/code/compose/cli/docopt_command.py", line 23, in sys_dispatch File "/code/compose/cli/docopt_command.py", line 26, in dispatch File "/code/compose/cli/main.py", line 170, in perform_command File "/code/compose/cli/main.py", line 588, in up File "/code/compose/cli/main.py", line 658, in attach_to_logs File "/code/compose/cli/log_printer.py", line 27, in run File "/code/compose/cli/multiplexer.py", line 35, in loop UnicodeDecodeError: 'utf8' codec can't decode byte 0xed in position 2: invalid continuation byte
4,713
docker/compose
docker__compose-2400
2f20dfe508382a124b9b2d2986076c5668d3314d
diff --git a/compose/utils.py b/compose/utils.py --- a/compose/utils.py +++ b/compose/utils.py @@ -102,7 +102,7 @@ def stream_as_text(stream): def line_splitter(buffer, separator=u'\n'): index = buffer.find(six.text_type(separator)) if index == -1: - return None, None + return None return buffer[:index + 1], buffer[index + 1:] @@ -120,11 +120,11 @@ def split_buffer(stream, splitter=None, decoder=lambda a: a): for data in stream_as_text(stream): buffered += data while True: - item, rest = splitter(buffered) - if not item: + buffer_split = splitter(buffered) + if buffer_split is None: break - buffered = rest + item, buffered = buffer_split yield item if buffered: @@ -140,7 +140,7 @@ def json_splitter(buffer): rest = buffer[json.decoder.WHITESPACE.match(buffer, index).end():] return obj, rest except ValueError: - return None, None + return None def json_stream(stream): @@ -148,7 +148,7 @@ def json_stream(stream): This handles streams which are inconsistently buffered (some entries may be newline delimited, and others are not). """ - return split_buffer(stream_as_text(stream), json_splitter, json_decoder.decode) + return split_buffer(stream, json_splitter, json_decoder.decode) def write_out_msg(stream, lines, msg_index, msg, status="done"):
docker-compose build fails on ADD url directive After updating to docker-compose 1.5.1 (latest from Debian Unstable), trying to build the following image with docker-compose throws an error. The same yml file with docker-compose 1.4.0 builds the image without errors. ``` $ docker-compose -f pep-wilma.yml build pepwilma Building pepwilma Step 0 : FROM node:0.10-slim ---> 04e511e59c2e [...] Step 16 : ADD https://raw.githubusercontent.com/Bitergia/docker/master/utils/entrypoint-common.sh / Traceback (most recent call last): File "/usr/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.5.1', 'console_scripts', 'docker-compose')() File "/usr/lib/python2.7/dist-packages/compose/cli/main.py", line 54, in main command.sys_dispatch() File "/usr/lib/python2.7/dist-packages/compose/cli/docopt_command.py", line 23, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/lib/python2.7/dist-packages/compose/cli/docopt_command.py", line 26, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/lib/python2.7/dist-packages/compose/cli/main.py", line 171, in perform_command handler(project, command_options) File "/usr/lib/python2.7/dist-packages/compose/cli/main.py", line 192, in build force_rm=bool(options.get('--force-rm', False))) File "/usr/lib/python2.7/dist-packages/compose/project.py", line 284, in build service.build(no_cache, pull, force_rm) File "/usr/lib/python2.7/dist-packages/compose/service.py", line 727, in build all_events = stream_output(build_output, sys.stdout) File "/usr/lib/python2.7/dist-packages/compose/progress_stream.py", line 15, in stream_output for event in utils.json_stream(output): File "/usr/lib/python2.7/dist-packages/compose/utils.py", line 131, in split_buffer yield decoder(buffered) File "/usr/lib/python2.7/json/decoder.py", line 367, in decode raise ValueError(errmsg("Extra data", s, end, len(s))) ValueError: Extra data: line 2 column 1 - line 15 column 1 (char 4 - 544) ``` The original yml can be found here: [pep-wilma.yml](https://github.com/Bitergia/fiware-chanchan-docker/blob/master/compose/pep-wilma.yml) To build the image, we've just replaced the image line with build: ``` diff diff --git a/compose/pep-wilma.yml b/compose/pep-wilma.yml index f91f349..3a46b13 100644 --- a/compose/pep-wilma.yml +++ b/compose/pep-wilma.yml @@ -43,7 +43,7 @@ idm: - "5000" pepwilma: - image: bitergia/pep-wilma:4.3.0 + build: ../images/pep-wilma/4.3.0 hostname: pepwilma links: - orion ``` Replacing the ADD line from the Dockerfile with a local file or COPY works as expected, as building the image directly with docker. ``` $ docker-compose version docker-compose version: 1.5.1 docker-py version: 1.5.0 CPython version: 2.7.10+ OpenSSL version: OpenSSL 1.0.2d 9 Jul 2015 ``` ``` $ docker version Client: Version: 1.8.3 API version: 1.20 Go version: go1.5.1 Git commit: f4bf5c7 Built: OS/Arch: linux/amd64 Server: Version: 1.8.3 API version: 1.20 Go version: go1.5.1 Git commit: f4bf5c7 Built: OS/Arch: linux/amd64 ```
2015-11-16T17:37:00Z
[]
[]
Traceback (most recent call last): File "/usr/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.5.1', 'console_scripts', 'docker-compose')() File "/usr/lib/python2.7/dist-packages/compose/cli/main.py", line 54, in main command.sys_dispatch() File "/usr/lib/python2.7/dist-packages/compose/cli/docopt_command.py", line 23, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/lib/python2.7/dist-packages/compose/cli/docopt_command.py", line 26, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/lib/python2.7/dist-packages/compose/cli/main.py", line 171, in perform_command handler(project, command_options) File "/usr/lib/python2.7/dist-packages/compose/cli/main.py", line 192, in build force_rm=bool(options.get('--force-rm', False))) File "/usr/lib/python2.7/dist-packages/compose/project.py", line 284, in build service.build(no_cache, pull, force_rm) File "/usr/lib/python2.7/dist-packages/compose/service.py", line 727, in build all_events = stream_output(build_output, sys.stdout) File "/usr/lib/python2.7/dist-packages/compose/progress_stream.py", line 15, in stream_output for event in utils.json_stream(output): File "/usr/lib/python2.7/dist-packages/compose/utils.py", line 131, in split_buffer yield decoder(buffered) File "/usr/lib/python2.7/json/decoder.py", line 367, in decode raise ValueError(errmsg("Extra data", s, end, len(s))) ValueError: Extra data: line 2 column 1 - line 15 column 1 (char 4 - 544)
4,732
docker/compose
docker__compose-255
feb8ad7b4cbfdba86aa56742835a877ffb53f420
diff --git a/fig/cli/log_printer.py b/fig/cli/log_printer.py --- a/fig/cli/log_printer.py +++ b/fig/cli/log_printer.py @@ -10,16 +10,17 @@ class LogPrinter(object): - def __init__(self, containers, attach_params=None): + def __init__(self, containers, attach_params=None, output=sys.stdout): self.containers = containers self.attach_params = attach_params or {} self.prefix_width = self._calculate_prefix_width(containers) self.generators = self._make_log_generators() + self.output = output def run(self): mux = Multiplexer(self.generators) for line in mux.loop(): - sys.stdout.write(line.encode(sys.__stdout__.encoding or 'utf-8')) + self.output.write(line) def _calculate_prefix_width(self, containers): """ @@ -45,12 +46,12 @@ def _make_log_generators(self): return generators def _make_log_generator(self, container, color_fn): - prefix = color_fn(self._generate_prefix(container)) + prefix = color_fn(self._generate_prefix(container)).encode('utf-8') # Attach to container before log printer starts running line_generator = split_buffer(self._attach(container), '\n') for line in line_generator: - yield prefix + line.decode('utf-8') + yield prefix + line exit_code = container.wait() yield color_fn("%s exited with code %s\n" % (container.name, exit_code)) diff --git a/fig/cli/socketclient.py b/fig/cli/socketclient.py --- a/fig/cli/socketclient.py +++ b/fig/cli/socketclient.py @@ -81,7 +81,7 @@ def recv(self, socket, stream): chunk = socket.recv(4096) if chunk: - stream.write(chunk.encode(stream.encoding or 'utf-8')) + stream.write(chunk) stream.flush() else: break diff --git a/fig/progress_stream.py b/fig/progress_stream.py new file mode 100644 --- /dev/null +++ b/fig/progress_stream.py @@ -0,0 +1,83 @@ +import json +import os +import codecs + + +class StreamOutputError(Exception): + pass + + +def stream_output(output, stream): + is_terminal = hasattr(stream, 'fileno') and os.isatty(stream.fileno()) + stream = codecs.getwriter('utf-8')(stream) + all_events = [] + lines = {} + diff = 0 + + for chunk in output: + event = json.loads(chunk) + all_events.append(event) + + if 'progress' in event or 'progressDetail' in event: + image_id = event['id'] + + if image_id in lines: + diff = len(lines) - lines[image_id] + else: + lines[image_id] = len(lines) + stream.write("\n") + diff = 0 + + if is_terminal: + # move cursor up `diff` rows + stream.write("%c[%dA" % (27, diff)) + + print_output_event(event, stream, is_terminal) + + if 'id' in event and is_terminal: + # move cursor back down + stream.write("%c[%dB" % (27, diff)) + + stream.flush() + + return all_events + + +def print_output_event(event, stream, is_terminal): + if 'errorDetail' in event: + raise StreamOutputError(event['errorDetail']['message']) + + terminator = '' + + if is_terminal and 'stream' not in event: + # erase current line + stream.write("%c[2K\r" % 27) + terminator = "\r" + pass + elif 'progressDetail' in event: + return + + if 'time' in event: + stream.write("[%s] " % event['time']) + + if 'id' in event: + stream.write("%s: " % event['id']) + + if 'from' in event: + stream.write("(from %s) " % event['from']) + + status = event.get('status', '') + + if 'progress' in event: + stream.write("%s %s%s" % (status, event['progress'], terminator)) + elif 'progressDetail' in event: + detail = event['progressDetail'] + if 'current' in detail: + percentage = float(detail['current']) / float(detail['total']) * 100 + stream.write('%s (%.1f%%)%s' % (status, percentage, terminator)) + else: + stream.write('%s%s' % (status, terminator)) + elif 'stream' in event: + stream.write("%s%s" % (event['stream'], terminator)) + else: + stream.write("%s%s\n" % (status, terminator)) diff --git a/fig/service.py b/fig/service.py --- a/fig/service.py +++ b/fig/service.py @@ -5,8 +5,8 @@ import re import os import sys -import json from .container import Container +from .progress_stream import stream_output, StreamOutputError log = logging.getLogger(__name__) @@ -343,84 +343,6 @@ def can_be_scaled(self): return True -class StreamOutputError(Exception): - pass - - -def stream_output(output, stream): - is_terminal = hasattr(stream, 'fileno') and os.isatty(stream.fileno()) - all_events = [] - lines = {} - diff = 0 - - for chunk in output: - event = json.loads(chunk) - all_events.append(event) - - if 'progress' in event or 'progressDetail' in event: - image_id = event['id'] - - if image_id in lines: - diff = len(lines) - lines[image_id] - else: - lines[image_id] = len(lines) - stream.write("\n") - diff = 0 - - if is_terminal: - # move cursor up `diff` rows - stream.write("%c[%dA" % (27, diff)) - - print_output_event(event, stream, is_terminal) - - if 'id' in event and is_terminal: - # move cursor back down - stream.write("%c[%dB" % (27, diff)) - - stream.flush() - - return all_events - -def print_output_event(event, stream, is_terminal): - if 'errorDetail' in event: - raise StreamOutputError(event['errorDetail']['message']) - - terminator = '' - - if is_terminal and 'stream' not in event: - # erase current line - stream.write("%c[2K\r" % 27) - terminator = "\r" - pass - elif 'progressDetail' in event: - return - - if 'time' in event: - stream.write("[%s] " % event['time']) - - if 'id' in event: - stream.write("%s: " % event['id']) - - if 'from' in event: - stream.write("(from %s) " % event['from']) - - status = event.get('status', '') - - if 'progress' in event: - stream.write("%s %s%s" % (status, event['progress'], terminator)) - elif 'progressDetail' in event: - detail = event['progressDetail'] - if 'current' in detail: - percentage = float(detail['current']) / float(detail['total']) * 100 - stream.write('%s (%.1f%%)%s' % (status, percentage, terminator)) - else: - stream.write('%s%s' % (status, terminator)) - elif 'stream' in event: - stream.write("%s%s" % (event['stream'], terminator)) - else: - stream.write("%s%s\n" % (status, terminator)) - - NAME_RE = re.compile(r'^([^_]+)_([^_]+)_(run_)?(\d+)$')
UnicodeEncodeError: 'ascii' codec can't encode character u'\u2023' When trying to install `RUN /bin/bash -l -c 'passenger-install-nginx-module --auto --auto-download --prefix=/opt/nginx'` from Docker, I get: ``` Traceback (most recent call last): File "<string>", line 3, in <module> File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.cli.main", line 39, in main File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 21, in sys_dispatch File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.cli.command", line 31, in dispatch File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 24, in dispatch File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.cli.command", line 50, in perform_command File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 27, in perform_command File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.cli.main", line 108, in build File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.project", line 123, in build File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.service", line 313, in build File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.service", line 374, in stream_output File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.service", line 419, in print_output_event UnicodeEncodeError: 'ascii' codec can't encode character u'\u2023' in position 2: ordinal not in range(128) ``` I guess the reason is that at this point nginx shows a list of languages and there is a unicode character used as a bullet point in front of each language. fig version is: 0.4.1 docker version is: Client version: 0.11.1 Client API version: 1.11 Go version (client): go1.2.1 Git commit (client): fb99f99 Server version: 0.11.1 Server API version: 1.11 Git commit (server): fb99f99 Go version (server): go1.2.1
Getting a similar problem when installing a PPA. https://gist.github.com/dginther/9b194951e60c1d3a0fb9 Could you try #233 and tell if it helps? Sure, I can try that. meanwhile I put a big ugly hack in that fixed things for me. ``` def print_output_event(event, stream, is_terminal): if 'errorDetail' in event: raise StreamOutputError(event['errorDetail']['message']) reload(sys) sys.setdefaultencoding('utf-8') terminator = '' ```
2014-06-18T14:35:03Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.cli.main", line 39, in main File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 21, in sys_dispatch File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.cli.command", line 31, in dispatch File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 24, in dispatch File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.cli.command", line 50, in perform_command File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 27, in perform_command File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.cli.main", line 108, in build File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.project", line 123, in build File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.service", line 313, in build File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.service", line 374, in stream_output File "/Users/nonsense/code/fig/build/fig/out00-PYZ.pyz/fig.service", line 419, in print_output_event UnicodeEncodeError: 'ascii' codec can't encode character u'\u2023' in position 2: ordinal not in range(128)
4,751
docker/compose
docker__compose-2877
7bae4cdbb25650274ec91a3598eb73e074fd3a9f
diff --git a/compose/config/config.py b/compose/config/config.py --- a/compose/config/config.py +++ b/compose/config/config.py @@ -874,6 +874,9 @@ def validate_paths(service_dict): build_path = build elif isinstance(build, dict) and 'context' in build: build_path = build['context'] + else: + # We have a build section but no context, so nothing to validate + return if ( not is_url(build_path) and diff --git a/compose/config/validation.py b/compose/config/validation.py --- a/compose/config/validation.py +++ b/compose/config/validation.py @@ -253,10 +253,9 @@ def handle_generic_service_error(error, path): msg_format = "{path} contains an invalid type, it should be {msg}" error_msg = _parse_valid_types_from_validator(error.validator_value) - # TODO: no test case for this branch, there are no config options - # which exercise this branch elif error.validator == 'required': - msg_format = "{path} is invalid, {msg}" + error_msg = ", ".join(error.validator_value) + msg_format = "{path} is invalid, {msg} is required." elif error.validator == 'dependencies': config_key = list(error.validator_value.keys())[0]
Obscure failure when build context is not specified When trying to use a `docker-compose.yml` file like this one: ``` version: "2" services: foo: build: dockerfile: "Dockerfile.foo" ``` The failure mode from docker-compose 1.6.0 is an unhandled `UnboundLocalError`: ``` Traceback (most recent call last): File "/usr/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.6.0', 'console_scripts', 'docker-compose')() File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 55, in main command.sys_dispatch() File "/usr/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 23, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 26, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 172, in perform_command project = project_from_options(self.base_dir, options) File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 52, in project_from_options verbose=options.get('--verbose'), File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 80, in get_project config_data = config.load(config_details) File "/usr/lib/python2.7/site-packages/compose/config/config.py", line 298, in load [file.get_service_dicts() for file in config_details.config_files]) File "/usr/lib/python2.7/site-packages/compose/config/config.py", line 376, in load_services return build_services(service_config) File "/usr/lib/python2.7/site-packages/compose/config/config.py", line 359, in build_services for name, service_dict in service_config.items() File "/usr/lib/python2.7/site-packages/compose/config/config.py", line 348, in build_service validate_service(service_config, service_names, config_file.version) File "/usr/lib/python2.7/site-packages/compose/config/config.py", line 531, in validate_service validate_paths(service_dict) File "/usr/lib/python2.7/site-packages/compose/config/config.py", line 881, in validate_paths not is_url(build_path) and UnboundLocalError: local variable 'build_path' referenced before assignment ``` This makes it difficult to understand the problem - which seems to be that the context has not been specified.
Thanks for the bug report!
2016-02-10T18:33:09Z
[]
[]
Traceback (most recent call last): File "/usr/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.6.0', 'console_scripts', 'docker-compose')() File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 55, in main command.sys_dispatch() File "/usr/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 23, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/lib/python2.7/site-packages/compose/cli/docopt_command.py", line 26, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 172, in perform_command project = project_from_options(self.base_dir, options) File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 52, in project_from_options verbose=options.get('--verbose'), File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 80, in get_project config_data = config.load(config_details) File "/usr/lib/python2.7/site-packages/compose/config/config.py", line 298, in load [file.get_service_dicts() for file in config_details.config_files]) File "/usr/lib/python2.7/site-packages/compose/config/config.py", line 376, in load_services return build_services(service_config) File "/usr/lib/python2.7/site-packages/compose/config/config.py", line 359, in build_services for name, service_dict in service_config.items() File "/usr/lib/python2.7/site-packages/compose/config/config.py", line 348, in build_service validate_service(service_config, service_names, config_file.version) File "/usr/lib/python2.7/site-packages/compose/config/config.py", line 531, in validate_service validate_paths(service_dict) File "/usr/lib/python2.7/site-packages/compose/config/config.py", line 881, in validate_paths not is_url(build_path) and UnboundLocalError: local variable 'build_path' referenced before assignment
4,782
docker/compose
docker__compose-2948
1f2c2942d7acab16bf45006f024054a7f2a8b6a1
diff --git a/compose/config/config.py b/compose/config/config.py --- a/compose/config/config.py +++ b/compose/config/config.py @@ -33,11 +33,11 @@ from .validation import match_named_volumes from .validation import validate_against_fields_schema from .validation import validate_against_service_schema +from .validation import validate_config_section from .validation import validate_depends_on from .validation import validate_extends_file_path from .validation import validate_network_mode from .validation import validate_top_level_object -from .validation import validate_top_level_service_objects from .validation import validate_ulimits @@ -388,22 +388,31 @@ def merge_services(base, override): return build_services(service_config) -def process_config_file(config_file, service_name=None): - service_dicts = config_file.get_service_dicts() - validate_top_level_service_objects(config_file.filename, service_dicts) +def interpolate_config_section(filename, config, section): + validate_config_section(filename, config, section) + return interpolate_environment_variables(config, section) + - interpolated_config = interpolate_environment_variables(service_dicts, 'service') +def process_config_file(config_file, service_name=None): + services = interpolate_config_section( + config_file.filename, + config_file.get_service_dicts(), + 'service') if config_file.version == V2_0: processed_config = dict(config_file.config) - processed_config['services'] = services = interpolated_config - processed_config['volumes'] = interpolate_environment_variables( - config_file.get_volumes(), 'volume') - processed_config['networks'] = interpolate_environment_variables( - config_file.get_networks(), 'network') + processed_config['services'] = services + processed_config['volumes'] = interpolate_config_section( + config_file.filename, + config_file.get_volumes(), + 'volume') + processed_config['networks'] = interpolate_config_section( + config_file.filename, + config_file.get_networks(), + 'network') if config_file.version == V1: - processed_config = services = interpolated_config + processed_config = services config_file = config_file._replace(config=processed_config) validate_against_fields_schema(config_file) diff --git a/compose/config/interpolation.py b/compose/config/interpolation.py --- a/compose/config/interpolation.py +++ b/compose/config/interpolation.py @@ -21,7 +21,7 @@ def process_item(name, config_dict): ) return dict( - (name, process_item(name, config_dict)) + (name, process_item(name, config_dict or {})) for name, config_dict in config.items() ) diff --git a/compose/config/validation.py b/compose/config/validation.py --- a/compose/config/validation.py +++ b/compose/config/validation.py @@ -91,29 +91,49 @@ def match_named_volumes(service_dict, project_volumes): ) -def validate_top_level_service_objects(filename, service_dicts): - """Perform some high level validation of the service name and value. - - This validation must happen before interpolation, which must happen - before the rest of validation, which is why it's separate from the - rest of the service validation. +def python_type_to_yaml_type(type_): + type_name = type(type_).__name__ + return { + 'dict': 'mapping', + 'list': 'array', + 'int': 'number', + 'float': 'number', + 'bool': 'boolean', + 'unicode': 'string', + 'str': 'string', + 'bytes': 'string', + }.get(type_name, type_name) + + +def validate_config_section(filename, config, section): + """Validate the structure of a configuration section. This must be done + before interpolation so it's separate from schema validation. """ - for service_name, service_dict in service_dicts.items(): - if not isinstance(service_name, six.string_types): + if not isinstance(config, dict): + raise ConfigurationError( + "In file '{filename}', {section} must be a mapping, not " + "{type}.".format( + filename=filename, + section=section, + type=anglicize_json_type(python_type_to_yaml_type(config)))) + + for key, value in config.items(): + if not isinstance(key, six.string_types): raise ConfigurationError( - "In file '{}' service name: {} needs to be a string, eg '{}'".format( - filename, - service_name, - service_name)) + "In file '{filename}', the {section} name {name} must be a " + "quoted string, i.e. '{name}'.".format( + filename=filename, + section=section, + name=key)) - if not isinstance(service_dict, dict): + if not isinstance(value, (dict, type(None))): raise ConfigurationError( - "In file '{}' service '{}' doesn\'t have any configuration options. " - "All top level keys in your docker-compose.yml must map " - "to a dictionary of configuration options.".format( - filename, service_name - ) - ) + "In file '{filename}', {section} '{name}' must be a mapping not " + "{type}.".format( + filename=filename, + section=section, + name=key, + type=anglicize_json_type(python_type_to_yaml_type(value)))) def validate_top_level_object(config_file): @@ -182,10 +202,10 @@ def get_unsupported_config_msg(path, error_key): return msg -def anglicize_validator(validator): - if validator in ["array", "object"]: - return 'an ' + validator - return 'a ' + validator +def anglicize_json_type(json_type): + if json_type.startswith(('a', 'e', 'i', 'o', 'u')): + return 'an ' + json_type + return 'a ' + json_type def is_service_dict_schema(schema_id): @@ -293,14 +313,14 @@ def _parse_valid_types_from_validator(validator): a valid type. Parse the valid types and prefix with the correct article. """ if not isinstance(validator, list): - return anglicize_validator(validator) + return anglicize_json_type(validator) if len(validator) == 1: - return anglicize_validator(validator[0]) + return anglicize_json_type(validator[0]) return "{}, or {}".format( - ", ".join([anglicize_validator(validator[0])] + validator[1:-1]), - anglicize_validator(validator[-1])) + ", ".join([anglicize_json_type(validator[0])] + validator[1:-1]), + anglicize_json_type(validator[-1])) def _parse_oneof_validator(error):
Cryptic error when volumes key has incorrect syntax In the doc, there is very little explanation for the named volumes. It should be at least said that using one implies also having a top-level volumes key. Nevermind, running `docker-compose build` gives an explicit message about this. So I added a top-level volumes key like this: ``` yaml version: '2' services: app: volumes: - db:/var/lib/mysql volumes: - db ``` But the `docker-compose build` failed miserably with the following error: ``` Traceback (most recent call last): File "/usr/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.6.0', 'console_scripts', 'docker-compose')() File "/usr/lib/python3.5/site-packages/compose/cli/main.py", line 55, in main command.sys_dispatch() File "/usr/lib/python3.5/site-packages/compose/cli/docopt_command.py", line 23, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/lib/python3.5/site-packages/compose/cli/docopt_command.py", line 26, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/lib/python3.5/site-packages/compose/cli/main.py", line 172, in perform_command project = project_from_options(self.base_dir, options) File "/usr/lib/python3.5/site-packages/compose/cli/command.py", line 52, in project_from_options verbose=options.get('--verbose'), File "/usr/lib/python3.5/site-packages/compose/cli/command.py", line 80, in get_project config_data = config.load(config_details) File "/usr/lib/python3.5/site-packages/compose/config/config.py", line 288, in load for config_file in config_details.config_files File "/usr/lib/python3.5/site-packages/compose/config/config.py", line 288, in <listcomp> for config_file in config_details.config_files File "/usr/lib/python3.5/site-packages/compose/config/config.py", line 389, in process_config_file config_file.get_volumes(), 'volume') File "/usr/lib/python3.5/site-packages/compose/config/interpolation.py", line 25, in interpolate_environment_variables for name, config_dict in config.items() AttributeError: 'list' object has no attribute 'items' ``` Instead, the expected syntax is (at least that's what's working for me): ``` yaml version: '2' services: app: volumes: - db:/var/lib/mysql volumes: db: driver: local ``` It would be nice to at least add an example in the docs. And maybe give a less cryptic error message. Anyway, thanks for the great work, docker-compose is awesome!
https://docs.docker.com/compose/compose-file/#version-2 has an example, but I guess it wouldn't hurt to have one in the volume section as well. We should have a better error message for that case.
2016-02-18T04:00:35Z
[]
[]
Traceback (most recent call last): File "/usr/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.6.0', 'console_scripts', 'docker-compose')() File "/usr/lib/python3.5/site-packages/compose/cli/main.py", line 55, in main command.sys_dispatch() File "/usr/lib/python3.5/site-packages/compose/cli/docopt_command.py", line 23, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/lib/python3.5/site-packages/compose/cli/docopt_command.py", line 26, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/lib/python3.5/site-packages/compose/cli/main.py", line 172, in perform_command project = project_from_options(self.base_dir, options) File "/usr/lib/python3.5/site-packages/compose/cli/command.py", line 52, in project_from_options verbose=options.get('--verbose'), File "/usr/lib/python3.5/site-packages/compose/cli/command.py", line 80, in get_project config_data = config.load(config_details) File "/usr/lib/python3.5/site-packages/compose/config/config.py", line 288, in load for config_file in config_details.config_files File "/usr/lib/python3.5/site-packages/compose/config/config.py", line 288, in <listcomp> for config_file in config_details.config_files File "/usr/lib/python3.5/site-packages/compose/config/config.py", line 389, in process_config_file config_file.get_volumes(), 'volume') File "/usr/lib/python3.5/site-packages/compose/config/interpolation.py", line 25, in interpolate_environment_variables for name, config_dict in config.items() AttributeError: 'list' object has no attribute 'items'
4,793
docker/compose
docker__compose-2975
c7ceacfeaea5e8c3bd0e91cb04ac1e139446ce75
diff --git a/compose/config/validation.py b/compose/config/validation.py --- a/compose/config/validation.py +++ b/compose/config/validation.py @@ -312,6 +312,10 @@ def _parse_oneof_validator(error): types = [] for context in error.context: + if context.validator == 'oneOf': + _, error_msg = _parse_oneof_validator(context) + return path_string(context.path), error_msg + if context.validator == 'required': return (None, context.message)
Empty args list results in unhelpful error Having the Dockerfile ``` Dockerfile FROM scratch ``` and the `docker-compose.yml` file ``` yml version: "2" services: web: build: context: . args: ``` Gives me the following error on any docker-compose command (`config`, `build`) ``` Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/compose/cli/main.py", line 55, in main File "/code/compose/cli/docopt_command.py", line 23, in sys_dispatch File "/code/compose/cli/docopt_command.py", line 26, in dispatch File "/code/compose/cli/main.py", line 169, in perform_command File "/code/compose/cli/main.py", line 210, in config File "/code/compose/config/config.py", line 288, in load File "/code/compose/config/config.py", line 397, in process_config_file File "/code/compose/config/validation.py", line 371, in validate_against_fields_schema File "/code/compose/config/validation.py", line 412, in _validate_against_schema File "/code/compose/config/validation.py", line 362, in process_errors File "/code/compose/config/validation.py", line 362, in <genexpr> File "/code/compose/config/validation.py", line 360, in format_error_message File "/code/compose/config/validation.py", line 248, in handle_generic_service_error File "/code/compose/config/validation.py", line 328, in _parse_oneof_validator File "/code/compose/config/validation.py", line 303, in _parse_valid_types_from_validator File "/code/compose/config/validation.py", line 188, in anglicize_validator TypeError: coercing to Unicode: need string or buffer, dict found docker-compose returned -1 ``` Using `docker-compose version 1.6.0, build d99cad6` on Ubuntu 15.10. I see the empty dictionary is a stupid mistake, it came up when I commented out all the arguments, but would it be possible to provide a helpful error message here?
2016-02-19T19:24:35Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/compose/cli/main.py", line 55, in main File "/code/compose/cli/docopt_command.py", line 23, in sys_dispatch File "/code/compose/cli/docopt_command.py", line 26, in dispatch File "/code/compose/cli/main.py", line 169, in perform_command File "/code/compose/cli/main.py", line 210, in config File "/code/compose/config/config.py", line 288, in load File "/code/compose/config/config.py", line 397, in process_config_file File "/code/compose/config/validation.py", line 371, in validate_against_fields_schema File "/code/compose/config/validation.py", line 412, in _validate_against_schema File "/code/compose/config/validation.py", line 362, in process_errors File "/code/compose/config/validation.py", line 362, in <genexpr> File "/code/compose/config/validation.py", line 360, in format_error_message File "/code/compose/config/validation.py", line 248, in handle_generic_service_error File "/code/compose/config/validation.py", line 328, in _parse_oneof_validator File "/code/compose/config/validation.py", line 303, in _parse_valid_types_from_validator File "/code/compose/config/validation.py", line 188, in anglicize_validator TypeError: coercing to Unicode: need string or buffer, dict found
4,794
docker/compose
docker__compose-3269
86530287d6fdc13e52ffbc3fe2447888ecc6e197
diff --git a/compose/container.py b/compose/container.py --- a/compose/container.py +++ b/compose/container.py @@ -39,7 +39,7 @@ def from_ps(cls, client, dictionary, **kwargs): @classmethod def from_id(cls, client, id): - return cls(client, client.inspect_container(id)) + return cls(client, client.inspect_container(id), has_been_inspected=True) @classmethod def create(cls, client, **options):
`docker-compose logs` gets confused when scaling down and then up again **What I do** - View logs of a service with `docker-compose logs --follow` - Scale up service (e.g. from 1 to 2) - Scale down service (e.g. from 2 to 1) - Scale up again (e.g. from 1 to 3) **What I expect to see** - Compose streams logs of service instance 1 - ... then instances 1+2 - ... then instance 1 only (possibly with a message telling me that instance 2 is gone) - ... the instances 1+2+3 **What I see instead** - Compose streams logs of service instance 1 - ... then instances 1+2 - ... then, when scaling down, I get the following traceback: ``` Traceback (most recent call last): File "threading.py", line 810, in __bootstrap_inner File "threading.py", line 763, in run File "compose/cli/log_printer.py", line 190, in watch_events File "compose/project.py", line 355, in events File "compose/container.py", line 71, in service File "compose/container.py", line 108, in labels File "compose/container.py", line 181, in get File "compose/container.py", line 236, in inspect_if_not_inspected File "compose/container.py", line 245, in inspect File ".tox/py27/lib/python2.7/site-packages/docker/utils/decorators.py", line 21, in wrapped File ".tox/py27/lib/python2.7/site-packages/docker/api/container.py", line 182, in inspect_container File ".tox/py27/lib/python2.7/site-packages/docker/client.py", line 158, in _result File ".tox/py27/lib/python2.7/site-packages/docker/client.py", line 153, in _raise_for_status NotFound: 404 Client Error: Not Found ("No such container: 751d68c04b7bef09e6b4b2bbfd74b6733736491621a6b131ac68beba8751970a") ``` - Compose then only displays logs for instance 1 and doesn't display other instances, even if I scale back up **Extra info** ``` $ docker-compose -v docker-compose version 1.7.0rc1, build 1ad8866 ``` To reproduce: ``` bash git clone git://github.com/jpetazzo/orchestration-workshop cd orchestration-workshop/dockercoins docker-compose up -d docker-compose logs --tail 1 --follow worker # in another terminal ... docker-compose scale worker=2 docker-compose scale worker=1 docker-compose scale worker=3 docker-compose down ``` Related to: #3075 #2227
Thanks for the bug report! I'm not able to reproduce this issue. I think this would only happen if the container was removed with `docker rm -f <conatiner id>`, but even with that I couldn't reproduce the issue. If scale stops a container it should still exist, so I'm not sure why it failed to find it. That said, I did find a bug while looking into it. I had attempted to handle this error, but I missed `Container.from_id()` was not setting `has_been_inspected` to `True`, which meant that a second API call was being made unnecessarily outside of the `try/except`, which is what lead to this crash. So even though I wasn't able to reproduce the problem, I think I have a fix for it.
2016-04-05T15:35:26Z
[]
[]
Traceback (most recent call last): File "threading.py", line 810, in __bootstrap_inner File "threading.py", line 763, in run File "compose/cli/log_printer.py", line 190, in watch_events File "compose/project.py", line 355, in events File "compose/container.py", line 71, in service File "compose/container.py", line 108, in labels File "compose/container.py", line 181, in get File "compose/container.py", line 236, in inspect_if_not_inspected File "compose/container.py", line 245, in inspect File ".tox/py27/lib/python2.7/site-packages/docker/utils/decorators.py", line 21, in wrapped File ".tox/py27/lib/python2.7/site-packages/docker/api/container.py", line 182, in inspect_container File ".tox/py27/lib/python2.7/site-packages/docker/client.py", line 158, in _result File ".tox/py27/lib/python2.7/site-packages/docker/client.py", line 153, in _raise_for_status NotFound: 404 Client Error: Not Found ("No such container: 751d68c04b7bef09e6b4b2bbfd74b6733736491621a6b131ac68beba8751970a")
4,815
docker/compose
docker__compose-3351
a0aea42f758038536ab984244a0a61f1097624dc
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -34,7 +34,7 @@ def find_version(*file_paths): 'requests >= 2.6.1, < 2.8', 'texttable >= 0.8.1, < 0.9', 'websocket-client >= 0.32.0, < 1.0', - 'docker-py > 1.7.2, < 2', + 'docker-py >= 1.8.0, < 2', 'dockerpty >= 0.4.1, < 0.5', 'six >= 1.3.0, < 2', 'jsonschema >= 2.5.1, < 3',
TypeError: kwargs_from_env() got an unexpected keyword argument 'environment' after 1.7 update Hello! After updating Docker got this error when building: ``` $ docker-compose build Traceback (most recent call last): File "/usr/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.7.0', 'console_scripts', 'docker-compose')() File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 57, in main command() File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 105, in perform_command project = project_from_options('.', options) File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 31, in project_from_options environment=environment File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 76, in get_project host=host, environment=environment File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 49, in get_client environment=environment File "/usr/lib/python2.7/site-packages/compose/cli/docker_client.py", line 52, in docker_client kwargs = kwargs_from_env(environment=environment) TypeError: kwargs_from_env() got an unexpected keyword argument 'environment' ``` I'm using Fedora 23. Docker from docker repository. ``` $ rpm -qa | grep -i docker docker-compose-1.7.0-1.fc23.noarch python-docker-py-1.7.2-1.fc23.noarch docker-engine-1.11.0-1.fc23.x86_64 python-dockerpty-0.4.1-2.fc23.noarch python3-docker-py-1.7.2-1.fc23.noarch docker-engine-selinux-1.11.0-1.fc23.noarch ``` ``` $ python --version Python 2.7.11 ``` Docker-compose 1.6 still works well.
Looks like we forgot to update `setup.py`, it requires `docker-py >= 1.8`, so you'll need to pip install that. setup.py specifies `docker-py > 1.7.2, < 2` which should theoretically install 1.8.0, but Fedora's repackaging/distribution might confuse it.
2016-04-20T00:39:56Z
[]
[]
Traceback (most recent call last): File "/usr/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.7.0', 'console_scripts', 'docker-compose')() File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 57, in main command() File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 105, in perform_command project = project_from_options('.', options) File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 31, in project_from_options environment=environment File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 76, in get_project host=host, environment=environment File "/usr/lib/python2.7/site-packages/compose/cli/command.py", line 49, in get_client environment=environment File "/usr/lib/python2.7/site-packages/compose/cli/docker_client.py", line 52, in docker_client kwargs = kwargs_from_env(environment=environment) TypeError: kwargs_from_env() got an unexpected keyword argument 'environment'
4,823
docker/compose
docker__compose-3436
c9fe8920c97384b02ea446d95a4d59df4a3d75a3
diff --git a/compose/cli/main.py b/compose/cli/main.py --- a/compose/cli/main.py +++ b/compose/cli/main.py @@ -334,6 +334,13 @@ def exec_command(self, options): """ index = int(options.get('--index')) service = self.project.get_service(options['SERVICE']) + detach = options['-d'] + + if IS_WINDOWS_PLATFORM and not detach: + raise UserError( + "Interactive mode is not yet supported on Windows.\n" + "Please pass the -d flag when using `docker-compose exec`." + ) try: container = service.get_container(number=index) except ValueError as e: @@ -350,7 +357,7 @@ def exec_command(self, options): exec_id = container.create_exec(command, **create_exec_options) - if options['-d']: + if detach: container.start_exec(exec_id, tty=tty) return
Windows: NameError: global name 'ExecOperation' is not defined If I try to run a `docker-compose exec` command it always gives me this error: ``` D:\Path\> docker-compose exec some-container some-command Traceback (most recent call last): File "<string>", line 3, in <module> File "compose\cli\main.py", line 57, in main File "compose\cli\main.py", line 108, in perform_command File "compose\cli\main.py", line 347, in exec_command NameError: global name 'ExecOperation' is not defined docker-compose returned -1 ``` Is there a known bug on this message? My guess is that this has something to do with this message `Interactive mode is not yet supported on Windows.` when you use `docker-compose run` but I'm not sure.. Thanks for any help in advance! Some more system information: ``` D:\> docker version Client: Version: 1.11.0 API version: 1.23 Go version: go1.5.4 Git commit: 4dc5990 Built: Wed Apr 13 18:13:28 2016 OS/Arch: windows/amd64 Server: Version: 1.11.0 API version: 1.23 Go version: go1.5.4 Git commit: 4dc5990 Built: Wed Apr 13 19:36:04 2016 OS/Arch: linux/amd64 D:\> docker-compose version docker-compose version 1.7.0, build 0d7bf73 docker-py version: 1.8.0 CPython version: 2.7.11 OpenSSL version: OpenSSL 1.0.2d 9 Jul 2015 ```
Yes, this is similar to the warning in `docker run`, with `exec` not handling the error quite as gracefully. `exec` on Windows currently only supports detached mode (`-d`)
2016-05-09T23:30:43Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "compose\cli\main.py", line 57, in main File "compose\cli\main.py", line 108, in perform_command File "compose\cli\main.py", line 347, in exec_command NameError: global name 'ExecOperation' is not defined
4,830
docker/compose
docker__compose-3615
72d3d5d84ba29b1fb6d8ac37dfc8eca76812b2e2
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -34,7 +34,7 @@ def find_version(*file_paths): 'requests >= 2.6.1, < 2.8', 'texttable >= 0.8.1, < 0.9', 'websocket-client >= 0.32.0, < 1.0', - 'docker-py >= 1.8.1, < 2', + 'docker-py == 1.9.0rc2', 'dockerpty >= 0.4.1, < 0.5', 'six >= 1.3.0, < 2', 'jsonschema >= 2.5.1, < 3',
Support npipe url values for DOCKER_HOST The Windows Docker Engine recently got support for named pipe (think of the unix socket on Linux) instead of listening on local http port 2375. See https://github.com/docker/docker/pull/19911 for the first PR. Now it seems that Docker Compose needs this support as well when the Docker Engine is no longer listening on tcp port. ``` C:\> docker-compose --verbose up compose.config.config.find: Using configuration files: .\docker-compose.yml docker.auth.auth.load_config: File doesn't exist Traceback (most recent call last): File "<string>", line 3, in <module> File "C:\projects\compose\compose\cli\main.py", line 55, in main File "C:\projects\compose\compose\cli\docopt_command.py", line 23, in sys_dispatch File "C:\projects\compose\compose\cli\docopt_command.py", line 26, in dispatch File "C:\projects\compose\compose\cli\main.py", line 172, in perform_command File "C:\projects\compose\compose\cli\command.py", line 52, in project_from_options File "C:\projects\compose\compose\cli\command.py", line 85, in get_project File "C:\projects\compose\compose\cli\command.py", line 68, in get_client File "c:\projects\compose\venv\lib\site-packages\docker\api\daemon.py", line 78, in version File "c:\projects\compose\venv\lib\site-packages\docker\utils\decorators.py", line 47, in inner File "c:\projects\compose\venv\lib\site-packages\docker\client.py", line 112, in _get File "c:\projects\compose\venv\lib\site-packages\requests\sessions.py", line 477, in get File "c:\projects\compose\venv\lib\site-packages\requests\sessions.py", line 465, in request File "c:\projects\compose\venv\lib\site-packages\requests\sessions.py", line 573, in send File "c:\projects\compose\venv\lib\site-packages\requests\adapters.py", line 415, in send requests.exceptions.ConnectionError: ('Connection aborted.', error(10061, 'No connection could be made because the target machine actively refused it')) docker-compose returned -1 C:\> docker-compose --version docker-compose version 1.6.0, build cdb920a C:\> docker version Client: Version: 1.11.0-dev API version: 1.23 Go version: go1.6 Git commit: 9e53024 Built: Sat Mar 19 14:35:55 2016 OS/Arch: windows/amd64 Server: Version: 1.11.0-dev API version: 1.23 Go version: go1.6 Git commit: 9e53024 Built: Sat Mar 19 14:35:55 2016 OS/Arch: windows/amd64 ``` Tested on a TP4 with the nightly docker.exe
2016-06-16T22:13:12Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "C:\projects\compose\compose\cli\main.py", line 55, in main File "C:\projects\compose\compose\cli\docopt_command.py", line 23, in sys_dispatch File "C:\projects\compose\compose\cli\docopt_command.py", line 26, in dispatch File "C:\projects\compose\compose\cli\main.py", line 172, in perform_command File "C:\projects\compose\compose\cli\command.py", line 52, in project_from_options File "C:\projects\compose\compose\cli\command.py", line 85, in get_project File "C:\projects\compose\compose\cli\command.py", line 68, in get_client File "c:\projects\compose\venv\lib\site-packages\docker\api\daemon.py", line 78, in version File "c:\projects\compose\venv\lib\site-packages\docker\utils\decorators.py", line 47, in inner File "c:\projects\compose\venv\lib\site-packages\docker\client.py", line 112, in _get File "c:\projects\compose\venv\lib\site-packages\requests\sessions.py", line 477, in get File "c:\projects\compose\venv\lib\site-packages\requests\sessions.py", line 465, in request File "c:\projects\compose\venv\lib\site-packages\requests\sessions.py", line 573, in send File "c:\projects\compose\venv\lib\site-packages\requests\adapters.py", line 415, in send requests.exceptions.ConnectionError: ('Connection aborted.', error(10061, 'No connection could be made because the target machine actively refused it'))
4,841
docker/compose
docker__compose-367
847ec5b559cf97296c80624774cc4210bbabcac6
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -22,8 +22,16 @@ def find_version(*file_paths): return version_match.group(1) raise RuntimeError("Unable to find version string.") -with open('requirements.txt') as f: - install_requires = f.read().splitlines() + +install_requires = [ + 'docopt >= 0.6.1, < 0.7', + 'PyYAML >= 3.10, < 4', + 'requests >= 2.2.1, < 3', + 'texttable >= 0.8.1, < 0.9', + 'websocket-client >= 0.11.0, < 0.12', + 'dockerpty >= 0.2.3, < 0.3', +] + with open('requirements-dev.txt') as f: tests_require = f.read().splitlines()
Using == in setup.py from requirements.txt will cause conflicts with other python packages requirements.txt with == operator is meant to be used when you need completely limit the versions of package. fig uses requirements.txt input in setup.py to fill install_requires. If any package installed after fig requires other version or updates the dependency things will break. ``` fig up -d Traceback (most recent call last): File "/usr/local/bin/fig", line 5, in <module> from pkg_resources import load_entry_point File "/usr/lib/python2.7/dist-packages/pkg_resources.py", line 2749, in <module> working_set = WorkingSet._build_master() File "/usr/lib/python2.7/dist-packages/pkg_resources.py", line 446, in _build_master return cls._build_from_requirements(__requires__) File "/usr/lib/python2.7/dist-packages/pkg_resources.py", line 459, in _build_from_requirements dists = ws.resolve(reqs, Environment()) File "/usr/lib/python2.7/dist-packages/pkg_resources.py", line 628, in resolve raise DistributionNotFound(req) pkg_resources.DistributionNotFound: requests==2.2.1 ``` The latest version of requests package available is 2.3.0 (which was installed in this case) and it's of course backward compatibile. You should either stop relaying on reading requirements.txt or start using >= operator.
`requirements.txt` should actually contain a full list of pinned versions (using ==). That way tests can run against a specific version each time, and you have a way to build a virtualenv with known working dependencies. I think the problem here is that setup.py uses the list from requirements.txt for `install_requires`, which is not correct. setup.py should use `>=, <` Reference: https://caremad.io/blog/setup-vs-requirement/
2014-07-29T17:30:45Z
[]
[]
Traceback (most recent call last): File "/usr/local/bin/fig", line 5, in <module> from pkg_resources import load_entry_point File "/usr/lib/python2.7/dist-packages/pkg_resources.py", line 2749, in <module> working_set = WorkingSet._build_master() File "/usr/lib/python2.7/dist-packages/pkg_resources.py", line 446, in _build_master return cls._build_from_requirements(__requires__) File "/usr/lib/python2.7/dist-packages/pkg_resources.py", line 459, in _build_from_requirements dists = ws.resolve(reqs, Environment()) File "/usr/lib/python2.7/dist-packages/pkg_resources.py", line 628, in resolve raise DistributionNotFound(req) pkg_resources.DistributionNotFound: requests==2.2.1
4,844
docker/compose
docker__compose-3830
d29f8e10222f3dbd3b57d8dfb9a15404a4302609
diff --git a/compose/cli/docker_client.py b/compose/cli/docker_client.py --- a/compose/cli/docker_client.py +++ b/compose/cli/docker_client.py @@ -11,15 +11,16 @@ from ..const import HTTP_TIMEOUT from .errors import UserError from .utils import generate_user_agent +from .utils import unquote_path log = logging.getLogger(__name__) def tls_config_from_options(options): tls = options.get('--tls', False) - ca_cert = options.get('--tlscacert') - cert = options.get('--tlscert') - key = options.get('--tlskey') + ca_cert = unquote_path(options.get('--tlscacert')) + cert = unquote_path(options.get('--tlscert')) + key = unquote_path(options.get('--tlskey')) verify = options.get('--tlsverify') skip_hostname_check = options.get('--skip-hostname-check', False) diff --git a/compose/cli/utils.py b/compose/cli/utils.py --- a/compose/cli/utils.py +++ b/compose/cli/utils.py @@ -122,3 +122,11 @@ def generate_user_agent(): else: parts.append("{}/{}".format(p_system, p_release)) return " ".join(parts) + + +def unquote_path(s): + if not s: + return s + if s[0] == '"' and s[-1] == '"': + return s[1:-1] + return s
docker-compose $(docker-machine config default) up Fails. docker and docker-compose handle arguments differently. `docker-compose $(docker-machine config default) up`causes an error. (see below) and `docker-compose $(docker-machine config default | xargs) up` works. joe@jsd-mbp ~ $ docker-compose $(docker-machine config default) -f docker-compose.yml up [2.2.3] Traceback (most recent call last): File "<string>", line 3, in <module> File "compose/cli/main.py", line 58, in main File "compose/cli/main.py", line 106, in perform_command File "compose/cli/command.py", line 33, in project_from_options File "compose/cli/docker_client.py", line 36, in tls_config_from_options File "site-packages/docker/tls.py", line 47, in **init** docker.errors.TLSParameterError: Path to a certificate and key files must be provided through the client_config param. TLS configurations should map the Docker CLI client configurations. See https://docs.docker.com/engine/articles/https/ for API details. docker-compose returned -1 joe@jsd-mbp ~ $
this issus is that tls.py is calling os.path.isfile(cert) where cert is a quoted path. so in python ```tls_cert=' "/Users/joe/.docker/machine/cert.pem" ' Makes sense. Thanks for the report! Compose doesn't like it, even if you strip the quotes ``` $ docker-compose $(docker-machine config default | sed 's/\"//g') ps ERROR: TLS configuration is invalid - make sure your DOCKER_TLS_VERIFY and DOCKER_CERT_PATH are set correctly. You might need to run `eval "$(docker-machine env default)"` ``` Try docker-compose $(docker-machine config default | xargs ) ps Sent from my iPad > On Jun 9, 2016, at 8:30 AM, Matt Wright notifications@github.com wrote: > > Compose doesn't like it, even if you strip the quotes > > $ docker-compose $(docker-machine config default | sed 's/\"//g') ps > ERROR: TLS configuration is invalid - make sure your DOCKER_TLS_VERIFY and DOCKER_CERT_PATH are set correctly. > You might need to run `eval "$(docker-machine env default)"` > — > You are receiving this because you authored the thread. > Reply to this email directly, view it on GitHub, or mute the thread. @jduhamel ah, that does work, thanks
2016-08-09T22:14:20Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "compose/cli/main.py", line 58, in main File "compose/cli/main.py", line 106, in perform_command File "compose/cli/command.py", line 33, in project_from_options File "compose/cli/docker_client.py", line 36, in tls_config_from_options File "site-packages/docker/tls.py", line 47, in **init** docker.errors.TLSParameterError: Path to a certificate and key files must be provided through the client_config param. TLS configurations should map the Docker CLI client configurations. See https://docs.docker.com/engine/articles/https/ for API details.
4,853
docker/compose
docker__compose-387
0a15e7fe9cae05bd1312db224d1499752fd41443
diff --git a/fig/cli/formatter.py b/fig/cli/formatter.py --- a/fig/cli/formatter.py +++ b/fig/cli/formatter.py @@ -4,11 +4,17 @@ import texttable +def get_tty_width(): + tty_size = os.popen('stty size', 'r').read().split() + if len(tty_size) != 2: + return 80 + _, width = tty_size + return width + + class Formatter(object): def table(self, headers, rows): - height, width = os.popen('stty size', 'r').read().split() - - table = texttable.Texttable(max_width=width) + table = texttable.Texttable(max_width=get_tty_width()) table.set_cols_dtype(['t' for h in headers]) table.add_rows([headers] + rows) table.set_deco(table.HEADER)
fig ps fails hellou I use fig (v0.5.2) to build my docker images on a jenkins server. my script builds all images, which looks like this: echo ----- BUILD fig up -d echo ----- DONE fig ps however the `fig ps`call just fails: ``` stty: standard input: Inappropriate ioctl for device Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/fig/out00-PYZ.pyz/fig.cli.main", line 30, in main File "/code/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 21, in sys_dispatch File "/code/build/fig/out00-PYZ.pyz/fig.cli.command", line 31, in dispatch File "/code/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 24, in dispatch File "/code/build/fig/out00-PYZ.pyz/fig.cli.command", line 50, in perform_command File "/code/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 27, in perform_command File "/code/build/fig/out00-PYZ.pyz/fig.cli.main", line 178, in ps File "/code/build/fig/out00-PYZ.pyz/fig.cli.formatter", line 9, in table ValueError: need more than 0 values to unpack ```
I've run into this same problems with another python project while running on jenkins. ``` os.popen('stty size', 'r').read() ``` Returns an empty string on jenkins, so the splitting into two variables fails. `fig.cli.formatter.Formatter` needs to check for empty list after `split()`, and default to some value (probably 80).
2014-08-07T16:30:53Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "/code/build/fig/out00-PYZ.pyz/fig.cli.main", line 30, in main File "/code/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 21, in sys_dispatch File "/code/build/fig/out00-PYZ.pyz/fig.cli.command", line 31, in dispatch File "/code/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 24, in dispatch File "/code/build/fig/out00-PYZ.pyz/fig.cli.command", line 50, in perform_command File "/code/build/fig/out00-PYZ.pyz/fig.cli.docopt_command", line 27, in perform_command File "/code/build/fig/out00-PYZ.pyz/fig.cli.main", line 178, in ps File "/code/build/fig/out00-PYZ.pyz/fig.cli.formatter", line 9, in table ValueError: need more than 0 values to unpack
4,854
docker/compose
docker__compose-3930
3dec600d8218833de11279e1dac6af048222dbc6
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -34,7 +34,7 @@ def find_version(*file_paths): 'requests >= 2.6.1, < 2.8', 'texttable >= 0.8.1, < 0.9', 'websocket-client >= 0.32.0, < 1.0', - 'docker-py >= 1.9.0, < 2.0', + 'docker-py >= 1.10.2, < 2.0', 'dockerpty >= 0.4.1, < 0.5', 'six >= 1.3.0, < 2', 'jsonschema >= 2.5.1, < 3',
Cannot use URL as build context According to https://docs.docker.com/compose/compose-file/#/context it should be possible to use URL to a git repository for build context, but I am having no luck docker-py version: 1.8.1 CPython version: 2.7.9 OpenSSL version: OpenSSL 1.0.1e 11 Feb 2013 docker-compose.yml ``` version: '2' services: test: build: context: https://github.com/docker/compose.git ``` docker-compose build ``` Building test Traceback (most recent call last): File "<string>", line 3, in <module> File "compose/cli/main.py", line 58, in main File "compose/cli/main.py", line 109, in perform_command File "compose/cli/main.py", line 215, in build File "compose/project.py", line 300, in build File "compose/service.py", line 719, in build File ".tox/py27/lib/python2.7/site-packages/docker/api/build.py", line 104, in build File ".tox/py27/lib/python2.7/site-packages/docker/utils/decorators.py", line 46, in inner AttributeError: 'NoneType' object has no attribute 'update' docker-compose returned -1 ``` Unable to pull private images when using credentials helper I'm using the secretservice credentials helper with docker 1.11.2 and docker-compose 1.8.0-rc1. If I try to run `docker-compose up`, it fails to pull private images from our gitlab registry. --verbose says: ``` docker.auth.auth.resolve_authconfig: No entry found docker.api.image.pull: No auth config found ``` Pulling the image manually using docker pull works.
I can't seem to be able to reproduce this on 1.7.1 or 1.8.0. What version of docker-compose are you using? Did you install it through pip, or are you using the binary distribution? Hi @shin- , I just tried with 1.8.0. ``` $ docker-compose version docker-compose version 1.8.0, build f3628c7 docker-py version: 1.9.0 CPython version: 2.7.9 OpenSSL version: OpenSSL 1.0.1e 11 Feb 2013 ``` Installed (or download) from https://github.com/docker/compose/releases: ``` curl -L https://github.com/docker/compose/releases/download/1.8.0/docker-compose-`uname -s`-`uname -m` > /usr/local/bin/docker-compose chmod +x /usr/local/bin/docker-compose ``` Running in Ubuntu: ``` $ uname -a Linux devmachine 4.4.0-31-generic #50-Ubuntu SMP Wed Jul 13 00:07:12 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux ``` ``` $ docker-compose build Building test Traceback (most recent call last): File "<string>", line 3, in <module> File "compose/cli/main.py", line 61, in main File "compose/cli/main.py", line 113, in perform_command File "compose/cli/main.py", line 219, in build File "compose/project.py", line 300, in build File "compose/service.py", line 727, in build File "site-packages/docker/api/build.py", line 104, in build File "site-packages/docker/utils/decorators.py", line 46, in inner AttributeError: 'NoneType' object has no attribute 'update' docker-compose returned -1 ``` @feltraco Care to share your compose file? I still can't reproduce the issue using a similar environment. ``` $ ls docker-compose docker-compose.yml $ ./docker-compose version docker-compose version 1.8.0, build f3628c7 docker-py version: 1.9.0 CPython version: 2.7.9 OpenSSL version: OpenSSL 1.0.1e 11 Feb 2013 $ cat docker-compose.yml version: '2' services: test: build: context: https://github.com/docker/compose.git $ uname -a Linux yuna 4.4.0-31-generic #50-Ubuntu SMP Wed Jul 13 00:07:12 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux $ ./docker-compose build Building test Step 1 : FROM debian:wheezy ---> 2e9c7e5da19c [...] Successfully built ade0143bebdb ``` Oh, nevermind - the issue triggers when you have a `HttpHeaders` entry in your `config.json` file. I've created issue https://github.com/docker/docker-py/issues/1148 to track its resolution in docker-py, after which we can port the fix in compose. Thank you for the report, and thanks for your patience. docker-compose version 1.8.0-rc1, build 9bf6bc6 docker-py version: 1.8.1 CPython version: 2.7.9 OpenSSL version: OpenSSL 1.0.1e 11 Feb 2013 same here, same docker-compose version. Docker pull works, of course - cause is the credentials helper, i use ``` "credsStore": "osxkeychain" ``` docker-compose version 1.8.0-rc1 Docker version 1.11.2, build b9f10c9 I think docker-py first needs to be updated to support the credential store API. See docker/docker-py/issues/1023. Then docker-compose can be updated to use the new docker-py. @grosskur docker is supporting credenial stores now, but compose is not. Is there any chance to move on?
2016-09-10T00:21:50Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "compose/cli/main.py", line 58, in main File "compose/cli/main.py", line 109, in perform_command File "compose/cli/main.py", line 215, in build File "compose/project.py", line 300, in build File "compose/service.py", line 719, in build File ".tox/py27/lib/python2.7/site-packages/docker/api/build.py", line 104, in build File ".tox/py27/lib/python2.7/site-packages/docker/utils/decorators.py", line 46, in inner AttributeError: 'NoneType' object has no attribute 'update'
4,859
docker/compose
docker__compose-3964
7687412e03b9c37cced49109d142de9d73472e8f
diff --git a/compose/service.py b/compose/service.py --- a/compose/service.py +++ b/compose/service.py @@ -1109,6 +1109,8 @@ def format_environment(environment): def format_env(key, value): if value is None: return key + if isinstance(value, six.binary_type): + value = value.decode('utf-8') return '{key}={value}'.format(key=key, value=value) return [format_env(*item) for item in environment.items()]
UnicodeDecodeError: 'ascii' codec can't decode byte 0xc5 in position 7: ordinal not in range(128) In `docker-compose.yml` I have something like that: ``` yaml app: environment: VARIABLE: ${VARIABLE} ``` When `VARIABLE` contains characters from Polish alphabet like `ą, ć, ę, ł, ń, ó, ś, ź, ż` it returns error: ``` Traceback (most recent call last): File "<string>", line 3, in <module> File "compose/cli/main.py", line 56, in main File "compose/cli/docopt_command.py", line 23, in sys_dispatch File "compose/cli/docopt_command.py", line 26, in dispatch File "compose/cli/main.py", line 191, in perform_command File "compose/cli/main.py", line 657, in up File "compose/project.py", line 318, in up File "compose/service.py", line 370, in execute_convergence_plan File "compose/service.py", line 410, in recreate_container File "compose/service.py", line 258, in create_container File "compose/service.py", line 625, in _get_container_create_options File "compose/service.py", line 1031, in format_environment File "compose/service.py", line 1030, in format_env UnicodeDecodeError: 'ascii' codec can't decode byte 0xc5 in position 7: ordinal not in range(128) docker-compose returned -1 make: *** [run_compose_up] Error 255 ``` Is it possible that more than ASCII will be handled?
Hi @szemek , what version of Compose are you running? I confirmed the issue still happens with master.
2016-09-21T01:08:34Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "compose/cli/main.py", line 56, in main File "compose/cli/docopt_command.py", line 23, in sys_dispatch File "compose/cli/docopt_command.py", line 26, in dispatch File "compose/cli/main.py", line 191, in perform_command File "compose/cli/main.py", line 657, in up File "compose/project.py", line 318, in up File "compose/service.py", line 370, in execute_convergence_plan File "compose/service.py", line 410, in recreate_container File "compose/service.py", line 258, in create_container File "compose/service.py", line 625, in _get_container_create_options File "compose/service.py", line 1031, in format_environment File "compose/service.py", line 1030, in format_env UnicodeDecodeError: 'ascii' codec can't decode byte 0xc5 in position 7: ordinal not in range(128)
4,863
docker/compose
docker__compose-4104
3b46a62f363fe3d2e59a264151f840b6ec8e54b4
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -34,7 +34,7 @@ def find_version(*file_paths): 'requests >= 2.6.1, != 2.11.0, < 2.12', 'texttable >= 0.8.1, < 0.9', 'websocket-client >= 0.32.0, < 1.0', - 'docker-py >= 1.10.5, < 2.0', + 'docker-py >= 1.10.6, < 2.0', 'dockerpty >= 0.4.1, < 0.5', 'six >= 1.3.0, < 2', 'jsonschema >= 2.5.1, < 3',
Working with multiple Windows containers: The semaphore timeout period has expired. Starting twelve containers with docker-compose 1.9.0-rc2 still shows a problem. Follow up of #4082 ``` PS C:\xxxxx> docker-compose up -d Creating xxxxx_consul_1 Creating xxxxx_db_1 Creating xxxxx_service1_1 Creating xxxxx_service2_1 Creating xxxxx_service3_1 Creating xxxxx_service4_1 Creating xxxxx_service5_1 Creating xxxxx_cli_1 ERROR: for service6 (121, 'WaitNamedPipe', 'The semaphore timeout period has expired.') ERROR: for service7 (121, 'WaitNamedPipe', 'The semaphore timeout period has expired.') ERROR: for service8 (121, 'WaitNamedPipe', 'The semaphore timeout period has expired.') ERROR: for service9 (121, 'WaitNamedPipe', 'The semaphore timeout period has expired.') ERROR: for service3 (121, 'WaitNamedPipe', 'The semaphore timeout period has expired.') Traceback (most recent call last): File "<string>", line 3, in <module> File "compose\cli\main.py", line 65, in main File "compose\cli\main.py", line 117, in perform_command File "compose\cli\main.py", line 849, in up File "compose\project.py", line 400, in up File "compose\parallel.py", line 64, in parallel_execute pywintypes.error: (121, 'WaitNamedPipe', 'The semaphore timeout period has expired.') docker-compose returned -1 PS C:\xxxxx> docker-compose --version docker-compose version 1.9.0-rc2, build d41f869 ```
Thanks for reporting. Haven't seen that one before. From a quick research, it seems that error usually appears when the client can not reach the server through the named pipe - any chance your Docker Engine might have errored out or been down during that test for some reason? Can you reproduce consistently? Thanks I can reproduce it. Tried today starting 6 services at once and got the same error. Docker engine does not show any errors in Event log. I also checked with ``` powershell iwr https://github.com/Microsoft/Virtualization-Documentation/raw/master/windows-server-container-tools/Debug-ContainerHost/Debug-ContainerHost.ps1 -UseBasicParsing | iex ``` Windows Defender is running with high CPU load in background. Trying to turn it off. @StefanScherer Thanks for the feedback. I'll try and reproduce. I identified the issue: docker/docker-py#1284 (Fix PR: docker/docker-py#1285)
2016-11-02T23:51:03Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "compose\cli\main.py", line 65, in main File "compose\cli\main.py", line 117, in perform_command File "compose\cli\main.py", line 849, in up File "compose\project.py", line 400, in up File "compose\parallel.py", line 64, in parallel_execute pywintypes.error: (121, 'WaitNamedPipe', 'The semaphore timeout period has expired.')
4,872
docker/compose
docker__compose-4113
9046e33ab2b303ee3103cdc694fff54614f58977
diff --git a/compose/config/config.py b/compose/config/config.py --- a/compose/config/config.py +++ b/compose/config/config.py @@ -771,7 +771,7 @@ def merge_service_dicts(base, override, version): for field in ['dns', 'dns_search', 'env_file', 'tmpfs']: md.merge_field(field, merge_list_or_string) - md.merge_field('logging', merge_logging) + md.merge_field('logging', merge_logging, default={}) for field in set(ALLOWED_KEYS) - set(md): md.merge_scalar(field)
docker-compose.yml merge broken when i want to merge two docker-compose.yml i get following error with the current beta 1.12.3-beta29.3 (619507e) docker-compose -f ../docker-compose.yml -f docker-compose.yml up myService ``` Traceback (most recent call last): File "<string>", line 3, in <module> File "compose/cli/main.py", line 65, in main File "compose/cli/main.py", line 114, in perform_command File "compose/cli/command.py", line 36, in project_from_options File "compose/cli/command.py", line 103, in get_project File "compose/config/config.py", line 323, in load File "compose/config/config.py", line 414, in load_services File "compose/config/config.py", line 405, in merge_services File "compose/config/config.py", line 405, in <dictcomp> File "compose/config/config.py", line 704, in merge_service_dicts_from_files File "compose/config/config.py", line 774, in merge_service_dicts File "compose/config/config.py", line 728, in merge_field File "compose/config/config.py", line 807, in merge_logging File "compose/config/config.py", line 749, in merge_scalar File "compose/config/config.py", line 720, in needs_merge TypeError: argument of type 'NoneType' is not iterable docker-compose returned -1 ``` diagnostic id: E9B25A88-ABE3-46AF-B578-39A426964940
Thanks for the report! Related: #4103
2016-11-03T23:30:22Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "compose/cli/main.py", line 65, in main File "compose/cli/main.py", line 114, in perform_command File "compose/cli/command.py", line 36, in project_from_options File "compose/cli/command.py", line 103, in get_project File "compose/config/config.py", line 323, in load File "compose/config/config.py", line 414, in load_services File "compose/config/config.py", line 405, in merge_services File "compose/config/config.py", line 405, in <dictcomp> File "compose/config/config.py", line 704, in merge_service_dicts_from_files File "compose/config/config.py", line 774, in merge_service_dicts File "compose/config/config.py", line 728, in merge_field File "compose/config/config.py", line 807, in merge_logging File "compose/config/config.py", line 749, in merge_scalar File "compose/config/config.py", line 720, in needs_merge TypeError: argument of type 'NoneType' is not iterable
4,874
docker/compose
docker__compose-4339
5ade097d74ddb192a5baefa75cfce6fbad5c20a7
diff --git a/compose/parallel.py b/compose/parallel.py --- a/compose/parallel.py +++ b/compose/parallel.py @@ -12,6 +12,8 @@ from six.moves.queue import Queue from compose.cli.signals import ShutdownException +from compose.errors import HealthCheckFailed +from compose.errors import NoHealthCheckConfigured from compose.errors import OperationFailedError from compose.utils import get_output_stream @@ -48,7 +50,7 @@ def parallel_execute(objects, func, get_name, msg, get_deps=None): elif isinstance(exception, APIError): errors[get_name(obj)] = exception.explanation writer.write(get_name(obj), 'error') - elif isinstance(exception, OperationFailedError): + elif isinstance(exception, (OperationFailedError, HealthCheckFailed, NoHealthCheckConfigured)): errors[get_name(obj)] = exception.msg writer.write(get_name(obj), 'error') elif isinstance(exception, UpstreamError): @@ -164,21 +166,27 @@ def feed_queue(objects, func, get_deps, results, state): for obj in pending: deps = get_deps(obj) - - if any(dep[0] in state.failed for dep in deps): - log.debug('{} has upstream errors - not processing'.format(obj)) - results.put((obj, None, UpstreamError())) - state.failed.add(obj) - elif all( - dep not in objects or ( - dep in state.finished and (not ready_check or ready_check(dep)) - ) for dep, ready_check in deps - ): - log.debug('Starting producer thread for {}'.format(obj)) - t = Thread(target=producer, args=(obj, func, results)) - t.daemon = True - t.start() - state.started.add(obj) + try: + if any(dep[0] in state.failed for dep in deps): + log.debug('{} has upstream errors - not processing'.format(obj)) + results.put((obj, None, UpstreamError())) + state.failed.add(obj) + elif all( + dep not in objects or ( + dep in state.finished and (not ready_check or ready_check(dep)) + ) for dep, ready_check in deps + ): + log.debug('Starting producer thread for {}'.format(obj)) + t = Thread(target=producer, args=(obj, func, results)) + t.daemon = True + t.start() + state.started.add(obj) + except (HealthCheckFailed, NoHealthCheckConfigured) as e: + log.debug( + 'Healthcheck for service(s) upstream of {} failed - ' + 'not processing'.format(obj) + ) + results.put((obj, None, e)) if state.is_done(): results.put(STOP)
Failed healthcheck raises uncaught exception and kills execution. Just testing out the new healthcheck feature in config v2.1 and ensuring my containers start in order via the `depends_on` however, intermittently, I get an error like this: ```sh $ docker-compose up Creating network "default" with the default driver Creating rabbitmq_1 Creating redis_1 Creating s3_1 Creating database_1 Creating elasticsearch_1 Traceback (most recent call last): File "<string>", line 3, in <module> File "compose/cli/main.py", line 64, in main File "compose/cli/main.py", line 116, in perform_command File "compose/cli/main.py", line 848, in up File "compose/project.py", line 410, in up File "compose/parallel.py", line 44, in parallel_execute File "compose/parallel.py", line 118, in parallel_execute_iter File "compose/parallel.py", line 175, in feed_queue File "compose/parallel.py", line 175, in <genexpr> File "compose/service.py", line 560, in <lambda> File "compose/service.py", line 920, in is_healthy compose.errors.HealthCheckFailed docker-compose returned -1 ``` This occurs when my database's healthcheck fails on startup. After doing some digging, it looks like exceptions raised in the [`is_healthy`](https://github.com/docker/compose/blob/bump-1.10.0-rc2/compose/service.py#L905-L921) method aren't being caught because they're called (as [`ready_check`](https://github.com/docker/compose/blob/bump-1.10.0-rc2/compose/parallel.py#L174)) outside the [`try..except` scope (inside `feed_queue`) in the `parralel_execute_iter` method](https://github.com/docker/compose/blob/bump-1.10.0-rc2/compose/parallel.py#L118-L126). Does that make sense? ``` $ docker --version Docker version 1.13.0-rc6, build 2f2d055 $ docker-compose --version docker-compose version 1.10.0-rc2, build fb241d0 ```
Thanks for the report! I'll look into it.
2017-01-17T21:23:15Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "compose/cli/main.py", line 64, in main File "compose/cli/main.py", line 116, in perform_command File "compose/cli/main.py", line 848, in up File "compose/project.py", line 410, in up File "compose/parallel.py", line 44, in parallel_execute File "compose/parallel.py", line 118, in parallel_execute_iter File "compose/parallel.py", line 175, in feed_queue File "compose/parallel.py", line 175, in <genexpr> File "compose/service.py", line 560, in <lambda> File "compose/service.py", line 920, in is_healthy compose.errors.HealthCheckFailed
4,884
docker/compose
docker__compose-4370
9a0962dacb786556e1b87b9c117010521eda2f85
diff --git a/compose/service.py b/compose/service.py --- a/compose/service.py +++ b/compose/service.py @@ -22,6 +22,7 @@ from .config import merge_environment from .config.types import VolumeSpec from .const import DEFAULT_TIMEOUT +from .const import IS_WINDOWS_PLATFORM from .const import LABEL_CONFIG_HASH from .const import LABEL_CONTAINER_NUMBER from .const import LABEL_ONE_OFF @@ -769,9 +770,9 @@ def build(self, no_cache=False, pull=False, force_rm=False): build_opts = self.options.get('build', {}) path = build_opts.get('context') - # python2 os.path() doesn't support unicode, so we need to encode it to - # a byte string - if not six.PY3: + # python2 os.stat() doesn't support unicode on some UNIX, so we + # encode it to a bytestring to be safe + if not six.PY3 and not IS_WINDOWS_PLATFORM: path = path.encode('utf8') build_output = self.client.build(
docker-compose cannot handle unicode file paths on Windows I've just hit this error: ``` Traceback (most recent call last): File "<string>", line 3, in <module> File "compose\cli\main.py", line 61, in main File "compose\cli\main.py", line 113, in perform_command File "compose\cli\main.py", line 219, in build File "compose\project.py", line 300, in build File "compose\service.py", line 727, in build File "site-packages\docker\api\build.py", line 54, in build File "site-packages\docker\utils\utils.py", line 103, in tar File "tarfile.py", line 2006, in add File "tarfile.py", line 1878, in gettarinfo WindowsError: [Error 123] The filename, directory name, or volume label syntax is incorrect: 'C:\\[redacted]\\?.jpg' ``` The actual file name is `∆.jpg` (don't ask why). This looks like the same issue in http://stackoverflow.com/questions/11545185/handling-utf-filenames-in-python
I've installed the project using Python 3.5.1 and it seems to be working without any issues. I am getting the above error in **Mac**: docker-compose -f /Users/saurabhsarkar/Documents/ASPNetCore/Console☃App/docker-compose.debug.yml down /var/folders/sj/k148ghk548v7b8lqf_4yd41m0000gn/T/_MEIpZB1OL/compose/config/config.py:222: UnicodeWarning: Unicode equal comparison failed to convert both arguments to Unicode - interpreting them as being unequal Traceback (most recent call last): File "<string>", line 3, in <module> File "compose/cli/main.py", line 61, in main File "compose/cli/main.py", line 110, in perform_command File "compose/cli/command.py", line 35, in project_from_options File "compose/cli/command.py", line 98, in get_project File "compose/config/config.py", line 234, in find UnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position 49: ordinal not in range(128) docker-compose returned -1 Installing Python 3.5.1 is not helping. _[using **docker-compose version 1.8.0, build f3628c7**]_ On **Windows** when I am running the command on Powershell or cmd: docker-compose -f "C:\Users\saurabsa\documents\visual studio 2015\Projects\WebApplication☃3\src\WebApplication☃3\docker-compose.yml" down its throwing the error: RROR: .IOError: [Errno 22] invalid mode ('r') or filename: 'C:\Users\saurabsa\documents\visual studio 2015\Projects\WeApplication?3\src\WebApplication?3\docker-compose.yml' _[using **docker-compose version 1.8.0, build d988a55**]_ same error here when I try build a project who mount a volume with a file with special character. `WindowsError: [Error 123] The filename, directory name, or volume label syntax is incorrect: 'C:\\Users\\marci\\Documents\\GitHub\\Walletsaver\\bestarnew\\test\\data\\suggestion?067927c0-ad49-11e6-84cd-b797ecf45ea3.json` @BYK For a work around, how do you "install the project using Python 3.5.1"? @happiness801 it's been a long time but I think I did `pip install git+https://github.com/docker/compose.git`
2017-01-20T23:07:34Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "compose\cli\main.py", line 61, in main File "compose\cli\main.py", line 113, in perform_command File "compose\cli\main.py", line 219, in build File "compose\project.py", line 300, in build File "compose\service.py", line 727, in build File "site-packages\docker\api\build.py", line 54, in build File "site-packages\docker\utils\utils.py", line 103, in tar File "tarfile.py", line 2006, in add File "tarfile.py", line 1878, in gettarinfo WindowsError: [Error 123] The filename, directory name, or volume label syntax is incorrect: 'C:\\[redacted]\\?.jpg'
4,886
docker/compose
docker__compose-4383
e05a9f4e62d94322c3b504f134627840c1547d27
diff --git a/compose/cli/main.py b/compose/cli/main.py --- a/compose/cli/main.py +++ b/compose/cli/main.py @@ -14,6 +14,30 @@ from inspect import getdoc from operator import attrgetter + +# Attempt to detect https://github.com/docker/compose/issues/4344 +try: + # A regular import statement causes PyInstaller to freak out while + # trying to load pip. This way it is simply ignored. + pip = __import__('pip') + pip_packages = pip.get_installed_distributions() + if 'docker-py' in [pkg.project_name for pkg in pip_packages]: + from .colors import red + print( + red('ERROR:'), + "Dependency conflict: an older version of the 'docker-py' package " + "is polluting the namespace. " + "Run the following command to remedy the issue:\n" + "pip uninstall docker docker-py; pip install docker", + file=sys.stderr + ) + sys.exit(1) +except ImportError: + # pip is not available, which indicates it's probably the binary + # distribution of Compose which is not affected + pass + + from . import errors from . import signals from .. import __version__
`AttributeError: 'module' object has no attribute 'get_config_header'` after upgrading to 1.10 I had to manually remove `docker-py` and then reinstall `docker` and `docker-compose`: ``` ubuntu@ip-10-3-0-103:~$ docker-compose --version docker-compose version 1.9.0, build 2585387 ubuntu@ip-10-3-0-103:~$ docker-compose pull Pulling moo (ubuntu:latest)... latest: Pulling from library/ubuntu b3e1c725a85f: Pull complete 4daad8bdde31: Pull complete 63fe8c0068a8: Pull complete 4a70713c436f: Pull complete bd842a2105a8: Pull complete Digest: sha256:7a64bc9c8843b0a8c8b8a7e4715b7615e4e1b0d8ca3c7e7a76ec8250899c397a Status: Downloaded newer image for ubuntu:latest ubuntu@ip-10-3-0-103:~$ sudo pip install -U docker-compose The directory '/home/ubuntu/.cache/pip/http' or its parent directory is not owned by the current user and the cache has been disabled. Please check the permissions and owner of that directory. If executing pip with sudo, you may want sudo's -H flag. The directory '/home/ubuntu/.cache/pip' or its parent directory is not owned by the current user and caching wheels has been disabled. check the permissions and owner of that directory. If executing pip with sudo, you may want sudo's -H flag. /usr/local/lib/python2.7/dist-packages/pip/_vendor/requests/packages/urllib3/util/ssl_.py:318: SNIMissingWarning: An HTTPS request has been made, but the SNI (Subject Name Indication) extension to TLS is not available on this platform. This may cause the server to present an incorrect TLS certificate, which can cause validation failures. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.org/en/latest/security.html#snimissingwarning. SNIMissingWarning /usr/local/lib/python2.7/dist-packages/pip/_vendor/requests/packages/urllib3/util/ssl_.py:122: InsecurePlatformWarning: A true SSLContext object is not available. This prevents urllib3 from configuring SSL appropriately and may cause certain SSL connections to fail. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.org/en/latest/security.html#insecureplatformwarning. InsecurePlatformWarning Collecting docker-compose Downloading docker_compose-1.10.0-py2.py3-none-any.whl (81kB) 100% |████████████████████████████████| 81kB 6.0MB/s Requirement already up-to-date: PyYAML<4,>=3.10 in /usr/local/lib/python2.7/dist-packages (from docker-compose) Requirement already up-to-date: websocket-client<1.0,>=0.32.0 in /usr/local/lib/python2.7/dist-packages (from docker-compose) Requirement already up-to-date: backports.ssl-match-hostname>=3.5; python_version < "3.5" in /usr/local/lib/python2.7/dist-packages (from docker-compose) Collecting docker<3.0,>=2.0.1 (from docker-compose) Downloading docker-2.0.1-py2.py3-none-any.whl (95kB) 100% |████████████████████████████████| 102kB 9.5MB/s Requirement already up-to-date: cached-property<2,>=1.2.0 in /usr/local/lib/python2.7/dist-packages (from docker-compose) Requirement already up-to-date: dockerpty<0.5,>=0.4.1 in /usr/local/lib/python2.7/dist-packages (from docker-compose) Requirement already up-to-date: requests!=2.11.0,<2.12,>=2.6.1 in /usr/local/lib/python2.7/dist-packages (from docker-compose) Requirement already up-to-date: ipaddress>=1.0.16; python_version < "3.3" in /usr/local/lib/python2.7/dist-packages (from docker-compose) Requirement already up-to-date: colorama<0.4,>=0.3.7 in /usr/local/lib/python2.7/dist-packages (from docker-compose) Requirement already up-to-date: enum34<2,>=1.0.4; python_version < "3.4" in /usr/local/lib/python2.7/dist-packages (from docker-compose) Requirement already up-to-date: six<2,>=1.3.0 in /usr/local/lib/python2.7/dist-packages (from docker-compose) Requirement already up-to-date: texttable<0.9,>=0.8.1 in /usr/local/lib/python2.7/dist-packages (from docker-compose) Requirement already up-to-date: jsonschema<3,>=2.5.1 in /usr/local/lib/python2.7/dist-packages (from docker-compose) Requirement already up-to-date: docopt<0.7,>=0.6.1 in /usr/local/lib/python2.7/dist-packages (from docker-compose) Requirement already up-to-date: docker-pycreds>=0.2.1 in /usr/local/lib/python2.7/dist-packages (from docker<3.0,>=2.0.1->docker-compose) Requirement already up-to-date: functools32; python_version == "2.7" in /usr/local/lib/python2.7/dist-packages (from jsonschema<3,>=2.5.1->docker-compose) Installing collected packages: docker, docker-compose Found existing installation: docker-compose 1.9.0 Uninstalling docker-compose-1.9.0: Successfully uninstalled docker-compose-1.9.0 Successfully installed docker-2.0.1 docker-compose-1.10.0 /usr/local/lib/python2.7/dist-packages/pip/_vendor/requests/packages/urllib3/util/ssl_.py:122: InsecurePlatformWarning: A true SSLContext object is not available. This prevents urllib3 from configuring SSL appropriately and may cause certain SSL connections to fail. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.org/en/latest/security.html#insecureplatformwarning. InsecurePlatformWarning You are using pip version 8.1.2, however version 9.0.1 is available. You should consider upgrading via the 'pip install --upgrade pip' command. ubuntu@ip-10-3-0-103:~$ docker-compose pull Pulling moo (ubuntu:latest)... Traceback (most recent call last): File "/usr/local/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/usr/local/lib/python2.7/dist-packages/compose/cli/main.py", line 64, in main command() File "/usr/local/lib/python2.7/dist-packages/compose/cli/main.py", line 116, in perform_command handler(command, command_options) File "/usr/local/lib/python2.7/dist-packages/compose/cli/main.py", line 607, in pull ignore_pull_failures=options.get('--ignore-pull-failures') File "/usr/local/lib/python2.7/dist-packages/compose/project.py", line 453, in pull service.pull(ignore_pull_failures) File "/usr/local/lib/python2.7/dist-packages/compose/service.py", line 878, in pull output = self.client.pull(repo, tag=tag, stream=True) File "/usr/local/lib/python2.7/dist-packages/docker/api/image.py", line 333, in pull header = auth.get_config_header(self, registry) AttributeError: 'module' object has no attribute 'get_config_header' ubuntu@ip-10-3-0-103:~$ pip freeze | grep docker /usr/local/lib/python2.7/dist-packages/pip/_vendor/requests/packages/urllib3/util/ssl_.py:318: SNIMissingWarning: An HTTPS request has been made, but the SNI (Subject Name Indication) extension to TLS is not available on this platform. This may cause the server to present an incorrect TLS certificate, which can cause validation failures. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.org/en/latest/security.html#snimissingwarning. SNIMissingWarning /usr/local/lib/python2.7/dist-packages/pip/_vendor/requests/packages/urllib3/util/ssl_.py:122: InsecurePlatformWarning: A true SSLContext object is not available. This prevents urllib3 from configuring SSL appropriately and may cause certain SSL connections to fail. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.org/en/latest/security.html#insecureplatformwarning. InsecurePlatformWarning You are using pip version 8.1.2, however version 9.0.1 is available. You should consider upgrading via the 'pip install --upgrade pip' command. docker==2.0.1 docker-compose==1.10.0 docker-py==1.10.6 docker-pycreds==0.2.1 dockerpty==0.4.1 ubuntu@ip-10-3-0-103:~$ sudo pip uninstall docker-py The directory '/home/ubuntu/.cache/pip/http' or its parent directory is not owned by the current user and the cache has been disabled. Please check the permissions and owner of that directory. If executing pip with sudo, you may want sudo's -H flag. Uninstalling docker-py-1.10.6: /usr/local/lib/python2.7/dist-packages/docker/__init__.py /usr/local/lib/python2.7/dist-packages/docker/__init__.pyc /usr/local/lib/python2.7/dist-packages/docker/api/__init__.py /usr/local/lib/python2.7/dist-packages/docker/api/__init__.pyc /usr/local/lib/python2.7/dist-packages/docker/api/build.py /usr/local/lib/python2.7/dist-packages/docker/api/build.pyc /usr/local/lib/python2.7/dist-packages/docker/api/container.py /usr/local/lib/python2.7/dist-packages/docker/api/container.pyc /usr/local/lib/python2.7/dist-packages/docker/api/daemon.py /usr/local/lib/python2.7/dist-packages/docker/api/daemon.pyc /usr/local/lib/python2.7/dist-packages/docker/api/exec_api.py /usr/local/lib/python2.7/dist-packages/docker/api/exec_api.pyc /usr/local/lib/python2.7/dist-packages/docker/api/image.py /usr/local/lib/python2.7/dist-packages/docker/api/image.pyc /usr/local/lib/python2.7/dist-packages/docker/api/network.py /usr/local/lib/python2.7/dist-packages/docker/api/network.pyc /usr/local/lib/python2.7/dist-packages/docker/api/service.py /usr/local/lib/python2.7/dist-packages/docker/api/service.pyc /usr/local/lib/python2.7/dist-packages/docker/api/swarm.py /usr/local/lib/python2.7/dist-packages/docker/api/swarm.pyc /usr/local/lib/python2.7/dist-packages/docker/api/volume.py /usr/local/lib/python2.7/dist-packages/docker/api/volume.pyc /usr/local/lib/python2.7/dist-packages/docker/auth/__init__.py /usr/local/lib/python2.7/dist-packages/docker/auth/__init__.pyc /usr/local/lib/python2.7/dist-packages/docker/auth/auth.py /usr/local/lib/python2.7/dist-packages/docker/auth/auth.pyc /usr/local/lib/python2.7/dist-packages/docker/client.py /usr/local/lib/python2.7/dist-packages/docker/client.pyc /usr/local/lib/python2.7/dist-packages/docker/constants.py /usr/local/lib/python2.7/dist-packages/docker/constants.pyc /usr/local/lib/python2.7/dist-packages/docker/errors.py /usr/local/lib/python2.7/dist-packages/docker/errors.pyc /usr/local/lib/python2.7/dist-packages/docker/ssladapter/__init__.py /usr/local/lib/python2.7/dist-packages/docker/ssladapter/__init__.pyc /usr/local/lib/python2.7/dist-packages/docker/ssladapter/ssladapter.py /usr/local/lib/python2.7/dist-packages/docker/ssladapter/ssladapter.pyc /usr/local/lib/python2.7/dist-packages/docker/tls.py /usr/local/lib/python2.7/dist-packages/docker/tls.pyc /usr/local/lib/python2.7/dist-packages/docker/transport/__init__.py /usr/local/lib/python2.7/dist-packages/docker/transport/__init__.pyc /usr/local/lib/python2.7/dist-packages/docker/transport/npipeconn.py /usr/local/lib/python2.7/dist-packages/docker/transport/npipeconn.pyc /usr/local/lib/python2.7/dist-packages/docker/transport/npipesocket.py /usr/local/lib/python2.7/dist-packages/docker/transport/npipesocket.pyc /usr/local/lib/python2.7/dist-packages/docker/transport/unixconn.py /usr/local/lib/python2.7/dist-packages/docker/transport/unixconn.pyc /usr/local/lib/python2.7/dist-packages/docker/types/__init__.py /usr/local/lib/python2.7/dist-packages/docker/types/__init__.pyc /usr/local/lib/python2.7/dist-packages/docker/types/base.py /usr/local/lib/python2.7/dist-packages/docker/types/base.pyc /usr/local/lib/python2.7/dist-packages/docker/types/containers.py /usr/local/lib/python2.7/dist-packages/docker/types/containers.pyc /usr/local/lib/python2.7/dist-packages/docker/types/services.py /usr/local/lib/python2.7/dist-packages/docker/types/services.pyc /usr/local/lib/python2.7/dist-packages/docker/types/swarm.py /usr/local/lib/python2.7/dist-packages/docker/types/swarm.pyc /usr/local/lib/python2.7/dist-packages/docker/utils/__init__.py /usr/local/lib/python2.7/dist-packages/docker/utils/__init__.pyc /usr/local/lib/python2.7/dist-packages/docker/utils/decorators.py /usr/local/lib/python2.7/dist-packages/docker/utils/decorators.pyc /usr/local/lib/python2.7/dist-packages/docker/utils/ports/__init__.py /usr/local/lib/python2.7/dist-packages/docker/utils/ports/__init__.pyc /usr/local/lib/python2.7/dist-packages/docker/utils/ports/ports.py /usr/local/lib/python2.7/dist-packages/docker/utils/ports/ports.pyc /usr/local/lib/python2.7/dist-packages/docker/utils/socket.py /usr/local/lib/python2.7/dist-packages/docker/utils/socket.pyc /usr/local/lib/python2.7/dist-packages/docker/utils/types.py /usr/local/lib/python2.7/dist-packages/docker/utils/types.pyc /usr/local/lib/python2.7/dist-packages/docker/utils/utils.py /usr/local/lib/python2.7/dist-packages/docker/utils/utils.pyc /usr/local/lib/python2.7/dist-packages/docker/version.py /usr/local/lib/python2.7/dist-packages/docker/version.pyc /usr/local/lib/python2.7/dist-packages/docker_py-1.10.6.dist-info/DESCRIPTION.rst /usr/local/lib/python2.7/dist-packages/docker_py-1.10.6.dist-info/INSTALLER /usr/local/lib/python2.7/dist-packages/docker_py-1.10.6.dist-info/METADATA /usr/local/lib/python2.7/dist-packages/docker_py-1.10.6.dist-info/RECORD /usr/local/lib/python2.7/dist-packages/docker_py-1.10.6.dist-info/WHEEL /usr/local/lib/python2.7/dist-packages/docker_py-1.10.6.dist-info/metadata.json /usr/local/lib/python2.7/dist-packages/docker_py-1.10.6.dist-info/top_level.txt Proceed (y/n)? y Successfully uninstalled docker-py-1.10.6 The directory '/home/ubuntu/.cache/pip/http' or its parent directory is not owned by the current user and the cache has been disabled. Please check the permissions and owner of that directory. If executing pip with sudo, you may want sudo's -H flag. You are using pip version 8.1.2, however version 9.0.1 is available. You should consider upgrading via the 'pip install --upgrade pip' command. ubuntu@ip-10-3-0-103:~$ sudo pip install --force -U docker-compose docker The directory '/home/ubuntu/.cache/pip/http' or its parent directory is not owned by the current user and the cache has been disabled. Please check the permissions and owner of that directory. If executing pip with sudo, you may want sudo's -H flag. The directory '/home/ubuntu/.cache/pip' or its parent directory is not owned by the current user and caching wheels has been disabled. check the permissions and owner of that directory. If executing pip with sudo, you may want sudo's -H flag. Collecting docker-compose /usr/local/lib/python2.7/dist-packages/pip/_vendor/requests/packages/urllib3/util/ssl_.py:318: SNIMissingWarning: An HTTPS request has been made, but the SNI (Subject Name Indication) extension to TLS is not available on this platform. This may cause the server to present an incorrect TLS certificate, which can cause validation failures. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.org/en/latest/security.html#snimissingwarning. SNIMissingWarning /usr/local/lib/python2.7/dist-packages/pip/_vendor/requests/packages/urllib3/util/ssl_.py:122: InsecurePlatformWarning: A true SSLContext object is not available. This prevents urllib3 from configuring SSL appropriately and may cause certain SSL connections to fail. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.org/en/latest/security.html#insecureplatformwarning. InsecurePlatformWarning Downloading docker_compose-1.10.0-py2.py3-none-any.whl (81kB) 100% |████████████████████████████████| 81kB 6.9MB/s Collecting docker Downloading docker-2.0.1-py2.py3-none-any.whl (95kB) 100% |████████████████████████████████| 102kB 10.3MB/s Collecting PyYAML<4,>=3.10 (from docker-compose) Downloading PyYAML-3.12.tar.gz (253kB) 100% |████████████████████████████████| 256kB 4.6MB/s Collecting websocket-client<1.0,>=0.32.0 (from docker-compose) Downloading websocket_client-0.40.0.tar.gz (196kB) 100% |████████████████████████████████| 204kB 6.0MB/s Collecting backports.ssl-match-hostname>=3.5; python_version < "3.5" (from docker-compose) Downloading backports.ssl_match_hostname-3.5.0.1.tar.gz Collecting cached-property<2,>=1.2.0 (from docker-compose) Downloading cached_property-1.3.0-py2.py3-none-any.whl Collecting dockerpty<0.5,>=0.4.1 (from docker-compose) Downloading dockerpty-0.4.1.tar.gz Collecting requests!=2.11.0,<2.12,>=2.6.1 (from docker-compose) Downloading requests-2.11.1-py2.py3-none-any.whl (514kB) 100% |████████████████████████████████| 522kB 2.5MB/s Collecting ipaddress>=1.0.16; python_version < "3.3" (from docker-compose) Downloading ipaddress-1.0.18-py2-none-any.whl Collecting colorama<0.4,>=0.3.7 (from docker-compose) Downloading colorama-0.3.7-py2.py3-none-any.whl Collecting enum34<2,>=1.0.4; python_version < "3.4" (from docker-compose) Downloading enum34-1.1.6-py2-none-any.whl Collecting six<2,>=1.3.0 (from docker-compose) Downloading six-1.10.0-py2.py3-none-any.whl Collecting texttable<0.9,>=0.8.1 (from docker-compose) Downloading texttable-0.8.7.tar.gz Collecting jsonschema<3,>=2.5.1 (from docker-compose) Downloading jsonschema-2.5.1-py2.py3-none-any.whl Collecting docopt<0.7,>=0.6.1 (from docker-compose) Downloading docopt-0.6.2.tar.gz Collecting docker-pycreds>=0.2.1 (from docker) Downloading docker_pycreds-0.2.1-py2.py3-none-any.whl Collecting functools32; python_version == "2.7" (from jsonschema<3,>=2.5.1->docker-compose) Downloading functools32-3.2.3-2.zip Installing collected packages: PyYAML, six, backports.ssl-match-hostname, websocket-client, ipaddress, requests, docker-pycreds, docker, cached-property, dockerpty, colorama, enum34, texttable, functools32, jsonschema, docopt, docker-compose Found existing installation: PyYAML 3.12 Uninstalling PyYAML-3.12: Successfully uninstalled PyYAML-3.12 Running setup.py install for PyYAML ... done Found existing installation: six 1.10.0 Uninstalling six-1.10.0: Successfully uninstalled six-1.10.0 Found existing installation: backports.ssl-match-hostname 3.5.0.1 Uninstalling backports.ssl-match-hostname-3.5.0.1: Successfully uninstalled backports.ssl-match-hostname-3.5.0.1 Running setup.py install for backports.ssl-match-hostname ... done Found existing installation: websocket-client 0.40.0 Uninstalling websocket-client-0.40.0: Successfully uninstalled websocket-client-0.40.0 Running setup.py install for websocket-client ... done Found existing installation: ipaddress 1.0.18 Uninstalling ipaddress-1.0.18: Successfully uninstalled ipaddress-1.0.18 Found existing installation: requests 2.11.1 Uninstalling requests-2.11.1: Successfully uninstalled requests-2.11.1 Found existing installation: docker-pycreds 0.2.1 Uninstalling docker-pycreds-0.2.1: Successfully uninstalled docker-pycreds-0.2.1 Found existing installation: docker 2.0.1 Uninstalling docker-2.0.1: Successfully uninstalled docker-2.0.1 Found existing installation: cached-property 1.3.0 Uninstalling cached-property-1.3.0: Successfully uninstalled cached-property-1.3.0 Found existing installation: dockerpty 0.4.1 Uninstalling dockerpty-0.4.1: Successfully uninstalled dockerpty-0.4.1 Running setup.py install for dockerpty ... done Found existing installation: colorama 0.3.7 Uninstalling colorama-0.3.7: Successfully uninstalled colorama-0.3.7 Found existing installation: enum34 1.1.6 Uninstalling enum34-1.1.6: Successfully uninstalled enum34-1.1.6 Found existing installation: texttable 0.8.7 Uninstalling texttable-0.8.7: Successfully uninstalled texttable-0.8.7 Running setup.py install for texttable ... done Found existing installation: functools32 3.2.3.post2 Uninstalling functools32-3.2.3.post2: Successfully uninstalled functools32-3.2.3.post2 Running setup.py install for functools32 ... done Found existing installation: jsonschema 2.5.1 Uninstalling jsonschema-2.5.1: Successfully uninstalled jsonschema-2.5.1 Found existing installation: docopt 0.6.2 Uninstalling docopt-0.6.2: Successfully uninstalled docopt-0.6.2 Running setup.py install for docopt ... done Found existing installation: docker-compose 1.10.0 Uninstalling docker-compose-1.10.0: Successfully uninstalled docker-compose-1.10.0 Successfully installed PyYAML-3.12 backports.ssl-match-hostname-3.5.0.1 cached-property-1.3.0 colorama-0.3.7 docker-2.0.1 docker-compose-1.10.0 docker-pycreds-0.2.1 dockerpty-0.4.1 docopt-0.6.2 enum34-1.1.6 functools32-3.2.3.post2 ipaddress-1.0.18 jsonschema-2.5.1 requests-2.11.1 six-1.10.0 texttable-0.8.7 websocket-client-0.40.0 You are using pip version 8.1.2, however version 9.0.1 is available. You should consider upgrading via the 'pip install --upgrade pip' command. ubuntu@ip-10-3-0-103:~$ docker-compose pull Pulling moo (ubuntu:latest)... latest: Pulling from library/ubuntu Digest: sha256:7a64bc9c8843b0a8c8b8a7e4715b7615e4e1b0d8ca3c7e7a76ec8250899c397a Status: Image is up to date for ubuntu:latest ubuntu@ip-10-3-0-103:~$ ```
I should note that this is on Ubuntu 14.04 You'll have to uninstall `docker-py` to avoid name conflict issues. For people in need of a solution: ``` pip uninstall docker docker-py ; pip install docker==2.0.1 ``` We should have a fix coming in the next few days.
2017-01-24T23:05:34Z
[]
[]
Traceback (most recent call last): File "/usr/local/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/usr/local/lib/python2.7/dist-packages/compose/cli/main.py", line 64, in main command() File "/usr/local/lib/python2.7/dist-packages/compose/cli/main.py", line 116, in perform_command handler(command, command_options) File "/usr/local/lib/python2.7/dist-packages/compose/cli/main.py", line 607, in pull ignore_pull_failures=options.get('--ignore-pull-failures') File "/usr/local/lib/python2.7/dist-packages/compose/project.py", line 453, in pull service.pull(ignore_pull_failures) File "/usr/local/lib/python2.7/dist-packages/compose/service.py", line 878, in pull output = self.client.pull(repo, tag=tag, stream=True) File "/usr/local/lib/python2.7/dist-packages/docker/api/image.py", line 333, in pull header = auth.get_config_header(self, registry) AttributeError: 'module' object has no attribute 'get_config_header'
4,887
docker/compose
docker__compose-442
6dab8c1b8932ec1a9de3c2fb22d97bc0aae0d9c4
diff --git a/fig/cli/formatter.py b/fig/cli/formatter.py --- a/fig/cli/formatter.py +++ b/fig/cli/formatter.py @@ -9,7 +9,7 @@ def get_tty_width(): if len(tty_size) != 2: return 80 _, width = tty_size - return width + return int(width) class Formatter(object): diff --git a/fig/service.py b/fig/service.py --- a/fig/service.py +++ b/fig/service.py @@ -1,5 +1,6 @@ from __future__ import unicode_literals from __future__ import absolute_import +from collections import namedtuple from .packages.docker.errors import APIError import logging import re @@ -39,6 +40,9 @@ class ConfigError(ValueError): pass +VolumeSpec = namedtuple('VolumeSpec', 'external internal mode') + + class Service(object): def __init__(self, name, client=None, project='default', links=None, volumes_from=None, **options): if not re.match('^%s+$' % VALID_NAME_CHARS, name): @@ -214,37 +218,22 @@ def start_container_if_stopped(self, container, **options): return self.start_container(container, **options) def start_container(self, container=None, intermediate_container=None, **override_options): - if container is None: - container = self.create_container(**override_options) - - options = self.options.copy() - options.update(override_options) - - port_bindings = {} + container = container or self.create_container(**override_options) + options = dict(self.options, **override_options) + ports = dict(split_port(port) for port in options.get('ports') or []) - if options.get('ports', None) is not None: - for port in options['ports']: - internal_port, external_port = split_port(port) - port_bindings[internal_port] = external_port - - volume_bindings = {} - - if options.get('volumes', None) is not None: - for volume in options['volumes']: - if ':' in volume: - external_dir, internal_dir = volume.split(':') - volume_bindings[os.path.abspath(external_dir)] = { - 'bind': internal_dir, - 'ro': False, - } + volume_bindings = dict( + build_volume_binding(parse_volume_spec(volume)) + for volume in options.get('volumes') or [] + if ':' in volume) privileged = options.get('privileged', False) net = options.get('net', 'bridge') dns = options.get('dns', None) container.start( - links=self._get_links(link_to_self=override_options.get('one_off', False)), - port_bindings=port_bindings, + links=self._get_links(link_to_self=options.get('one_off', False)), + port_bindings=ports, binds=volume_bindings, volumes_from=self._get_volumes_from(intermediate_container), privileged=privileged, @@ -256,7 +245,7 @@ def start_container(self, container=None, intermediate_container=None, **overrid def start_or_create_containers(self): containers = self.containers(stopped=True) - if len(containers) == 0: + if not containers: log.info("Creating %s..." % self.next_container_name()) new_container = self.create_container() return [self.start_container(new_container)] @@ -338,7 +327,9 @@ def _get_container_create_options(self, override_options, one_off=False): container_options['ports'] = ports if 'volumes' in container_options: - container_options['volumes'] = dict((split_volume(v)[1], {}) for v in container_options['volumes']) + container_options['volumes'] = dict( + (parse_volume_spec(v).internal, {}) + for v in container_options['volumes']) if 'environment' in container_options: if isinstance(container_options['environment'], list): @@ -433,32 +424,47 @@ def get_container_name(container): return name[1:] -def split_volume(v): - """ - If v is of the format EXTERNAL:INTERNAL, returns (EXTERNAL, INTERNAL). - If v is of the format INTERNAL, returns (None, INTERNAL). - """ - if ':' in v: - return v.split(':', 1) - else: - return (None, v) +def parse_volume_spec(volume_config): + parts = volume_config.split(':') + if len(parts) > 3: + raise ConfigError("Volume %s has incorrect format, should be " + "external:internal[:mode]" % volume_config) + + if len(parts) == 1: + return VolumeSpec(None, parts[0], 'rw') + + if len(parts) == 2: + parts.append('rw') + + external, internal, mode = parts + if mode not in ('rw', 'ro'): + raise ConfigError("Volume %s has invalid mode (%s), should be " + "one of: rw, ro." % (volume_config, mode)) + + return VolumeSpec(external, internal, mode) + + +def build_volume_binding(volume_spec): + internal = {'bind': volume_spec.internal, 'ro': volume_spec.mode == 'ro'} + external = os.path.expanduser(volume_spec.external) + return os.path.abspath(os.path.expandvars(external)), internal def split_port(port): - port = str(port) - external_ip = None - if ':' in port: - external_port, internal_port = port.rsplit(':', 1) - if ':' in external_port: - external_ip, external_port = external_port.split(':', 1) - else: - external_port, internal_port = (None, port) - if external_ip: - if external_port: - external_port = (external_ip, external_port) - else: - external_port = (external_ip,) - return internal_port, external_port + parts = str(port).split(':') + if not 1 <= len(parts) <= 3: + raise ConfigError('Invalid port "%s", should be ' + '[[remote_ip:]remote_port:]port[/protocol]' % port) + + if len(parts) == 1: + internal_port, = parts + return internal_port, None + if len(parts) == 2: + external_port, internal_port = parts + return internal_port, external_port + + external_ip, external_port, internal_port = parts + return internal_port, (external_ip, external_port or None) def split_env(env):
Error when specifying permission in volume ``` root@dev-stats-web:~# fig up Recreating root_statssymfony_1... Traceback (most recent call last): File "/usr/local/bin/fig", line 9, in <module> load_entry_point('fig==0.4.2', 'console_scripts', 'fig')() File "/usr/local/lib/python2.7/dist-packages/fig/cli/main.py", line 39, in main command.sys_dispatch() File "/usr/local/lib/python2.7/dist-packages/fig/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/local/lib/python2.7/dist-packages/fig/cli/command.py", line 31, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/local/lib/python2.7/dist-packages/fig/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/local/lib/python2.7/dist-packages/fig/cli/command.py", line 50, in perform_command return super(Command, self).perform_command(options, *args, **kwargs) File "/usr/local/lib/python2.7/dist-packages/fig/cli/docopt_command.py", line 27, in perform_command handler(command_options) File "/usr/local/lib/python2.7/dist-packages/fig/cli/main.py", line 306, in up to_attach = self.project.up(service_names=options['SERVICE']) File "/usr/local/lib/python2.7/dist-packages/fig/project.py", line 131, in up for (_, new) in service.recreate_containers(): File "/usr/local/lib/python2.7/dist-packages/fig/service.py", line 173, in recreate_containers tuples.append(self.recreate_container(c, **override_options)) File "/usr/local/lib/python2.7/dist-packages/fig/service.py", line 195, in recreate_container self.start_container(new_container, volumes_from=intermediate_container.id) File "/usr/local/lib/python2.7/dist-packages/fig/service.py", line 225, in start_container external_dir, internal_dir = volume.split(':') ValueError: too many values to unpack ``` Took me a while to figure out I'd added `:rw` to the fig.yml ``` volumes: - stats-symfony/logs:/var/log/nginx:rw ``` Would be good idea to add a warning, or support it.
Added to #129. How do we specify :ro to the fig.yml for volumes How do we specify :ro to the fig.yml for volumes http://stackoverflow.com/questions/25385164/how-to-specify-write-access-to-volume-with-fig +1
2014-08-24T23:05:39Z
[]
[]
Traceback (most recent call last): File "/usr/local/bin/fig", line 9, in <module> load_entry_point('fig==0.4.2', 'console_scripts', 'fig')() File "/usr/local/lib/python2.7/dist-packages/fig/cli/main.py", line 39, in main command.sys_dispatch() File "/usr/local/lib/python2.7/dist-packages/fig/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/usr/local/lib/python2.7/dist-packages/fig/cli/command.py", line 31, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/usr/local/lib/python2.7/dist-packages/fig/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/usr/local/lib/python2.7/dist-packages/fig/cli/command.py", line 50, in perform_command return super(Command, self).perform_command(options, *args, **kwargs) File "/usr/local/lib/python2.7/dist-packages/fig/cli/docopt_command.py", line 27, in perform_command handler(command_options) File "/usr/local/lib/python2.7/dist-packages/fig/cli/main.py", line 306, in up to_attach = self.project.up(service_names=options['SERVICE']) File "/usr/local/lib/python2.7/dist-packages/fig/project.py", line 131, in up for (_, new) in service.recreate_containers(): File "/usr/local/lib/python2.7/dist-packages/fig/service.py", line 173, in recreate_containers tuples.append(self.recreate_container(c, **override_options)) File "/usr/local/lib/python2.7/dist-packages/fig/service.py", line 195, in recreate_container self.start_container(new_container, volumes_from=intermediate_container.id) File "/usr/local/lib/python2.7/dist-packages/fig/service.py", line 225, in start_container external_dir, internal_dir = volume.split(':') ValueError: too many values to unpack
4,892
docker/compose
docker__compose-4469
1fe24437350d11a9e1b6483839cb72a6f8f2fbf2
diff --git a/compose/config/config.py b/compose/config/config.py --- a/compose/config/config.py +++ b/compose/config/config.py @@ -763,6 +763,11 @@ def finalize_service(service_config, service_names, version, environment): if 'restart' in service_dict: service_dict['restart'] = parse_restart_spec(service_dict['restart']) + if 'secrets' in service_dict: + service_dict['secrets'] = [ + types.ServiceSecret.parse(s) for s in service_dict['secrets'] + ] + normalize_build(service_dict, service_config.working_dir, environment) service_dict['name'] = service_config.name diff --git a/compose/config/types.py b/compose/config/types.py --- a/compose/config/types.py +++ b/compose/config/types.py @@ -253,3 +253,11 @@ def parse(cls, spec): @property def merge_field(self): return self.source + + def repr(self): + return dict( + source=self.source, + target=self.target, + uid=self.uid, + gid=self.gid, + mode=self.mode) diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -1,10 +1,10 @@ #!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import +from __future__ import print_function from __future__ import unicode_literals import codecs -import logging import os import re import sys @@ -64,11 +64,9 @@ def find_version(*file_paths): for key, value in extras_require.items(): if key.startswith(':') and pkg_resources.evaluate_marker(key[1:]): install_requires.extend(value) -except Exception: - logging.getLogger(__name__).exception( - 'Failed to compute platform dependencies. All dependencies will be ' - 'installed as a result.' - ) +except Exception as e: + print("Failed to compute platform dependencies: {}. ".format(e) + + "All dependencies will be installed as a result.", file=sys.stderr) for key, value in extras_require.items(): if key.startswith(':'): install_requires.extend(value)
docker-compose fails if using secrets file ## Environment ``` $ docker --version Docker version 1.13.1, build 092cba3 $ docker-compose --version docker-compose version 1.11.1, build 7c5d5e4 ``` ## Steps to reproduce ``` $ ls docker-compose.yml my_file_secret.txt $ cat my_file_secret.txt This is a file secret. $ cat docker-compose.yml version: '3.1' services: a: image: tutum/hello-world b: image: tutum/hello-world secrets: - my_file_secret secrets: my_file_secret: file: my_file_secret.txt ``` ``` $ docker-compose up Traceback (most recent call last): File "docker-compose", line 3, in <module> File "compose/cli/main.py", line 88, in main File "compose/cli/main.py", line 137, in perform_command File "compose/cli/command.py", line 36, in project_from_options File "compose/cli/command.py", line 115, in get_project File "compose/project.py", line 110, in from_config File "compose/project.py", line 566, in get_secrets AttributeError: 'str' object has no attribute 'source' Failed to execute script docker-compose ```
@arielelkin Does it work if you rewrite `b` as below? ```yaml b: image: tutum/hello-world secrets: - source: my_file_secret ``` cc @dnephin
2017-02-10T22:13:04Z
[]
[]
Traceback (most recent call last): File "docker-compose", line 3, in <module> File "compose/cli/main.py", line 88, in main File "compose/cli/main.py", line 137, in perform_command File "compose/cli/command.py", line 36, in project_from_options File "compose/cli/command.py", line 115, in get_project File "compose/project.py", line 110, in from_config File "compose/project.py", line 566, in get_secrets AttributeError: 'str' object has no attribute 'source'
4,893
docker/compose
docker__compose-4480
bb5d7b2433eec26c1bbd0880648cedfb903d45ad
diff --git a/compose/config/serialize.py b/compose/config/serialize.py --- a/compose/config/serialize.py +++ b/compose/config/serialize.py @@ -102,4 +102,7 @@ def denormalize_service_dict(service_dict, version): service_dict['healthcheck']['timeout'] ) + if 'secrets' in service_dict: + service_dict['secrets'] = map(lambda s: s.repr(), service_dict['secrets']) + return service_dict diff --git a/compose/config/types.py b/compose/config/types.py --- a/compose/config/types.py +++ b/compose/config/types.py @@ -256,8 +256,5 @@ def merge_field(self): def repr(self): return dict( - source=self.source, - target=self.target, - uid=self.uid, - gid=self.gid, - mode=self.mode) + [(k, v) for k, v in self._asdict().items() if v is not None] + )
Secrets combined with override file results in stacktrace I have a `docker-compose.yml` and `docker-compose.override.yml` both using the `3.1` compose format. Adding secrets to the first yml seems to break the yaml interpolation so any attempt to bring up the stack or check the config results in a stack trace ``` $ docker-compose config Traceback (most recent call last): File "/usr/local/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/usr/local/lib/python2.7/site-packages/compose/cli/main.py", line 88, in main command() File "/usr/local/lib/python2.7/site-packages/compose/cli/main.py", line 134, in perform_command handler(command, options, command_options) File "/usr/local/lib/python2.7/site-packages/compose/cli/main.py", line 330, in config compose_config = get_config_from_options(self.project_dir, config_options) File "/usr/local/lib/python2.7/site-packages/compose/cli/command.py", line 46, in get_config_from_options config.find(base_dir, config_path, environment) File "/usr/local/lib/python2.7/site-packages/compose/config/config.py", line 333, in load service_dicts = load_services(config_details, main_file) File "/usr/local/lib/python2.7/site-packages/compose/config/config.py", line 455, in load_services service_config = merge_services(service_config, next_config) File "/usr/local/lib/python2.7/site-packages/compose/config/config.py", line 446, in merge_services for name in all_service_names File "/usr/local/lib/python2.7/site-packages/compose/config/config.py", line 446, in <dictcomp> for name in all_service_names File "/usr/local/lib/python2.7/site-packages/compose/config/config.py", line 797, in merge_service_dicts_from_files new_service = merge_service_dicts(base, override, version) File "/usr/local/lib/python2.7/site-packages/compose/config/config.py", line 856, in merge_service_dicts md.merge_sequence('secrets', types.ServiceSecret.parse) File "/usr/local/lib/python2.7/site-packages/compose/config/config.py", line 839, in merge_sequence self[field] = [item.repr() for item in sorted(merged.values())] AttributeError: 'ServiceSecret' object has no attribute 'repr' $ ``` I am running `docker-compose version 1.11.1, build 7c5d5e4`
cc @dnephin
2017-02-14T00:07:11Z
[]
[]
Traceback (most recent call last): File "/usr/local/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/usr/local/lib/python2.7/site-packages/compose/cli/main.py", line 88, in main command() File "/usr/local/lib/python2.7/site-packages/compose/cli/main.py", line 134, in perform_command handler(command, options, command_options) File "/usr/local/lib/python2.7/site-packages/compose/cli/main.py", line 330, in config compose_config = get_config_from_options(self.project_dir, config_options) File "/usr/local/lib/python2.7/site-packages/compose/cli/command.py", line 46, in get_config_from_options config.find(base_dir, config_path, environment) File "/usr/local/lib/python2.7/site-packages/compose/config/config.py", line 333, in load service_dicts = load_services(config_details, main_file) File "/usr/local/lib/python2.7/site-packages/compose/config/config.py", line 455, in load_services service_config = merge_services(service_config, next_config) File "/usr/local/lib/python2.7/site-packages/compose/config/config.py", line 446, in merge_services for name in all_service_names File "/usr/local/lib/python2.7/site-packages/compose/config/config.py", line 446, in <dictcomp> for name in all_service_names File "/usr/local/lib/python2.7/site-packages/compose/config/config.py", line 797, in merge_service_dicts_from_files new_service = merge_service_dicts(base, override, version) File "/usr/local/lib/python2.7/site-packages/compose/config/config.py", line 856, in merge_service_dicts md.merge_sequence('secrets', types.ServiceSecret.parse) File "/usr/local/lib/python2.7/site-packages/compose/config/config.py", line 839, in merge_sequence self[field] = [item.repr() for item in sorted(merged.values())] AttributeError: 'ServiceSecret' object has no attribute 'repr'
4,894
docker/compose
docker__compose-4484
40c26ca6766cc78ac0f4d24f3faa3c8f9e89a7ba
diff --git a/compose/cli/__init__.py b/compose/cli/__init__.py --- a/compose/cli/__init__.py +++ b/compose/cli/__init__.py @@ -0,0 +1,37 @@ +from __future__ import absolute_import +from __future__ import print_function +from __future__ import unicode_literals + +import subprocess +import sys + +# Attempt to detect https://github.com/docker/compose/issues/4344 +try: + # We don't try importing pip because it messes with package imports + # on some Linux distros (Ubuntu, Fedora) + # https://github.com/docker/compose/issues/4425 + # https://github.com/docker/compose/issues/4481 + # https://github.com/pypa/pip/blob/master/pip/_vendor/__init__.py + s_cmd = subprocess.Popen( + ['pip', 'freeze'], stderr=subprocess.PIPE, stdout=subprocess.PIPE + ) + packages = s_cmd.communicate()[0].splitlines() + dockerpy_installed = len( + list(filter(lambda p: p.startswith(b'docker-py=='), packages)) + ) > 0 + if dockerpy_installed: + from .colors import red + print( + red('ERROR:'), + "Dependency conflict: an older version of the 'docker-py' package " + "is polluting the namespace. " + "Run the following command to remedy the issue:\n" + "pip uninstall docker docker-py; pip install docker", + file=sys.stderr + ) + sys.exit(1) + +except OSError: + # pip command is not available, which indicates it's probably the binary + # distribution of Compose which is not affected + pass diff --git a/compose/cli/main.py b/compose/cli/main.py --- a/compose/cli/main.py +++ b/compose/cli/main.py @@ -14,30 +14,6 @@ from inspect import getdoc from operator import attrgetter - -# Attempt to detect https://github.com/docker/compose/issues/4344 -try: - # A regular import statement causes PyInstaller to freak out while - # trying to load pip. This way it is simply ignored. - pip = __import__('pip') - pip_packages = pip.get_installed_distributions() - if 'docker-py' in [pkg.project_name for pkg in pip_packages]: - from .colors import red - print( - red('ERROR:'), - "Dependency conflict: an older version of the 'docker-py' package " - "is polluting the namespace. " - "Run the following command to remedy the issue:\n" - "pip uninstall docker docker-py; pip install docker", - file=sys.stderr - ) - sys.exit(1) -except ImportError: - # pip is not available, which indicates it's probably the binary - # distribution of Compose which is not affected - pass - - from . import errors from . import signals from .. import __version__
Exception on 404 when using the HTTP interface directly Hi guys. I've been struggling with a weird issue: ``` compose.cli.verbose_proxy.proxy_callable: docker inspect_network <- (u'mytest_default') Traceback (most recent call last): File "/usr/local/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/usr/local/lib/python2.7/dist-packages/compose/cli/main.py", line 88, in main command() File "/usr/local/lib/python2.7/dist-packages/compose/cli/main.py", line 140, in perform_command handler(command, command_options) File "/usr/local/lib/python2.7/dist-packages/compose/cli/main.py", line 900, in up remove_orphans=remove_orphans) File "/usr/local/lib/python2.7/dist-packages/compose/project.py", line 387, in up self.initialize() File "/usr/local/lib/python2.7/dist-packages/compose/project.py", line 431, in initialize self.networks.initialize() File "/usr/local/lib/python2.7/dist-packages/compose/network.py", line 203, in initialize network.ensure() File "/usr/local/lib/python2.7/dist-packages/compose/network.py", line 59, in ensure data = self.inspect() File "/usr/local/lib/python2.7/dist-packages/compose/network.py", line 91, in inspect return self.client.inspect_network(self.full_name) File "/usr/local/lib/python2.7/dist-packages/compose/cli/verbose_proxy.py", line 55, in proxy_callable result = getattr(self.obj, call_name)(*args, **kwargs) File "/usr/local/lib/python2.7/dist-packages/docker/utils/decorators.py", line 35, in wrapper return f(self, *args, **kwargs) File "/usr/local/lib/python2.7/dist-packages/docker/api/network.py", line 158, in inspect_network return self._result(res, json=True) File "/usr/local/lib/python2.7/dist-packages/docker/api/client.py", line 216, in _result self._raise_for_status(response) File "/usr/local/lib/python2.7/dist-packages/docker/api/client.py", line 210, in _raise_for_status response.raise_for_status() File "/usr/share/python-wheels/requests-2.9.1-py2.py3-none-any.whl/requests/models.py", line 840, in raise_for_status requests.exceptions.HTTPError: 404 Client Error: Not Found for url: http://docker:2375/v1.25/networks/baseagent_default ``` From the looks of it, the client fails to recover from a 404 (an expected one, the network `mytest_default` doesn't exist yet. I have been able to create a more self-contained test to reproduce the error (spoiler alert it relies on dind, but that should be irrelevant): https://github.com/ColinHebert/compose-404 It may sound a bit convoluted but here is what's happening: - docker-compose.yaml starts two containers, the first one is dind (the one we're going to test against through http). It uses privileges to run docker properly. - The second container runs a docker container "client" which is an ubuntu box with docker-compose installed. The client container is linked to dind container and has the env variable `DOCKER_HOST=docker:2375` (meaning that docker commands run inside this container are going straight to dind - When starting the client container, we attempt to run docker-compose which ends up failing on the 404. I'm not 100% across the reason why the HTTP 404 bubbles up in my case. When running against a socket (albeit a different environment), the 404 gets caught properly ending with: ```pip._vendor.requests.packages.urllib3.connectionpool._make_request: "GET /v1.25/networks/mytest_default HTTP/1.1" 404 50```
After more testing, it seems that the problem was introduce between `1.10.0` and `1.10.1`. (Tested changing `RUN pip install docker-compose` into `RUN pip install docker-compose==1.10.0` in the Dockerfile) which is when docker-py 2.0.2 was introduced... The search continues... EDIT: checking the dependencies when it works and when it doesn't docker-py's version doesn't have any effect. Actually the version of the dependencies are the same between 1.10.0 and 1.10.1 (due to the use of ranges). That was a red herring. So something else in https://github.com/docker/compose/compare/1.10.0...1.10.1 is the cause of this :/ @ColinHebert If you use 1.10.0 in combination with `docker==2.0.2`, you don't get the issue? Indeed, I thought it was docker-py 2.0.2 as well, but it clearly isn't the case. I've been staring at the diff between 1.10.0 and 1.10.1 in github and nothing in there seems to be relevant to the problem Here's a pip freeze with 1.10.1 (failing with 404): ``` client_1 | backports.ssl-match-hostname==3.5.0.1 client_1 | cached-property==1.3.0 client_1 | colorama==0.3.7 client_1 | docker==2.0.2 client_1 | docker-compose==1.10.1 client_1 | docker-pycreds==0.2.1 client_1 | dockerpty==0.4.1 client_1 | docopt==0.6.2 client_1 | enum34==1.1.6 client_1 | functools32==3.2.3.post2 client_1 | ipaddress==1.0.18 client_1 | jsonschema==2.6.0 client_1 | PyYAML==3.12 client_1 | requests==2.11.1 client_1 | six==1.10.0 client_1 | texttable==0.8.7 client_1 | websocket-client==0.40.0 ``` And with 1.10.0 (successfully running): ``` client_1 | backports.ssl-match-hostname==3.5.0.1 client_1 | cached-property==1.3.0 client_1 | colorama==0.3.7 client_1 | docker==2.0.2 client_1 | docker-compose==1.10.0 client_1 | docker-pycreds==0.2.1 client_1 | dockerpty==0.4.1 client_1 | docopt==0.6.2 client_1 | enum34==1.1.6 client_1 | functools32==3.2.3.post2 client_1 | ipaddress==1.0.18 client_1 | jsonschema==2.6.0 client_1 | PyYAML==3.12 client_1 | requests==2.11.1 client_1 | six==1.10.0 client_1 | texttable==0.8.7 client_1 | websocket-client==0.40.0 ``` Okay, after attempts to reintroduce the problem by adding changes one by one, it seems that the cause is this: ``` # Attempt to detect https://github.com/docker/compose/issues/4344 try: # A regular import statement causes PyInstaller to freak out while # trying to load pip. This way it is simply ignored. pip = __import__('pip') pip_packages = pip.get_installed_distributions() if 'docker-py' in [pkg.project_name for pkg in pip_packages]: from .colors import red print( red('ERROR:'), "Dependency conflict: an older version of the 'docker-py' package " "is polluting the namespace. " "Run the following command to remedy the issue:\n" "pip uninstall docker docker-py; pip install docker", file=sys.stderr ) sys.exit(1) except ImportError: # pip is not available, which indicates it's probably the binary # distribution of Compose which is not affected pass ``` EDIT: Attempted to add this change and only that change. Indeed this is the bit of code causing trouble. I narrowed it down to the `pip = __import__('pip')` alone. This is outside of my domain of expertise now, but this line alone is causing some trouble :/ I found a temporary "fix" submitted to your reproduction case: ColinHebert/compose-404#1 It looks like the pip import confuses things, and this issue looks like another symptom of the one described in #4425 If you have a chance, please test the potential fix in #4484 The fix doesn't seem to work. I tried to reduce it to just running ``` original_sys_path = sys.path[:] print(sys.path[:]) pip = __import__('pip') print(sys.path[:]) sys.path[:] = original_sys_path print(sys.path[:]) ``` And it prints out the following: ``` ['/usr/local/bin', '/usr/lib/python2.7', '/usr/lib/python2.7/plat-x86_64-linux-gnu', '/usr/lib/python2.7/lib-tk', '/usr/lib/python2.7/lib-old', '/usr/lib/python2.7/lib-dynload', '/usr/local/lib/python2.7/dist-packages', '/usr/lib/python2.7/dist-packages'] ['/usr/share/python-wheels/ipaddress-0.0.0-py2.py3-none-any.whl', '/usr/share/python-wheels/pyparsing-2.0.3-py2.py3-none-any.whl', '/usr/share/python-wheels/packaging-16.6-py2.py3-none-any.whl', '/usr/share/python-wheels/chardet-2.3.0-py2.py3-none-any.whl', '/usr/share/python-wheels/retrying-1.3.3-py2.py3-none-any.whl', '/usr/share/python-wheels/urllib3-1.13.1-py2.py3-none-any.whl', '/usr/share/python-wheels/setuptools-20.7.0-py2.py3-none-any.whl', '/usr/share/python-wheels/wheel-0.29.0-py2.py3-none-any.whl', '/usr/share/python-wheels/requests-2.9.1-py2.py3-none-any.whl', '/usr/share/python-wheels/CacheControl-0.11.5-py2.py3-none-any.whl', '/usr/share/python-wheels/progress-1.2-py2.py3-none-any.whl', '/usr/share/python-wheels/colorama-0.3.7-py2.py3-none-any.whl', '/usr/share/python-wheels/distlib-0.2.2-py2.py3-none-any.whl', '/usr/share/python-wheels/six-1.10.0-py2.py3-none-any.whl', '/usr/share/python-wheels/pkg_resources-0.0.0-py2.py3-none-any.whl', '/usr/share/python-wheels/html5lib-0.999-py2.py3-none-any.whl', '/usr/share/python-wheels/pip-8.1.1-py2.py3-none-any.whl', '/usr/share/python-wheels/lockfile-0.12.2-py2.py3-none-any.whl', '/usr/local/bin', '/usr/lib/python2.7', '/usr/lib/python2.7/plat-x86_64-linux-gnu', '/usr/lib/python2.7/lib-tk', '/usr/lib/python2.7/lib-old', '/usr/lib/python2.7/lib-dynload', '/usr/local/lib/python2.7/dist-packages', '/usr/lib/python2.7/dist-packages'] ['/usr/local/bin', '/usr/lib/python2.7', '/usr/lib/python2.7/plat-x86_64-linux-gnu', '/usr/lib/python2.7/lib-tk', '/usr/lib/python2.7/lib-old', '/usr/lib/python2.7/lib-dynload', '/usr/local/lib/python2.7/dist-packages', '/usr/lib/python2.7/dist-packages'] ``` I think we're on the right tracks, but there might be another side effect.
2017-02-14T23:51:05Z
[]
[]
Traceback (most recent call last): File "/usr/local/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/usr/local/lib/python2.7/dist-packages/compose/cli/main.py", line 88, in main command() File "/usr/local/lib/python2.7/dist-packages/compose/cli/main.py", line 140, in perform_command handler(command, command_options) File "/usr/local/lib/python2.7/dist-packages/compose/cli/main.py", line 900, in up remove_orphans=remove_orphans) File "/usr/local/lib/python2.7/dist-packages/compose/project.py", line 387, in up self.initialize() File "/usr/local/lib/python2.7/dist-packages/compose/project.py", line 431, in initialize self.networks.initialize() File "/usr/local/lib/python2.7/dist-packages/compose/network.py", line 203, in initialize network.ensure() File "/usr/local/lib/python2.7/dist-packages/compose/network.py", line 59, in ensure data = self.inspect() File "/usr/local/lib/python2.7/dist-packages/compose/network.py", line 91, in inspect return self.client.inspect_network(self.full_name) File "/usr/local/lib/python2.7/dist-packages/compose/cli/verbose_proxy.py", line 55, in proxy_callable result = getattr(self.obj, call_name)(*args, **kwargs) File "/usr/local/lib/python2.7/dist-packages/docker/utils/decorators.py", line 35, in wrapper return f(self, *args, **kwargs) File "/usr/local/lib/python2.7/dist-packages/docker/api/network.py", line 158, in inspect_network return self._result(res, json=True) File "/usr/local/lib/python2.7/dist-packages/docker/api/client.py", line 216, in _result self._raise_for_status(response) File "/usr/local/lib/python2.7/dist-packages/docker/api/client.py", line 210, in _raise_for_status response.raise_for_status() File "/usr/share/python-wheels/requests-2.9.1-py2.py3-none-any.whl/requests/models.py", line 840, in raise_for_status requests.exceptions.HTTPError: 404 Client Error: Not Found for url: http://docker:2375/v1.25/networks/baseagent_default
4,895
docker/compose
docker__compose-4504
aa8fb8f708a96ba9415cd657d8ac1b1808d63615
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -37,7 +37,7 @@ def find_version(*file_paths): 'requests >= 2.6.1, != 2.11.0, < 2.12', 'texttable >= 0.8.1, < 0.9', 'websocket-client >= 0.32.0, < 1.0', - 'docker >= 2.0.2, < 3.0', + 'docker >= 2.1.0, < 3.0', 'dockerpty >= 0.4.1, < 0.5', 'six >= 1.3.0, < 2', 'jsonschema >= 2.5.1, < 3',
AttributeError: 'NoneType' object has no attribute 'tobuf' I upgraded to the latest docker version (1.13) for mac os x and when I try to run docker-compose I get the following error: Traceback (most recent call last): File "<string>", line 3, in <module> File "compose/cli/main.py", line 64, in main File "compose/cli/main.py", line 116, in perform_command File "compose/cli/main.py", line 848, in up File "compose/project.py", line 389, in up File "compose/service.py", line 302, in ensure_image_exists File "compose/service.py", line 786, in build File "site-packages/docker/api/build.py", line 139, in build File "site-packages/docker/utils/utils.py", line 105, in tar File "tarfile.py", line 2046, in addfile AttributeError: 'NoneType' object has no attribute 'tobuf' Everything was working fine on all previous versions of docker for osx until this recent upgrade
I encountered this in my projects today. It turned out to be something compose didn't like in the`tmp/sockets` directory of my rails application. Adding it to my .dockerignore has resolved the issue so far. Got the same error message for any docker-compose command, just after the update. (centos image with nginx and php) I am also getting this error after updating to Docker for Mac 1.13.0. **Update:** I could build other projects just fine. I looked in the non-working project, there was a `uwsgi.sock` file left over that didn't need to be there. I deleted it, now I can build images again. So the workaround seems to be, find socket files that are lying around in the project, and either delete or add to `.dockerignore`. I had to downgrade to 1.12.5 to get it working again. I have the same problem here. And there is no socket file inside my project. I will make a downgrade to make it work again. This happened to me after update to Docker for Mac 1.13.0 too. **Update:** fixed removing my project and git cloning again. Maybe could be a socket file in any node package, I am not sure, I know that on my project there were not socket files that's why I am guessing it was in some node package. I had the same issue for composing a Python project after upgrading to 1.13.0 on a Mac. It got resolved by cloning the project directory again. Thanks for the report everyone. For those who have encountered the bug, could you provide your OS (It seems like everyone here is on OSX?) as well as the output of `docker-compose version`? Thanks again. ``` Mac OS X El Capitan 10.11.6 docker-compose version 1.10.0, build 4bd6f1a docker-py version: 2.0.1 CPython version: 2.7.12 OpenSSL version: OpenSSL 1.0.2j 26 Sep 2016 ``` For my project, it seems to dislike the socket I have under `tmp/sockets` ``` ls -la tmp/sockets total 0 drwxr-xr-x 4 me staff 136 Dec 12 12:37 . drwxr-xr-x 7 me staff 238 Jan 23 08:15 .. srwxrwxrwx 1 me staff 0 Oct 3 09:40 server.sock ``` OSX Sierra docker-compose version 1.9.0, build 2585387 docker-py version: 1.10.6 CPython version: 2.7.12 OpenSSL version: OpenSSL 1.0.2j 26 Sep 2016 ``` MacOS 10.11.6 docker-compose version 1.10.0, build 4bd6f1a docker-py version: 2.0.1 CPython version: 2.7.12 OpenSSL version: OpenSSL 1.0.2j 26 Sep 2016 ``` ```Shell macOS Sierra: version 10.12.2 docker-compose version 1.10.0, build 4bd6f1a docker-py version: 2.0.1 CPython version: 2.7.12 OpenSSL version: OpenSSL 1.0.2j 26 Sep 2016 ``` The issue came back for me. It happens when `docker-compose up` is trying to rebuild an image. I get it when I try `docker-compose up --build` or `docker-compose build`. Any other solution instead of cloning again every time?!? @atamahjoubfar Do you have any socket file in your build context? What does `find . -name '*.sock'` output? What about `find -type s` ? @shin- Yes, I do have a socket file in my build. Both of the outputs are the same socket file: ```shell ata@macbook:~$ find . -type s ./socket/mysocket.sock ata@macbook:~$ find . -name '*.sock' ./socket/mysocket.sock ``` @atamahjoubfar Add it to your `.dockerignore`, that should solve the issue for the time being. Had the same issue on OSX with Docker 1.13.0, `docker-compose version`: ```docker-compose version 1.10.0, build 4bd6f1a docker-py version: 2.0.1 CPython version: 2.7.12 OpenSSL version: OpenSSL 1.0.2j 26 Sep 2016 ``` There was indeed a stray `stats.socket` file in the build directory; removing it fixed the issue. 👍 > Linux 4.7.0-0.bpo.1-amd64 x86_64 GNU/Linux > docker-compose version 1.11.1, build 7c5d5e4 > docker-py version: 2.0.2 > CPython version: 2.7.13 > OpenSSL version: OpenSSL 1.0.1t 3 May 2016 There was also a socket in ``.ppm/run/controller.sock``, also in my case after removing it fixed the issue. Thanks all, the issue was identified, a fix is in `docker-py`'s master branch https://github.com/docker/docker-py/pull/1408 and will be part of an upcoming release.
2017-02-17T01:06:13Z
[]
[]
Traceback (most recent call last): File "<string>", line 3, in <module> File "compose/cli/main.py", line 64, in main File "compose/cli/main.py", line 116, in perform_command File "compose/cli/main.py", line 848, in up File "compose/project.py", line 389, in up File "compose/service.py", line 302, in ensure_image_exists File "compose/service.py", line 786, in build File "site-packages/docker/api/build.py", line 139, in build File "site-packages/docker/utils/utils.py", line 105, in tar File "tarfile.py", line 2046, in addfile AttributeError: 'NoneType' object has no attribute 'tobuf'
4,896
docker/compose
docker__compose-4621
a1823007068b734a0a50229ff2e91e8ddf544f12
diff --git a/compose/cli/main.py b/compose/cli/main.py --- a/compose/cli/main.py +++ b/compose/cli/main.py @@ -61,6 +61,7 @@ def main(): + signals.ignore_sigpipe() try: command = dispatch() command() diff --git a/compose/cli/signals.py b/compose/cli/signals.py --- a/compose/cli/signals.py +++ b/compose/cli/signals.py @@ -3,6 +3,8 @@ import signal +from ..const import IS_WINDOWS_PLATFORM + class ShutdownException(Exception): pass @@ -19,3 +21,10 @@ def set_signal_handler(handler): def set_signal_handler_to_shutdown(): set_signal_handler(shutdown) + + +def ignore_sigpipe(): + # Restore default behavior for SIGPIPE instead of raising + # an exception when encountered. + if not IS_WINDOWS_PLATFORM: + signal.signal(signal.SIGPIPE, signal.SIG_DFL)
piping data out of Docker crash I have a log file inside a container ``` $ sudo docker-compose exec syslog ls -lh /var/log/messages -rw-r--r-- 1 root root 80.8M Jan 26 18:49 /var/log/messages ``` its about 600k lines ``` $ sudo docker-compose exec syslog wc -l /var/log/messages 639629 /var/log/messages ``` The following command works: ``` sudo docker-compose exec syslog cat /var/log/messages ``` Also the following (there are few results for this command): ``` sudo docker-compose exec syslog cat /var/log/messages | grep "asdfasdf" ``` But greping for http (almost all the lines), crash after printing a few hundred lines ``` sudo docker-compose exec syslog cat /var/log/messages | grep "http" ``` ``` Traceback (most recent call last):.1 <redacted. just a random line from messages file> File "<string>", line 3, in <module> File "compose/cli/main.py", line 64, in main File "compose/cli/main.py", line 116, in perform_command File "compose/cli/main.py", line 461, in exec_command File "site-packages/dockerpty/pty.py", line 334, in start File "site-packages/dockerpty/pty.py", line 370, in _hijack_tty File "site-packages/dockerpty/io.py", line 164, in do_write OSError: [Errno 32] Broken pipe docker-compose returned -1 ``` - I can reproduce the bug everytime. - Yes, I can avoid the bug using grep http messages instead of the useless use of cat. But this is just for the sake of reporting the bug. - Seems related to #1509 #3352 # Info ``` $ sudo docker-compose --version docker-compose version 1.10.0, build 4bd6f1a ``` ``` $ sudo docker version Client: Version: 1.12.3 API version: 1.24 Go version: go1.6.3 Git commit: 6b644ec Built: Wed Oct 26 21:44:32 2016 OS/Arch: linux/amd64 Server: Version: 1.12.3 API version: 1.24 Go version: go1.6.3 Git commit: 6b644ec Built: Wed Oct 26 21:44:32 2016 OS/Arch: linux/amd64 ``` ``` $ uname -a Linux codexServer 4.2.0-42-generic #49~14.04.1-Ubuntu SMP Wed Jun 29 20:22:11 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux ``` ``` $ grep -i desc /etc/lsb-release DISTRIB_DESCRIPTION="Ubuntu 14.04.5 LTS" ```
I'm not able to reproduce this locally with a similar setup (I don't have a 600k lines syslog file, so I wrote a quick python script that continuously output random message lines). Do you have `ndg-httpsclient` installed? Can you reproduce with `docker-compose 1.11.2`? ```ndg-httpsclient``` not installed. I just found out a way to reproduce this 100% of the time, instantly. Tested with latest docker-engine and docker-compose across 3 different environments. ``` docker-compose exec alpine sh -c "cat /dev/urandom | tr -dc "[:alnum:][:punct:]" | tr '[:digits:]' '\n' | head -c 10000000" | xargs -0 ``` ``` # cat docker-compose.yml version: '2' services: alpine: image: alpine:latest command: sleep 3600 ``` With the container running: ```docker-compose up -d```. Some thoughts: If instead of ```head -c 10000000``` I execute it with ```head -c 1000``` it never crashes. With 120925 it crash about 50% of the time. The other 50% either prints lots of chars, or prints: ``` xargs: argument line too long ``` It seems to me that xargs it exiting or crashing and that makes the compose command crash. So in my original post, its probably grep that is crashing and not compose. The strange thing is that if I place the ```xargs``` inside the command, it does not crash, like this: ``` docker-compose exec alpine sh -c "cat /dev/urandom | tr -dc "[:alnum:][:punct:]" | tr '[:digits:]' '\n' | head -c 1000000 | xargs -0" ``` So piping out from ```docker-compose exec``` large amounts of data makes some gnu utils crash. The easiest workaround is doing ```docker exec -ti``` instead of ```docker-compose exec```. Ok - I can indeed reproduce consistently with that command. I'm not 100% sure what we can do about it though. Just catch the OSError and print out a nice message I guess? I'm not sure either. But a good start would be finding out why this command never crash: ``` docker exec -ti composetest_alpine_1 sh -c "cat /dev/urandom | tr -dc "[:alnum:][:punct:]" | tr '[:digits:]' '\n' | head -c 100000000" | grep -E -e "^." ``` while this one crashes in about 10-20 seconds (in my setup): ``` docker-compose exec alpine sh -c "cat /dev/urandom | tr -dc "[:alnum:][:punct:]" | tr '[:digits:]' '\n' | head -c 100000000" | grep -E -e "^." ```
2017-03-14T21:23:20Z
[]
[]
Traceback (most recent call last):.1 <redacted. just a random line from messages file> File "<string>", line 3, in <module> File "compose/cli/main.py", line 64, in main File "compose/cli/main.py", line 116, in perform_command File "compose/cli/main.py", line 461, in exec_command File "site-packages/dockerpty/pty.py", line 334, in start File "site-packages/dockerpty/pty.py", line 370, in _hijack_tty File "site-packages/dockerpty/io.py", line 164, in do_write OSError: [Errno 32] Broken pipe
4,911
docker/compose
docker__compose-4714
7aa55120db74ef79c8ba4e1c5dee0abaf98224ed
diff --git a/compose/config/serialize.py b/compose/config/serialize.py --- a/compose/config/serialize.py +++ b/compose/config/serialize.py @@ -111,9 +111,9 @@ def denormalize_service_dict(service_dict, version, image_digest=None): ) if 'ports' in service_dict and version not in (V3_2,): - service_dict['ports'] = map( - lambda p: p.legacy_repr() if isinstance(p, types.ServicePort) else p, - service_dict['ports'] - ) + service_dict['ports'] = [ + p.legacy_repr() if isinstance(p, types.ServicePort) else p + for p in service_dict['ports'] + ] return service_dict
`config` doesn't work with extended `ports` with Python3 With these files: `base.yml`: ```yaml version: '2' services: _nginx: image: nginx:1.11.12 ``` `docker-compose.yml`: ```yaml version: '2' services: nginx: extends: file: base.yml service: _nginx ports: - "8080:80" ``` Running `config`: ```text $ docker-compose config Traceback (most recent call last): File "/home/ken/.virtualenvs/compose3_clean/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/compose/cli/main.py", line 67, in main command() File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/compose/cli/main.py", line 111, in perform_command handler(command, options, command_options) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/compose/cli/main.py", line 306, in config print(serialize_config(compose_config, image_digests)) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/compose/config/serialize.py", line 62, in serialize_config width=80) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/__init__.py", line 216, in safe_dump return dump_all([data], stream, Dumper=SafeDumper, **kwds) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/__init__.py", line 188, in dump_all dumper.represent(data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 26, in represent node = self.represent_data(data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 47, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 205, in represent_dict return self.represent_mapping('tag:yaml.org,2002:map', data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 116, in represent_mapping node_value = self.represent_data(item_value) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 47, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 205, in represent_dict return self.represent_mapping('tag:yaml.org,2002:map', data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 116, in represent_mapping node_value = self.represent_data(item_value) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 47, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 205, in represent_dict return self.represent_mapping('tag:yaml.org,2002:map', data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 116, in represent_mapping node_value = self.represent_data(item_value) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 57, in represent_data node = self.yaml_representers[None](self, data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 229, in represent_undefined raise RepresenterError("cannot represent an object: %s" % data) yaml.representer.RepresenterError: cannot represent an object: <map object at 0x7f3caace4470> ``` ```txt $ docker-compose --version docker-compose version 1.12.0, build b31ff33 $ python -V Python 3.5.2 $ pip list DEPRECATION: The default format will switch to columns in the future. You can use --format=(legacy|columns) (or define a format=(legacy|columns) in your pip.conf under the [list] section) to disable this warning. appdirs (1.4.3) cached-property (1.3.0) click (6.7) colorama (0.3.7) docker (2.2.1) docker-compose (1.12.0) docker-pycreds (0.2.1) dockerpty (0.4.1) docopt (0.6.2) first (2.0.1) jsonschema (2.6.0) packaging (16.8) pip (9.0.1) pip-tools (1.8.2) pyparsing (2.2.0) PyYAML (3.12) requests (2.11.1) setuptools (34.3.3) six (1.10.0) texttable (0.8.8) websocket-client (0.40.0) wheel (0.29.0) $ docker info Containers: 3 Running: 0 Paused: 0 Stopped: 3 Images: 181 Server Version: 17.04.0-ce Storage Driver: aufs Root Dir: /media/r0/var/lib/docker/aufs Backing Filesystem: extfs Dirs: 577 Dirperm1 Supported: true Logging Driver: json-file Cgroup Driver: cgroupfs Plugins: Volume: local Network: bridge host macvlan null overlay Swarm: inactive Runtimes: runc Default Runtime: runc Init Binary: containerd version: 422e31ce907fd9c3833a38d7b8fdd023e5a76e73 runc version: 9c2d8d184e5da67c95d601382adf14862e4f2228 init version: 949e6fa Security Options: apparmor seccomp Profile: default Kernel Version: 4.4.0-71-generic Operating System: Ubuntu 16.04.2 LTS OSType: linux Architecture: x86_64 CPUs: 8 Total Memory: 15.56GiB Name: oblako ID: V6CL:5652:TVAP:Z4Q2:DOBE:CHDS:G3MY:Y6TU:BWGK:5B6E:REYP:PNGR Docker Root Dir: /media/r0/var/lib/docker Debug Mode (client): false Debug Mode (server): false Username: oeuftete Registry: https://index.docker.io/v1/ Experimental: false Insecure Registries: 127.0.0.0/8 Live Restore Enabled: false WARNING: No swap limit support ```
Similar cases that do work for me: * Same files as above, compose 1.11.2 * Same files as above, python 2.7.12 * Single file definition of `nginx` with `ports` (no `extends`) There is also a case where extended `ports` also ends up being `None` when using `up`, but I haven't been able to narrow that one down: `ERROR: for nginx Cannot create container for service nginx: b'invalid port specification: "None"'` Maybe #4653 is related, although it sounds like the opposite case where already defined `ports` go away when using `extends`, whereas here something is wrong with `ports` when being extended. Thank you for the report. We'll take a look ASAP This seems to do the trick: ```diff diff --git a/compose/config/serialize.py b/compose/config/serialize.py index 5b36124..b06922a 100644 --- a/compose/config/serialize.py +++ b/compose/config/serialize.py @@ -111,9 +111,9 @@ def denormalize_service_dict(service_dict, version, image_digest=None): ) if 'ports' in service_dict and version not in (V3_2,): - service_dict['ports'] = map( + service_dict['ports'] = list(map( lambda p: p.legacy_repr() if isinstance(p, types.ServicePort) else p, service_dict['ports'] - ) + )) return service_dict ```
2017-04-07T23:55:04Z
[]
[]
Traceback (most recent call last): File "/home/ken/.virtualenvs/compose3_clean/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/compose/cli/main.py", line 67, in main command() File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/compose/cli/main.py", line 111, in perform_command handler(command, options, command_options) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/compose/cli/main.py", line 306, in config print(serialize_config(compose_config, image_digests)) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/compose/config/serialize.py", line 62, in serialize_config width=80) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/__init__.py", line 216, in safe_dump return dump_all([data], stream, Dumper=SafeDumper, **kwds) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/__init__.py", line 188, in dump_all dumper.represent(data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 26, in represent node = self.represent_data(data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 47, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 205, in represent_dict return self.represent_mapping('tag:yaml.org,2002:map', data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 116, in represent_mapping node_value = self.represent_data(item_value) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 47, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 205, in represent_dict return self.represent_mapping('tag:yaml.org,2002:map', data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 116, in represent_mapping node_value = self.represent_data(item_value) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 47, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 205, in represent_dict return self.represent_mapping('tag:yaml.org,2002:map', data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 116, in represent_mapping node_value = self.represent_data(item_value) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 57, in represent_data node = self.yaml_representers[None](self, data) File "/home/ken/.virtualenvs/compose3_clean/lib/python3.5/site-packages/yaml/representer.py", line 229, in represent_undefined raise RepresenterError("cannot represent an object: %s" % data) yaml.representer.RepresenterError: cannot represent an object: <map object at 0x7f3caace4470>
4,921
docker/compose
docker__compose-4721
94defc159a2dcdc548f2c5295f29acfad6122684
diff --git a/compose/cli/command.py b/compose/cli/command.py --- a/compose/cli/command.py +++ b/compose/cli/command.py @@ -49,14 +49,17 @@ def get_config_from_options(base_dir, options): def get_config_path_from_options(base_dir, options, environment): + def unicode_paths(paths): + return [p.decode('utf-8') if isinstance(p, six.binary_type) else p for p in paths] + file_option = options.get('--file') if file_option: - return file_option + return unicode_paths(file_option) config_files = environment.get('COMPOSE_FILE') if config_files: pathsep = environment.get('COMPOSE_PATH_SEPARATOR', os.pathsep) - return config_files.split(pathsep) + return unicode_paths(config_files.split(pathsep)) return None
Support unicode characters in -f paths On Ubuntu 16.04: ``` $ docker-compose -f 就吃饭/docker-compose.yml config/home/joffrey/work/compose/compose/config/config.py:234: UnicodeWarning: Unicode equal comparison failed to convert both arguments to Unicode - interpreting them as being unequal if filenames == ['-']: Traceback (most recent call last): File "/home/joffrey/.envs/compose/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.11.0.dev0', 'console_scripts', 'docker-compose')() File "/home/joffrey/work/compose/compose/cli/main.py", line 64, in main command() File "/home/joffrey/work/compose/compose/cli/main.py", line 110, in perform_command handler(command, options, command_options) File "/home/joffrey/work/compose/compose/cli/main.py", line 305, in config compose_config = get_config_from_options(self.project_dir, config_options) File "/home/joffrey/work/compose/compose/cli/command.py", line 46, in get_config_from_options config.find(base_dir, config_path, environment) File "/home/joffrey/work/compose/compose/config/config.py", line 242, in find filenames = [os.path.join(base_dir, f) for f in filenames] File "/home/joffrey/.envs/compose/lib/python2.7/posixpath.py", line 73, in join path += '/' + b UnicodeDecodeError: 'ascii' codec can't decode byte 0xe5 in position 1: ordinal not in range(128) ``` On Windows: ``` docker-compose -f "C:\Users\husun\documents\visual studio 2017\Projects\测试中文\docker-compose.yml" up -d --build ERROR: compose.cli.main.main: .IOError: [Errno 22] invalid mode ('r') or filename: 'C:\\Users\\husun\\documents\\visual studio 2017\\Projects\\????\\docker-compose.yml' ```
2017-04-11T00:40:02Z
[]
[]
Traceback (most recent call last): File "/home/joffrey/.envs/compose/bin/docker-compose", line 9, in <module> load_entry_point('docker-compose==1.11.0.dev0', 'console_scripts', 'docker-compose')() File "/home/joffrey/work/compose/compose/cli/main.py", line 64, in main command() File "/home/joffrey/work/compose/compose/cli/main.py", line 110, in perform_command handler(command, options, command_options) File "/home/joffrey/work/compose/compose/cli/main.py", line 305, in config compose_config = get_config_from_options(self.project_dir, config_options) File "/home/joffrey/work/compose/compose/cli/command.py", line 46, in get_config_from_options config.find(base_dir, config_path, environment) File "/home/joffrey/work/compose/compose/config/config.py", line 242, in find filenames = [os.path.join(base_dir, f) for f in filenames] File "/home/joffrey/.envs/compose/lib/python2.7/posixpath.py", line 73, in join path += '/' + b UnicodeDecodeError: 'ascii' codec can't decode byte 0xe5 in position 1: ordinal not in range(128)
4,924
docker/compose
docker__compose-4770
59dbeac4c1f4941e631df2ab02ae2a2c4958008f
diff --git a/compose/network.py b/compose/network.py --- a/compose/network.py +++ b/compose/network.py @@ -158,8 +158,8 @@ def check_remote_ipam_config(remote, local): if sorted(lc.get('AuxiliaryAddresses')) != sorted(rc.get('AuxiliaryAddresses')): raise NetworkConfigChangedError(local.full_name, 'IPAM config aux_addresses') - remote_opts = remote_ipam.get('Options', {}) - local_opts = local.ipam.get('options', {}) + remote_opts = remote_ipam.get('Options') or {} + local_opts = local.ipam.get('options') or {} for k in set.union(set(remote_opts.keys()), set(local_opts.keys())): if remote_opts.get(k) != local_opts.get(k): raise NetworkConfigChangedError(local.full_name, 'IPAM option "{}"'.format(k))
AttributeError: 'NoneType' object has no attribute 'keys' hey, i want to install mailcow-dockerized on my debian jessie machine and after `docker pull` I need to run `docker-compose up -d` but it throws following error: ```sh Traceback (most recent call last): File "bin/docker-compose", line 3, in <module> File "compose/cli/main.py", line 68, in main File "compose/cli/main.py", line 118, in perform_command File "compose/cli/main.py", line 924, in up File "compose/project.py", line 389, in up File "compose/project.py", line 437, in initialize File "compose/network.py", line 252, in initialize File "compose/network.py", line 60, in ensure File "compose/network.py", line 180, in check_remote_network_config File "compose/network.py", line 163, in check_remote_ipam_config AttributeError: 'NoneType' object has no attribute 'keys' Failed to execute script docker-compose ``` since I dont know python I dont know wheres the problem.
What version of the Docker Engine are you using - can you share the output of `docker version`? Sure! Look here: ``` Client: Version: 17.03.1-ce API version: 1.27 Go version: go1.7.5 Git commit: c6d412e Built: Mon Mar 27 17:07:28 2017 OS/Arch: linux/amd64 Server: Version: 17.03.1-ce API version: 1.27 (minimum version 1.12) Go version: go1.7.5 Git commit: c6d412e Built: Mon Mar 27 17:07:28 2017 OS/Arch: linux/amd64 Experimental: false ``` Thanks - I believe I identified the issue as a bug in Compose. As a workaround, you might want to try running `docker-compose down && docker-compose up -d` instead. That worked! Thanks!
2017-04-26T20:35:32Z
[]
[]
Traceback (most recent call last): File "bin/docker-compose", line 3, in <module> File "compose/cli/main.py", line 68, in main File "compose/cli/main.py", line 118, in perform_command File "compose/cli/main.py", line 924, in up File "compose/project.py", line 389, in up File "compose/project.py", line 437, in initialize File "compose/network.py", line 252, in initialize File "compose/network.py", line 60, in ensure File "compose/network.py", line 180, in check_remote_network_config File "compose/network.py", line 163, in check_remote_ipam_config AttributeError: 'NoneType' object has no attribute 'keys'
4,930
docker/compose
docker__compose-4955
c38eaeaba342d586c8986482b745006efb801665
diff --git a/compose/config/config.py b/compose/config/config.py --- a/compose/config/config.py +++ b/compose/config/config.py @@ -959,7 +959,7 @@ def parse_sequence_func(seq): merged = parse_sequence_func(md.base.get(field, [])) merged.update(parse_sequence_func(md.override.get(field, []))) - md[field] = [item for item in sorted(merged.values())] + md[field] = [item for item in sorted(merged.values(), key=lambda x: x.target)] def merge_build(output, base, override):
TypeError: unorderable types: NoneType() < str() Hi guys! I just updated to version 1.14 via pip and got error: ``` (master) hexlet$ make compose-bash docker-compose run web bash Traceback (most recent call last): File "/usr/local/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/usr/local/lib/python3.5/dist-packages/compose/cli/main.py", line 68, in main command() File "/usr/local/lib/python3.5/dist-packages/compose/cli/main.py", line 115, in perform_command project = project_from_options('.', options) File "/usr/local/lib/python3.5/dist-packages/compose/cli/command.py", line 37, in project_from_options override_dir=options.get('--project-directory'), File "/usr/local/lib/python3.5/dist-packages/compose/cli/command.py", line 91, in get_project config_data = config.load(config_details) File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 379, in load service_dicts = load_services(config_details, main_file) File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 475, in load_services service_config = merge_services(service_config, next_config) File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 466, in merge_services for name in all_service_names File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 466, in <dictcomp> for name in all_service_names File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 853, in merge_service_dicts_from_files new_service = merge_service_dicts(base, override, version) File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 929, in merge_service_dicts merge_ports(md, base, override) File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 962, in merge_ports md[field] = [item for item in sorted(merged.values())] TypeError: unorderable types: NoneType() < str() Makefile:72: recipe for target 'compose-bash' failed make: *** [compose-bash] Error 1 (master) hexlet$ cat ^C (master) hexlet$ docker-compose --version docker-compose version 1.14.0, build c7bdf9e ```
Thanks for the report, can you share your Compose file? Sure! ``` version: '3.1' services: web-nginx: build: context: services/web_nginx dockerfile: Dockerfile.production ports: - "80:80" depends_on: - web web: dns: - 169.254.1.1 - 8.8.8.8 user: internal build: context: services/web dockerfile: Dockerfile.production command: bin/rails s -b 0.0.0.0 volumes: - ./.irb_history:/home/internal/.irb_history ports: - "3000:3000" consul: network_mode: host build: context: services/consul dockerfile: Dockerfile environment: CONSUL_LOCAL_CONFIG: '{ "node_name": "exercise.hexlet.dev" }' registrator: image: gliderlabs/registrator network_mode: host command: consul://localhost:8500 volumes: - '/var/run/docker.sock:/tmp/docker.sock' depends_on: - consul evaluator-nginx: build: context: services/evaluator_nginx dockerfile: Dockerfile environment: SERVICE_TAGS: evaluator ports: - "443:443" dnsmasq: network_mode: host build: context: services/dnsmasq dockerfile: Dockerfile ports: - "53:53/udp" - "53:53" cap_add: - NET_ADMIN ``` Hm, I can't reproduce locally. Do you have a `docker-compose.override.yml` as well, or do you provide multiple files using the `-f` option? If so, having the content of those files would be useful as well. docker-compose.override.yml ``` version: '3.1' services: spring: user: internal build: context: services/web dockerfile: Dockerfile.development volumes: - ./services/web:/usr/src/app env_file: .env.development command: bin/spring server pid: host web-nginx: build: context: services/web_nginx dockerfile: Dockerfile.development frontend: build: context: services/web dockerfile: Dockerfile.development command: make frontend_watch # ports: # - '8080:80' volumes: - ./services/web:/usr/src/app web: build: context: services/web dockerfile: Dockerfile.development env_file: .env.development volumes: - ~/.bash_history:/internal/.bash_history - ./services/web:/usr/src/app command: sh -c "rm -f tmp/pids/server.pid && bin/rails s -b 0.0.0.0" activejob-web: build: context: services/web dockerfile: Dockerfile.development env_file: .env.development volumes: - ~/.bash_history:/internal/.bash_history - ./services/web:/usr/src/app command: bundle exec shoryuken -R -C config/shoryuken.web.yml activejob-fast: build: context: services/web dockerfile: Dockerfile.development env_file: .env.development volumes: - ~/.bash_history:/internal/.bash_history - ./services/web:/usr/src/app command: bundle exec shoryuken -R -C config/shoryuken.fast.yml activejob-slow: build: context: services/web dockerfile: Dockerfile.development env_file: .env.development volumes: - ~/.bash_history:/internal/.bash_history - ./services/web:/usr/src/app command: bundle exec shoryuken -R -C config/shoryuken.slow.yml ``` any docker-compose calls throws the same error... ``` (feature/consul-template) hexlet$ make compose-build docker-compose build Traceback (most recent call last): File "/usr/local/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/usr/local/lib/python3.5/dist-packages/compose/cli/main.py", line 68, in main command() File "/usr/local/lib/python3.5/dist-packages/compose/cli/main.py", line 115, in perform_command project = project_from_options('.', options) File "/usr/local/lib/python3.5/dist-packages/compose/cli/command.py", line 37, in project_from_options override_dir=options.get('--project-directory'), File "/usr/local/lib/python3.5/dist-packages/compose/cli/command.py", line 91, in get_project config_data = config.load(config_details) File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 379, in load service_dicts = load_services(config_details, main_file) File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 475, in load_services service_config = merge_services(service_config, next_config) File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 466, in merge_services for name in all_service_names File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 466, in <dictcomp> for name in all_service_names File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 853, in merge_service_dicts_from_files new_service = merge_service_dicts(base, override, version) File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 929, in merge_service_dicts merge_ports(md, base, override) File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 962, in merge_ports md[field] = [item for item in sorted(merged.values())] TypeError: unorderable types: str() < NoneType() Makefile:57: recipe for target 'compose-build' failed make: *** [compose-build] Error 1 ```
2017-06-23T22:05:14Z
[]
[]
Traceback (most recent call last): File "/usr/local/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/usr/local/lib/python3.5/dist-packages/compose/cli/main.py", line 68, in main command() File "/usr/local/lib/python3.5/dist-packages/compose/cli/main.py", line 115, in perform_command project = project_from_options('.', options) File "/usr/local/lib/python3.5/dist-packages/compose/cli/command.py", line 37, in project_from_options override_dir=options.get('--project-directory'), File "/usr/local/lib/python3.5/dist-packages/compose/cli/command.py", line 91, in get_project config_data = config.load(config_details) File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 379, in load service_dicts = load_services(config_details, main_file) File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 475, in load_services service_config = merge_services(service_config, next_config) File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 466, in merge_services for name in all_service_names File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 466, in <dictcomp> for name in all_service_names File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 853, in merge_service_dicts_from_files new_service = merge_service_dicts(base, override, version) File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 929, in merge_service_dicts merge_ports(md, base, override) File "/usr/local/lib/python3.5/dist-packages/compose/config/config.py", line 962, in merge_ports md[field] = [item for item in sorted(merged.values())] TypeError: unorderable types: NoneType() < str()
4,944
docker/compose
docker__compose-4990
e33041582fc7dda54a1412f1e56209aa8671c280
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -56,6 +56,7 @@ def find_version(*file_paths): ':python_version < "3.4"': ['enum34 >= 1.0.4, < 2'], ':python_version < "3.5"': ['backports.ssl_match_hostname >= 3.5'], ':python_version < "3.3"': ['ipaddress >= 1.0.16'], + 'socks': ['PySocks >= 1.5.6, != 1.5.7, < 2'], }
Compose exits with "Missing dependencies for SOCKS support" Hi there! Ran into a bit of an issue with docker-compose 1.13.0, build 1719ceb. In my test the issue goes back to at least 1.11.0. A script I'm running does `docker-compose down`, which results in an uncaught exception that ends with the line: "requests.exceptions.InvalidSchema: Missing dependencies for SOCKS support.": ``` Traceback (most recent call last): File "docker-compose", line 3, in <module> File "compose/cli/main.py", line 68, in main File "compose/cli/main.py", line 118, in perform_command File "compose/cli/main.py", line 358, in down File "compose/project.py", line 286, in down File "compose/project.py", line 252, in stop File "compose/project.py", line 498, in containers File "compose/project.py", line 489, in _labeled_containers File "site-packages/docker/api/container.py", line 189, in containers File "site-packages/docker/utils/decorators.py", line 47, in inner File "site-packages/docker/api/client.py", line 183, in _get File "site-packages/requests/sessions.py", line 488, in get File "site-packages/requests/sessions.py", line 475, in request File "site-packages/requests/sessions.py", line 596, in send File "site-packages/requests/adapters.py", line 390, in send File "site-packages/docker/transport/ssladapter.py", line 56, in get_connection File "site-packages/requests/adapters.py", line 290, in get_connection File "site-packages/requests/adapters.py", line 184, in proxy_manager_for File "site-packages/requests/adapters.py", line 43, in SOCKSProxyManager requests.exceptions.InvalidSchema: Missing dependencies for SOCKS support. ``` I'm running through docker machine and nothing in the docker compose yaml even mentions socks. The script that wraps the docker-compose does set `ALL_PROXY`, which I think may be causing `requests` or some other library to eagerly load socks support. My recommendation would be to add `requests[socks]` to the setup.py and call it good.
Thanks for the report! The majority of people do not need those additional dependencies - adding extra dependencies indiscriminately seems like a poor solution. The onus is on the script provider to ensure additional requirements are included, or at least properly advertised. docker-compose is being distributed as a frozen binary, which means adding requirements after the fact is not really possible. Yeah, this is for the frozen command, not the pip installed version. And this would affect anyone with the rather innocuous `ALL_PROXY` environment variable configured. In any case the PySocks module is the only dependency required, and it is really very tiny ([literally one file](https://github.com/Anorov/PySocks)). The version is managed automatically by simply using the socks requests extra (`requests[socks]`). It seems unfriendly for users to have to figure out what extra packages they need to use the features supported by `docker-compose` - maybe `setup.py` can have a `extras` option that includes all of these packages, and the distribution channels for the frozen version of `docker-compose` can install `docker-compose` with that option instead? That would be a reasonable addition.
2017-07-03T22:36:09Z
[]
[]
Traceback (most recent call last): File "docker-compose", line 3, in <module> File "compose/cli/main.py", line 68, in main File "compose/cli/main.py", line 118, in perform_command File "compose/cli/main.py", line 358, in down File "compose/project.py", line 286, in down File "compose/project.py", line 252, in stop File "compose/project.py", line 498, in containers File "compose/project.py", line 489, in _labeled_containers File "site-packages/docker/api/container.py", line 189, in containers File "site-packages/docker/utils/decorators.py", line 47, in inner File "site-packages/docker/api/client.py", line 183, in _get File "site-packages/requests/sessions.py", line 488, in get File "site-packages/requests/sessions.py", line 475, in request File "site-packages/requests/sessions.py", line 596, in send File "site-packages/requests/adapters.py", line 390, in send File "site-packages/docker/transport/ssladapter.py", line 56, in get_connection File "site-packages/requests/adapters.py", line 290, in get_connection File "site-packages/requests/adapters.py", line 184, in proxy_manager_for File "site-packages/requests/adapters.py", line 43, in SOCKSProxyManager requests.exceptions.InvalidSchema: Missing dependencies for SOCKS support.
4,946
docker/compose
docker__compose-5113
390ba801a3299b1a73d31e739afb0992115a3901
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -36,7 +36,7 @@ def find_version(*file_paths): 'requests >= 2.6.1, != 2.11.0, < 2.12', 'texttable >= 0.9.0, < 0.10', 'websocket-client >= 0.32.0, < 1.0', - 'docker >= 2.4.2, < 3.0', + 'docker >= 2.5.0, < 3.0', 'dockerpty >= 0.4.1, < 0.5', 'six >= 1.3.0, < 2', 'jsonschema >= 2.5.1, < 3',
Up a service cause MemoryError I am consistently getting a MemoryError on container: ``` bdhameli@debian:~/Docker$ docker-compose up openvpn Starting dmzpi_openvpn_1 Attaching to dmzpi_openvpn_1 Traceback (most recent call last): File "bin/docker-compose", line 3, in <module> File "compose/cli/main.py", line 67, in main File "compose/cli/main.py", line 117, in perform_command File "compose/cli/main.py", line 937, in up File "compose/cli/log_printer.py", line 87, in run File "compose/cli/log_printer.py", line 235, in consume_queue MemoryError ``` At this point, the container is running normally. The docker-compose.yml file contains: ``` services: openvpn: build: ./openvpn cap_add: - net_admin dns: - 8.8.4.4 - 8.8.8.8 read_only: true tmpfs: - /tmp restart: unless-stopped stdin_open: true tty: true volumes: - ./openvpn/vpn:/vpn devices: - /dev/net/tun:/dev/net/tun ``` Docker and Docker-Compose versions are: ``` :~/Docker$ docker --version Docker version 17.03.1-ce, build c6d412e :~/Docker/$ docker-compose version docker-compose version 1.12.0, build b31ff33 docker-py version: 2.2.1 CPython version: 2.7.13 OpenSSL version: OpenSSL 1.0.1t 3 May 2016 ``` Compose ignores .dockerignore when containing exclusion Following on from this comment, https://github.com/docker/compose/issues/1607#issuecomment-190654346 I've created a repository demonstrating the issue, https://github.com/multimac/compose-dockerignore When including a file already excluded by a prior rule in a .dockerignore file, Compose stops ignoring any other files. **Expected Results** ``` ignore_1 | /context/ ignore_1 | /context/.dockerignore ignore_1 | /context/docker-compose.yml ignore_1 | /context/Dockerfile ignore_1 | /context/folder ignore_1 | /context/folder/included ``` **Actual Results** - (/context/folder/ignored shouldn't be there) ``` ignore_1 | /context/ ignore_1 | /context/.dockerignore ignore_1 | /context/docker-compose.yml ignore_1 | /context/Dockerfile ignore_1 | /context/folder ignore_1 | /context/folder/included ignore_1 | /context/folder/ignored ``` **Versions** ``` Windows 10 Pro Docker version 1.12.0, build 8eab29e docker-compose version 1.8.0, build d988a55 ```
Thanks for the report. Do you get the same error if you run `docker-compose logs -f openvpn`? What about `docker-compose logs --tail=20 -f openvpn`? Hi, These two commands (`docker-compose logs -f openvpn` and `docker-compose logs --tail=20 -f openvpn`) return the same content: ``` openvpn_1 | Wed Apr 26 07:55:26 2017 OpenVPN 2.3.4 x86_64-pc-linux-gnu [SSL (OpenSSL)] [LZO] [EPOLL] [PKCS11] [MH] [IPv6] built on Nov 12 2015 openvpn_1 | Wed Apr 26 07:55:26 2017 library versions: OpenSSL 1.0.1t 3 May 2016, LZO 2.08 openvpn_1 | Wed Apr 26 07:55:26 2017 Control Channel Authentication: tls-auth using INLINE static key file openvpn_1 | Wed Apr 26 07:55:26 2017 Outgoing Control Channel Authentication: Using 160 bit message hash 'SHA1' for HMAC authentication openvpn_1 | Wed Apr 26 07:55:26 2017 Incoming Control Channel Authentication: Using 160 bit message hash 'SHA1' for HMAC authentication openvpn_1 | Wed Apr 26 07:55:26 2017 Socket Buffers: R=[212992->131072] S=[212992->131072] openvpn_1 | Wed Apr 26 07:55:46 2017 RESOLVE: Cannot resolve host address: #####.########.###: Temporary failure in name resolution openvpn_1 | Wed Apr 26 07:56:06 2017 RESOLVE: Cannot resolve host address: #####.########.###: Temporary failure in name resolution openvpn_1 | Wed Apr 26 07:56:06 2017 SIGUSR1[soft,init_instance] received, process restarting openvpn_1 | Wed Apr 26 07:56:06 2017 Restart pause, 2 second(s) ``` If helping, the DNS server define in `docker-compose.yml` for openvpn service can't be reached. I've got the same problem. It is working when I run docker-compose in daemon mode. // not working `docker-compose up` // working `docker-compose up -d` ``` ubuntu@myuser:~/api $ docker version Client: Version: 1.13.0 API version: 1.25 Go version: go1.7.3 Git commit: 49bf474 Built: Tue Jan 17 09:58:26 2017 OS/Arch: linux/amd64 Server: Version: 1.13.0 API version: 1.25 (minimum version 1.12) Go version: go1.7.3 Git commit: 49bf474 Built: Tue Jan 17 09:58:26 2017 OS/Arch: linux/amd64 Experimental: false ubuntu@myuser:~/api $ docker-compose -v docker-compose version 1.11.2, build dfed245 ``` I'm also getting the same error in running `docker-compose -f docker-compose-1.yml -f docker-compose-2.yml up` in a gcloud server. But, on my local system, it is running just fine. ``` Traceback (most recent call last): File "bin/docker-compose", line 3, in <module> File "compose/cli/main.py", line 64, in main File "compose/cli/main.py", line 116, in perform_command File "compose/cli/main.py", line 889, in up File "compose/cli/log_printer.py", line 87, in run File "compose/cli/log_printer.py", line 229, in consume_queue MemoryError Failed to execute script docker-compose ``` docker-compose-1.yml ``` version: '2' services: drupal: build: context: . dockerfile: Dockerfile.drupal ports: - "3700:80" restart: always networks: main: aliases: - drupal networks: main: ``` docker-compose-2.yml ``` version: '2' services: redisMain: image: redis:alpine command: redis-server /usr/local/etc/redis/redis.conf volumes: - "./redis/redisMain.conf:/usr/local/etc/redis/redis.conf" restart: always networks: main: aliases: - redisMain rabbitmq: image: rabbitmq:management-alpine restart: always hostname: rabbitmq environment: RABBITMQ_DEFAULT_USER: app RABBITMQ_DEFAULT_PASS: app ports: - "15672:15672" networks: main: aliases: - rabbitmq mongodbMain: image: mongo:latest restart: always networks: main: aliases: - mongodbMain mysqlMain: image: mysql:latest volumes: - "~/Office/app-db/mysql-data/db:/var/lib/mysql" restart: always environment: MYSQL_ROOT_PASSWORD: app MYSQL_DATABASE: test_db MYSQL_USER: app MYSQL_PASSWORD: app networks: main: aliases: - mysqlMain haproxymain: image: haproxy:alpine volumes: - "./haproxy/haproxy.cfg:/usr/local/etc/haproxy/haproxy.cfg" links: - app ports: - "9000:80" restart: always networks: main: aliases: - haproxymain app: depends_on: - redisMain - mysqlMain - mongodbMain - rabbitmq build: context: . dockerfile: Dockerfile args: branch: ${BRANCH} # expire: ${EXPIRE} tty: true restart: always networks: main: aliases: - app networks: main: ``` I'm trying to clone a git repository and do `npm install` inside it as an entrypoint script in one of my containers. Also, I tried running it with `-d` flag. But that doesn't seemed to clone the repo in my container. Docker version: ``` dev@dev:~$ docker --version Docker version 17.03.1-ce, build c6d412e dev@dev:~$ docker-compose --version docker-compose version 1.11.2, build dfed245 ``` Could it be the same issue as #409 ? What is the output of `ulimit -m` in your different environments? I don't think so : ``` ubuntu@myuser:~$ ulimit -m unlimited ``` Idem: ``` :~/Docker/dmz-pi$ ulimit -m unlimited ``` Output of `ulimit -m` is same on both environments - glcoud server and my local system. ``` dev@dev:~$ ulimit -m unlimited ``` Also, if run `docker-compose up` again without doing `docker-compose down`, then I'm getting a `COMPOSE_HTTP_TIMEOUT` error and not the `MemoryError`. What could be the reason behind this? ``` ERROR: An HTTP request took too long to complete. Retry with --verbose to obtain debug information. If you encounter this issue regularly because of slow network conditions, consider setting COMPOSE_HTTP_TIMEOUT to a higher value (current value: 60). ``` i'm also getting the same error. is there any workaround to this problem? edit: I just removed tty:true and it worked +1 Anyone have an update on this one? This bug is making it impossible to tune the context size of my monolithic repository. +1 EDIT: im using docker-compose version 1.11.2 I have same issue on mac. ``` Containers: 3 Running: 0 Paused: 0 Stopped: 3 Images: 30 Server Version: 17.06.0-ce Storage Driver: aufs Root Dir: /var/lib/docker/aufs Backing Filesystem: extfs Dirs: 44 Dirperm1 Supported: true Logging Driver: json-file Cgroup Driver: cgroupfs Plugins: Volume: local Network: bridge host ipvlan macvlan null overlay Log: awslogs fluentd gcplogs gelf journald json-file logentries splunk syslog Swarm: inactive Runtimes: runc Default Runtime: runc Init Binary: docker-init containerd version: cfb82a876ecc11b5ca0977d1733adbe58599088a runc version: 2d41c047c83e09a6d61d464906feb2a2f3c52aa4 init version: 949e6fa Security Options: seccomp Profile: default Kernel Version: 4.9.36-moby Operating System: Alpine Linux v3.5 OSType: linux Architecture: x86_64 CPUs: 4 Total Memory: 3.855GiB Name: moby ID: LE4E:ZN7A:OKB5:G65M:Y25C:LDQ4:MNNF:E6JR:TJ7D:CX7T:32EN:SJY2 Docker Root Dir: /var/lib/docker Debug Mode (client): false Debug Mode (server): true File Descriptors: 18 Goroutines: 30 System Time: 2017-07-21T12:59:47.84242558Z EventsListeners: 1 No Proxy: *.local, 169.254/16 Registry: https://index.docker.io/v1/ Experimental: true Insecure Registries: 127.0.0.0/8 Live Restore Enabled: false ```
2017-08-17T21:33:18Z
[]
[]
Traceback (most recent call last): File "bin/docker-compose", line 3, in <module> File "compose/cli/main.py", line 67, in main File "compose/cli/main.py", line 117, in perform_command File "compose/cli/main.py", line 937, in up File "compose/cli/log_printer.py", line 87, in run File "compose/cli/log_printer.py", line 235, in consume_queue MemoryError
4,960
docker/compose
docker__compose-5217
432dffc7106849b4d143f99059ee284f5afaacf6
diff --git a/compose/config/validation.py b/compose/config/validation.py --- a/compose/config/validation.py +++ b/compose/config/validation.py @@ -325,7 +325,6 @@ def _parse_oneof_validator(error): """ types = [] for context in error.context: - if context.validator == 'oneOf': _, error_msg = _parse_oneof_validator(context) return path_string(context.path), error_msg @@ -337,6 +336,13 @@ def _parse_oneof_validator(error): invalid_config_key = parse_key_from_error_msg(context) return (None, "contains unsupported option: '{}'".format(invalid_config_key)) + if context.validator == 'uniqueItems': + return ( + path_string(context.path) if context.path else None, + "contains non-unique items, please remove duplicates from {}".format( + context.instance), + ) + if context.path: return ( path_string(context.path), @@ -345,13 +351,6 @@ def _parse_oneof_validator(error): _parse_valid_types_from_validator(context.validator_value)), ) - if context.validator == 'uniqueItems': - return ( - None, - "contains non unique items, please remove duplicates from {}".format( - context.instance), - ) - if context.validator == 'type': types.append(context.validator_value)
compose crashes on duplicates in cache_from when a compose file has duplicate entries in `build.cache_from` section any interaction with compose crashes on reporting validation error. Example compose file: ```yml version: '3.3' services: nginx: build: context: nginx cache_from: - "example/my-nginx:develop" - "example/my-nginx:master" - "example/my-nginx:develop" image: example/my-nginx:new ``` Example output: ``` $ docker-compose -f example.yml build Traceback (most recent call last): File "/home/pupssman/venv/py2/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/cli/main.py", line 68, in main command() File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/cli/main.py", line 118, in perform_command project = project_from_options('.', options) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/cli/command.py", line 37, in project_from_options override_dir=options.get('--project-directory'), File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/cli/command.py", line 91, in get_project config_data = config.load(config_details) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/config.py", line 368, in load for config_file in config_details.config_files File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/config.py", line 534, in process_config_file validate_against_config_schema(config_file) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/validation.py", line 393, in validate_against_config_schema config_file.filename) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/validation.py", line 454, in handle_errors error_msg = '\n'.join(format_error_func(error) for error in errors) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/validation.py", line 454, in <genexpr> error_msg = '\n'.join(format_error_func(error) for error in errors) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/validation.py", line 380, in process_config_schema_errors return handle_generic_error(error, path) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/validation.py", line 252, in handle_generic_error config_key, error_msg = _parse_oneof_validator(error) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/validation.py", line 335, in _parse_oneof_validator _parse_valid_types_from_validator(context.validator_value)), File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/validation.py", line 300, in _parse_valid_types_from_validator return anglicize_json_type(validator) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/validation.py", line 216, in anglicize_json_type if json_type.startswith(('a', 'e', 'i', 'o', 'u')): AttributeError: 'bool' object has no attribute 'startswith' ``` Versions: ``` $ pip freeze | grep docker docker==2.5.1 docker-compose==1.16.1 ``` Expected behaviour: * do not crash on error reporting * extra credit: actually allow duplicates there as: * that should not pose any trouble * it is convenient to have one of those parametrized and one default to something (like `develop` here`) and pass the var from build script
Thank you for the report!
2017-09-26T23:27:15Z
[]
[]
Traceback (most recent call last): File "/home/pupssman/venv/py2/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/cli/main.py", line 68, in main command() File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/cli/main.py", line 118, in perform_command project = project_from_options('.', options) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/cli/command.py", line 37, in project_from_options override_dir=options.get('--project-directory'), File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/cli/command.py", line 91, in get_project config_data = config.load(config_details) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/config.py", line 368, in load for config_file in config_details.config_files File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/config.py", line 534, in process_config_file validate_against_config_schema(config_file) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/validation.py", line 393, in validate_against_config_schema config_file.filename) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/validation.py", line 454, in handle_errors error_msg = '\n'.join(format_error_func(error) for error in errors) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/validation.py", line 454, in <genexpr> error_msg = '\n'.join(format_error_func(error) for error in errors) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/validation.py", line 380, in process_config_schema_errors return handle_generic_error(error, path) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/validation.py", line 252, in handle_generic_error config_key, error_msg = _parse_oneof_validator(error) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/validation.py", line 335, in _parse_oneof_validator _parse_valid_types_from_validator(context.validator_value)), File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/validation.py", line 300, in _parse_valid_types_from_validator return anglicize_json_type(validator) File "/home/pupssman/venv/py2/local/lib/python2.7/site-packages/compose/config/validation.py", line 216, in anglicize_json_type if json_type.startswith(('a', 'e', 'i', 'o', 'u')): AttributeError: 'bool' object has no attribute 'startswith'
4,967
docker/compose
docker__compose-5224
7f82a2857264b4d0929c931c0ae6d6d318ce4bbd
diff --git a/compose/config/config.py b/compose/config/config.py --- a/compose/config/config.py +++ b/compose/config/config.py @@ -762,7 +762,10 @@ def process_blkio_config(service_dict): for field in ['device_read_bps', 'device_write_bps']: if field in service_dict['blkio_config']: for v in service_dict['blkio_config'].get(field, []): - v['rate'] = parse_bytes(v.get('rate', 0)) + rate = v.get('rate', 0) + v['rate'] = parse_bytes(rate) + if v['rate'] is None: + raise ConfigurationError('Invalid format for bytes value: "{}"'.format(rate)) for field in ['device_read_iops', 'device_write_iops']: if field in service_dict['blkio_config']: diff --git a/compose/utils.py b/compose/utils.py --- a/compose/utils.py +++ b/compose/utils.py @@ -12,7 +12,6 @@ from docker.errors import DockerException from docker.utils import parse_bytes as sdk_parse_bytes -from .config.errors import ConfigurationError from .errors import StreamParseError from .timeparse import MULTIPLIERS from .timeparse import timeparse @@ -143,4 +142,4 @@ def parse_bytes(n): try: return sdk_parse_bytes(n) except DockerException: - raise ConfigurationError('Invalid format for bytes value: {}'.format(n)) + return None
splitdrive import issue on docker-compose==1.16.0 Using `Python 2.7.5` with `docker-compose==1.16.0` through `pip install --upgrade docker-compose==1.16.0` ``` Python 2.7.5 (default, Nov 20 2015, 02:00:19) [GCC 4.8.5 20150623 (Red Hat 4.8.5-4)] on linux2 Type "help", "copyright", "credits" or "license" for more information. >>> from compose.utils import splitdrive Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/usr/lib/python2.7/site-packages/compose/utils.py", line 15, in <module> from .config.errors import ConfigurationError File "/usr/lib/python2.7/site-packages/compose/config/__init__.py", line 6, in <module> from .config import ConfigurationError File "/usr/lib/python2.7/site-packages/compose/config/config.py", line 15, in <module> from . import types File "/usr/lib/python2.7/site-packages/compose/config/types.py", line 17, in <module> from compose.utils import splitdrive ``` Whereas using `docker-compose==1.15.0` through `pip install --upgrade docker-compose==1.15.0` **(tested using `docker-compose==1.15.0` and `docker-compose==1.14.0` both are working)** ``` Python 2.7.5 (default, Nov 20 2015, 02:00:19) [GCC 4.8.5 20150623 (Red Hat 4.8.5-4)] on linux2 Type "help", "copyright", "credits" or "license" for more information. >>> from compose.utils import splitdrive >>> ``` --- PS: I'm not using `docker-compose` python by myself, but it used by [_Ansible_](https://github.com/ansible/ansible) [docker-service module](http://docs.ansible.com/ansible/latest/docker_service_module.html) and with `docker-compose==1.16.0` I got that message ``` "msg": "Unable to load docker-compose. Try `pip install docker-compose`. Error: cannot import name splitdrive" ```
[As I said before](https://github.com/docker/compose/issues/4542#issuecomment-283191533), we make no guarantee of a stable API for developers attempting to use Compose as a library. For those of you who are trying to use boot2docker image (docker-machine, Tiny Core Linux) on macos with ansible through its docker module... Downgrade docker-compose pip library to 1.9.0. If you're only downgrading to 1.15.0, you'll hit #4401. @shin- This isn't about importing from Ansible, I don't think. That traceback is a circular import. Importing `compose.utils` begins an import series that breaks when it tries to import `compose.utils` again. Looking to see if I can identify the breaking change.... The import that kicks off the chain was introduced recently: https://github.com/docker/compose/commit/6e802df80948696739d6aaf721d8bcf8c3bbe6a1#diff-7345a7e448db63ef1a5b786900e2b7b4R15 For those in Ansible land, this issue does not manifest in devel or the 2.4 rc's. So either limit yourself to Docker Compose <1.16 or upgrade from Ansible 2.3.x into 2.4/devel. I'm aware it is a circular import. It just does not affect the software when used as intended. I'll work on a fix when I return from vacation. Thank you for your patience. Enjoy your time off! Thanks!
2017-09-28T01:26:00Z
[]
[]
Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/usr/lib/python2.7/site-packages/compose/utils.py", line 15, in <module> from .config.errors import ConfigurationError File "/usr/lib/python2.7/site-packages/compose/config/__init__.py", line 6, in <module> from .config import ConfigurationError File "/usr/lib/python2.7/site-packages/compose/config/config.py", line 15, in <module> from . import types File "/usr/lib/python2.7/site-packages/compose/config/types.py", line 17, in <module> from compose.utils import splitdrive ``` Whereas using `docker-compose==1.15.0` through `pip install --upgrade docker-compose==1.15.0`
4,969
docker/compose
docker__compose-5274
f855ed405b8497718dcbd4e0b9a84c87bd9d8a32
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -55,7 +55,7 @@ def find_version(*file_paths): ':python_version < "3.4"': ['enum34 >= 1.0.4, < 2'], ':python_version < "3.5"': ['backports.ssl_match_hostname >= 3.5'], ':python_version < "3.3"': ['ipaddress >= 1.0.16'], - ':sys_platform == "win32"': ['colorama >= 0.3.7, < 0.4'], + ':sys_platform == "win32"': ['colorama >= 0.3.9, < 0.4'], 'socks': ['PySocks >= 1.5.6, != 1.5.7, < 2'], }
Unable to mount volumes on Windows 10: UnboundLocalError: local variable 'cells_to_erase' referenced before assignment I am getting a persistent error on `docker-compose up`. I am getting the same error on two different Windows 10 Pro machines. ``` Traceback (most recent call last): File "docker-compose", line 6, in <module> File "compose\cli\main.py", line 68, in main File "compose\cli\main.py", line 121, in perform_command File "compose\cli\main.py", line 954, in up File "compose\cli\log_printer.py", line 105, in run File "compose\cli\log_printer.py", line 109, in write File "codecs.py", line 370, in write File "site-packages\colorama\ansitowin32.py", line 40, in write File "site-packages\colorama\ansitowin32.py", line 141, in write File "site-packages\colorama\ansitowin32.py", line 167, in write_and_convert File "site-packages\colorama\ansitowin32.py", line 181, in convert_ansi File "site-packages\colorama\ansitowin32.py", line 212, in call_win32 File "site-packages\colorama\winterm.py", line 132, in erase_screen UnboundLocalError: local variable 'cells_to_erase' referenced before assignment Failed to execute script docker-compose ``` After the error the port is open and the application is running, but the volumes are not mounted. Here is my setup: - Edition: Windows 10 Pro - Version: 1703 - OS Build 15063.674 - Docker version 17.09.0-ce, build afdb6d4 - docker-compose version 1.16.1, build 6d1ac219 **Dockerfile** ``` # Use an official Ubuntu Xenial as a parent image FROM ubuntu:16.04 # Install Node.js 8 and npm 5 RUN apt-get update RUN apt-get -qq upgrade RUN apt-get install -y build-essential RUN apt-get install -y curl RUN curl -sL https://deb.nodesource.com/setup_8.x | bash RUN apt-get install -y nodejs # Set the working directory to /pretty-prism WORKDIR /pretty-prism # Copy the package.json into the container at /pretty-prism COPY package.json /pretty-prism # Install any needed packages specified in package.json RUN npm install # Define environment variable # Make port 8080 available to the world outside this container EXPOSE 8080 # Run `npm start` when the container launches CMD ["npm", "start"] ``` **docker-compose.yml** ``` version: '3' services: preact: build: . command: npm start ports: - "8080:8080" volumes: - .:/pretty-prism ```
This looks to be a bug in one of our dependencies. I created an [issue report](https://github.com/tartley/colorama/issues/147) upstream. In the meantime, You might be able to circumvent the issue using the `--no-ansi` flag. Let me know if that helps! Thanks, @shin- ! Nope, getting the same error with the flag. First, the application in the container starts and the exposed port works fine too. But then after any change in files, I get that error.
2017-10-16T23:57:47Z
[]
[]
Traceback (most recent call last): File "docker-compose", line 6, in <module> File "compose\cli\main.py", line 68, in main File "compose\cli\main.py", line 121, in perform_command File "compose\cli\main.py", line 954, in up File "compose\cli\log_printer.py", line 105, in run File "compose\cli\log_printer.py", line 109, in write File "codecs.py", line 370, in write File "site-packages\colorama\ansitowin32.py", line 40, in write File "site-packages\colorama\ansitowin32.py", line 141, in write File "site-packages\colorama\ansitowin32.py", line 167, in write_and_convert File "site-packages\colorama\ansitowin32.py", line 181, in convert_ansi File "site-packages\colorama\ansitowin32.py", line 212, in call_win32 File "site-packages\colorama\winterm.py", line 132, in erase_screen UnboundLocalError: local variable 'cells_to_erase' referenced before assignment
4,979
docker/compose
docker__compose-5341
2780559a14796ff7c34e4ce6edb6afffaccc0bf1
diff --git a/compose/config/config.py b/compose/config/config.py --- a/compose/config/config.py +++ b/compose/config/config.py @@ -709,14 +709,7 @@ def process_service(service_config): ] if 'build' in service_dict: - if isinstance(service_dict['build'], six.string_types): - service_dict['build'] = resolve_build_path(working_dir, service_dict['build']) - elif isinstance(service_dict['build'], dict): - if 'context' in service_dict['build']: - path = service_dict['build']['context'] - service_dict['build']['context'] = resolve_build_path(working_dir, path) - if 'labels' in service_dict['build']: - service_dict['build']['labels'] = parse_labels(service_dict['build']['labels']) + process_build_section(service_dict, working_dir) if 'volumes' in service_dict and service_dict.get('volume_driver') is None: service_dict['volumes'] = resolve_volume_paths(working_dir, service_dict) @@ -724,6 +717,9 @@ def process_service(service_config): if 'sysctls' in service_dict: service_dict['sysctls'] = build_string_dict(parse_sysctls(service_dict['sysctls'])) + if 'labels' in service_dict: + service_dict['labels'] = parse_labels(service_dict['labels']) + service_dict = process_depends_on(service_dict) for field in ['dns', 'dns_search', 'tmpfs']: @@ -737,6 +733,17 @@ def process_service(service_config): return service_dict +def process_build_section(service_dict, working_dir): + if isinstance(service_dict['build'], six.string_types): + service_dict['build'] = resolve_build_path(working_dir, service_dict['build']) + elif isinstance(service_dict['build'], dict): + if 'context' in service_dict['build']: + path = service_dict['build']['context'] + service_dict['build']['context'] = resolve_build_path(working_dir, path) + if 'labels' in service_dict['build']: + service_dict['build']['labels'] = parse_labels(service_dict['build']['labels']) + + def process_ports(service_dict): if 'ports' not in service_dict: return service_dict
Unable to create container with label With the following container we are getting an error creating the container: ``` dynamodb: image: instructure/dynamo-local-admin labels: - "com.acuris.service.group=companyhub" ports: - "8000:8000" ``` When we run the command: ``` docker-compose run dynamodb ``` We get the error: ``` Traceback (most recent call last): File "/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 68, in main command() File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 121, in perform_command handler(command, command_options) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 768, in run run_one_off_container(container_options, self.project, service, options) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 1178, in run_one_off_container **container_options) File "/usr/lib/python2.7/site-packages/compose/service.py", line 288, in create_container previous_container=previous_container, File "/usr/lib/python2.7/site-packages/compose/service.py", line 795, in _get_container_create_options self.config_hash if add_config_hash else None) File "/usr/lib/python2.7/site-packages/compose/service.py", line 1338, in build_container_labels labels = dict(label_options or {}) ValueError: dictionary update sequence element #0 has length 35; 2 is required ``` However, if we change the container definition to the following then it works: ``` dynamodb: image: instructure/dynamo-local-admin labels: com.acuris.service.group: companyhub ports: - "8000:8000" ``` This has only started failing since we upgraded to docker-compose version 1.17.0, build ac53b73.
Same problem here with latest release (1.17.0) ``` ubuntu@master:~$ docker-compose up ubuntu_traefik_1 is up-to-date Creating ubuntu_rancher-server_1 ... ERROR: for ubuntu_rancher-server_1 dictionary update sequence element #0 has length 23; 2 is required ERROR: for rancher-server dictionary update sequence element #0 has length 23; 2 is required Traceback (most recent call last): File "bin/docker-compose", line 6, in <module> File "compose/cli/main.py", line 68, in main File "compose/cli/main.py", line 121, in perform_command File "compose/cli/main.py", line 952, in up File "compose/project.py", line 455, in up File "compose/parallel.py", line 70, in parallel_execute ValueError: dictionary update sequence element #0 has length 23; 2 is required Failed to execute script docker-compose ``` Thanks for the report! I'll look into it asap.
2017-11-03T21:05:45Z
[]
[]
Traceback (most recent call last): File "/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 68, in main command() File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 121, in perform_command handler(command, command_options) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 768, in run run_one_off_container(container_options, self.project, service, options) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 1178, in run_one_off_container **container_options) File "/usr/lib/python2.7/site-packages/compose/service.py", line 288, in create_container previous_container=previous_container, File "/usr/lib/python2.7/site-packages/compose/service.py", line 795, in _get_container_create_options self.config_hash if add_config_hash else None) File "/usr/lib/python2.7/site-packages/compose/service.py", line 1338, in build_container_labels labels = dict(label_options or {}) ValueError: dictionary update sequence element #0 has length 35; 2 is required
4,987
docker/compose
docker__compose-5356
7208aeb002c9ecb1668729064681872481499f25
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -36,7 +36,7 @@ def find_version(*file_paths): 'requests >= 2.6.1, != 2.11.0, != 2.12.2, != 2.18.0, < 2.19', 'texttable >= 0.9.0, < 0.10', 'websocket-client >= 0.32.0, < 1.0', - 'docker >= 2.6.0, < 3.0', + 'docker >= 2.6.1, < 3.0', 'dockerpty >= 0.4.1, < 0.5', 'six >= 1.3.0, < 2', 'jsonschema >= 2.5.1, < 3',
AttributeError: 'SocketIO' object has no attribute 'raw' docker-composeup redis Creating procensus_redis_1 ... Creating procensus_redis_1 ERROR: for procensus_redis_1 'SocketIO' object has no attribute 'raw' ERROR: for redis 'SocketIO' object has no attribute 'raw' ```pytb Traceback (most recent call last): File "/home/graingert/.virtualenvs/gitlab/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/cli/main.py", line 68, in main command() File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/cli/main.py", line 121, in perform_command handler(command, command_options) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/cli/main.py", line 952, in up start=not no_start File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/project.py", line 455, in up get_deps, File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/parallel.py", line 70, in parallel_execute raise error_to_reraise File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/parallel.py", line 168, in producer result = func(obj) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/project.py", line 441, in do start=start File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/service.py", line 471, in execute_convergence_plan scale, detached, start File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/service.py", line 396, in _execute_convergence_create "Creating", File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/parallel.py", line 70, in parallel_execute raise error_to_reraise File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/parallel.py", line 168, in producer result = func(obj) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/service.py", line 394, in <lambda> lambda n: create_and_start(self, n), File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/service.py", line 387, in create_and_start container.attach_log_stream() File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/container.py", line 177, in attach_log_stream self.log_stream = self.attach(stdout=True, stderr=True, stream=True) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/container.py", line 254, in attach return self.client.attach(self.id, *args, **kwargs) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/docker/utils/decorators.py", line 19, in wrapped return f(self, resource_id, *args, **kwargs) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/docker/api/container.py", line 56, in attach response, stream, self._check_is_tty(container) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/docker/api/client.py", line 368, in _read_from_socket socket = self._get_raw_response_socket(response) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/docker/api/client.py", line 276, in _get_raw_response_socket sock = response.raw._fp.fp.raw AttributeError: 'SocketIO' object has no attribute 'raw' ```
probably an issue with docker==2.6.0 Issue raised over there: https://github.com/docker/docker-py/issues/1799
2017-11-09T00:42:17Z
[]
[]
Traceback (most recent call last): File "/home/graingert/.virtualenvs/gitlab/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/cli/main.py", line 68, in main command() File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/cli/main.py", line 121, in perform_command handler(command, command_options) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/cli/main.py", line 952, in up start=not no_start File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/project.py", line 455, in up get_deps, File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/parallel.py", line 70, in parallel_execute raise error_to_reraise File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/parallel.py", line 168, in producer result = func(obj) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/project.py", line 441, in do start=start File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/service.py", line 471, in execute_convergence_plan scale, detached, start File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/service.py", line 396, in _execute_convergence_create "Creating", File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/parallel.py", line 70, in parallel_execute raise error_to_reraise File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/parallel.py", line 168, in producer result = func(obj) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/service.py", line 394, in <lambda> lambda n: create_and_start(self, n), File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/service.py", line 387, in create_and_start container.attach_log_stream() File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/container.py", line 177, in attach_log_stream self.log_stream = self.attach(stdout=True, stderr=True, stream=True) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/compose/container.py", line 254, in attach return self.client.attach(self.id, *args, **kwargs) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/docker/utils/decorators.py", line 19, in wrapped return f(self, resource_id, *args, **kwargs) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/docker/api/container.py", line 56, in attach response, stream, self._check_is_tty(container) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/docker/api/client.py", line 368, in _read_from_socket socket = self._get_raw_response_socket(response) File "/home/graingert/.virtualenvs/gitlab/lib/python3.6/site-packages/docker/api/client.py", line 276, in _get_raw_response_socket sock = response.raw._fp.fp.raw AttributeError: 'SocketIO' object has no attribute 'raw'
4,989
docker/compose
docker__compose-5476
e8d4616723dce3946bb323314547d51f8c84d298
diff --git a/compose/cli/main.py b/compose/cli/main.py --- a/compose/cli/main.py +++ b/compose/cli/main.py @@ -511,7 +511,10 @@ def images(self, options): rows = [] for container in containers: image_config = container.image_config - repo_tags = image_config['RepoTags'][0].rsplit(':', 1) + repo_tags = ( + image_config['RepoTags'][0].rsplit(':', 1) if image_config['RepoTags'] + else ('<none>', '<none>') + ) image_id = image_config['Id'].split(':')[1][:12] size = human_readable_file_size(image_config['Size']) rows.append([
Stacktrace / traceback on `sudo docker-compose images` sudo docker-compose images Traceback (most recent call last): File "/usr/bin/docker-compose", line 11, in <module> load_entry_point('docker-compose==1.17.1', 'console_scripts', 'docker-compose')() File "/usr/lib/python3.6/site-packages/compose/cli/main.py", line 68, in main command() File "/usr/lib/python3.6/site-packages/compose/cli/main.py", line 121, in perform_command handler(command, command_options) File "/usr/lib/python3.6/site-packages/compose/cli/main.py", line 517, in images repo_tags = image_config['RepoTags'][0].rsplit(':', 1) IndexError: list index out of range
Thanks for the report! Could you share your Compose file as well? I've got this too, going to see if I can make a minimal compose file that'll trigger it e: Hmm It's not immediately obvious whats triggering this from my end ok the offending container has a RepoTags entry in its dict that's empty. I don't know if thats out of spec or? `u'RepoTags': []` Yeah, that makes sense given the error, but I'm curious because pulled images should be tagged, and Compose automatically tags images it builds. So there's a scenario I must be missing. Are you using a sha ID as the `image` value? I'm not using a sha as the image value. it's built by compose I'm pretty sure though I'm having some difficulties locating the container outside of inspect https://gist.github.com/jrabbit/dab888257a06656f0f6e0fd7d4e8327e
2017-12-12T22:46:19Z
[]
[]
Traceback (most recent call last): File "/usr/bin/docker-compose", line 11, in <module> load_entry_point('docker-compose==1.17.1', 'console_scripts', 'docker-compose')() File "/usr/lib/python3.6/site-packages/compose/cli/main.py", line 68, in main command() File "/usr/lib/python3.6/site-packages/compose/cli/main.py", line 121, in perform_command handler(command, command_options) File "/usr/lib/python3.6/site-packages/compose/cli/main.py", line 517, in images repo_tags = image_config['RepoTags'][0].rsplit(':', 1) IndexError: list index out of range
5,007
docker/compose
docker__compose-548
6580c5609c11cf90253a8a71088586c161916564
diff --git a/setup.py b/setup.py --- a/setup.py +++ b/setup.py @@ -30,7 +30,7 @@ def find_version(*file_paths): 'requests >= 2.2.1, < 3', 'texttable >= 0.8.1, < 0.9', 'websocket-client >= 0.11.0, < 0.12', - 'docker-py >= 0.5, < 0.6', + 'docker-py >= 0.5.3, < 0.6', 'six >= 1.3.0, < 2', ]
fig can not start with tls env I have installed fig 1.0.0 on my mac, but I can not start fig now. My env: Boot2Docker 1.3 on Mac ➜ figtest env |grep DOCKER DOCKER_HOST=tcp://192.168.59.103:2376 DOCKER_CERT_PATH=/Users/linux_china/.boot2docker/certs/boot2docker-vm DOCKER_TLS_VERIFY=1 ➜ figtest fig up Traceback (most recent call last): File "/usr/local/bin/fig", line 8, in <module> load_entry_point('fig==1.0.0', 'console_scripts', 'fig')() File "/Library/Python/2.7/site-packages/fig/cli/main.py", line 31, in main command.sys_dispatch() File "/Library/Python/2.7/site-packages/fig/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/Library/Python/2.7/site-packages/fig/cli/command.py", line 28, in dispatch super(Command, self).dispatch(_args, *_kwargs) File "/Library/Python/2.7/site-packages/fig/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/Library/Python/2.7/site-packages/fig/cli/command.py", line 49, in perform_command verbose=options.get('--verbose')) File "/Library/Python/2.7/site-packages/fig/cli/command.py", line 78, in get_project self.get_client(verbose=verbose)) File "/Library/Python/2.7/site-packages/fig/cli/command.py", line 54, in get_client client = docker_client() File "/Library/Python/2.7/site-packages/fig/cli/docker_client.py", line 31, in docker_client ca_cert=ca_cert, TypeError: **init**() got an unexpected keyword argument 'assert_hostname' ➜ figtest
This looks like it's using an older version of `docker-py` somehow. This feature required `docker-py>=0.5.3` How did you install fig? with pip or the binary ? What is the version in `/Library/Python/2.7/site-packages/docker/version.py` ? I think we need to bump the minimum version in setup.py to be docker-py==0.5.3. @dnephin you are right. I installed fig 1.0.0 with pip and found 0.5.3 is absent on https://pypi.python.org/pypi . Now I install the docker-py manually and fig is ok with tls support.
2014-10-17T10:21:00Z
[]
[]
Traceback (most recent call last): File "/usr/local/bin/fig", line 8, in <module> load_entry_point('fig==1.0.0', 'console_scripts', 'fig')() File "/Library/Python/2.7/site-packages/fig/cli/main.py", line 31, in main command.sys_dispatch() File "/Library/Python/2.7/site-packages/fig/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/Library/Python/2.7/site-packages/fig/cli/command.py", line 28, in dispatch super(Command, self).dispatch(_args, *_kwargs) File "/Library/Python/2.7/site-packages/fig/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/Library/Python/2.7/site-packages/fig/cli/command.py", line 49, in perform_command verbose=options.get('--verbose')) File "/Library/Python/2.7/site-packages/fig/cli/command.py", line 78, in get_project self.get_client(verbose=verbose)) File "/Library/Python/2.7/site-packages/fig/cli/command.py", line 54, in get_client client = docker_client() File "/Library/Python/2.7/site-packages/fig/cli/docker_client.py", line 31, in docker_client ca_cert=ca_cert, TypeError: **init**() got an unexpected keyword argument 'assert_hostname'
5,008
docker/compose
docker__compose-5490
9355d35c6bbdfce7e35260bad091f97bb0ae0c0a
diff --git a/compose/service.py b/compose/service.py --- a/compose/service.py +++ b/compose/service.py @@ -785,6 +785,35 @@ def _get_container_create_options( self.options.get('labels'), override_options.get('labels')) + container_options, override_options = self._build_container_volume_options( + previous_container, container_options, override_options + ) + + container_options['image'] = self.image_name + + container_options['labels'] = build_container_labels( + container_options.get('labels', {}), + self.labels(one_off=one_off), + number, + self.config_hash if add_config_hash else None) + + # Delete options which are only used in HostConfig + for key in HOST_CONFIG_KEYS: + container_options.pop(key, None) + + container_options['host_config'] = self._get_container_host_config( + override_options, + one_off=one_off) + + networking_config = self.build_default_networking_config() + if networking_config: + container_options['networking_config'] = networking_config + + container_options['environment'] = format_environment( + container_options['environment']) + return container_options + + def _build_container_volume_options(self, previous_container, container_options, override_options): container_volumes = [] container_mounts = [] if 'volumes' in container_options: @@ -801,7 +830,11 @@ def _get_container_create_options( container_options['environment'].update(affinity) container_options['volumes'] = dict((v.internal, {}) for v in container_volumes or {}) - override_options['mounts'] = [build_mount(v) for v in container_mounts] or None + if version_gte(self.client.api_version, '1.30'): + override_options['mounts'] = [build_mount(v) for v in container_mounts] or None + else: + override_options['binds'].extend(m.legacy_repr() for m in container_mounts) + container_options['volumes'].update((m.target, {}) for m in container_mounts) secret_volumes = self.get_secret_volumes() if secret_volumes: @@ -814,29 +847,7 @@ def _get_container_create_options( override_options['mounts'] = override_options.get('mounts') or [] override_options['mounts'].extend([build_mount(v) for v in secret_volumes]) - container_options['image'] = self.image_name - - container_options['labels'] = build_container_labels( - container_options.get('labels', {}), - self.labels(one_off=one_off), - number, - self.config_hash if add_config_hash else None) - - # Delete options which are only used in HostConfig - for key in HOST_CONFIG_KEYS: - container_options.pop(key, None) - - container_options['host_config'] = self._get_container_host_config( - override_options, - one_off=one_off) - - networking_config = self.build_default_networking_config() - if networking_config: - container_options['networking_config'] = networking_config - - container_options['environment'] = format_environment( - container_options['environment']) - return container_options + return container_options, override_options def _get_container_host_config(self, override_options, one_off=False): options = dict(self.options, **override_options)
API Version mismatch between docker-compose and system docker-ce **Relevant `docker-compose --verbose up` output** ``` urllib3.connectionpool._make_request: http://localhost:None "GET /v1.25/images/ormuco-vpn:compose-build/json HTTP/1.1" 200 2041 ``` Then a little bit later... ``` compose.cli.verbose_proxy.proxy_callable: docker create_host_config <- (device_read_iops=None, mem_swappiness=None, links=[], oom_score_adj=None, blkio_weight=None, cpu_count=None, cpuset_cpus=None, dns_search=None, pid_mode=None, init_path=None, log_config={'Type': u'', 'Config': {}}, cpu_quota=None, read_only=None, cpu_percent=None, device_read_bps=None, storage_opt=None, init=None, dns=None, volumes_from=[], ipc_mode=None, mem_reservation=None, security_opt=None, shm_size=None, device_write_iops=None, dns_opt=None, cgroup_parent=None, group_add=None, network_mode=u'cerebrom_default', volume_driver=None, oom_kill_disable=None, userns_mode=None, tmpfs=None, nano_cpus=None, port_bindings={'9600/tcp': [None]}, isolation=None, memswap_limit=None, restart_policy=None, blkio_weight_device=None, devices=None, extra_hosts=None, binds=[], sysctls=None, pids_limit=None, device_write_bps=None, cap_add=None, mounts=[{'Source': u'/home/wayne/projects/ormuco/cerebrom/docker/ormuco-vpn/credentials/vpnc', 'ReadOnly': None, 'Type': 'bind', 'Target': '/etc/vpnc'}, {'Source': u'/home/wayne/projects/ormuco/cerebrom/docker/ormuco-vpn/credentials/ssh', 'ReadOnly': None, 'Type': 'bind', 'Target': '/ssh'}], mem_limit=None, cap_drop=None, privileged=False, ulimits=None, cpu_shares=None) compose.parallel.parallel_execute_iter: Failed: <Container: 21d4214f35f5_cerebrom_ormuco-vpn_1 (21d421)> compose.parallel.feed_queue: Pending: set([]) ERROR: for 21d4214f35f5_cerebrom_ormuco-vpn_1 mounts param is not supported in API versions < 1.30 compose.parallel.parallel_execute_iter: Failed: <Service: ormuco-vpn> compose.parallel.feed_queue: Pending: set([<Service: proctor>]) compose.parallel.feed_queue: <Service: proctor> has upstream errors - not processing compose.parallel.parallel_execute_iter: Failed: <Service: proctor> compose.parallel.feed_queue: Pending: set([]) ERROR: for ormuco-vpn mounts param is not supported in API versions < 1.30 Traceback (most recent call last): File "/home/wayne/.virtualenvs/cerebrom-5qiep0Vx/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/home/wayne/.virtualenvs/cerebrom-5qiep0Vx/local/lib/python2.7/site-packages/compose/cli/main.py", line 71, in main command() File "/home/wayne/.virtualenvs/cerebrom-5qiep0Vx/local/lib/python2.7/site-packages/compose/cli/main.py", line 124, in perform_command handler(command, command_options) File "/home/wayne/.virtualenvs/cerebrom-5qiep0Vx/local/lib/python2.7/site-packages/compose/cli/main.py", line 956, in up start=not no_start File "/home/wayne/.virtualenvs/cerebrom-5qiep0Vx/local/lib/python2.7/site-packages/compose/project.py", line 479, in up get_deps, File "/home/wayne/.virtualenvs/cerebrom-5qiep0Vx/local/lib/python2.7/site-packages/compose/parallel.py", line 80, in parallel_execute raise error_to_reraise docker.errors.InvalidVersion: mounts param is not supported in API versions < 1.30 ``` **`docker version` output** ``` Client: Version: 17.09.0-ce API version: 1.32 Go version: go1.8.3 Git commit: afdb6d4 Built: Tue Sep 26 22:42:09 2017 OS/Arch: linux/amd64 Server: Version: 17.09.0-ce API version: 1.32 (minimum version 1.12) Go version: go1.8.3 Git commit: afdb6d4 Built: Tue Sep 26 22:40:48 2017 OS/Arch: linux/amd64 Experimental: false ``` **`docker info` output** ``` Containers: 13 Running: 4 Paused: 0 Stopped: 9 Images: 245 Server Version: 17.09.0-ce Storage Driver: overlay Backing Filesystem: extfs Supports d_type: true Logging Driver: json-file Cgroup Driver: cgroupfs Plugins: Volume: local Network: bridge host macvlan null overlay Log: awslogs fluentd gcplogs gelf journald json-file logentries splunk syslog Swarm: inactive Runtimes: runc Default Runtime: runc Init Binary: docker-init containerd version: 06b9cb35161009dcb7123345749fef02f7cea8e0 runc version: 3f2f8b84a77f73d38244dd690525642a72156c64 init version: 949e6fa Security Options: seccomp Profile: default Kernel Version: 4.9.0-3-amd64 Operating System: Debian GNU/Linux 9 (stretch) OSType: linux Architecture: x86_64 CPUs: 12 Total Memory: 23.5GiB Name: agora ID: UW3Z:67B4:Y2UM:YQS6:RD6G:7IMS:NE5R:3P32:27YZ:ZJQD:44ZA:QVX5 Docker Root Dir: /var/lib/docker Debug Mode (client): false Debug Mode (server): false Registry: https://index.docker.io/v1/ Experimental: false Insecure Registries: 127.0.0.0/8 Live Restore Enabled: false ``` **`docker-compose --version` output** ``` docker-compose version 1.18.0-rc2, build 189468b ``` **my thoughts** It seems strange to me that `docker-compose` attempts to use API version `1.25` rather than `1.32` as reported by `docker version` (see above). Is it because `docker-compose` `1.18.0-rc2` simply doesn't support API version `1.32`? Or am I doing something incorrectly that causes `docker-compose` to use API version `1.25`? Also it's worth mentioning that I am attempting to use https://docs.docker.com/compose/compose-file/#long-syntax-3; whenever I exclude the `volumes` key from the service this problem affects the api version conflict doesn't occur. I guess the next thing I'll try is https://docs.docker.com/compose/compose-file/#short-syntax-3.
Thanks for the report - do you mind sharing your Compose file as well? @shin- I don't think I'm allowed to share the compose file i've been working with but here is a minimal reproduction case: ``` version: "3.2" services: mysql: image: "mysql:5.7" volumes: - type: bind source: ./mysql_data target: /var/lib/mysql environment: - MYSQL_USER=whatever - MYSQL_PASSWORD=changeme - MYSQL_ROOT_PASSWORD=changeme - MYSQL_DATABASE=dbname ``` However, the following does not reproduce: ``` version: "3.2" services: mysql: image: "mysql:5.7" volumes: - './mysql_data:/var/lib/mysql' environment: - MYSQL_USER=whatever - MYSQL_PASSWORD=changeme - MYSQL_ROOT_PASSWORD=changeme - MYSQL_DATABASE=dbname ``` So in other words, using the short-syntax as I mentioned in my original comment solves the problem I was seeing. Thanks! I'll make sure to get that fixed in time for GA. I'd also be happy to take a stab at it after work some night this week if you think it's something a new contributor could take on with a few pointers. Thanks @waynr - I really appreciate the offer! We're a bit far into the release cycle this time around and with the holiday coming up, I want to take care of this ASAP, but in general I'm always elated to see bugfix PRs coming in :+1: Also, as a workaround, the error you're seeing should clear up if you use version `3.3` or `3.4` instead. HTH!
2017-12-18T20:41:23Z
[]
[]
Traceback (most recent call last): File "/home/wayne/.virtualenvs/cerebrom-5qiep0Vx/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/home/wayne/.virtualenvs/cerebrom-5qiep0Vx/local/lib/python2.7/site-packages/compose/cli/main.py", line 71, in main command() File "/home/wayne/.virtualenvs/cerebrom-5qiep0Vx/local/lib/python2.7/site-packages/compose/cli/main.py", line 124, in perform_command handler(command, command_options) File "/home/wayne/.virtualenvs/cerebrom-5qiep0Vx/local/lib/python2.7/site-packages/compose/cli/main.py", line 956, in up start=not no_start File "/home/wayne/.virtualenvs/cerebrom-5qiep0Vx/local/lib/python2.7/site-packages/compose/project.py", line 479, in up get_deps, File "/home/wayne/.virtualenvs/cerebrom-5qiep0Vx/local/lib/python2.7/site-packages/compose/parallel.py", line 80, in parallel_execute raise error_to_reraise docker.errors.InvalidVersion: mounts param is not supported in API versions < 1.30
5,009
docker/compose
docker__compose-5537
bcc13d7fae223c49322bb1218e62c9bcecc2f35d
diff --git a/compose/cli/command.py b/compose/cli/command.py --- a/compose/cli/command.py +++ b/compose/cli/command.py @@ -10,6 +10,7 @@ from . import errors from . import verbose_proxy from .. import config +from .. import parallel from ..config.environment import Environment from ..const import API_VERSIONS from ..project import Project @@ -23,6 +24,8 @@ def project_from_options(project_dir, options): environment = Environment.from_env_file(project_dir) + set_parallel_limit(environment) + host = options.get('--host') if host is not None: host = host.lstrip('=') @@ -38,6 +41,22 @@ def project_from_options(project_dir, options): ) +def set_parallel_limit(environment): + parallel_limit = environment.get('COMPOSE_PARALLEL_LIMIT') + if parallel_limit: + try: + parallel_limit = int(parallel_limit) + except ValueError: + raise errors.UserError( + 'COMPOSE_PARALLEL_LIMIT must be an integer (found: "{}")'.format( + environment.get('COMPOSE_PARALLEL_LIMIT') + ) + ) + if parallel_limit <= 1: + raise errors.UserError('COMPOSE_PARALLEL_LIMIT can not be less than 2') + parallel.GlobalLimit.set_global_limit(parallel_limit) + + def get_config_from_options(base_dir, options): environment = Environment.from_env_file(base_dir) config_path = get_config_path_from_options( diff --git a/compose/const.py b/compose/const.py --- a/compose/const.py +++ b/compose/const.py @@ -18,6 +18,7 @@ LABEL_VOLUME = 'com.docker.compose.volume' LABEL_CONFIG_HASH = 'com.docker.compose.config-hash' NANOCPUS_SCALE = 1000000000 +PARALLEL_LIMIT = 64 SECRETS_PATH = '/run/secrets' diff --git a/compose/parallel.py b/compose/parallel.py --- a/compose/parallel.py +++ b/compose/parallel.py @@ -15,6 +15,7 @@ from compose.cli.colors import green from compose.cli.colors import red from compose.cli.signals import ShutdownException +from compose.const import PARALLEL_LIMIT from compose.errors import HealthCheckFailed from compose.errors import NoHealthCheckConfigured from compose.errors import OperationFailedError @@ -26,6 +27,20 @@ STOP = object() +class GlobalLimit(object): + """Simple class to hold a global semaphore limiter for a project. This class + should be treated as a singleton that is instantiated when the project is. + """ + + global_limiter = Semaphore(PARALLEL_LIMIT) + + @classmethod + def set_global_limit(cls, value): + if value is None: + value = PARALLEL_LIMIT + cls.global_limiter = Semaphore(value) + + def parallel_execute(objects, func, get_name, msg, get_deps=None, limit=None, parent_objects=None): """Runs func on objects in parallel while ensuring that func is ran on object only after it is ran on all its dependencies. @@ -173,7 +188,7 @@ def producer(obj, func, results, limiter): The entry point for a producer thread which runs func on a single object. Places a tuple on the results queue once func has either returned or raised. """ - with limiter: + with limiter, GlobalLimit.global_limiter: try: result = func(obj) results.put((obj, result, None))
Using "scale" exhausts system limits and then errors out # Problem Its possible to exhaust system resources by doing something like ``` $ docker-compose scale web=1000 ``` It appears that we try to execute this request in parallel, causing the system limit of open files (in this case) to be exceed. ``` Exception in thread Thread-5: Traceback (most recent call last): File "/compose/build/docker-compose/out00-PYZ.pyz/threading", line 810, in __bootstrap_inner File "/compose/build/docker-compose/out00-PYZ.pyz/threading", line 763, in run File "/compose/build/docker-compose/out00-PYZ.pyz/compose.utils", line 31, in inner_execute_function File "/compose/build/docker-compose/out00-PYZ.pyz/compose.service", line 220, in <lambda> File "/compose/build/docker-compose/out00-PYZ.pyz/compose.service", line 173, in create_and_start File "/compose/build/docker-compose/out00-PYZ.pyz/compose.service", line 268, in create_container File "/compose/build/docker-compose/out00-PYZ.pyz/compose.service", line 579, in _get_container_create_options File "/compose/build/docker-compose/out00-PYZ.pyz/compose.service", line 464, in config_hash File "/compose/build/docker-compose/out00-PYZ.pyz/compose.service", line 469, in config_dict File "/compose/build/docker-compose/out00-PYZ.pyz/compose.service", line 295, in image File "/compose/build/docker-compose/out00-PYZ.pyz/docker.utils.decorators", line 20, in wrapped File "/compose/build/docker-compose/out00-PYZ.pyz/docker.client", line 489, in inspect_image File "/compose/build/docker-compose/out00-PYZ.pyz/docker.clientbase", line 86, in _get File "/compose/build/docker-compose/out00-PYZ.pyz/requests.sessions", line 477, in get File "/compose/build/docker-compose/out00-PYZ.pyz/requests.sessions", line 465, in request File "/compose/build/docker-compose/out00-PYZ.pyz/requests.sessions", line 573, in send File "/compose/build/docker-compose/out00-PYZ.pyz/requests.adapters", line 370, in send File "/compose/build/docker-compose/out00-PYZ.pyz/requests.packages.urllib3.connectionpool", line 544, in urlopen File "/compose/build/docker-compose/out00-PYZ.pyz/requests.packages.urllib3.connectionpool", line 341, in _make_request File "/compose/build/docker-compose/out00-PYZ.pyz/requests.packages.urllib3.connectionpool", line 761, in _validate_conn File "/compose/build/docker-compose/out00-PYZ.pyz/requests.packages.urllib3.connection", line 238, in connect File "/compose/build/docker-compose/out00-PYZ.pyz/requests.packages.urllib3.util.ssl_", line 277, in ssl_wrap_socket IOError: [Errno 24] Too many open files ``` # Solution Possible solutions - inspect the system limits and work within those limits - limit the parallelization that compose uses # Workaround Batch the scaling from the command line e.g. `docker-compose scale web=10` # Environment ``` $ docker-compose --version docker-compose version: 1.4.0rc3 ``` ``` $ uname -a Darwin vero.local 14.4.0 Darwin Kernel Version 14.4.0: Thu May 28 11:35:04 PDT 2015; root:xnu-2782.30.5~1/RELEASE_X86_64 x86_64 ``` ``` $ ulimit -a core file size (blocks, -c) 0 data seg size (kbytes, -d) unlimited file size (blocks, -f) unlimited max locked memory (kbytes, -l) unlimited max memory size (kbytes, -m) unlimited open files (-n) 256 pipe size (512 bytes, -p) 1 stack size (kbytes, -s) 8192 cpu time (seconds, -t) unlimited max user processes (-u) 709 virtual memory (kbytes, -v) unlimited ```
Rather than dynamically pick a limit, it'd be simpler to arbitrarily hardcode one (with an environment variable to override it, as was done in an older implementation). 64 concurrent operations should be enough for anyone. As per #5422, I have another use case: when a developer is starting all these containers on a laptop. In that case, it makes sense to stagger the spawning of the containers so they don't thrash competing for resources.
2018-01-04T20:20:29Z
[]
[]
Traceback (most recent call last): File "/compose/build/docker-compose/out00-PYZ.pyz/threading", line 810, in __bootstrap_inner File "/compose/build/docker-compose/out00-PYZ.pyz/threading", line 763, in run File "/compose/build/docker-compose/out00-PYZ.pyz/compose.utils", line 31, in inner_execute_function File "/compose/build/docker-compose/out00-PYZ.pyz/compose.service", line 220, in <lambda> File "/compose/build/docker-compose/out00-PYZ.pyz/compose.service", line 173, in create_and_start File "/compose/build/docker-compose/out00-PYZ.pyz/compose.service", line 268, in create_container File "/compose/build/docker-compose/out00-PYZ.pyz/compose.service", line 579, in _get_container_create_options File "/compose/build/docker-compose/out00-PYZ.pyz/compose.service", line 464, in config_hash File "/compose/build/docker-compose/out00-PYZ.pyz/compose.service", line 469, in config_dict File "/compose/build/docker-compose/out00-PYZ.pyz/compose.service", line 295, in image File "/compose/build/docker-compose/out00-PYZ.pyz/docker.utils.decorators", line 20, in wrapped File "/compose/build/docker-compose/out00-PYZ.pyz/docker.client", line 489, in inspect_image File "/compose/build/docker-compose/out00-PYZ.pyz/docker.clientbase", line 86, in _get File "/compose/build/docker-compose/out00-PYZ.pyz/requests.sessions", line 477, in get File "/compose/build/docker-compose/out00-PYZ.pyz/requests.sessions", line 465, in request File "/compose/build/docker-compose/out00-PYZ.pyz/requests.sessions", line 573, in send File "/compose/build/docker-compose/out00-PYZ.pyz/requests.adapters", line 370, in send File "/compose/build/docker-compose/out00-PYZ.pyz/requests.packages.urllib3.connectionpool", line 544, in urlopen File "/compose/build/docker-compose/out00-PYZ.pyz/requests.packages.urllib3.connectionpool", line 341, in _make_request File "/compose/build/docker-compose/out00-PYZ.pyz/requests.packages.urllib3.connectionpool", line 761, in _validate_conn File "/compose/build/docker-compose/out00-PYZ.pyz/requests.packages.urllib3.connection", line 238, in connect File "/compose/build/docker-compose/out00-PYZ.pyz/requests.packages.urllib3.util.ssl_", line 277, in ssl_wrap_socket IOError: [Errno 24] Too many open files
5,014
docker/compose
docker__compose-554
2efb4f5be0e4f0df9340b4769524cd470e53017c
diff --git a/fig/cli/docker_client.py b/fig/cli/docker_client.py --- a/fig/cli/docker_client.py +++ b/fig/cli/docker_client.py @@ -11,7 +11,7 @@ def docker_client(): """ cert_path = os.environ.get('DOCKER_CERT_PATH', '') if cert_path == '': - cert_path = os.path.join(os.environ.get('HOME'), '.docker') + cert_path = os.path.join(os.environ.get('HOME', ''), '.docker') base_url = os.environ.get('DOCKER_HOST') tls_config = None diff --git a/fig/progress_stream.py b/fig/progress_stream.py --- a/fig/progress_stream.py +++ b/fig/progress_stream.py @@ -19,7 +19,9 @@ def stream_output(output, stream): all_events.append(event) if 'progress' in event or 'progressDetail' in event: - image_id = event['id'] + image_id = event.get('id') + if not image_id: + continue if image_id in lines: diff = len(lines) - lines[image_id]
Error in stream_output during ADD command in Dockerfile ``` Step 5 : ADD https://rtdata.dtcc.com/payload.tgz /tmp/ Traceback (most recent call last): File "/usr/local/bin/fig", line 9, in <module> load_entry_point('fig==0.5.2', 'console_scripts', 'fig')() File "/home/vagrant/src/fig/fig/cli/main.py", line 31, in main command.sys_dispatch() File "/home/vagrant/src/fig/fig/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/home/vagrant/src/fig/fig/cli/command.py", line 28, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/home/vagrant/src/fig/fig/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/home/vagrant/src/fig/fig/cli/command.py", line 49, in perform_command handler(project, command_options) File "/home/vagrant/src/fig/fig/cli/main.py", line 119, in build project.build(service_names=options['SERVICE'], no_cache=no_cache) File "/home/vagrant/src/fig/fig/project.py", line 166, in build service.build(no_cache) File "/home/vagrant/src/fig/fig/service.py", line 404, in build all_events = stream_output(build_output, sys.stdout) File "/home/vagrant/src/fig/fig/progress_stream.py", line 22, in stream_output image_id = event['id'] KeyError: 'id' ``` branch master More details available or forthcoming (still investigating).
It looks like the output from the docker remote is not what fig expected. What version of Docker are you using? I believe fig expects 1.0 or later. Docker version 1.3.0, build c78088f I tried an `ADD https://...tar.gz` and fig build worked with both docker 1.0 and 1.2. It's possible the api changed in 1.3 (this part of the api is not well documented). Would it be possible to add a `print event` before this line (line 22) to get some more debug information? ``` event is: {u'status': u'Downloading', u'progressDetail': {u'current': 15988, u'start': 1413495811, u'total': 1545876}, u'progress': u'[> ] 15.99 kB/1.546 MB 8s'} ``` Other idea, maybe it's docker-py's fault? Further idea: switch to `RUN wget`. I think the reason I can't reproduce is because the tarball I was using was small enough that it never send a "status: downloading" line. I will have to try again with a larger file, it may not be docker 1.3 related at all I can confirm this is a change in behaviour with docker 1.3. I tried with a large `.tar.gz` on docker 1.2, and there was no crash. In 1.3 the API for this endpoint has changed, and it's sending back a different response. I think there are actually two things to do here: 1. prevent the crash 2. actually use the new output to display a progress status bar for these downloads (possibly create a new issue for this)
2014-10-18T17:57:53Z
[]
[]
Traceback (most recent call last): File "/usr/local/bin/fig", line 9, in <module> load_entry_point('fig==0.5.2', 'console_scripts', 'fig')() File "/home/vagrant/src/fig/fig/cli/main.py", line 31, in main command.sys_dispatch() File "/home/vagrant/src/fig/fig/cli/docopt_command.py", line 21, in sys_dispatch self.dispatch(sys.argv[1:], None) File "/home/vagrant/src/fig/fig/cli/command.py", line 28, in dispatch super(Command, self).dispatch(*args, **kwargs) File "/home/vagrant/src/fig/fig/cli/docopt_command.py", line 24, in dispatch self.perform_command(*self.parse(argv, global_options)) File "/home/vagrant/src/fig/fig/cli/command.py", line 49, in perform_command handler(project, command_options) File "/home/vagrant/src/fig/fig/cli/main.py", line 119, in build project.build(service_names=options['SERVICE'], no_cache=no_cache) File "/home/vagrant/src/fig/fig/project.py", line 166, in build service.build(no_cache) File "/home/vagrant/src/fig/fig/service.py", line 404, in build all_events = stream_output(build_output, sys.stdout) File "/home/vagrant/src/fig/fig/progress_stream.py", line 22, in stream_output image_id = event['id'] KeyError: 'id'
5,015
docker/compose
docker__compose-5552
c4fda0834d0fd70d9a9f4334ce94b8d88f8eb03c
diff --git a/compose/config/interpolation.py b/compose/config/interpolation.py --- a/compose/config/interpolation.py +++ b/compose/config/interpolation.py @@ -137,6 +137,8 @@ def convert(mo): if named is not None: val = mapping[named] + if isinstance(val, six.binary_type): + val = val.decode('utf-8') return '%s' % (val,) if mo.group('escaped') is not None: return self.delimiter diff --git a/compose/config/serialize.py b/compose/config/serialize.py --- a/compose/config/serialize.py +++ b/compose/config/serialize.py @@ -27,6 +27,9 @@ def serialize_string(dumper, data): """ Ensure boolean-like strings are quoted in the output and escape $ characters """ representer = dumper.represent_str if six.PY3 else dumper.represent_unicode + if isinstance(data, six.binary_type): + data = data.decode('utf-8') + data = data.replace('$', '$$') if data.lower() in ('y', 'n', 'yes', 'no', 'on', 'off', 'true', 'false'): @@ -90,7 +93,8 @@ def serialize_config(config, image_digests=None): denormalize_config(config, image_digests), default_flow_style=False, indent=2, - width=80 + width=80, + allow_unicode=True )
docker-compose config fails with stack trace It works on my machine (docker-compose version 1.17.1, build 6d101fb) but not on my CI (docker-compose version 1.18.0, build 8dd22a9). `docker-compose -f docker-compose.yml -f docker-compose.env.yml config > docker-compose.deploy.yml` `docker-compose.yml` ``` version: '3.3' services: botrouter: environment: - BOTS image: ${IMAGE:-botrouter}:${TAG:-latest} labels: - "traefik.backend=production-bot-router" - "traefik.enable=true" - "traefik.frontend.rule=Host: bot.domain.com" - "traefik.port=80" restart: always ``` `docker-compose.env.yml` ``` version: '3.3' services: botrouter: labels: traefik.backend: "${ENV}-bot-router" traefik.frontend.rule: "Host: ${ENV}.bot.domain.com" ``` Output: ``` Traceback (most recent call last): File "/usr/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 71, in main command() File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 118, in perform_command handler(command, options, command_options) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 322, in config print(serialize_config(compose_config, image_digests)) File "/usr/lib/python2.7/site-packages/compose/config/serialize.py", line 93, in serialize_config width=80 File "/usr/lib/python2.7/site-packages/yaml/__init__.py", line 218, in safe_dump return dump_all([data], stream, Dumper=SafeDumper, **kwds) File "/usr/lib/python2.7/site-packages/yaml/__init__.py", line 190, in dump_all dumper.represent(data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 28, in represent node = self.represent_data(data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 57, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 225, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 123, in represent_mapping node_value = self.represent_data(item_value) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 57, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 225, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 123, in represent_mapping node_value = self.represent_data(item_value) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 57, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 225, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 123, in represent_mapping node_value = self.represent_data(item_value) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 57, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 225, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 123, in represent_mapping node_value = self.represent_data(item_value) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 57, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/usr/lib/python2.7/site-packages/compose/config/serialize.py", line 30, in serialize_string data = data.replace('$', '$$') UnicodeDecodeError: ‘ascii’ codec can't decode byte 0xe2 in position 1: ordinal not in range(128) ```
@yangm97 thanks! have you tried downgrading to 1.17.1 on your CI? Do you still see the issue? Downgrading to 1.171 fixes the issue indeed.
2018-01-10T00:51:00Z
[]
[]
Traceback (most recent call last): File "/usr/bin/docker-compose", line 11, in <module> sys.exit(main()) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 71, in main command() File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 118, in perform_command handler(command, options, command_options) File "/usr/lib/python2.7/site-packages/compose/cli/main.py", line 322, in config print(serialize_config(compose_config, image_digests)) File "/usr/lib/python2.7/site-packages/compose/config/serialize.py", line 93, in serialize_config width=80 File "/usr/lib/python2.7/site-packages/yaml/__init__.py", line 218, in safe_dump return dump_all([data], stream, Dumper=SafeDumper, **kwds) File "/usr/lib/python2.7/site-packages/yaml/__init__.py", line 190, in dump_all dumper.represent(data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 28, in represent node = self.represent_data(data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 57, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 225, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 123, in represent_mapping node_value = self.represent_data(item_value) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 57, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 225, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 123, in represent_mapping node_value = self.represent_data(item_value) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 57, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 225, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 123, in represent_mapping node_value = self.represent_data(item_value) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 57, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 225, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 123, in represent_mapping node_value = self.represent_data(item_value) File "/usr/lib/python2.7/site-packages/yaml/representer.py", line 57, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/usr/lib/python2.7/site-packages/compose/config/serialize.py", line 30, in serialize_string data = data.replace('$', '$$') UnicodeDecodeError: ‘ascii’ codec can't decode byte 0xe2 in position 1: ordinal not in range(128)
5,017