-
-
Notifications
You must be signed in to change notification settings - Fork 1.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Not all markers added to output of pipenv lock --requirements #2748
Comments
I think this is the same as #2662? |
techalchemy
added a commit
that referenced
this issue
Sep 3, 2018
- Fixes #2748 Signed-off-by: Dan Ryan <dan@danryan.co>
techalchemy
added a commit
that referenced
this issue
Sep 4, 2018
- Fixes #2748 Signed-off-by: Dan Ryan <dan@danryan.co>
uranusjr
pushed a commit
that referenced
this issue
Sep 5, 2018
- Fixes #2748 Signed-off-by: Dan Ryan <dan@danryan.co>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Issue description
If an installed package has a marker consisting of multiple expressions connected with and/or then only the last expression will be added to the output of
pipenv lock -r
.Excerpt from Pipfile.lock:
Output of
pipenv lock -r
:It should be noted that due to #2662 it is possible for the markers to change order in Pipfile.lock, which in turn changes which marker appears in this output.
Expected result
Actual result
Extra info
This is the relevant line from the METADATA file of urllib3:
$ pipenv --support
Pipenv version:
'2018.7.1'
Pipenv location:
'/usr/lib/python3.4/site-packages/pipenv'
Python location:
'/usr/bin/python3.4'
Other Python installations in
PATH
:2.6
:/usr/bin/python2.6
2.6
:/usr/bin/python2.6
3.4
:/usr/bin/python3.4m
3.4
:/usr/bin/python3.4
3.6
:/home/gecko/.pyenv/shims/python3.6
2.6.6
:/home/gecko/.pyenv/shims/python
2.6.6
:/usr/bin/python
2.6.6
:/usr/bin/python2
3.4.5
:/home/gecko/.pyenv/shims/python3
3.4.5
:/usr/bin/python3
PEP 508 Information:
System environment variables:
PYENV_VIRTUALENV_INIT
JAVA_HOME
LESSOPEN
SSH_CLIENT
PYTHONDONTWRITEBYTECODE
SHLVL
SSH_CONNECTION
CVS_RSH
PYTHON_VERSION
PWD
LANG
PYTHONPATH
DISPLAY
QTINC
G_BROKEN_FILENAMES
HISTCONTROL
MAIL
HOME
TERM
PYENV_SHELL
LS_COLORS
PATH
OLDPWD
TMUX_PANE
QTDIR
SSH_TTY
HISTSIZE
PIP_PYTHON_PATH
USER
SHELL
TMUX
HOSTNAME
EDITOR
SPARK_HOME
QTLIB
LOGNAME
_
Pipenv–specific environment variables:
Debug–specific environment variables:
PATH
:/home/gecko/.pyenv/plugins/pyenv-virtualenv/shims:/home/gecko/.pyenv/shims:/home/gecko/.pyenv/bin:/home/gecko/.cargo/bin:/usr/lib/jvm/java/bin:/usr/lib64/qt-3.3/bin:/usr/lib/jvm/java/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/usr/hdp/current/falcon-client/bin:/usr/hdp/current/hadoop-mapreduce-historyserver/bin:/usr/hdp/current/oozie-client/bin:/usr/hdp/current/falcon-server/bin:/usr/hdp/current/hadoop-yarn-client/bin:/usr/hdp/current/oozie-server/bin:/usr/hdp/current/flume-client/bin:/usr/hdp/current/hadoop-yarn-nodemanager/bin:/usr/hdp/current/pig-client/bin:/usr/hdp/current/flume-server/bin:/usr/hdp/current/hadoop-yarn-resourcemanager/bin:/usr/hdp/current/slider-client/bin:/usr/hdp/current/hadoop-client/bin:/usr/hdp/current/hadoop-yarn-timelineserver/bin:/usr/hdp/current/sqoop-client/bin:/usr/hdp/current/hadoop-hdfs-client/bin:/usr/hdp/current/hbase-client/bin:/usr/hdp/current/sqoop-server/bin:/usr/hdp/current/hadoop-hdfs-datanode/bin:/usr/hdp/current/hbase-master/bin:/usr/hdp/current/storm-client/bin:/usr/hdp/current/hadoop-hdfs-journalnode/bin:/usr/hdp/current/hbase-regionserver/bin:/usr/hdp/current/storm-nimbus/bin:/usr/hdp/current/hadoop-hdfs-namenode/bin:/usr/hdp/current/hive-client/bin:/usr/hdp/current/storm-supervisor/bin:/usr/hdp/current/hadoop-hdfs-nfs3/bin:/usr/hdp/current/hive-metastore/bin:/usr/hdp/current/zookeeper-client/bin:/usr/hdp/current/hadoop-hdfs-portmap/bin:/usr/hdp/current/hive-server2/bin:/usr/hdp/current/zookeeper-server/bin:/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin:/usr/hdp/current/hive-webhcat/bin:/usr/hdp/current/hadoop-mapreduce-client/bin:/usr/hdp/current/knox-server/bin:/usr/hdp/current/hadoop-client/sbin:/usr/hdp/current/hadoop-hdfs-nfs3/sbin:/usr/hdp/current/hadoop-yarn-client/sbin:/usr/hdp/current/hadoop-hdfs-client/sbin:/usr/hdp/current/hadoop-hdfs-portmap/sbin:/usr/hdp/current/hadoop-yarn-nodemanager/sbin:/usr/hdp/current/hadoop-hdfs-datanode/sbin:/usr/hdp/current/hadoop-hdfs-secondarynamenode/sbin:/usr/hdp/current/hadoop-yarn-resourcemanager/sbin:/usr/hdp/current/hadoop-hdfs-journalnode/sbin:/usr/hdp/current/hadoop-mapreduce-client/sbin:/usr/hdp/current/hadoop-yarn-timelineserver/sbin:/usr/hdp/current/hadoop-hdfs-namenode/sbin:/usr/hdp/current/hadoop-mapreduce-historyserver/sbin:/usr/hdp/current/hive-webhcat/sbin:/usr/hdp/current/falcon-client/bin:/usr/hdp/current/hadoop-mapreduce-historyserver/bin:/usr/hdp/current/oozie-client/bin:/usr/hdp/current/falcon-server/bin:/usr/hdp/current/hadoop-yarn-client/bin:/usr/hdp/current/oozie-server/bin:/usr/hdp/current/flume-client/bin:/usr/hdp/current/hadoop-yarn-nodemanager/bin:/usr/hdp/current/pig-client/bin:/usr/hdp/current/flume-server/bin:/usr/hdp/current/hadoop-yarn-resourcemanager/bin:/usr/hdp/current/slider-client/bin:/usr/hdp/current/hadoop-client/bin:/usr/hdp/current/hadoop-yarn-timelineserver/bin:/usr/hdp/current/sqoop-client/bin:/usr/hdp/current/hadoop-hdfs-client/bin:/usr/hdp/current/hbase-client/bin:/usr/hdp/current/sqoop-server/bin:/usr/hdp/current/hadoop-hdfs-datanode/bin:/usr/hdp/current/hbase-master/bin:/usr/hdp/current/storm-client/bin:/usr/hdp/current/hadoop-hdfs-journalnode/bin:/usr/hdp/current/hbase-regionserver/bin:/usr/hdp/current/storm-nimbus/bin:/usr/hdp/current/hadoop-hdfs-namenode/bin:/usr/hdp/current/hive-client/bin:/usr/hdp/current/storm-supervisor/bin:/usr/hdp/current/hadoop-hdfs-nfs3/bin:/usr/hdp/current/hive-metastore/bin:/usr/hdp/current/zookeeper-client/bin:/usr/hdp/current/hadoop-hdfs-portmap/bin:/usr/hdp/current/hive-server2/bin:/usr/hdp/current/zookeeper-server/bin:/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin:/usr/hdp/current/hive-webhcat/bin:/usr/hdp/current/hadoop-mapreduce-client/bin:/usr/hdp/current/knox-server/bin:/usr/hdp/current/hadoop-client/sbin:/usr/hdp/current/hadoop-hdfs-nfs3/sbin:/usr/hdp/current/hadoop-yarn-client/sbin:/usr/hdp/current/hadoop-hdfs-client/sbin:/usr/hdp/current/hadoop-hdfs-portmap/sbin:/usr/hdp/current/hadoop-yarn-nodemanager/sbin:/usr/hdp/current/hadoop-hdfs-datanode/sbin:/usr/hdp/current/hadoop-hdfs-secondarynamenode/sbin:/usr/hdp/current/hadoop-yarn-resourcemanager/sbin:/usr/hdp/current/hadoop-hdfs-journalnode/sbin:/usr/hdp/current/hadoop-mapreduce-client/sbin:/usr/hdp/current/hadoop-yarn-timelineserver/sbin:/usr/hdp/current/hadoop-hdfs-namenode/sbin:/usr/hdp/current/hadoop-mapreduce-historyserver/sbin:/usr/hdp/current/hive-webhcat/sbin:/home/gecko/bin
SHELL
:/bin/bash
EDITOR
:vim
LANG
:en_US.UTF-8
PWD
:/home/gecko/git/transform_pyspark
Contents of
Pipfile
('/home/gecko/git/transform_pyspark/Pipfile'):Contents of
Pipfile.lock
('/home/gecko/git/transform_pyspark/Pipfile.lock'):The text was updated successfully, but these errors were encountered: