有谁知道是否有一种简单的方法可以发现在何处pip install git+ssh:///...
搜索 ssh 密钥?
我在 Windows 10 上。我需要在 conda 虚拟环境中使用 pip 从 github 安装一个私有远程存储库。我是回购所有者。我通过 github.com 设置了一个公共/私有 ssh 密钥对。我的本地密钥存储在C:\Users\MyName\\.ssh\id_rsa
. 使用此密钥,我可以使用我的 IDE Eclipse 从 github 毫无问题地推拉。
但是,当我使用激活的 conda 环境执行以下命令时:
pip install git+ssh://github.com/USER_NAME/REPO_NAME.git
Run Code Online (Sandbox Code Playgroud)
我收到以下错误:
Collecting git+ssh://github.com/USER_NAME/REPO_NAME.git
Cloning ssh://github.com/USER_NAME/REPO_NAME.git to
c:\users\USER_NAME\appdata\local\temp\pip-ghf3ts-build
Warning: Permanently added 'github.com,IP_ADDRESS' (RSA) to the list of
known hosts.
Permission denied (publickey).
fatal: Could not read from remote repository.
Please make sure you have the correct access rights
and the repository exists.
Run Code Online (Sandbox Code Playgroud)
我知道 repo …
我有一个数据框:
df = pd.DataFrame(
{'number': ['10', '20' , '30', '40'], 'condition': ['A', 'B', 'A', 'B']})
df =
number condition
0 10 A
1 20 B
2 30 A
3 40 B
Run Code Online (Sandbox Code Playgroud)
我想将一个函数应用于number列中的每个元素,如下所示:
df['number'] = df['number'].apply(lambda x: func(x))
Run Code Online (Sandbox Code Playgroud)
但是,即使我将函数应用于数字列,我也希望函数也引用该condition
列,即使用伪代码:
func(n):
#if the value in corresponding condition column is equal to some set of values:
# do some stuff to n using the value in condition
# return new value for n
Run Code Online (Sandbox Code Playgroud)
对于单个数字和示例函数,我将编写:
number = 10
condition = A
def …
Run Code Online (Sandbox Code Playgroud) 我定义了以下docker镜像:
FROM python:3.6
RUN pip install --upgrade pip
RUN pip install --upgrade mlflow
ENTRYPOINT mlflow server --host 0.0.0.0 --file-store /mnt/mlruns/
Run Code Online (Sandbox Code Playgroud)
并建立一个名为的图像mlflow-server
.接下来,我从本地计算机启动此服务器:
docker run --rm -it -p 5000:5000 -v ${PWD}/mlruns/:/mnt/mlruns mlflow-server
Run Code Online (Sandbox Code Playgroud)
接下来,我定义以下函数:
def foo(x, with_af=False):
mlflow.start_run()
mlflow.log_param("x", x)
print(x)
if with_af:
with open(str(x), 'wb') as fout:
fout.write(os.urandom(1024))
mlflow.log_artifact(str(x))
mlflow.log_artifact('./foo.data')
mlflow.end_run()
Run Code Online (Sandbox Code Playgroud)
从我运行的同一目录中foo(10)
,正确记录参数.但是,会foo(10, True)
产生以下错误:PermissionError: [Errno 13] Permission denied: '/mnt'
.似乎log_artifact
试图直接将文件保存在本地文件系统上.
知道我做错了什么吗?
将数据从一个 s3 连续复制到另一个 s3 存储桶的最佳方法是什么?我了解 S3 支持事件通知并将其发送到 lambda、SNS 和 SQS,但我不确定可以选择做什么?我应该触发 lambda 函数来接收来自 S3 的记录并将它们复制到另一个 s3 中吗?我应该使用 SNS 还是 SQS 来做到这一点?
使用kafka-python
包构建 KafkaProducer 时出现此错误:
[ERROR] UnrecognizedBrokerVersion: UnrecognizedBrokerVersion
Traceback (most recent call last):
File "/var/lang/lib/python3.7/imp.py", line 234, in load_module
return load_source(name, filename, file)
File "/var/lang/lib/python3.7/imp.py", line 171, in load_source
module = _load(spec)
File "<frozen importlib._bootstrap>", line 696, in _load
File "<frozen importlib._bootstrap>", line 677, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 728, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/var/task/kafka/producer/kafka.py", line 381, in __init__
**self.config)
File "/var/task/kafka/client_async.py", line 240, in __init__
self.config['api_version'] = self.check_version(timeout=check_timeout)
File "/var/task/kafka/client_async.py", line 908, …
Run Code Online (Sandbox Code Playgroud) 我在推送到Github时遇到503错误:
$ git push github develop
Counting objects: 22, done.
Delta compression using up to 8 threads.
Compressing objects: 100% (22/22), done.
Writing objects: 100% (22/22), 4.16 KiB | 0 bytes/s, done.
Total 22 (delta 16), reused 0 (delta 0)
error: RPC failed; HTTP 503 curl 22 The requested URL returned error: 503 Service Unavailable
fatal: The remote end hung up unexpectedly
fatal: The remote end hung up unexpectedly
Everything up-to-date
Run Code Online (Sandbox Code Playgroud)
我已经检查了他们的状态页和“所有系统都在运行”,所以我认为它必须与我的配置有关。我的.gitconfig
文件只有我的名字和电子邮件:
[user]
name = Bradley …
Run Code Online (Sandbox Code Playgroud) 我正在对我拥有的数据集运行拼写校正功能。我曾经from pathos.multiprocessing import ProcessingPool as Pool
做过这项工作。处理完成后,我想实际访问结果。这是我的代码:
import codecs
import nltk
from textblob import TextBlob
from nltk.tokenize import sent_tokenize
from pathos.multiprocessing import ProcessingPool as Pool
class SpellCorrect():
def load_data(self, path_1):
with codecs.open(path_1, "r", "utf-8") as file:
data = file.read()
return sent_tokenize(data)
def correct_spelling(self, data):
data = TextBlob(data)
return str(data.correct())
def run_clean(self, path_1):
pool = Pool()
data = self.load_data(path_1)
return pool.amap(self.correct_spelling, data)
if __name__ == "__main__":
path_1 = "../Data/training_data/training_corpus.txt"
SpellCorrect = SpellCorrect()
result = SpellCorrect.run_clean(path_1)
print(result)
result = " ".join(temp …
Run Code Online (Sandbox Code Playgroud) python multiprocessing python-3.x python-multiprocessing pathos
如何查看 AWS Lamdba 函数使用的存储的已用大小?
例如,可以这样检查文件是否存在:
import os
os.path.isfile('/tmp/' + filename)
Run Code Online (Sandbox Code Playgroud)
但是,我需要知道何时达到 512 MB 限制。
将代码放入 docker 容器的最佳实践是什么?
以下是一些可能的方法:
plot.density()
Pandas和function之间有区别吗plot.kde()
?
python ×7
docker ×2
git ×2
pandas ×2
amazon-s3 ×1
amazon-sns ×1
amazon-sqs ×1
apache-kafka ×1
aws-lambda ×1
aws-msk ×1
bash ×1
github ×1
http ×1
kafka-python ×1
mlflow ×1
numpy ×1
pathos ×1
pip ×1
python-3.x ×1
ssh ×1