Skip to content

Commit 074a8a8

Browse files
authored
Merge pull request #128 from afermg/fix-ci
ci: change zenodo root id, fix hash check.
2 parents 1c2e38f + d96e4e1 commit 074a8a8

File tree

2 files changed

+9
-5
lines changed

2 files changed

+9
-5
lines changed

README.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,8 @@ This collection comprises 4 datasets:
2424
- Different subsets of data in the principal dataset, assembled into single parquet files. The URLs to the subsets are [here](https://github.com/jump-cellpainting/datasets/blob/main/manifests/profile_index.csv). The corresponding folders for each contain all the data levels (e.g. this [folder](https://cellpainting-gallery.s3.amazonaws.com/index.html#cpg0016-jump-assembled/source_all/workspace/profiles/jump-profiling-recipe_2024_a917fa7/ORF/profiles_wellpos_cc_var_mad_outlier_featselect_sphering_harmony/)). Snakemake workflows for producing these assembled profiles are available [here](https://github.com/broadinstitute/jump-profiling-recipe/releases/tag/v0.1.0). We recommend working with the the `all` or `all_interpretable` subsets -- they contain all three data modalities in single dataframe. Note that cross-modality matching is still poor (ORF-CRISPR, COMPOUND-CRISPR, COMPOUND-ORF), but within modality generally works well.
2525
- A [tutorial](https://broadinstitute.github.io/2023_12_JUMP_data_only_vignettes/howto/1_retrieve_profiles.html) to load these subsets of data.
2626
- Other [tutorials](https://broad.io/jump) to work with `cpg0016`.
27+
- The datasets and their DOI can be found on this [Zenodo](https://zenodo.org/records/13892061/latest) record.
28+
- Multiple datasets of interest for JUMP are available on our [Zenodo](https://zenodo.org/communities/broad-imaging/records?q=&l=list&p=1&s=10&sort=newest) community.
2729

2830
### What’s coming up
2931

manifests/src/upload_index.sh

Lines changed: 7 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,9 @@
11
# Find the latest version of the dataset
22
ZENODO_ENDPOINT="https://zenodo.org"
33
DEPOSITION_PREFIX="${ZENODO_ENDPOINT}/api/deposit/depositions"
4-
ORIGINAL_ID="13146273"
4+
ORIGINAL_ID="13892061"
55
FILE_TO_VERSION="manifests/profile_index.csv"
6+
FILENAME=$(echo ${FILE_TO_VERSION} | sed 's+.*/++g')
67

78
echo "Checking that S3 ETags match their local counterpart"
89
S3_ETAGS=$(cat ${FILE_TO_VERSION} | tail -n +2 | cut -f2 -d',' | xargs -I {} -- curl -I --silent "{}" | grep ETag | awk '{print $2}' | sed 's/\r$//' | md5sum | cut -f1 -d" ")
@@ -27,7 +28,7 @@ else # Update existing dataset
2728
LOCAL_HASH=$(md5sum ${FILE_TO_VERSION} | cut -f1 -d" ")
2829

2930
echo "Checking for changes in file contents: Remote ${REMOTE_HASH} vs Local ${LOCAL_HASH}"
30-
if [ "${REMOTE_HASH}" = "${LOCAL_HASH}" ]; then
31+
if [ "${REMOTE_HASH}" == "${LOCAL_HASH}" ]; then
3132
echo "The urls and md5sums have not changed"
3233
exit 0
3334
fi
@@ -54,7 +55,7 @@ DEPOSITION=$(curl -H "Content-Type: application/json" \
5455
echo "New deposition ID is ${DEPOSITION}"
5556

5657
# Variables
57-
BUCKET_DATA=$(curl "${DEPOSITION_PREFIX}/$DEPOSITION?access_token=$ZENODO_TOKEN")
58+
BUCKET_DATA=$(curl "${DEPOSITION_PREFIX}/${DEPOSITION}?access_token=${ZENODO_TOKEN}")
5859
BUCKET=$(echo "${BUCKET_DATA}" | jq --raw-output .links.bucket)
5960

6061
if [ "${BUCKET}" = "null" ]; then
@@ -64,10 +65,11 @@ if [ "${BUCKET}" = "null" ]; then
6465
fi
6566

6667
# Upload file
67-
echo "Uploading file to bucket ${BUCKET}"
68+
echo "Uploading file ${FILE_TO_VERSION} to bucket ${BUCKET}"
69+
cat ${FILE_TO_VERSION}
6870
curl -o /dev/null \
6971
--upload-file ${FILE_TO_VERSION} \
70-
${BUCKET}/${FILE_TO_VERSION}?access_token="${ZENODO_TOKEN}"
72+
${BUCKET}/${FILENAME}?access_token="${ZENODO_TOKEN}"
7173

7274

7375
# Upload Metadata

0 commit comments

Comments
 (0)