Globus is a fast, reliable file transfer service that makes it easy for users to move data between two GridFTP servers or between a GridFTP server and a user's machine (Windows, Mac or Linux). It is increasingly used by researchers to transfer large data sets because it provides faster and more reliable transfers than previous methods.
For a quick intro to Globus take a look at Globus for newbies.
There are currently two EC2 Servers serving up the exact same globus structure. Need to find out why that is... Current Globus collections for dataspace
There is a single globus enpoint: Princeton Data Repository Endpoint
Storage gateways are created and configured in Globus (not in AWS). There are currently four storage gateways:
- Princeton DSS Gateway - S3 Gateway
dss-server-data3
(defunct) - Princeton Dataspace Posix Gateway - POSIX Gateway (defunct)
- Princeton Dataspace s3 Gateway - S3 Gateway
prds-globus
- Princeton Research Data Dataspace Gateway - S3 Gateway
prds-dataspace
There are currently six collections:
- another_test | fkayiwa@princeton.edu | guest | Princeton Research Data Dataspace Gateway
- Princeton University Library Data and Statistical Services | fkayiwa@princeton.edu | mapped | Princeton DSS Gateway
- Princeton Dataspace Collection | fkayiwa@princeton.edu | mapped | Princeton Research Data Dataspace Gateway
- Princeton Dataspace Posix Collection | fkayiwa@princeton.edu | mapped | Princeton Dataspace Posix Gateway
- Princeton Research Data Repository | mjc12@princeton.edu | guest | Princeton Research Data Dataspace Gateway
- CG | fkayiwa@princeton.edu | guest | Princeton Research Data Dataspace Gateway
Note: There seem to be collections for PDC, DSS, Dataspace and the POSIX collection. We should figure out what these collections are all for...
Our globus endpoints are being monitored by Honey Badger and Monit
Additional sites can be configured for monitoring by adding the sites GCS Manager URL
to honey badger.
You must be on VPN to ssh onto the Globus EC2 instances
- To access the aws console first got to https://princeton.edu/aws and log in
- Access the EC2 instances to view the public IP address for the endpoint you would like to monitor
- SSH onto the EC2 instance by runnning
pulsys@<public ip address>
- On the EC2 server run
globus-connect-server endpoint show
. This will display theGCS Manager URL
A new URL once it is determined via the steps above can be added to Honey Badger by:
- visiting Honey Badger
- click
Add uptime check
- fill in the Name appropriately for the endpoint
- fill in the url with
<GCS Manager URL>/api/info
- Choose
Page includes string
from match type - Fill in Match with
"code":"success"
- Click Save Changes
Our globus setup consists of four globus collections connected to three s3 buckets on two globus endpoints.
We are creating a separate Endpoint and EC2 instance to make sure that Upload does not impact download/ deposit and vice versa. We also believe this might scale better into the future.
graph LR;
S31[("Precuration S3 (private) [pdc-describe-*-precuration]")]-->SG1
S32[("Postcuration S3 (private) [pdc-describe-*-postcuration]")]-->SG2
S33[("Deposit S3 (private) [pdc-describe-deposit]")]-->SG3
subgraph project ["Princeton Data Commons * Project for Staging or Production"]
subgraph project_space [" "]
style project fill:#fff,stroke:#000,stroke-width:4px,color:#000,stroke-dasharray: 5 5
style project_space fill:#fff,stroke:#000,stroke-width:0px
subgraph EC2 ["Precuration EC2 [pdc-globus-*-precuration]]"]
subgraph ec2_sp [" "]
subgraph "PreCuration Globus Endpoint [pdc precuration]]"
SG1[["Precuration Storage Gateway [pdc s3 storage gateway precuration]"]]-->B(["Pre Curation Collection (private) [Princeton Data Commons * Precuration]"])
end
end
end
subgraph EC2a ["Postcuration EC2 [pdc-globus-*-postcuration]]"]
subgraph ec2a_sp [" "]
subgraph "Postcuration Globus Endpoint [pdc postcuration]]"
SG2[["Post Curation Storage Gateway [pdc s3 storage gateway postcuration]"]]-->D(["Curation Collection(curator only read/write) [Princeton Data Commons * Postcuration]"]);
SG2-->E(["Public Guest Globus Collection (public read) [Princeton Data Commons *]"]);
end
end
end
subgraph EC2b ["Deposit EC2 [pdc-globus-deposit]]"]
subgraph ec2b_sp [" "]
subgraph "Deposit Globus Endpoint [pdc deposit]]"
SG3[["Deposit Storage Gateway [pdc s3 storage gateway deposit]"]]-->DE(["Curation Collection(curator controlled read/write) [Princeton Data Commons Deposit]"]);
end
end
end
end
end
classDef ecclass fill:#00f,stroke:#00f,stroke-width:0px,color:#fff;
class EC2,EC2a,EC2b,ec2_sp,ec2a_sp,ec2b_sp ecclass;
Pre Curation collections should be private (Not show up in the collections search on globus) Postcuration guest collection - Should be public
-
There should be two globus endpoints for each environment (staging & prod) and a deposit endpoint
- staging
- pdc precuration
- pdc postcuration
- prod
- pdc precuration
- pdc postcuration
- pdc deposit
- staging
-
The globus endpoints should exist on five separate EC2 instances
- staging
- pdc-globus-staging-precuration
- pdc-globus-staging-postcuration
- prod
- pdc-globus-prod-precuration
- pdc-globus-prod-postcuration
- deposit
- pdc-globus-deposit
- staging
-
There will be two s3 buckets for each environment and a deposit bucket
- staging
- pdc-describe-staging-precuration
- pdc-describe-staging-postcuration
- prod
- pdc-describe-prod-precuration
- pdc-describe-prod-postcuration
- deposit
- pdc-describe-deposit
- staging
-
Each bucket will have a storage gateway attached to the correct globus endpoint
- staging
- pdc s3 storage gateway precuration
- pdc s3 storage gateway postcuration
- prod
- pdc s3 storage gateway precuration
- pdc s3 storage gateway postcuration
- deposit
- pdc s3 storage gateway deposit
- staging
-
Each storage gateway will have a mapped collection
- staging
- Princeton Data Commons Staging Precuration
- Princeton Data Commons Staging Postcuration
- prod
- Princeton Data Commons Precuration
- Princeton Data Commons Postcuration
- deposit
- Princeton Data Commons Deposit
- staging
-
Each Mapped post-curation collection will have a guest collection attached to the corresponding mapped collection. This will give public read access to the post-curated data in Globus.
- staging
- Princeton Data Commons Staging
- prod
- Princeton Data Commons
- staging
Our Pre-Curation Globus instance is for research data that has not yet gone through curation. It should have a DOI that matches the draft one created by PDC Describe.
Our legacy Globus instance does not have the concept of pre curated data sets.
Our Globus post curation instance is for research data that has gone through curation. Because it has gone through curation, it will always have a DOI, and it is organized under its DOI number.
Data is stored in S3 buckets. We are in the process of building a new Globus production system. (Project plan here.)
Our legacy Globus instance has its data in an S3 bucket called prds-dataspace
.
Our new PDC branded globus instance will have its data in a bucket called pdc-globus-prod
.
In order to help maintain our Globus infrastructure you'll need an AWS account with the right permission. See the onboarding documentation for how to set that up.
To contact globus for support either email them (SUPPORT@GLOBUS.ORG), or contact them via their Help Page
When using the Help page click "Check Support Ticket Status". You'll need to create an account on their support system (zendesk). You can submit tickets and keep track of them via this option. You can also include attachments. Try to request that Francis is added to the thread when you submit a support ticket.
It is helpful to include the output of globus-connect-server self-diagnostic
in the ticket to Globus.
See also Globus Troubleshooting.
- To build an ami please install the prequisites from https://github.com/pulibrary/vm-builds#prerequisite-software
- The following steps require awscli which should be installed as part of vm-builds
- You will be utilizing princeton-ansible. Please make sure the environment is set up on your local machine
This section documents steps that were taken to setup items that will be utilized in the following sections like the service account and the Amazon Machine Instance (AMI). They do not need to be run to set up each Globus Endpoint
In the event that a Princeton service account needs to be created, please refer to these steps:
Setting up Princeton service account
- run
packer build aws-globus.pkr.hcl
from the vm-buids directory on your local machine - see your new ami in AWS
Note This section is untested
- Go to developers.globus.org, sign in as rdss-globus TODO does rdss-globus have this ability??
- Click on
Register a new Globus Connect Server v5
- Click
Add another project
To make a new S3 bucket with a copy of the PDC Globus data:
-
To access the aws console first got to https://princeton.edu/aws and log in
-
Then access https://s3.console.aws.amazon.com/s3/buckets?region=us-east-1# for the s3 buckets
-
Click create new bucket
-
Name the bucket For PDC describe the name should be one of the five choices below:
- pdc-describe-staging-precuration
- pdc-describe-staging-postcuration
- pdc-describe-prod-precuration
- pdc-describe-prod-postcuration
- pdc-describe-deposit
-
Under
Copy settings from existing bucket
choosepdc-describe-staging-precuration
(assumes it has been set up) -
Under
Object Ownership
choose enable ACLs -
click
Create Bucket
-
click
Properties
tab on the bucket you just created -
Enable Versions
- click on the Edit button to the far right of
Bucket Versioning
section - choose
Enable
and click onSave
- click
Management
tab - click on
Create lifecycle rule
- Fill in
Remove old deleted versions
for the name - choose
Apply to all objects in the bucket
- check
I acknowledge...
- check
Permanently delete noncurrent versions of objects
- check
Delete expired object delete markers or incomplete multipart uploads
- Fill in 30 (14 for deposit) in
Days after objects become noncurrent
- check
Delete expired object delete markers
- check
Delete incomplete multipart uploads
- Fill in 14 in
Number of days
- Click on
Create Rule
- click on the Edit button to the far right of
-
(Optionally) Copy the data using the aws cli:
aws s3 sync s3://prds-dataspace s3://pdc-globus-prod
Note: that the copy will take about five hours.
- Download https://github.com/pulibrary/pdc_describe/blob/main/s3-checksum.yaml to your local machine OR if you have PDC Describe cloned to your machine use it there.
- Visit princeton.edu/aws to login to AWS
- then visit the AWS Stacks
- Click on
Create Stack
and choosewith new resources
from the drop down - Choose
Upload a template file
and select the file from step 1 - Click
Next
- Specify a name like
prod-precuration-sha-256-checksum
which includes the name of the bucket this stack will apply to - Specify the bucket for example:
pdc-describe-prod-precuration
- Click
Next
- Leave defaults in place on the next page and click
Next
- Acknowledge that this gives the stack access to the data
- Click Submit
- Wait until you see
CREATE_COMPLETE
in the events tab - Click on the Resources tab
- Click on the link next to ChecksumLambdaRole (which will open a new window)
- Click on "Add Permission" and select "Create inline policy"
- Choose S3 from the Service section
- Click
All S3 actions (s3:*)
in the Actions section - Choose s3 bucket in the resources and put in your bucket name
- Click
Review Policy
at the bottom - Fill in
Allow-S3-Access
in the Name - Click on
Create Policy
- Close the window that was opened
- In the Stack Window on your new Stack Click on the Resources tab and make note of the Physical ID of the
ChecksumLambdaFunction
(You will need it below) For examplesha-256-checksum-precuratio-ChecksumLambdaFunction-IsK4yDcshh1C
- Go to your bucket (specified above)
- Click on properties (https://s3.console.aws.amazon.com/s3/buckets/pdc-describe-staging-precuration?region=us-east-1&tab=properties)
- Scroll down to Event Notifications and click
Create event notification
- Fill in event name with
Create SHA256 on Upload
- Choose
All object create events
underEvent Types
- Under
Destination choose
Lambda function
- fill in the lambda function Physical ID you just created ex
sha-256-checksum-precuratio-ChecksumLambdaFunction-IsK4yDcshh1C
and sleect the arn - Test the lambda by uploading a file to the bucket
-
In princeton-ansible check the group_vars/globus/common.yml to make sure the image_id matches the ami you either created or are using
-
Run the playbook with either staging (default) or production (
-e runtime_env=production
). For example the command below will create a staging EC2 instance named pdc-globus-stagingansible-playbook playbooks/aws_ec2.yml -e ec2_name=<name>
should be one of the following:
- pdc-globus-staging-precuration
- pdc-globus-staging-postcuration
- pdc-globus-prod-precuration
- pdc-globus-prod-postcuration
-
To access the aws console first go to AWS login
-
See your new instance in AWS Note the IP address you will need it to configure Globus in following steps
-
Create the Elastic IP
To register a Globus Endpoint do the following:
-
Go to developers.globus.org, sign in as rdssglobus
-
Add a new Globus Connect Server and fill out the form in the Princeton Data Commmons Production or Princeton Data Commmons Staging Projects.
- Type in the new name
The name should be one of the following:
- PDC Precuration
- PDC Postcuration
- click Register Globus Connect Server
- Type in the new name
The name should be one of the following:
-
Generate a client secret. Be ready to save the value off, you will only see it once.
- Click Generate a New Client Secret within the new section on the website created by the step above
- Fill in the description of the secret (can be anything)
- Click Save
- Save the Client ID and Client Secret values in a note in "Shared-ITMS-Passwords" in Lastpass (They are needed in the next step)
-
Log into the virtual machine and configure globus:
-
ssh pulsys@<public IP>
-
On the server set the endpoint name you utilized above and generated client id. This will allow a copy and paste of the commands in the next step to work without editing
export endpoint_name=<name from above (include staging or production)> export client_id=<client id from above step>
-
On the server run
mkdir app_config cd app_config globus-connect-server endpoint setup "$endpoint_name" --organization "Princeton University Library" --client-id $client_id --owner rdssglobus@princeton.edu --contact-email lsupport@princeton.edu
-
you will be prompted for the client secret
-
Accept Let's Encrypt ToS
-
It will then run for some time (10 minutes about) and then give you a URL to paste into a browser You will get a response close to this
Please authenticate with Globus to set the owner of the endpoint: ---------------------------------------------------------------- https://auth.globus.org/v2/oauth2/authorize?client_id=auuid-&redirect_uri=https%3A%2F%2Fauth.globus.org%2Fv2%2Fweb%2Fauth-code&... ----------------------------------------------------------------
-
Paste the results from your browser to complete endpoint creation. Utilize the browser where you logged into Globus as rdssglobus for this step
-
click on the link rddglobus@princeton.edu on the page and it will create a code
-
-
paste the above code in on the command line
-
It will then finish giving you a url for google cloud storage
-
This will also create a
deployment-key.json
file in theapp_config
directory. This is an important file that connects the localhost to the globus web network
-
-
-
ssh pulsys@<public IP>
-
On the server set the endpoint name you utilized above and generated client id. This will allow a copy and paste of the commands in the next step to work without editing
cd app_config export endpoint_name=<name from above (include staging or production)> export client_id=<client id from above step>
-
run the following command to configure your node: (note: use of
sudo
) You will be prompted for the client secret you saved in lastpass.sudo globus-connect-server node setup \ --client-id $client_id sudo systemctl restart apache2
-
Finish the setup by running this on the VM
sudo globus-connect-server login localhost
NOTE: Depending on the browser you paste the url into, you will be logged in as that person. If you put it into an incognito and login with rdssglobus you will be rdssglobus. If you use a browser that you are logged into then you will be logged into the connect server as yourself. If you receive messages like
None of your identities have been granted a role to access this resource
you may need to get administrator access for your account, or login as rdssglobus instead.
This is a manual step.
-
ssh onto the EC2 instance
ssh pulsys@<public IP>
-
run
sudo globus-connect-server endpoint show
-
email
pglobus@princeton.edu
with subjectRegister Globus Endpoint
Hello, Can you please register the following Princeton Research Data Globus endpoint to be managed by Research Computing's Globus account: <copy and paste from step 1> We will be connecting this endpoint to an S3 bucket. Thank you! -- <your name> Princeton University Library
-
go to Globus Endpoints and sign in as rdssglobus (see last pass)
-
click on New endpoint you created above
-
click on the
Roles
tab and Continue for Authentication Consent- you may need to add the aws key here
- In princeton_ansible directory on your local machine, see the aws key information needed
ansible-vault view group_vars/globus/vault.yml
- In princeton_ansible directory on your local machine, see the aws key information needed
- you may need to add the aws key here
-
Add administrators via the command line
ssh pulsys@<public IP>
and runsudo globus-connect-server endpoint role create administrator rdssglobus@princeton.edu sudo globus-connect-server endpoint role create administrator cac9@princeton.edu sudo globus-connect-server endpoint role create administrator fkayiwa@princeton.edu sudo globus-connect-server endpoint role create administrator bs3097@princeton.edu sudo globus-connect-server endpoint role create administrator hc8719@princeton.edu sudo globus-connect-server endpoint role create administrator jrg5@princeton.edu sudo globus-connect-server endpoint role create administrator kl37@princeton.edu
- After rebooting the VM take note of the new AWS provided IP. Retrieve the UUID with the following.
- login following the prompts
globus-connect-server login localhost
- see the node UUID
globus-connect-server node list --use-explicit-host localhost
- login following the prompts
- Update your Nodes (IP)s
globus-connect-server node update ${yourNodesUUID} -i ${yourNewIP} --use-explicit-host localhost
Utilize the rdssglobus
AWS IAM user by logging in in as rdssglobus in an incognito window. Login information is in lastpass.
-
login to globus
sudo globus-connect-server login localhost
- paste the url in to your browser
NOTE: Depending on the browser you paste the url into, you will be logged in as that person. If you put it into an incognito and login with rdssglobus you will be rdssglobus. If you use a browser that you are logged into then you will be logged into the connect server as yourself. If you receive messages like
None of your identities have been granted a role to access this resource
you may need to get administrator access for your account, or login as rdssglobus instead.
- paste the url in to your browser
NOTE: Depending on the browser you paste the url into, you will be logged in as that person. If you put it into an incognito and login with rdssglobus you will be rdssglobus. If you use a browser that you are logged into then you will be logged into the connect server as yourself. If you receive messages like
-
Put the bucket name in a variable. Choose one of the following:
- pdc-describe-staging-precuration
- pdc-describe-staging-postcuration
- pdc-describe-prod-precuration
- pdc-describe-prod-postcuration
- pdc-describe-deposit
export bucket_name=<bucket name>
-
Decide on a Gateway name and put it in a variable. For pdc*, choose one of the following:
- pdc s3 storage gateway precuration
- pdc s3 storage gateway postcuration
- pdc s3 storage gateway precuration
- pdc s3 storage gateway postcuration
- pdc s3 storage gateway deposit
export gateway_name=<gateway name>
-
Create the S3 gateway and connect the bucket to it:
sudo globus-connect-server storage-gateway create s3 "$gateway_name" --domain princeton.edu --s3-endpoint https://s3.amazonaws.com --s3-user-credential --bucket $bucket_name
Note the id returned by this command, you will need it in the following
- Assign the gateway id to a variable
export gateway_id=<output from above command>
- Assign the gateway id to a variable
-
Create a collection for the bucket
-
Utilize one of the following for the
<name>
:- Princeton Data Commons Staging Precuration
- Princeton Data Commons Staging Postcuration
- Princeton Data Commons Precuration
- Princeton Data Commons Postcuration
- Princeton Data Commons Deposit
export collection_name=<name>
-
Utilize one of the following for the
<info link>
- https://pdc-describe-staging.princeton.edu/about
- https://pdc-describe.princeton.edu/about
export info_link=<info link>
-
Utilize one of the following for the
<description>
-
if this is a staging system add
--user-message "Staging data! Please do not store production data here"
sudo globus-connect-server collection create $gateway_id / "$collection_name" --organization 'princeton.edu' --contact-email prds@princeton.edu --info-link $info_link --description "$collection_name Globus Collection for curators" --keywords princeton.edu,RDOS,research --allow-guest-collections --enable-https
Note the id returned by this command, you will need it in the next step
-
-
Assign the collection id to a variable
export collection_id=<output from above command>
-
Add administrators via the command line
sudo globus-connect-server collection role create $collection_id administrator rdssglobus@princeton.edu sudo globus-connect-server collection role create $collection_id administrator cac9@princeton.edu sudo globus-connect-server collection role create $collection_id administrator fkayiwa@princeton.edu sudo globus-connect-server collection role create $collection_id administrator bs3097@princeton.edu sudo globus-connect-server collection role create $collection_id administrator hc8719@princeton.edu sudo globus-connect-server collection role create $collection_id administrator jrg5@princeton.edu sudo globus-connect-server collection role create $collection_id administrator kl37@princeton.edu
-
In princeton_ansible directory on your local machine, see the aws key information needed
ansible-vault view group_vars/globus/vault.yml
-
visit the collection on https://app.globus.org/collections?scope=administered-by-me
- Click on the credentials tab and click continue to setup the credentials for accessing s3 (utilize the IAM key from above)
Guest collections are public and readonly, we use them for the general public to be able to download the data.
** this is a manual step ** Choose one of the following for the name:
- Princeton Data Commons Staging
- Princeton Data Commons
- visit https://app.globus.org/collections?scope=administered-by-me
- click on the collection you just created
- click on the collections tab
- click on "Add a guest Collection"
- You may need to authorize (click rdssglobus@princeton.edu and allow)
- click browse and select the S3 bucket top path
- Select folder
pdc-describe-staging-postcuration
for stating - Select folder
pdc-describe-prod-postcuration
for production
- Select folder
- Fill in the name and description
- Click create collection
- Click "Add Permissions - Share With"
- Choose "Public"
- Click Add Permissions
- visit https://app.globus.org (and login as rdssglobus)
- click on "Groups" on the left hand side
- click on "Create New Group" on the left hand side
- Type in "Princeton Curators" for the group name
- For each Curator invote them into the group
- Utilize one of the following for the
<name>
:- Princeton Data Commons Staging Precuration
- Princeton Data Commons Staging Postcuration
- Princeton Data Commons Precuration
- Princeton Data Commons Postcuration
- visit https://app.globus.org (and login as rdssglobus)
- Find the collection you need to update in Collections
- Click on the Roles Tab
- Click on "Assign new Role"
- Choose Group
- Enter "Princeton Curators"
- Choose "Administrator"
- Click "Add Role"
- We have a Globus Certs Playbook set up to automatically update the certificates for Globus VMs. Run this Playbook.
- When completed check the
/var/lib/globus-connect-server/grid-security/certificates
to see if old certificates still exist.
For ca-update-2022 the following files should be removed if the playbook fails to remove them
7a42187f.* 4b828555.* c7ab88a4.* globus_transfer_ca_2.* globus_transfer_ca_2_int.*
- visit developers.globus.org (and login as rdssglobus)
- click on the endpoint you would like to rotate the client secret for
- click on
Generate new Client Secret
- copy the client secret to last pass or somewhere safe
- ssh onto the EC2 instance
ssh pulsys@<public IP>
- run
sudo globus-connect-server node set-configuration client_secret
- enter the new client secret when prompted
sudo systemctl restart gcs_manager_assistant
sudo systemctl restart gcs_manager
globus-connect-server endpoint show
(should now work, you may need to login)
- run
- go back to developers.globus.org (and login as rdssglobus if not already logged in)
- click on the endpoint you just rotated
- Delete the client secret that should now be invalidated
-
Run the following command to list all volumes with volume id and name below:
aws ec2 describe-instances --query 'Reservations[*].Instances[*].[Tags[?Key==`Name`].Value,InstanceId,BlockDeviceMappings[*].Ebs.VolumeId]' --output text
you will see results with the following:
i-043b67b2b215d180b prds-dataspace-endpoint1 vol-0c02aca34398ea8e5 vol-0e8522f82fd721b25 i-03d6cd75ba1fca824 ... ...
-
Select the volume you wish to create a snapshot of in the event something goes wrong. Then create a snapshot of it with the following. In my example below I will create a snapshot of
prds-dataspace-endpoint1
with the volume IDvol-0c02aca34398ea8e5
and instance ID ofi-043b67b2b215d180b
aws ec2 create-snapshot --volume-id vol-0c02aca34398ea8e5 --description "This is the prds-dataspace-endpoint root volume snapshot"
-
Modify the volume of your instance. We will continue to use the VM in step two. Our goal is to resize it to have 100GB. We do that with the following (this can take a while so grab some ☕ ):
aws ec2 modify-volume --volume-type gp2 --size 100 --volume-id vol-0c02aca34398ea8e5
-
Check to see if the results if your server has a larger disk size by running the following:
sudo lsblk
You will see results similar to this:
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
loop0 7:0 0 24.4M 1 loop /snap/amazon-ssm-agent/6312
loop1 7:1 0 25.1M 1 loop /snap/amazon-ssm-agent/5656
loop2 7:2 0 55.6M 1 loop /snap/core18/2714
loop3 7:3 0 63.3M 1 loop /snap/core20/1852
loop4 7:4 0 63.3M 1 loop
loop6 7:6 0 91.9M 1 loop /snap/lxd/24061
loop7 7:7 0 55.6M 1 loop /snap/core18/2721
loop8 7:8 0 49.9M 1 loop /snap/snapd/18357
loop9 7:9 0 91.8M 1 loop /snap/lxd/23991
loop10 7:10 0 63.3M 1 loop /snap/core20/1828
loop11 7:11 0 49.9M 1 loop /snap/snapd/18596
nvme0n1 259:0 0 100G 0 disk
├─nvme0n1p1 259:1 0 7.9G 0 part /
├─nvme0n1p14 259:2 0 4M 0 part
└─nvme0n1p15 259:3 0 106M 0 part /boot/efi
- Expand the volume so the Operating System can use it with the following command (note the space between the ones which matter):
sudo growpart /dev/nvme0n1 1
- You can now delete your snapshot in step 1 if successful.