I recently had the opportunity to configure Alfresco with ATMOS using the Alfresco S3 connector.
The company I work for has an ATMOS cloud setup across multiple data centers.
ATMOS is an object-based cloud storage platform to store, archive and access unstructured content at scale.
We had an existing instance of Alfresco leveraging NetApp storage that I had to migrate to ATMOS.
In order to do this I needed to download and install 2 tools.
1. Stop your Alfresco and Solr servers
2. Create your S3 bucket with ATMOS FOX.
3. After you create the bucket you must add a non listable meta tag (bucket-mapping-type=one_to_one) to the bucket folder.
4. Copy your files in your contentstore and contentstore.deleted to Atmos using AtmosSync.jar.
5. Update your alfresco-global.properties with your S3 configuration.
### S3 Config ###
s3.accessKey=xxxxxxxxxxxx/xxxx
s3.secretKey=xxxxxxxxs3.bucketName=bucketNAME
#s3.bucketLocation=US
s3.flatRoot=falses3service.https-only=false
s3service.s3-endpoint=ATMOSHOST
s3service.s3-endpoint-http-port=8080#
s3service.disable-dns-buckets=false
dir.contentstore=contentstore
dir.contentstore.deleted=contentstore.deleted
#Maximum disk usage for the cache in MB
system.content.caching.maxUsageMB=51200
#Maximum size of files which can be stored in the cache in MB (zero implies no limit)
system.content.caching.maxFileSizeMB=0
6. Back up your DB if you haven’t already
7. Update all records in the ALF_CONTENT_URL table (store:// to s3://)
UPDATE alf_content_url SET content_url = replace(content_url, 'store:', 's3:’)
8. Startup Alfresco and Solr servers and you should be good to go.
Once you have verified that Alfresco is functioning properly you can repurpose the filesystem storage.
The company I work for has an ATMOS cloud setup across multiple data centers.
ATMOS is an object-based cloud storage platform to store, archive and access unstructured content at scale.
We had an existing instance of Alfresco leveraging NetApp storage that I had to migrate to ATMOS.
In order to do this I needed to download and install 2 tools.
- ATMOS FOX - a firefox plugging that allows you to browse you ATMOS storage. https://addons.mozilla.org/en-US/firefox/addon/atmosfox/
- AtmosSync.jar - a tool to copy files existing on a filesystem to ATMOS. https://code.google.com/p/atmos-java/downloads/detail?name=AtmosSync.jar&can=2&q=label%3AOpSys-All
1. Stop your Alfresco and Solr servers
2. Create your S3 bucket with ATMOS FOX.
3. After you create the bucket you must add a non listable meta tag (bucket-mapping-type=one_to_one) to the bucket folder.
4. Copy your files in your contentstore and contentstore.deleted to Atmos using AtmosSync.jar.
5. Update your alfresco-global.properties with your S3 configuration.
### S3 Config ###
s3.accessKey=xxxxxxxxxxxx/xxxx
s3.secretKey=xxxxxxxxs3.bucketName=bucketNAME
#s3.bucketLocation=US
s3.flatRoot=falses3service.https-only=false
s3service.s3-endpoint=ATMOSHOST
s3service.s3-endpoint-http-port=8080#
s3service.disable-dns-buckets=false
dir.contentstore=contentstore
dir.contentstore.deleted=contentstore.deleted
#Maximum disk usage for the cache in MB
system.content.caching.maxUsageMB=51200
#Maximum size of files which can be stored in the cache in MB (zero implies no limit)
system.content.caching.maxFileSizeMB=0
6. Back up your DB if you haven’t already
7. Update all records in the ALF_CONTENT_URL table (store:// to s3://)
UPDATE alf_content_url SET content_url = replace(content_url, 'store:', 's3:’)
8. Startup Alfresco and Solr servers and you should be good to go.
Once you have verified that Alfresco is functioning properly you can repurpose the filesystem storage.
0 comments:
Post a Comment