manpagez: man pages & more
man S3(n)
Home | html | info | man
S3(n)                   Amazon S3 Web Service Interface                  S3(n)



______________________________________________________________________________


NAME

       S3 - Amazon S3 Web Service Interface


SYNOPSIS

       package require Tcl  8.5

       package require sha1  1.0

       package require md5  2.0

       package require base64  2.3

       package require xsxp  1.0

       S3::Configure   ?-reset   boolean?   ?-retries  integer?  ?-accesskeyid
       idstring?  ?-secretaccesskey  idstring?  ?-service-access-point   FQDN?
       ?-use-tls    boolean?    ?-default-compare    always|never|exists|miss-
       ing|newer|date|checksum|different?     ?-default-separator      string?
       ?-default-acl      private|public-read|public-read-write|authenticated-
       read|keep|calc? ?-default-bucket bucketname?

       S3::SuggestBucket ?name?

       S3::REST dict

       S3::ListAllMyBuckets   ?-blocking   boolean?   ?-parse-xml   xmlstring?
       ?-result-type REST|xml|pxml|dict|names|owner?

       S3::PutBucket  ?-bucket  bucketname?  ?-blocking boolean? ?-acl {}|pri-
       vate|public-read|public-read-write|authenticated-read?

       S3::DeleteBucket ?-bucket bucketname? ?-blocking boolean?

       S3::GetBucket ?-bucket bucketname? ?-blocking boolean? ?-parse-xml xml-
       string?  ?-max-count integer? ?-prefix prefixstring? ?-delimiter delim-
       iterstring? ?-result-type REST|xml|pxml|names|dict?

       S3::Put ?-bucket bucketname? -resource resourcename ?-blocking boolean?
       ?-file   filename?   ?-content   contentstring?  ?-acl  private|public-
       read|public-read-write|authenticated-read|calc|keep?     ?-content-type
       contenttypestring?  ?-x-amz-meta-* metadatatext? ?-compare comparemode?

       S3::Get ?-bucket bucketname? -resource resourcename ?-blocking boolean?
       ?-compare   comparemode?  ?-file  filename?  ?-content  contentvarname?
       ?-timestamp aws|now? ?-headers headervarname?

       S3::Head  ?-bucket  bucketname?   -resource   resourcename   ?-blocking
       boolean? ?-dict dictvarname? ?-headers headersvarname? ?-status status-
       varname?

       S3::GetAcl ?-blocking boolean? ?-bucket bucketname? -resource resource-
       name ?-result-type REST|xml|pxml?

       S3::PutAcl ?-blocking boolean? ?-bucket bucketname? -resource resource-
       name ?-acl new-acl?

       S3::Delete  ?-bucket  bucketname?  -resource  resourcename   ?-blocking
       boolean? ?-status statusvar?

       S3::Push  ?-bucket  bucketname?  -directory directoryname ?-prefix pre-
       fixstring?  ?-compare  comparemode?  ?-x-amz-meta-*  metastring?  ?-acl
       aclcode?  ?-delete  boolean?  ?-error  throw|break|continue? ?-progress
       scriptprefix?

       S3::Pull ?-bucket bucketname? -directory  directoryname  ?-prefix  pre-
       fixstring? ?-blocking boolean? ?-compare comparemode? ?-delete boolean?
       ?-timestamp aws|now? ?-error throw|break|continue?  ?-progress  script-
       prefix?

       S3::Toss  ?-bucket bucketname? -prefix prefixstring ?-blocking boolean?
       ?-error throw|break|continue? ?-progress scriptprefix?

_________________________________________________________________


DESCRIPTION

       This package provides access to Amazon's Simple  Storage  Solution  web
       service.

       As  a  quick summary, Amazon Simple Storage Solution provides a for-fee
       web service allowing the  storage  of  arbitrary  data  as  "resources"
       within  "buckets" online.  See http://www.amazonaws.com/ for details on
       that system.  Access to the service is via HTTP (SOAP or  REST).   Much
       of  this  documentation will not make sense if you're not familiar with
       the terms and functionality of the Amazon S3 service.

       This package provides services for reading and writing the  data  items
       via the REST interface.  It also provides some higher-level operations.
       Other packages in the same distribution provide for even more function-
       ality.

       Copyright  2006  Darren New. All Rights Reserved.  NO WARRANTIES OF ANY
       TYPE ARE PROVIDED.  COPYING OR USE INDEMNIFIES THE AUTHOR IN ALL  WAYS.
       This  software is licensed under essentially the same terms as Tcl. See
       LICENSE.txt for the terms.


ERROR REPORTING

       The error reporting from this package makes use of $errorCode  to  pro-
       vide  more details on what happened than simply throwing an error.  Any
       error caught by the S3 package (and we try  to  catch  them  all)  will
       return  with an $errorCode being a list having at least three elements.
       In all cases, the first element will be "S3". The second  element  will
       take  on one of six values, with that element defining the value of the
       third and subsequent elements. S3::REST does not throw  an  error,  but
       rather  returns  a  dictionary  with the keys "error", "errorInfo", and
       "errorCode" set. This allows for reliable background use. The  possible
       second elements are these:

       usage  The  usage  of  the package is incorrect. For example, a command
              has been invoked which requires the  library  to  be  configured
              before  the  library has been configured, or an invalid combina-
              tion of options has been specified. The third element of $error-
              Code  supplies  the  name  of  the parameter that was wrong. The
              fourth usually provides the arguments that  were  actually  sup-
              plied  to  the  throwing proc, unless the usage error isn't con-
              fined to a single proc.

       local  Something happened on the local system which threw an error. For
              example, a request to upload or download a file was made and the
              file permissions denied that sort of access. The  third  element
              of $errorCode is the original $errorCode.

       socket Something  happened  with  the socket. It closed prematurely, or
              some other condition of  failure-to-communicate-with-Amazon  was
              detected.  The  third  element  of  $errorCode  is  the original
              $errorCode, or sometimes the message from fcopy, or ...?

       remote The Amazon web service returned an error code  outside  the  2xx
              range  in  the  HTTP  header. In other words, everything went as
              documented, except this particular case was  documented  not  to
              work.   The  third  element  is  the  dictionary  returned  from
              ::S3::REST.  Note that S3::REST itself never throws this  error,
              but  just  returns the dictionary. Most of the higher-level com-
              mands throw for convenience, unless an argument  indicates  they
              should  not.  If  something is documented as "not throwing an S3
              remote error", it means a  status  return  is  set  rather  than
              throwing  an error if Amazon returns a non-2XX HTTP result code.

       notyet The user obeyed the documentation, but the author  has  not  yet
              gotten around to implementing this feature. (Right now, only TLS
              support and sophisticated permissions fall into  this  category,
              as well as the S3::Acl command.)

       xml    The  service  has  returned  invalid XML, or XML whose schema is
              unexpected. For the high-level commands that accept service  XML
              as input for parsing, this may also be thrown.



COMMANDS

       This package provides several separate levels of complexity.

       o      The  lowest  level simply takes arguments to be sent to the ser-
              vice, sends them, retrieves the result, and provides it  to  the
              caller.   Note:  This  layer  allows both synchronous and event-
              driven processing. It depends on the MD5  and  SHA1  and  base64
              packages   from   Tcllib   (available  at  http://tcllib.source-
              forge.net/).  Note that S3::Configure is required  for  S3::REST
              to work due to the authentication portion, so we put that in the
              "lowest level."

       o      The next layer parses the results of calls, allowing  for  func-
              tionality  such  as  uploading only changed files, synchronizing
              directories, and so on.  This layer depends on the TclXML  pack-
              age  as  well  as  the included xsxp package. These packages are
              package required  when  these  more-sophisticated  routines  are
              called, so nothing breaks if they are not correctly installed.

       o      Also  included  is a separate program that uses the library.  It
              provides code to parse $argv0 and $argv from the  command  line,
              allowing invocation as a tclkit, etc.  (Not yet implmented.)

       o      Another separate program provides a GUI interface allowing drag-
              and-drop and other such functionality. (Not yet implemented.)

       o      Also built on this package is the OddJob program. It is a  sepa-
              rate  program  designed  to  allow distribution of computational
              work units over Amazon's Elastic Compute Cloud web service.


       The goal is to have at least the bottom-most layers implemented in pure
       Tcl  using only that which comes from widely-available sources, such as
       Tcllib.


LOW LEVEL COMMANDS

       These commands do not require any packages not listed above.  They talk
       directly to the service, or they are utility or configuration routines.
       Note that the "xsxp" package was written to support this package, so it
       should be available wherever you got this package.

       S3::Configure   ?-reset   boolean?   ?-retries  integer?  ?-accesskeyid
       idstring?  ?-secretaccesskey  idstring?  ?-service-access-point   FQDN?
       ?-use-tls    boolean?    ?-default-compare    always|never|exists|miss-
       ing|newer|date|checksum|different?     ?-default-separator      string?
       ?-default-acl      private|public-read|public-read-write|authenticated-
       read|keep|calc? ?-default-bucket bucketname?
              There  is one command for configuration, and that is S3::Config-
              ure.  If called with no arguments, it returns  a  dictionary  of
              key/value  pairs  listing  all current settings.  If called with
              one argument, it returns the value of that single argument.   If
              called  with two or more arguments, it must be called with pairs
              of arguments, and it applies the changes  in  order.   There  is
              only one set of configuration information per interpreter.

              The following options are accepted:

              -reset boolean
                     By default, false.  If true, any previous changes and any
                     changes on the same call before the reset option will  be
                     returned to default values.

              -retries integer
                     Default  value  is  3.   If Amazon returns a 500 error, a
                     retry after an exponential backoff delay  will  be  tried
                     this  many  times  before finally throwing the 500 error.
                     This applies to each call to S3::REST  from  the  higher-
                     level  commands,  but  not  to S3::REST itself.  That is,
                     S3::REST will always return httpstatus 500 if that's what
                     it  receives.  Functions  like S3::Put will retry the PUT
                     call, and will also retry the GET and HEAD calls used  to
                     do  content  comparison.  Changing this to 0 will prevent
                     retries and their associated delays.  In addition, socket
                     errors  (i.e.,  errors  whose  errorCode  starts with "S3
                     socket") will be similarly retried after backoffs.

              -accesskeyid idstring

              -secretaccesskey idstring
                     Each defaults to an empty  string.   These  must  be  set
                     before  any calls are made. This is your S3 ID.  Once you
                     sign up for an account, go to  http://www.amazonaws.com/,
                     sign  in,  go  to the "Your Web Services Account" button,
                     pick "AWS Access Identifiers", and your access key ID and
                     secret  access keys will be available. All S3::REST calls
                     are authenticated.  Blame Amazon for the poor  choice  of
                     names.

              -service-access-point FQDN
                     Defaults  to "s3.amazonaws.com". This is the fully-quali-
                     fied domain name of the server to  contact  for  S3::REST
                     calls.  You  should  probably  never  need to touch this,
                     unless someone else implements a compatible  service,  or
                     you  wish  to  test  something by pointing the library at
                     your own service.

              -slop-seconds integer
                     When  comparing  dates  between  Amazon  and  the   local
                     machine, two dates within this many seconds of each other
                     are considered the same. Useful for clock  drift  correc-
                     tion, processing overhead time, and so on.

              -use-tls boolean
                     Defaults  to false. This is not yet implemented. If true,
                     S3::REST will negotiate a TLS connection  to  Amazon.  If
                     false, unencrypted connections are used.

              -bucket-prefix string
                     Defaults to "TclS3".  This string is used by S3::Suggest-
                     BucketName if that command is passed an empty  string  as
                     an argument. It is used to distinguish different applica-
                     tions using the Amazon service.  Your application  should
                     always set this to keep from interfering with the buckets
                     of other users of Amazon S3 or with other buckets of  the
                     same user.

              -default-compare   always|never|exists|missing|newer|date|check-
              sum|different
                     Defaults  to  "always."  If  no  -compare is specified on
                     S3::Put, S3::Get, or S3::Delete, this comparison is used.
                     See those commands for a description of the meaning.

              -default-separator string
                     Defaults  to "/". This is currently unused. It might make
                     sense to use this for S3::Push and S3::Pull, but allowing
                     resources  to  have  slashes  in  their names that aren't
                     marking directories would  be  problematic.  Hence,  this
                     currently does nothing.

              -default-acl     private|public-read|public-read-write|authenti-
              cated-read|keep|calc
                     Defaults  to an empty string. If no -acl argument is pro-
                     vided to S3::Put or S3::Push, this string is used  (given
                     as  the x-amz-acl header if not keep or calc). If this is
                     also empty, no x-amz-acl header is  generated.   This  is
                     not used by S3::REST.

              -default-bucket bucketname
                     If  no  bucket  is given to S3::GetBucket, S3::PutBucket,
                     S3::Get,   S3::Put,   S3::Head,   S3::Acl,    S3::Delete,
                     S3::Push,  S3::Pull,  or S3::Toss, and if this configura-
                     tion variable is not an  empty  string  (and  not  simply
                     "/"),  then  this value will be used for the bucket. This
                     is useful if one program does a large amount of  resource
                     manipulation within a single bucket.


       S3::SuggestBucket ?name?
              The  S3::SuggestBucket  command  accepts an optional string as a
              prefix and returns a valid bucket containing the  name  argument
              and  the  Access Key ID. This makes the name unique to the owner
              and to the application (assuming the application  picks  a  good
              name  argument).  If no name is provided, the name from S3::Con-
              figure -bucket-prefix is used.  If that too is empty  (which  is
              not the default), an error is thrown.

       S3::REST dict
              The  S3::REST  command  takes  as  an  argument a dictionary and
              returns a dictionary.  The return dictionary has the  same  keys
              as  the  input  dictionary,  and includes additional keys as the
              result.  The presence or absence of keys in the input dictionary
              can  control  the  behavior  of the routine.  It never throws an
              error directly, but  includes  keys  "error",  "errorInfo",  and
              "errorCode"   if   necessary.   Some  keys  are  required,  some
              optional. The routine can run either in blocking or non-blocking
              mode,  based  on  the presense of resultvar in the input dictio-
              nary. This requires the -accesskeyid and -secretaccesskey to  be
              configured via S3::Configure before being called.

              The possible input keys are these:

              verb GET|PUT|DELETE|HEAD
                     This required item indicates the verb to be used.

              resource string
                     This required item indicates the resource to be accessed.
                     A leading / is added if not there  already.  It  will  be
                     URL-encoded  for  you  if  necessary.  Do  not  supply  a
                     resource name that is already URL-encoded.

              ?rtype torrent|acl?
                     This indicates a torrent or acl resource is being manipu-
                     lated.   Do  not include this in the resource key, or the
                     "?" separator will get URL-encoded.

              ?parameters dict?
                     This optional dictionary provides parameters added to the
                     URL  for the transaction. The keys must be in the correct
                     case (which is confusing in the Amazon documentation) and
                     the values must be valid. This can be an empty dictionary
                     or omitted entirely if  no  parameters  are  desired.  No
                     other error checking on parameters is performed.

              ?headers dict?
                     This  optional dictionary provides headers to be added to
                     the HTTP request. The keys must be in lower case for  the
                     authentication  to  work.  The  values  must  not contain
                     embedded newlines or carriage returns. This is  primarily
                     useful  for  adding x-amz-* headers. Since authentication
                     is calculated by S3::REST, do not add that  header  here.
                     Since content-type gets its own key, also do not add that
                     header here.

              ?inbody contentstring?
                     This optional item, if provided, gives the  content  that
                     will  be  sent.  It  is  sent  with a tranfer encoding of
                     binary, and only the low bytes are used, so use [encoding
                     convertto utf-8] if the string is a utf-8 string. This is
                     written all in one blast, so if you are using  non-block-
                     ing mode and the inbody is especially large, you may wind
                     up blocking on the write socket.

              ?infile filename?
                     This optional item, if provided, and  if  inbody  is  not
                     provided,  names the file from which the body of the HTTP
                     message will be constructed. The file is opened for read-
                     ing  and  sent progressively by [fcopy], so it should not
                     block in non-blocking mode  even  if  the  file  is  very
                     large.  The  file  is  transfered  in binary mode, so the
                     bytes on your disk will match the bytes in your resource.
                     Due  to  HTTP  restrictions,  it  must be possible to use
                     [file size] on this file to determine  the  size  at  the
                     start of the transaction.

              ?S3chan channel?
                     This  optional  item, if provided, indicates the already-
                     open socket over which the  transaction  should  be  con-
                     ducted. If not provided, a connection is made to the ser-
                     vice access point specified via S3::Configure,  which  is
                     normally s3.amazonaws.com. If this is provided, the chan-
                     nel is not closed at the end of the transaction.

              ?outchan channel?
                     This optional item, if provided, indicates  the  already-
                     open channel to which the body returned from S3 should be
                     written. That is, to retrieve a large  resource,  open  a
                     file,  set  the translation mode, and pass the channel as
                     the value of the key outchan. Output will be  written  to
                     the channel in pieces so memory does not fill up unneces-
                     sarily. The channel is not  closed  at  the  end  of  the
                     transaction.

              ?resultvar varname?
                     This  optional item, if provided, indicates that S3::REST
                     should run in non-blocking mode. The  varname  should  be
                     fully  qualified with respect to namespaces and cannot be
                     local to a proc. If provided, the result of the  S3::REST
                     call  is  assigned  to  this variable once everything has
                     completed; use trace or vwait to know when this has  hap-
                     pened.  If this key is not provided, the result is simply
                     returned from the call to S3::REST and no  calls  to  the
                     eventloop are invoked from within this call.

              ?throwsocket throw|return?
                     This  optional item, if provided, indicates that S3::REST
                     should throw an error if throwmode is throw and a  socket
                     error  is encountered.  It indicates that S3::REST should
                     return the error code in the  returned  dictionary  if  a
                     socket error is encountered and this is set to return. If
                     throwsocket is set to return or if the call is not block-
                     ing, then a socket error (i.e., an error whose error code
                     starts with "S3 socket" will be returned in  the  dictio-
                     nary as error, errorInfo, and errorCode.  If a foreground
                     call is made (i.e., resultvar is not provided), and  this
                     option  is  not  provided  or is set to throw, then error
                     will be invoked instead.

       Once the call to S3::REST completes, a new dict is returned, either  in
       the resultvar or as the result of execution. This dict is a copy of the
       original dict with the results added as new keys. The possible new keys
       are these:

              error errorstring

              errorInfo errorstring

              errorCode errorstring
                     If  an  error  is caught, these three keys will be set in
                     the result.  Note  that  S3::REST  does  not  consider  a
                     non-2XX HTTP return code as an error. The errorCode value
                     will  be  formatted  according  to  the  ERROR  REPORTING
                     description.   If these are present, other keys described
                     here might not be.

              httpstatus threedigits
                     The three-digit code from the HTTP transaction.  2XX  for
                     good, 5XX for server error, etc.

              httpmessage text
                     The  textual  result after the status code. "OK" or "For-
                     bidden" or etc.

              outbody contentstring
                     If outchan was not specified, this key will hold a refer-
                     ence  to  the  (unencoded) contents of the body returned.
                     If Amazon returned an error (a la the  httpstatus  not  a
                     2XX value), the error message will be in outbody or writ-
                     ten to outchan as appropriate.

              outheaders dict
                     This contains a dictionary of headers returned by Amazon.
                     The  keys  are  always lower case. It's mainly useful for
                     finding the x-amz-meta-* headers, if any, although things
                     like last-modified and content-type are also useful.  The
                     keys of this dictionary are always lower case.  Both keys
                     and values are trimmed of extraneous whitespace.



HIGH LEVEL COMMANDS

       The  routines  in  this  section  all  make use of one or more calls to
       S3::REST to do their work, then parse and manage the data in  a  conve-
       nient  way.   All  these  commands  throw  errors as described in ERROR
       REPORTING unless otherwise noted.

       In all these commands, all arguments are presented as name/value pairs,
       in any order. All the argument names start with a hyphen.

       There  are  a  few options that are common to many of the commands, and
       those common options are documented here.

       -blocking boolean
              If provided and specified as false, then any  calls  to  S3:REST
              will  be  non-blocking,  and internally these routines will call
              [vwait] to get the results. In other words, these routines  will
              return  the  same  value,  but  they'll have event loops running
              while waiting for Amazon.

       -parse-xml xmlstring
              If provided, the routine skips actually communicating with  Ama-
              zon,  and  instead  behaves  as  if  the XML string provided was
              returned as the body of the call. Since several  of  these  rou-
              tines allow the return of data in various formats, this argument
              can be used to parse existing XML to extract the bits of  infor-
              mation that are needed. It's also helpful for testing.

       -bucket bucketname
              Almost  every  high-level  command needs to know what bucket the
              resources are in. This option specifies that. (Only the  command
              to  list  available  buckets  does  not require this parameter.)
              This does not need to be URL-encoded, even if it  contains  spe-
              cial  or non-ASCII characters. May or may not contain leading or
              trailing spaces - commands normalize the bucket. If this is  not
              supplied,  the value is taken from S3::Configure -default-bucket
              if that string isn't empty. Note that  spaces  and  slashes  are
              always  trimmed  from  both ends and the rest must leave a valid
              bucket.

       -resource resourcename
              This specifies the resource of interest within the  bucket.   It
              may  or  may  not  start  with a slash - both cases are handled.
              This does not need to be URL-encoded, even if it  contains  spe-
              cial or non-ASCII characters.

       -compare always|never|exists|missing|newer|date|checksum|different
              When commands copy resources to files or files to resources, the
              caller may specify that the copy should be skipped if  the  con-
              tents are the same. This argument specifies the conditions under
              which the files should be copied.  If  it  is  not  passed,  the
              result  of S3::Configure -default-compare is used, which in turn
              defaults to "always." The meanings of  the  various  values  are
              these:

              always Always copy the data. This is the default.

              never  Never  copy the data. This is essentially a no-op, except
                     in S3::Push and S3::Pull where  the  -delete  flag  might
                     make a difference.

              exists Copy the data only if the destination already exists.

              missing
                     Copy  the  data  only if the destination does not already
                     exist.

              newer  Copy the data if the destination is missing,  or  if  the
                     date on the source is newer than the date on the destina-
                     tion by at least S3::Configure -slop-seconds seconds.  If
                     the  source  is  Amazon, the date is taken from the Last-
                     Modified header. If the source is local, it is  taken  as
                     the mtime of the file. If the source data is specified in
                     a string rather than a file, it is taken  as  right  now,
                     via [clock seconds].

              date   Like newer, except copy if the date is newer or older.

              checksum
                     Calculate  the  MD5 checksum on the local file or string,
                     ask Amazon for the eTag of the  resource,  and  copy  the
                     data if they're different. Copy the data also if the des-
                     tination is missing. Note that  this  can  be  slow  with
                     large local files unless the C version of the MD5 support
                     is available.

              different
                     Copy the data if the destination does not exist.  If  the
                     destination  exists and an actual file name was specified
                     (rather than a content string), and the date on the  file
                     differs from the date on the resource, copy the data.  If
                     the data is provided as a content string, the  "date"  is
                     treated  as  "right now", so it will likely always differ
                     unless slop-seconds is large.  If the dates are the same,
                     the MD5 checksums are compared, and the data is copied if
                     the checksums differ.

       Note that "newer" and "date" don't care about the contents, and "check-
       sum" doesn't care about the dates, but "different" checks both.

       S3::ListAllMyBuckets   ?-blocking   boolean?   ?-parse-xml   xmlstring?
       ?-result-type REST|xml|pxml|dict|names|owner?
              This  routine  performs a GET on the Amazon S3 service, which is
              defined to return a list of buckets owned by the account identi-
              fied  by  the  authorization  header. (Blame Amazon for the dumb
              names.)

              -blocking boolean
                     See above for standard definition.

              -parse-xml xmlstring
                     See above for standard definition.

              -result-type REST
                     The dictionary returned by S3::REST is the  return  value
                     of S3::ListAllMyBuckets. In this case, a non-2XX httpsta-
                     tus will not throw an error. You  may  not  combine  this
                     with -parse-xml.

              -result-type xml
                     The  raw  XML of the body is returned as the result (with
                     no encoding applied).

              -result-type pxml
                     The XML of the body as parsed by xsxp::parse is returned.

              -result-type dict
                     A  dictionary  of  interesting  portions  of  the  XML is
                     returned. The dictionary contains the following keys:

                     Owner/ID
                            The Amazon AWS ID (in hex) of  the  owner  of  the
                            bucket.

                     Owner/DisplayName
                            The Amazon AWS ID's Display Name.

                     Bucket/Name
                            A list of names, one for each bucket.

                     Bucket/CreationDate
                            A  list of dates, one for each bucket, in the same
                            order as Bucket/Name, in ISO format  (as  returned
                            by Amazon).


              -result-type names
                     A  list of bucket names is returned with all other infor-
                     mation stripped out.  This is the default result type for
                     this command.

              -result-type owner
                     A  list  containing  two  elements is returned. The first
                     element is the owner's ID, and the second is the  owner's
                     display name.


       S3::PutBucket  ?-bucket  bucketname?  ?-blocking boolean? ?-acl {}|pri-
       vate|public-read|public-read-write|authenticated-read?
              This  command  creates  a  bucket  if it does not already exist.
              Bucket names are globally unique, so you may get  a  "Forbidden"
              error  from  Amazon  even  if  you  cannot  see  the  bucket  in
              S3::ListAllMyBuckets. See S3::SuggestBucket for ways to minimize
              this  risk.  The x-amz-acl header comes from the -acl option, or
              from S3::Configure -default-acl if not specified.

       S3::DeleteBucket ?-bucket bucketname? ?-blocking boolean?
              This command deletes a bucket if it is empty and you  have  such
              permission.   Note  that  Amazon's  list  of buckets is a global
              resource, requiring far-flung synchronization. If you  delete  a
              bucket,  it may be quite a few minutes (or hours) before you can
              recreate it, yielding "Conflict" errors until then.

       S3::GetBucket ?-bucket bucketname? ?-blocking boolean? ?-parse-xml xml-
       string?  ?-max-count integer? ?-prefix prefixstring? ?-delimiter delim-
       iterstring? ?-result-type REST|xml|pxml|names|dict?
              This lists the contents of a  bucket.  That  is,  it  returns  a
              directory  listing  of  resources  within  a bucket, rather than
              transfering any user data.

              -bucket bucketname
                     The standard bucket argument.

              -blocking boolean
                     The standard blocking argument.

              -parse-xml xmlstring
                     The standard parse-xml argument.

              -max-count integer
                     If supplied, this is the most number  of  records  to  be
                     returned.   If  not supplied, the code will iterate until
                     all records have been found.  Not compatible with -parse-
                     xml.  Note  that  if  this  is supplied, only one call to
                     S3::REST will be made. Otherwise, enough  calls  will  be
                     made to exhaust the listing, buffering results in memory,
                     so take care if you may have huge buckets.

              -prefix prefixstring
                     If present, restricts listing to resources with a partic-
                     ular prefix. One leading / is stripped if present.

              -delimiter delimiterstring
                     If  present,  specifies a delimiter for the listing.  The
                     presence of this will summarize multiple  resources  into
                     one entry, as if S3 supported directories. See the Amazon
                     documentation for details.

              -result-type REST|xml|pxml|names|dict
                     This indicates the format of the  return  result  of  the
                     command.

                     REST   If   -max-count   is   specified,  the  dictionary
                            returned from S3::REST is returned. If  -max-count
                            is  not  specified, a list of all the dictionaries
                            returned from the one or more calls to S3::REST is
                            returned.

                     xml    If -max-count is specified, the body returned from
                            S3::REST is returned. If -max-count is not  speci-
                            fied,  a  list of all the bodies returned from the
                            one or more calls to S3::REST is returned.

                     pxml   If -max-count is specified, the body returned from
                            S3::REST  is  passed throught xsxp::parse and then
                            returned.  If -max-count is not specified, a  list
                            of  all  the  bodies returned from the one or more
                            calls  to  S3::REST  are   each   passed   through
                            xsxp::parse and then returned.

                     names  Returns  a  list  of all names found in either the
                            Contents/Key fields or  the  CommonPrefixes/Prefix
                            fields. If no -delimiter is specified and no -max-
                            count is specified, this returns  a  list  of  all
                            resources with the specified -prefix.

                     dict   Returns a dictionary. (Returns only one dictionary
                            even if -max-count wasn't specified.) The keys  of
                            the dictionary are as follows:

                            Name   The name of the bucket (from the final call
                                   to S3::REST).

                            Prefix From the final call to S3::REST.

                            Marker From the final call to S3::REST.

                            MaxKeys
                                   From the final call to S3::REST.

                            IsTruncated
                                   From the final call to S3::REST, so  always
                                   false if -max-count is not specified.

                            NextMarker
                                   Always provided if IsTruncated is true, and
                                   calculated of Amazon does not  provide  it.
                                   May be empty if IsTruncated is false.

                            Key    A  list of names of resources in the bucket
                                   matching   the   -prefix   and   -delimiter
                                   restrictions.

                            LastModified
                                   A list of times of resources in the bucket,
                                   in the same order as  Key,  in  the  format
                                   returned by Amazon. (I.e., it is not parsed
                                   into a seconds-from-epoch.)

                            ETag   A list of entity tags  (a.k.a.  MD5  check-
                                   sums) in the same order as Key.

                            Size   A  list of sizes in bytes of the resources,
                                   in the same order as Key.

                            Owner/ID
                                   A list of owners of the  resources  in  the
                                   bucket, in the same order as Key.

                            Owner/DisplayName
                                   A  list  of  owners of the resources in the
                                   bucket, in the same order as Key. These are
                                   the display names.

                            CommonPrefixes/Prefix
                                   A list of prefixes common to multiple enti-
                                   ties. This is present  only  if  -delimiter
                                   was supplied.

       S3::Put ?-bucket bucketname? -resource resourcename ?-blocking boolean?
       ?-file  filename?  ?-content   contentstring?   ?-acl   private|public-
       read|public-read-write|authenticated-read|calc|keep?     ?-content-type
       contenttypestring? ?-x-amz-meta-* metadatatext? ?-compare  comparemode?
              This  command  sends  data to a resource on Amazon's servers for
              storage, using the HTTP PUT command. It returns 0 if  the  -com-
              pare  mode  prevented the transfer, 1 if the transfer worked, or
              throws an error  if  the  transfer  was  attempted  but  failed.
              Server  5XX errors and S3 socket errors are retried according to
              S3:Configure -retries settings before throwing an  error;  other
              errors throw immediately.

              -bucket
                     This specifies the bucket into which the resource will be
                     written.  Leading and/or trailing slashes are removed for
                     you, as are spaces.

              -resource
                     This  is the full name of the resource within the bucket.
                     A single leading slash is removed,  but  not  a  trailing
                     slash.  Spaces are not trimmed.

              -blocking
                     The standard blocking flag.

              -file  If  this  is  specified, the filename must exist, must be
                     readable, and must not be a special  or  directory  file.
                     [file  size] must apply to it and must not change for the
                     lifetime of the call.  The default content-type is calcu-
                     lated  based  on  the  name  and/or contents of the file.
                     Specifying this is an error if -content  is  also  speci-
                     fied, but at least one of -file or -content must be spec-
                     ified. (The file is allowed to not exist or not be  read-
                     able if -compare never is specified.)

              -content
                     If  this  is  specified, the contentstring is sent as the
                     body  of  the  resource.  The  content-type  defaults  to
                     "application/octet-string".  Only the low bytes are sent,
                     so non-ASCII should use the appropriate encoding (such as
                     [encoding  convertto  utf-8])  before  passing it to this
                     routine, if necessary. Specifying this  is  an  error  if
                     -file  is  also  specified,  but at least one of -file or
                     -content must be specified.

              -acl   This defaults to S3::Configure -default-acl if not speci-
                     fied.  It sets the x-amz-acl header on the PUT operation.
                     If the value provided is calc, the  x-amz-acl  header  is
                     calculated based on the I/O permissions of the file to be
                     uploaded; it is an error to specify  calc  and  -content.
                     If the value provided is keep, the acl of the resource is
                     read before the PUT  (or  the  default  is  used  if  the
                     resource  does  not  exist), then set back to what it was
                     after the PUT (if it existed). An error will occur if the
                     resource  is successfully written but the kept ACL cannot
                     be then applied. This should never happen.   Note:   calc
                     is not currently fully implemented.

              -x-amz-meta-*
                     If  any  header  starts with "-x-amz-meta-", its contents
                     are added to the PUT command to  be  stored  as  metadata
                     with  the  resource. Again, no encoding is performed, and
                     the metadata should not contain characters like newlines,
                     carriage  returns,  and  so  on. It is best to stick with
                     simple ASCII strings, or to fix the  library  in  several
                     places.

              -content-type
                     This  overrides  the  content-type calculated by -file or
                     sets the content-type for -content.

              -compare
                     This is  the  standard  compare  mode  argument.  S3::Put
                     returns  1  if  the  data was copied or 0 if the data was
                     skipped due to  the  comparison  mode  so  indicating  it
                     should be skipped.


       S3::Get ?-bucket bucketname? -resource resourcename ?-blocking boolean?
       ?-compare  comparemode?  ?-file  filename?  ?-content   contentvarname?
       ?-timestamp aws|now? ?-headers headervarname?
              This  command  retrieves  data  from  a  resource on Amazon's S3
              servers, using the HTTP GET command. It returns 0 if  the  -com-
              pare  mode  prevented the transfer, 1 if the transfer worked, or
              throws an error if the transfer was attempted but failed. Server
              5XX  errors  and  S3  socket errors are are retried according to
              S3:Configure settings before throwing  an  error;  other  errors
              throw immediately. Note that this is always authenticated as the
              user configured  in  via  S3::Configure  -accesskeyid.  Use  the
              Tcllib http for unauthenticated GETs.

              -bucket
                     This specifies the bucket from which the resource will be
                     read.  Leading and/or trailing slashes  are  removed  for
                     you, as are spaces.

              -resource
                     This  is the full name of the resource within the bucket.
                     A single leading slash is removed,  but  not  a  trailing
                     slash.  Spaces are not trimmed.

              -blocking
                     The standard blocking flag.

              -file  If  this  is  specified, the body of the resource will be
                     read into this file,  incrementally  without  pulling  it
                     entirely  into  memory  first.  The parent directory must
                     already exist. If the file already  exists,  it  must  be
                     writable.  If  an  error  is  thrown part-way through the
                     process and the file already existed,  it  may  be  clob-
                     bered. If an error is thrown part-way through the process
                     and the file did not already exist, any partial bits will
                     be  deleted.  Specifying  this is an error if -content is
                     also specified, but at least one  of  -file  or  -content
                     must be specified.

              -timestamp
                     This  is  only valid in conjunction with -file. It may be
                     specified as now or aws. The default is now. If now,  the
                     file's  modification  date  is  left up to the system. If
                     aws, the file's mtime is set to match  the  Last-Modified
                     header  on  the resource, synchronizing the two appropri-
                     ately for -compare date or -compare newer.

              -content
                     If this is specified, the contentvarname is a variable in
                     the caller's scope (not necessarily global) that receives
                     the value of the body of the  resource.  No  encoding  is
                     done, so if the resource (for example) represents a UTF-8
                     byte sequence, use [encoding convertfrom utf-8] to get  a
                     valid UTF-8 string. If this is specified, the -compare is
                     ignored unless it is never, in which case  no  assignment
                     to  contentvarname  is  performed.  Specifying this is an
                     error if -file is also specified, but  at  least  one  of
                     -file or -content must be specified.

              -compare
                     This  is  the  standard  compare  mode  argument. S3::Get
                     returns 1 if the data was copied or 0  if  the  data  was
                     skipped  due  to  the  comparison  mode  so indicating it
                     should be skipped.

              -headers
                     If this is specified,  the  headers  resulting  from  the
                     fetch  are  stored in the provided variable, as a dictio-
                     nary. This will  include  content-type  and  x-amz-meta-*
                     headers,  as well as the usual HTTP headers, the x-amz-id
                     debugging headers, and so on. If no file is fetched  (due
                     to -compare or other errors), no assignment to this vari-
                     able is performed.


       S3::Head  ?-bucket  bucketname?   -resource   resourcename   ?-blocking
       boolean? ?-dict dictvarname? ?-headers headersvarname? ?-status status-
       varname?
              This command  requests  HEAD  from  the  resource.   It  returns
              whether  a  2XX  code  was  returned as a result of the request,
              never throwing an S3 remote error.  That is, if this returns  1,
              the  resource exists and is accessible. If this returns 0, some-
              thing went wrong, and the -status result can  be  consulted  for
              details.

              -bucket
                     This specifies the bucket from which the resource will be
                     read.  Leading and/or trailing slashes  are  removed  for
                     you, as are spaces.

              -resource
                     This  is the full name of the resource within the bucket.
                     A single leading slash is removed,  but  not  a  trailing
                     slash.  Spaces are not trimmed.

              -blocking
                     The standard blocking flag.

              -dict  If  specified, the resulting dictionary from the S3::REST
                     call  is  assigned  to  the  indicated  (not  necessarily
                     global) variable in the caller's scope.

              -headers
                     If  specified,  the dictionary of headers from the result
                     are assigned to the indicated  (not  necessarily  global)
                     variable in the caller's scope.

              -status
                     If  specified,  the  indicated  (not  necessarily global)
                     variable in the caller's scope is  assigned  a  2-element
                     list.  The first element is the 3-digit HTTP status code,
                     while the second element is the  HTTP  message  (such  as
                     "OK" or "Forbidden").

       S3::GetAcl ?-blocking boolean? ?-bucket bucketname? -resource resource-
       name ?-result-type REST|xml|pxml?
              This command gets the ACL of the indicated resource or throws an
              error if it is unavailable.

              -blocking boolean
                     See above for standard definition.

              -bucket
                     This specifies the bucket from which the resource will be
                     read.   Leading  and/or  trailing slashes are removed for
                     you, as are spaces.

              -resource
                     This is the full name of the resource within the  bucket.
                     A  single  leading  slash  is removed, but not a trailing
                     slash.  Spaces are not trimmed.

              -parse-xml xml
                     The XML from a previous GetACL can be  passed  in  to  be
                     parsed  into dictionary form.  In this case, -result-type
                     must be pxml or dict.

              -result-type REST
                     The dictionary returned by S3::REST is the  return  value
                     of  S3::GetAcl.   In this case, a non-2XX httpstatus will
                     not throw an error.

              -result-type xml
                     The raw XML of the body is returned as the  result  (with
                     no encoding applied).

              -result-type pxml
                     The XML of the body as parsed by xsxp::parse is returned.

              -result-type dict
                     This fetches the ACL, parses it, and returns a dictionary
                     of two elements.

                     The  first element has the key "owner" whose value is the
                     canonical ID of the owner of the resource.

                     The second element has the key "acl"  whose  value  is  a
                     dictionary.   Each  key  in the dictionary is one of Ama-
                     zon's permissions, namely  "READ",  "WRITE",  "READ_ACP",
                     "WRITE_ACP",  or  "FULL_CONTROL".  Each value of each key
                     is a list of canonical IDs or group URLs that  have  that
                     permission.  Elements are not in the list in any particu-
                     lar order, and not  all  keys  are  necessarily  present.
                     Display  names  are  not  returned, as they are not espe-
                     cially useful; use pxml to obtain them if necessary.

       S3::PutAcl ?-blocking boolean? ?-bucket bucketname? -resource resource-
       name ?-acl new-acl?
              This sets the ACL on the indicated resource. It returns the  XML
              written to the ACL, or throws an error if anything went wrong.

              -blocking boolean
                     See above for standard definition.

              -bucket
                     This specifies the bucket from which the resource will be
                     read.  Leading and/or trailing slashes  are  removed  for
                     you, as are spaces.

              -resource
                     This  is the full name of the resource within the bucket.
                     A single leading slash is removed,  but  not  a  trailing
                     slash.  Spaces are not trimmed.

              -owner If  this is provided, it is assumed to match the owner of
                     the resource.  Otherwise, a GET may  need  to  be  issued
                     against  the  resource  to find the owner. If you already
                     have the owner (such as from a call  to  S3::GetAcl,  you
                     can  pass  the  value  of the "owner" key as the value of
                     this option, and it will be used in the  construction  of
                     the XML.

              -acl   If  this  option  is  specified,  it provides the ACL the
                     caller wishes to write to the resource. If  this  is  not
                     supplied or is empty, the value is taken from S3::Config-
                     ure -default-acl.  The ACL is written with a PUT  to  the
                     ?acl resource.

                     If the value passed to this option starts with "<", it is
                     taken to be a body to be PUT to the ACL resource.

                     If the value matches one of the standard Amazon x-amz-acl
                     headers  (i.e.,  a  canned access policy), that header is
                     translated to XML and then  applied.  The  canned  access
                     policies are private, public-read, public-read-write, and
                     authenticated-read (in lower case).

                     Otherwise, the value is assumed to be a  dictionary  for-
                     matted  as the "acl" sub-entry within the dict returns by
                     S3::GetAcl -result-type dict.  The proper XML  is  gener-
                     ated and applied to the resource.  Note that a value con-
                     taining "//" is assumed to be a group, a value containing
                     "@" is assumed to be an AmazonCustomerByEmail, and other-
                     wise the value is assumed to be a canonical Amazon ID.

                     Note that you cannot change the owner, so calling  GetAcl
                     on  a  resource  owned  by  one  user and applying it via
                     PutAcl on a resource owned by another  user  may  not  do
                     exactly what you expect.

       S3::Delete   ?-bucket  bucketname?  -resource  resourcename  ?-blocking
       boolean? ?-status statusvar?
              This  command  deletes the specified resource from the specified
              bucket.  It returns 1 if the resource was deleted  successfully,
              0  otherwise.   It  returns  0 rather than throwing an S3 remote
              error.

              -bucket
                     This specifies the bucket from which the resource will be
                     deleted.  Leading and/or trailing slashes are removed for
                     you, as are spaces.

              -resource
                     This is the full name of the resource within the  bucket.
                     A  single  leading  slash  is removed, but not a trailing
                     slash.  Spaces are not trimmed.

              -blocking
                     The standard blocking flag.

              -status
                     If specified,  the  indicated  (not  necessarily  global)
                     variable  in  the  caller's scope is set to a two-element
                     list. The first element is the 3-digit HTTP status  code.
                     The  second  element is the HTTP message (such as "OK" or
                     "Forbidden"). Note that Amazon's DELETE result is 204  on
                     success, that being the code indicating no content in the
                     returned body.


       S3::Push ?-bucket bucketname? -directory  directoryname  ?-prefix  pre-
       fixstring?  ?-compare  comparemode?  ?-x-amz-meta-*  metastring?  ?-acl
       aclcode? ?-delete  boolean?  ?-error  throw|break|continue?  ?-progress
       scriptprefix?
              This synchronises a local directory  with  a  remote  bucket  by
              pushing  the  differences  using S3::Put. Note that if something
              has changed in the bucket but not locally, those  changes  could
              be  lost.  Thus,  this  is not a general two-way synchronization
              primitive. (See S3::Sync for that.) Note too that resource names
              are  case sensitive, so changing the case of a file on a Windows
              machine may lead to otherwise-unnecessary transfers.  Note  that
              only  regular files are considered, so devices, pipes, symlinks,
              and directories are not copied.

              -bucket
                     This names the bucket into which data will be pushed.

              -directory
                     This names the local directory from which files  will  be
                     taken.   It must exist, be readable via [glob] and so on.
                     If only some of the files therein are readable,  S3::Push
                     will  PUT those files that are readable and return in its
                     results the list of files that could not be opened.

              -prefix
                     This  names  the  prefix  that  will  be  added  to   all
                     resources.   That  is,  it  is  the  remote equivalent of
                     -directory.  If it is not  specified,  the  root  of  the
                     bucket  will be treated as the remote directory. An exam-
                     ple may clarify.

                     S3::Push -bucket test -directory /tmp/xyz -prefix hello/world

                     In this example,  /tmp/xyz/pdq.html  will  be  stored  as
                     http://s3.amazonaws.com/test/hello/world/pdq.html in Ama-
                     zon's  servers.  Also,  /tmp/xyz/abc/def/Hello  will   be
                     stored                as                http://s3.amazon-
                     aws.com/test/hello/world/abc/def/Hello    in     Amazon's
                     servers.   Without  the -prefix option, /tmp/xyz/pdq.html
                     would be stored as http://s3.amazonaws.com/test/pdq.html.

              -blocking
                     This is the standard blocking option.

              -compare
                     If present, this is passed to each invocation of S3::Put.
                     Naturally, S3::Configure -default-compare is used if this
                     is not specified.

              -x-amz-meta-*
                     If present, this is passed to each invocation of S3::Put.
                     All copied files will have the same metadata.

              -acl   If present, this is passed to each invocation of S3::Put.

              -delete
                     This  defaults to false. If true, resources in the desti-
                     nation that are not in the source directory  are  deleted
                     with  S3::Delete.   Since  only regular files are consid-
                     ered, the existance of a symlink, pipe, device, or direc-
                     tory in the local source will not prevent the deletion of
                     a remote resource with a corresponding name.

              -error This controls the behavior of S3::Push in the event  that
                     S3::Put  throws an error. Note that errors encountered on
                     the local file system or in reading the list of resources
                     in  the  remote  bucket always throw errors.  This option
                     allows control over "partial"  errors,  when  some  files
                     were  copied and some were not. S3::Delete is always fin-
                     ished up, with  errors  simply  recorded  in  the  return
                     result.

                     throw  The error is rethrown with the same errorCode.

                     break  Processing  stops  without  throwing an error, the
                            error is recorded in the  return  value,  and  the
                            command  returns  with a normal return.  The calls
                            to S3::Delete are not started.

                     continue
                            This is the default. Processing continues  without
                            throwing,   recording  the  error  in  the  return
                            result, and resuming with the  next  file  in  the
                            local directory to be copied.

              -progress
                     If  this  is specified and the indicated script prefix is
                     not empty, the indicated script prefix  will  be  invoked
                     several  times  in  the  caller's context with additional
                     arguments at various  points  in  the  processing.   This
                     allows  progress  reporting  without  backgrounding.  The
                     provided prefix will be  invoked  with  additional  argu-
                     ments, with the first additional argument indicating what
                     part of the process is being reported on.  The prefix  is
                     initially invoked with args as the first additional argu-
                     ment and a dictionary representing the  normalized  argu-
                     ments to the S3::Push call as the second additional argu-
                     ment.  Then the prefix is invoked with local as the first
                     additional  argument  and a list of suffixes of the files
                     to be considered as the second argument.  Then the prefix
                     is  invoked  with remote as the first additional argument
                     and a list of suffixes existing in the remote  bucket  as
                     the  second  additional argument.  Then, for each file in
                     the local list, the prefix will be invoked with start  as
                     the  first  additional  argument and the common suffix as
                     the second additional argument.  When S3::Put returns for
                     that  file,  the  prefix will be invoked with copy as the
                     first additional argument, the common suffix as the  sec-
                     ond  additional  argument, and a third argument that will
                     be "copied" (if S3::Put sent the resource), "skipped" (if
                     S3::Put  decided not to based on -compare), or the error-
                     Code that S3::Put threw  due  to  unexpected  errors  (in
                     which  case the third argument is a list that starts with
                     "S3"). When all files have been  transfered,  the  prefix
                     may  be  invoked  zero  or  more times with delete as the
                     first additional argument and the suffix of the  resource
                     being  deleted  as the second additional argument, with a
                     third argument being  either  an  empty  string  (if  the
                     delete  worked)  or  the  errorCode from S3::Delete if it
                     failed. Finally, the prefix will be invoked with finished
                     as  the first additional argument and the return value as
                     the second additional argument.
       The return result from this command is a dictionary. They keys are  the
       suffixes (i.e., the common portion of the path after the -directory and
       -prefix), while the values are either "copied", "skipped" (if  -compare
       indicated not to copy the file), or the errorCode thrown by S3::Put, as
       appropriate. If -delete was true, there may also be  entries  for  suf-
       fixes  with the value "deleted" or "notdeleted", indicating whether the
       attempted S3::Delete worked or not, respectively. There  is  one  addi-
       tional  pair  in  the  return result, whose key is the empty string and
       whose value is a nested dictionary.  The keys of this nested dictionary
       include  "filescopied"  (the  number  of  files  successfully  copied),
       "bytescopied" (the number of data bytes in the files copied,  excluding
       headers,  metadata,  etc),  "compareskipped"  (the  number of files not
       copied due to -compare mode), "errorskipped" (the number of  files  not
       copied  due  to thrown errors), "filesdeleted" (the number of resources
       deleted due to not having corresponding files locally, or 0 if  -delete
       is  false),  and "filesnotdeleted" (the number of resources whose dele-
       tion was attempted but failed).

       Note that this is currently  implemented  somewhat  inefficiently.   It
       fetches the bucket listing (including timestamps and eTags), then calls
       S3::Put, which uses HEAD to find the timestamps and eTags  again.  Cor-
       recting this with no API change is planned for a future upgrade.


       S3::Pull  ?-bucket  bucketname?  -directory directoryname ?-prefix pre-
       fixstring? ?-blocking boolean? ?-compare comparemode? ?-delete boolean?
       ?-timestamp  aws|now?  ?-error throw|break|continue? ?-progress script-
       prefix?
              This  synchronises  a  remote  bucket  with a local directory by
              pulling the differences using  S3::Get  If  something  has  been
              changed  locally  but not in the bucket, those difference may be
              lost. This is not a general two-way  synchronization  mechanism.
              (See  S3::Sync  for  that.)  This creates directories if needed;
              new directories are created with default permissions. Note  that
              resource  names  are  case  sensitive, so changing the case of a
              file on a Windows  machine  may  lead  to  otherwise-unnecessary
              transfers.  Also,  try  not  to store data in resources that end
              with a slash, or which are prefixes of resources that  otherwise
              would start with a slash; i.e., don't use this if you store data
              in resources whose names have to be directories locally.

              Note that this is currently implemented somewhat  inefficiently.
              It  fetches the bucket listing (including timestamps and eTags),
              then calls S3::Get, which uses HEAD to find the  timestamps  and
              eTags again. Correcting this with no API change is planned for a
              future upgrade.

              -bucket
                     This names the bucket from which data will be pulled.

              -directory
                     This names the local directory into which files  will  be
                     written  It  must exist, be readable via [glob], writable
                     for file creation, and so on. If only some of  the  files
                     therein  are writable, S3::Pull will GET those files that
                     are writable and return in its results the list of  files
                     that could not be opened.

              -prefix
                     The  prefix  of  resources  that  will  be considered for
                     retrieval.  See S3::Push for more details, examples, etc.
                     (Of  course,  S3::Pull  reads rather than writes, but the
                     prefix is treated similarly.)

              -blocking
                     This is the standard blocking option.

              -compare
                     This is passed to each invocation of S3::Get if provided.
                     Naturally, S3::Configure -default-compare is used if this
                     is not provided.

              -timestamp
                     This is passed to each invocation of S3::Get if provided.

              -delete
                     If  this  is  specified and true, files that exist in the
                     -directory that are not in the -prefix  will  be  deleted
                     after  all resources have been copied. In addition, empty
                     directories (other than the top-level -directory) will be
                     deleted,  as  Amazon S3 has no concept of an empty direc-
                     tory.

              -error See S3::Push for a description of this option.

              -progress
                     See S3::Push for a description of this option.   It  dif-
                     fers  slightly  in that local directories may be included
                     with a trailing slash to indicate they are directories.
       The return value from this command is a dictionary. It is identical  in
       form  and  meaning to the description of the return result of S3::Push.
       It differs only in that directories may be included,  with  a  trailing
       slash in their name, if they are empty and get deleted.

       S3::Toss  ?-bucket bucketname? -prefix prefixstring ?-blocking boolean?
       ?-error throw|break|continue? ?-progress scriptprefix?
              This  deletes some or all resources within a bucket. It would be
              considered a "recursive delete" had  Amazon  implemented  actual
              directories.

              -bucket
                     The bucket from which resources will be deleted.

              -blocking
                     The standard blocking option.

              -prefix
                     The prefix for resources to be deleted. Any resource that
                     starts  with  this  string  will  be  deleted.  This   is
                     required.   To  delete  everything in the bucket, pass an
                     empty string for the prefix.

              -error If this is  "throw",  S3::Toss  rethrows  any  errors  it
                     encounters.   If this is "break", S3::Toss returns with a
                     normal return after the first error, recording that error
                     in the return result. If this is "continue", which is the
                     default, S3::Toss continues on and lists  all  errors  in
                     the return result.

              -progress
                     If  this is specified and not an empty string, the script
                     prefix will be invoked several times in  the  context  of
                     the  caller  with  additional  arguments  appended.  Ini-
                     tially, it will be  invoked  with  the  first  additional
                     argument  being  args  and the second being the processed
                     list of arguments to S3::Toss. Then it  is  invoked  with
                     remote  as  the first additional argument and the list of
                     suffixes in the bucket to be deleted as the second  addi-
                     tional  argument. Then it is invoked with the first addi-
                     tional argument being delete and  the  second  additional
                     argument  being  the  suffix  deleted and the third addi-
                     tional argument being "deleted" or "notdeleted" depending
                     on  whether  S3::Delete  threw  an  error.   Finally, the
                     script prefix is invoked with a first additional argument
                     of  "finished"  and  a  second additional argument of the
                     return value.
       The return value is a dictionary. The keys are the  suffixes  of  files
       that  S3::Toss  attempted  to  delete,  and whose values are either the
       string "deleted" or "notdeleted". There is also  one  additional  pair,
       whose  key  is  the empty string and whose value is an embedded dictio-
       nary. The keys of this embedded dictionary include  "filesdeleted"  and
       "filesnotdeleted", each of which has integer values.



LIMITATIONS

       o      The  pure-Tcl  MD5 checking is slow. If you are processing files
              in the megabyte  range,  consider  ensuring  binary  support  is
              available.

       o      The  commands  S3::Pull  and  S3::Push fetch a directory listing
              which includes timestamps and MD5 hashes,  then  invoke  S3::Get
              and  S3::Put.  If  a complex -compare mode is specified, S3::Get
              and S3::Put will invoke a HEAD operation for each file to  fetch
              timestamps and MD5 hashes of each resource again. It is expected
              that a future release of this package will  solve  this  without
              any API changes.

       o      The  commands  S3::Pull  and  S3::Push fetch a directory listing
              without using -max-count. The entire directory  is  pulled  into
              memory  at once. For very large buckets, this could be a perfor-
              mance problem. The author, at this time, does not plan to change
              this behavior. Welcome to Open Source.

       o      S3::Sync is neither designed nor implemented yet.  The intention
              would be to keep changes synchronised, so changes could be  made
              to  both  the  bucket  and  the local directory and be merged by
              S3::Sync.

       o      Nor is -compare calc fully implemented. This is primarily due to
              Windows  not  providing  a  convenient method for distinguishing
              between local files  that  are  "public-read"  or  "public-read-
              write". Assistance figuring out TWAPI for this would be appreci-
              ated. The U**X semantics are difficult to map directly as  well.
              See  the  source for details.  Note that there are not tests for
              calc, since it isn't done yet.

       o      The HTTP processing is implemented within  the  library,  rather
              than  using a "real" HTTP package. Hence, multi-line headers are
              not (yet) handled correctly. Do not include carriage returns  or
              linefeeds  in  x-amz-meta-* headers, content-type values, and so
              on.  The author does not at this time expect to improve this.

       o      Internally, S3::Push and S3::Pull and S3::Toss are all very sim-
              ilar and should be refactored.

       o      The  idea  of  using -compare never -delete true to delete files
              that have been deleted from one place but not the other yet  not
              copying changed files is untested.



USAGE SUGGESTIONS

       To  fetch  a  "directory"  out  of a bucket, make changes, and store it
       back:

       file mkdir ./tempfiles
       S3::Pull -bucket sample -prefix of/interest -directory ./tempfiles \
         -timestamp aws
       do_my_process ./tempfiles other arguments
       S3::Push -bucket sample -prefix of/interest -directory ./tempfiles \
         -compare newer -delete true


       To delete files locally that were deleted off of S3 but  not  otherwise
       update files:

       S3::Pull -bucket sample -prefix of/interest -directory ./myfiles \
         -compare never -delete true



FUTURE DEVELOPMENTS

       The  author  intends  to work on several additional projects related to
       this package, in addition to finishing the unfinished features.

       First, a command-line program allowing browsing of buckets and transfer
       of files from shell scripts and command prompts is useful.

       Second,  a GUI-based program allowing visual manipulation of bucket and
       resource trees not unlike Windows Explorer would be useful.

       Third, a command-line (and perhaps a GUI-based) program called "OddJob"
       that  will  use  S3 to synchronize computation amongst multiple servers
       running OddJob. An S3 bucket will be set up with a number of scripts to
       run,  and the OddJob program can be invoked on multiple machines to run
       scripts on all the machines, each moving on to the next unstarted  task
       as  it finishes each.  This is still being designed, and it is intended
       primarily to be run on Amazon's Elastic Compute Cloud.


BUGS, IDEAS, FEEDBACK

       This document, and the package it describes, will  undoubtedly  contain
       bugs  and other problems.  Please report such in the category amazon-s3
       of       the       Tcllib       SF       Trackers       [http://source-
       forge.net/tracker/?group_id=12883].   Please  also report any ideas for
       enhancements you may have for either package and/or documentation.


KEYWORDS

       amazon, cloud, s3


CATEGORY

       Networking


COPYRIGHT

       Copyright (c) Copyright 2006,2008 Darren New. All Rights Reserved. See LICENSE.TXT for terms.




amazon-s3                            1.0.0                               S3(n)

Mac OS X 10.8 - Generated Wed Sep 5 05:20:48 CDT 2012
© manpagez.com 2000-2025
Individual documents may contain additional copyright information.