Documentation ¶
Overview ¶
Copyright © Microsoft <wastore@microsoft.com>
Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
Index ¶
- Constants
- Variables
- func Execute(azsAppPathFolder, logPathFolder string, jobPlanFolder string, ...)
- func GetAccountRoot(resource common.ResourceString, location common.Location) (string, error)
- func GetContainerName(path string, location common.Location) (string, error)
- func GetCredTypeFromEnvVar() common.CredentialType
- func GetCredentialInfoForLocation(ctx context.Context, location common.Location, resource, resourceSAS string, ...) (credInfo common.CredentialInfo, isPublic bool, err error)
- func GetOAuthTokenManagerInstance() (*common.UserOAuthTokenManager, error)
- func GetResourceRoot(resource string, location common.Location) (resourceBase string, err error)
- func GetUserOAuthTokenManagerInstance() *common.UserOAuthTokenManager
- func HandleListJobsCommand(jobStatus common.JobStatus) error
- func HandlePauseCommand(jobIdString string)
- func HandleShowCommand(listRequest common.ListRequest) error
- func InferArgumentLocation(arg string) common.Location
- func InitPipeline(ctx context.Context, location common.Location, ...) (p pipeline.Pipeline, err error)
- func IsDestinationCaseInsensitive(fromTo common.FromTo) bool
- func ParseSizeString(s string, name string) (int64, error)
- func PrintExistingJobIds(listJobResponse common.ListJobsResponse) error
- func PrintJobProgressSummary(summary common.ListJobSummaryResponse)
- func PrintJobTransfers(listTransfersResponse common.ListJobTransfersResponse)
- func SplitResourceString(raw string, loc common.Location) (common.ResourceString, error)
- func UnfurlSymlinks(symlinkPath string) (result string, err error)
- func ValidateFromTo(src, dst string, userSpecifiedFromTo string) (common.FromTo, error)
- func WalkWithSymlinks(fullPath string, walkFunc filepath.WalkFunc, followSymlinks bool) (err error)
- func WarnStdoutAndScanningLog(toLog string)
- func WrapFolder(fullpath string, stat os.FileInfo) (os.FileInfo, error)
- type AccountTraverser
- type BlobFSAccountTraverser
- type BucketToContainerNameResolver
- type CookedCopyCmdArgs
- func (cca *CookedCopyCmdArgs) Cancel(lcm common.LifecycleMgr)
- func (cca *CookedCopyCmdArgs) InitModularFilters() []ObjectFilter
- func (cca *CookedCopyCmdArgs) MakeEscapedRelativePath(source bool, dstIsDir bool, object StoredObject) (relativePath string)
- func (cca *CookedCopyCmdArgs) ReportProgressOrExit(lcm common.LifecycleMgr) (totalKnownCount uint32)
- type CopyEnumerator
- type FilterSet
- type GCPBucketNameToAzureResourcesResolver
- type HTTPClient
- type IncludeAfterDateFilter
- func (f *IncludeAfterDateFilter) AppliesOnlyToFiles() bool
- func (f *IncludeAfterDateFilter) DoesPass(storedObject StoredObject) bool
- func (f *IncludeAfterDateFilter) DoesSupportThisOS() (msg string, supported bool)
- func (_ IncludeAfterDateFilter) FormatAsUTC(t time.Time) string
- func (_ IncludeAfterDateFilter) ParseISO8601(s string, chooseEarliest bool) (time.Time, error)
- type IncludeBeforeDateFilter
- func (f *IncludeBeforeDateFilter) AppliesOnlyToFiles() bool
- func (f *IncludeBeforeDateFilter) DoesPass(storedObject StoredObject) bool
- func (f *IncludeBeforeDateFilter) DoesSupportThisOS() (msg string, supported bool)
- func (_ IncludeBeforeDateFilter) FormatAsUTC(t time.Time) string
- func (_ IncludeBeforeDateFilter) ParseISO8601(s string, chooseEarliest bool) (time.Time, error)
- type IncludeFilter
- type ListReq
- type ListResponse
- type LocationLevel
- type ObjectFilter
- type ResourceTraverser
- type S3BucketNameToAzureResourcesResolver
- type StoredObject
- type Version
Constants ¶
const (
NumOfFilesPerDispatchJobPart = 10000
)
const PreservePermissionsFlag = "preserve-permissions"
Variables ¶
var AzcopyAppPathFolder string
var EnumerationParallelStatFiles = false
var EnumerationParallelism = 1
var FinalPartCreatedMessage = "Final job part has been created"
var IPv4Regex = regexp.MustCompile(`\d+\.\d+\.\d+\.\d+`) // simple regex
var NothingScheduledError = errors.New("no transfers were scheduled because no files matched the specified criteria")
var NothingToRemoveError = errors.New("nothing found to remove")
var Rpc = func(cmd common.RpcCmd, request interface{}, response interface{}) { err := inprocSend(cmd, request, response) common.PanicIfErr(err) }
Global singleton for sending RPC requests from the frontend to the STE
Functions ¶
func Execute ¶
func Execute(azsAppPathFolder, logPathFolder string, jobPlanFolder string, maxFileAndSocketHandles int, jobID common.JobID)
Execute adds all child commands to the root command and sets flags appropriately. This is called by main.main(). It only needs to happen once to the rootCmd.
func GetAccountRoot ¶
func GetContainerName ¶
func GetCredTypeFromEnvVar ¶
func GetCredTypeFromEnvVar() common.CredentialType
GetCredTypeFromEnvVar tries to get credential type from environment variable defined by envVarCredentialType.
func GetOAuthTokenManagerInstance ¶
func GetOAuthTokenManagerInstance() (*common.UserOAuthTokenManager, error)
* GetInstanceOAuthTokenInfo returns OAuth token, obtained by auto-login, * for current instance of AzCopy.
func GetResourceRoot ¶
GetResourceRoot should eliminate wildcards and error out in invalid scenarios. This is intended for the jobPartOrder.SourceRoot.
func GetUserOAuthTokenManagerInstance ¶
func GetUserOAuthTokenManagerInstance() *common.UserOAuthTokenManager
GetUserOAuthTokenManagerInstance gets or creates OAuthTokenManager for current user. Note: Currently, only support to have TokenManager for one user mapping to one tenantID.
func HandleListJobsCommand ¶
HandleListJobsCommand sends the ListJobs request to transfer engine Print the Jobs in the history of Azcopy
func HandlePauseCommand ¶
func HandlePauseCommand(jobIdString string)
handles the pause command dispatches the pause Job order to the storage engine
func HandleShowCommand ¶
func HandleShowCommand(listRequest common.ListRequest) error
handles the list command dispatches the list order to the transfer engine
func InferArgumentLocation ¶
func InitPipeline ¶
func PrintExistingJobIds ¶
func PrintExistingJobIds(listJobResponse common.ListJobsResponse) error
PrintExistingJobIds prints the response of listOrder command when listOrder command requested the list of existing jobs
func PrintJobProgressSummary ¶
func PrintJobProgressSummary(summary common.ListJobSummaryResponse)
PrintJobProgressSummary prints the response of listOrder command when listOrder command requested the progress summary of an existing job
func PrintJobTransfers ¶
func PrintJobTransfers(listTransfersResponse common.ListJobTransfersResponse)
PrintJobTransfers prints the response of listOrder command when list Order command requested the list of specific transfer of an existing job
func SplitResourceString ¶
func UnfurlSymlinks ¶
func ValidateFromTo ¶
func WalkWithSymlinks ¶
WalkWithSymlinks is a symlinks-aware, parallelized, version of filePath.Walk. Separate this from the traverser for two purposes: 1) Cleaner code 2) Easier to test individually than to test the entire traverser.
func WarnStdoutAndScanningLog ¶
func WarnStdoutAndScanningLog(toLog string)
Types ¶
type AccountTraverser ¶
type AccountTraverser interface { ResourceTraverser // contains filtered or unexported methods }
type BlobFSAccountTraverser ¶
type BlobFSAccountTraverser struct {
// contains filtered or unexported fields
}
We don't allow S2S from BlobFS, but what this gives us is the ability for users to download entire accounts at once. This is just added to create that feature parity. Enumerates an entire blobFS account, looking into each matching filesystem as it goes
func (*BlobFSAccountTraverser) IsDirectory ¶
func (t *BlobFSAccountTraverser) IsDirectory(isSource bool) bool
func (*BlobFSAccountTraverser) Traverse ¶
func (t *BlobFSAccountTraverser) Traverse(preprocessor objectMorpher, processor objectProcessor, filters []ObjectFilter) error
type CookedCopyCmdArgs ¶
type CookedCopyCmdArgs struct { // from arguments Source common.ResourceString Destination common.ResourceString FromTo common.FromTo // new include/exclude only apply to file names // implemented for remove (and sync) only // includePathPatterns are handled like a list-of-files. Do not panic. This is not a bug that it is not present here. IncludePatterns []string ExcludePatterns []string ExcludePathPatterns []string IncludeFileAttributes []string ExcludeFileAttributes []string IncludeBefore *time.Time IncludeAfter *time.Time // list of version ids ListOfVersionIDs chan string // filters from flags ListOfFilesChannel chan string // Channels are nullable. Recursive bool StripTopDir bool FollowSymlinks bool ForceWrite common.OverwriteOption // says whether we should try to overwrite ForceIfReadOnly bool // says whether we should _force_ any overwrites (triggered by forceWrite) to work on Azure Files objects that are set to read-only CheckLength bool LogVerbosity common.LogLevel // To specify whether user wants to preserve the blob index tags during service to service transfer. S2sPreserveBlobTags bool // whether to include blobs that have metadata 'hdi_isfolder = true' IncludeDirectoryStubs bool CpkOptions common.CpkOptions // contains filtered or unexported fields }
represents the processed copy command input from the user
func (*CookedCopyCmdArgs) Cancel ¶
func (cca *CookedCopyCmdArgs) Cancel(lcm common.LifecycleMgr)
func (*CookedCopyCmdArgs) InitModularFilters ¶
func (cca *CookedCopyCmdArgs) InitModularFilters() []ObjectFilter
Initialize the modular filters outside of copy to increase readability.
func (*CookedCopyCmdArgs) MakeEscapedRelativePath ¶
func (cca *CookedCopyCmdArgs) MakeEscapedRelativePath(source bool, dstIsDir bool, object StoredObject) (relativePath string)
func (*CookedCopyCmdArgs) ReportProgressOrExit ¶
func (cca *CookedCopyCmdArgs) ReportProgressOrExit(lcm common.LifecycleMgr) (totalKnownCount uint32)
type CopyEnumerator ¶
type CopyEnumerator struct { Traverser ResourceTraverser // general filters apply to the objects returned by the traverser Filters []ObjectFilter // receive objects from the traverser and dispatch them for transferring ObjectDispatcher objectProcessor // a finalizer that is always called if the enumeration finishes properly Finalize func() error }
func NewCopyEnumerator ¶
func NewCopyEnumerator(traverser ResourceTraverser, filters []ObjectFilter, objectDispatcher objectProcessor, finalizer func() error) *CopyEnumerator
type FilterSet ¶
type FilterSet []ObjectFilter
func (FilterSet) GetEnumerationPreFilter ¶
GetEnumerationPreFilter returns a prefix that is common to all the include filters, or "" if no such prefix can be found. (The implementation may return "" even in cases where such a prefix does exist, but in at least the simplest cases, it should return a non-empty prefix.) The result can be used to optimize enumeration, since anything without this prefix will fail the FilterSet
type GCPBucketNameToAzureResourcesResolver ¶
type GCPBucketNameToAzureResourcesResolver struct {
// contains filtered or unexported fields
}
func NewGCPBucketNameToAzureResourcesResolver ¶
func NewGCPBucketNameToAzureResourcesResolver(gcpBucketNames []string) *GCPBucketNameToAzureResourcesResolver
func (*GCPBucketNameToAzureResourcesResolver) ResolveName ¶
func (resolver *GCPBucketNameToAzureResourcesResolver) ResolveName(bucketName string) (string, error)
type HTTPClient ¶
type HTTPClient struct {
// contains filtered or unexported fields
}
todo : use url in case of string
func NewHttpClient ¶
func NewHttpClient(url string) *HTTPClient
NewHttpClient returns the instance of struct containing an instance of http.client and url
type IncludeAfterDateFilter ¶
includeAfterDateFilter includes files with Last Modified Times >= the specified threshold Used for copy, but doesn't make conceptual sense for sync
func (*IncludeAfterDateFilter) AppliesOnlyToFiles ¶
func (f *IncludeAfterDateFilter) AppliesOnlyToFiles() bool
func (*IncludeAfterDateFilter) DoesPass ¶
func (f *IncludeAfterDateFilter) DoesPass(storedObject StoredObject) bool
func (*IncludeAfterDateFilter) DoesSupportThisOS ¶
func (f *IncludeAfterDateFilter) DoesSupportThisOS() (msg string, supported bool)
func (IncludeAfterDateFilter) FormatAsUTC ¶
func (_ IncludeAfterDateFilter) FormatAsUTC(t time.Time) string
func (IncludeAfterDateFilter) ParseISO8601 ¶
type IncludeBeforeDateFilter ¶
IncludeBeforeDateFilter includes files with Last Modified Times <= the specified Threshold Used for copy, but doesn't make conceptual sense for sync
func (*IncludeBeforeDateFilter) AppliesOnlyToFiles ¶
func (f *IncludeBeforeDateFilter) AppliesOnlyToFiles() bool
func (*IncludeBeforeDateFilter) DoesPass ¶
func (f *IncludeBeforeDateFilter) DoesPass(storedObject StoredObject) bool
func (*IncludeBeforeDateFilter) DoesSupportThisOS ¶
func (f *IncludeBeforeDateFilter) DoesSupportThisOS() (msg string, supported bool)
func (IncludeBeforeDateFilter) FormatAsUTC ¶
func (_ IncludeBeforeDateFilter) FormatAsUTC(t time.Time) string
func (IncludeBeforeDateFilter) ParseISO8601 ¶
type IncludeFilter ¶
type IncludeFilter struct {
// contains filtered or unexported fields
}
design explanation: include filters are different from the exclude ones, which work together in the "AND" manner meaning and if an StoredObject is rejected by any of the exclude filters, then it is rejected by all of them as a result, the exclude filters can be in their own struct, and work correctly on the other hand, include filters work in the "OR" manner meaning that if an StoredObject is accepted by any of the include filters, then it is accepted by all of them consequently, all the include Patterns must be stored together
func (*IncludeFilter) AppliesOnlyToFiles ¶
func (f *IncludeFilter) AppliesOnlyToFiles() bool
func (*IncludeFilter) DoesPass ¶
func (f *IncludeFilter) DoesPass(storedObject StoredObject) bool
func (*IncludeFilter) DoesSupportThisOS ¶
func (f *IncludeFilter) DoesSupportThisOS() (msg string, supported bool)
type ListResponse ¶
type ListResponse struct {
ErrorMsg string
}
type LocationLevel ¶
type LocationLevel uint8
----- LOCATION LEVEL HANDLING -----
var ELocationLevel LocationLevel = 0
func DetermineLocationLevel ¶
func DetermineLocationLevel(location string, locationType common.Location, source bool) (LocationLevel, error)
Uses syntax to assume the "level" of a location. This is typically used to
func (LocationLevel) Container ¶
func (LocationLevel) Container() LocationLevel
func (LocationLevel) Object ¶
func (LocationLevel) Object() LocationLevel
func (LocationLevel) Service ¶
func (LocationLevel) Service() LocationLevel
type ObjectFilter ¶
type ObjectFilter interface { DoesSupportThisOS() (msg string, supported bool) DoesPass(storedObject StoredObject) bool AppliesOnlyToFiles() bool }
given a StoredObject, verify if it satisfies the defined conditions if yes, return true
type ResourceTraverser ¶
type ResourceTraverser interface { Traverse(preprocessor objectMorpher, processor objectProcessor, filters []ObjectFilter) error IsDirectory(isSource bool) bool }
capable of traversing a structured resource like container or local directory pass each StoredObject to the given objectProcessor if it passes all the filters
func InitResourceTraverser ¶
func InitResourceTraverser(resource common.ResourceString, location common.Location, ctx *context.Context, credential *common.CredentialInfo, followSymlinks *bool, listOfFilesChannel chan string, recursive, getProperties, includeDirectoryStubs bool, incrementEnumerationCounter enumerationCounterFunc, listOfVersionIds chan string, s2sPreserveBlobTags bool, logLevel pipeline.LogLevel, cpkOptions common.CpkOptions) (ResourceTraverser, error)
type S3BucketNameToAzureResourcesResolver ¶
type S3BucketNameToAzureResourcesResolver struct {
// contains filtered or unexported fields
}
S3BucketNameToAzureResourcesResolver resolves s3 bucket name to Azure Blob container/ADLS Gen2 filesystem/File share. For Azure, container/filesystem/share's naming follows: 1. Lower case letters, numbers and hyphen. 2. 3-63 length. 3. Name should not contain two consecutive hyphens. 4. Name should not start or end with hyphen. For S3, bucket's naming follows: 1. The bucket name can be between 3 and 63 characters long, and can contain only lower-case characters, numbers, periods, and dashes. 2. Each label in the bucket name must start with a lowercase letter or number. 3. The bucket name cannot contain underscores, end with a dash or period, have consecutive periods, or use dashes adjacent to periods. 4. The bucket name cannot be formatted as an IP address (198.51.100.24). Two common cases need be solved are: 1. bucket name with period. In this case, AzCopy try to replace period with hyphen. e.g. bucket.with.period -> bucket-with-period 2. bucket name with consecutive hyphens. In this case, AzCopy try to replace consecutive hyphen, with -[numberOfHyphens]-. e.g. bucket----hyphens -> bucket-4-hyphens The resolver checks if there are naming collision with other existing bucket names, and try to add suffix when there is any collision. e.g. There is buckets with name: bucket-name, bucket.name. Azcopy will resolve bucket.name -> bucket-name -> bucket-name-2 All the resolving should be logged and warned to user.
func NewS3BucketNameToAzureResourcesResolver ¶
func NewS3BucketNameToAzureResourcesResolver(s3BucketNames []string) *S3BucketNameToAzureResourcesResolver
NewS3BucketNameToAzureResourcesResolver creates S3BucketNameToAzureResourcesResolver. Users can provide bucket names upfront and on-demand via ResolveName. Previously resolved names will be returned outright by ResolveName.
func (*S3BucketNameToAzureResourcesResolver) ResolveName ¶
func (s3Resolver *S3BucketNameToAzureResourcesResolver) ResolveName(bucketName string) (string, error)
ResolveName returns resolved name for given bucket name.
type StoredObject ¶
type StoredObject struct { // container source, only included by account traversers. ContainerName string // destination container name. Included in the processor after resolving container names. DstContainerName string // metadata, included in S2S transfers Metadata common.Metadata // contains filtered or unexported fields }
represent a local or remote resource object (ex: local file, blob, etc.) we can add more properties if needed, as this is easily extensible ** DO NOT instantiate directly, always use newStoredObject ** (to make sure its fully populated and any preprocessor method runs)
func (*StoredObject) ToNewCopyTransfer ¶
func (s *StoredObject) ToNewCopyTransfer( steWillAutoDecompress bool, Source string, Destination string, preserveBlobTier bool, folderPropertiesOption common.FolderPropertyOption) (transfer common.CopyTransfer, shouldSendToSte bool)
type Version ¶
type Version struct {
// contains filtered or unexported fields
}
func NewVersion ¶
To keep the code simple, we assume we only use a simple subset of semantic versions. Namely, the version is either a normal stable version, or a pre-release version with '-preview' attached. Examples: 10.1.0, 11.2.0-preview
Source Files ¶
- benchmark.go
- cancel.go
- copy.go
- copyEnumeratorHelper.go
- copyEnumeratorInit.go
- copyUtil.go
- credentialUtil.go
- doc.go
- env.go
- gcpNameResolver.go
- helpMessages.go
- jobs.go
- jobsClean.go
- jobsList.go
- jobsRemove.go
- jobsResume.go
- jobsShow.go
- list.go
- load.go
- loadCLFS.go
- loadOutputTranslation.go
- login.go
- loginStatus.go
- logout.go
- make.go
- pathUtils.go
- pause.go
- remove.go
- removeEnumerator.go
- removeProcessor.go
- root.go
- rpc.go
- s3NameResolver.go
- sync.go
- syncComparator.go
- syncEnumerator.go
- syncIndexer.go
- syncProcessor.go
- validators.go
- versionChecker.go
- zc_attr_filter_notwin.go
- zc_enumerator.go
- zc_filter.go
- zc_newobjectadapters.go
- zc_pipeline_init.go
- zc_processor.go
- zc_traverser_benchmark.go
- zc_traverser_blob.go
- zc_traverser_blob_account.go
- zc_traverser_blob_versions.go
- zc_traverser_blobfs.go
- zc_traverser_blobfs_account.go
- zc_traverser_file.go
- zc_traverser_file_account.go
- zc_traverser_gcp.go
- zc_traverser_gcp_service.go
- zc_traverser_list.go
- zc_traverser_local.go
- zc_traverser_local_other.go
- zc_traverser_s3.go
- zc_traverser_s3_service.go