| Age | Commit message (Collapse) | Author | Files | Lines |
|
|
|
chrislusf/dependabot/maven/other/java/hdfs3/org.apache.hadoop-hadoop-common-3.1.4
Bump hadoop-common from 3.1.1 to 3.1.4 in /other/java/hdfs3
|
|
chrislusf/dependabot/maven/other/java/hdfs-over-ftp/org.apache.hadoop-hadoop-common-3.2.2
Bump hadoop-common from 3.2.1 to 3.2.2 in /other/java/hdfs-over-ftp
|
|
chrislusf/dependabot/maven/other/java/hdfs2/org.apache.hadoop-hadoop-common-2.10.1
Bump hadoop-common from 2.9.2 to 2.10.1 in /other/java/hdfs2
|
|
chrislusf/dependabot/maven/other/java/examples/org.apache.hadoop-hadoop-common-2.10.1
Bump hadoop-common from 2.9.2 to 2.10.1 in /other/java/examples
|
|
Bumps hadoop-common from 2.9.2 to 2.10.1.
---
updated-dependencies:
- dependency-name: org.apache.hadoop:hadoop-common
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
|
|
Bumps hadoop-common from 2.9.2 to 2.10.1.
---
updated-dependencies:
- dependency-name: org.apache.hadoop:hadoop-common
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
|
|
Bumps hadoop-common from 3.2.1 to 3.2.2.
---
updated-dependencies:
- dependency-name: org.apache.hadoop:hadoop-common
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
|
|
Bumps hadoop-common from 3.1.1 to 3.1.4.
---
updated-dependencies:
- dependency-name: org.apache.hadoop:hadoop-common
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
|
|
|
|
|
|
fix https://github.com/chrislusf/seaweedfs/issues/2642
|
|
|
|
|
|
related to https://github.com/chrislusf/seaweedfs/issues/2637
|
|
|
|
fix https://github.com/chrislusf/seaweedfs/issues/2637
|
|
graceful stop cluster http server default timeout 60 sec
|
|
https://github.com/chrislusf/seaweedfs/issues/2630
|
|
|
|
fix https://github.com/chrislusf/seaweedfs/issues/2641
|
|
fix preconditions
|
|
https://tools.ietf.org/id/draft-ietf-httpbis-p4-conditional-26.html#preconditions
|
|
|
|
fix https://github.com/chrislusf/seaweedfs/issues/2636
|
|
updated needle with fsync
|
|
|
|
|
|
|
|
Run two servers with volumes and fillers:
server -dir=Server1alpha -master.port=11000 -filer -filer.port=11001 -volume.port=11002
server -dir=Server1sigma -master.port=11006 -filer -filer.port=11007 -volume.port=11008
Run Active-Passive filler.sync:
filer.sync -a localhost:11007 -b localhost:11001 -isActivePassive
Upload file to 11007 port:
curl -F file=@/Desktop/9.xml "http://localhost:11007/testFacebook/"
If we request a file on two servers now, everything will be correct, even if we add data to the file and upload it again:
curl "http://localhost:11007/testFacebook/9.xml"
EQUALS
curl "http://localhost:11001/testFacebook/9.xml"
However, if we change the already existing data in the file (for example, we change the first line in the file, reducing its length), then this file on the second server will not be valid and will not be equivalent to the first file
Снимок экрана 2022-02-07 в 14 21 11
This problem occurs on line 202 in the filer_sink.go file. In particular, this is due to incorrect mapping of chunk names in the DoMinusChunks function. The names of deletedChunks do not match the chunks of existingEntry.Chunks, since the first chunks come from another server and have a different addressing (name) compared to the addressing on the server where the file is being overwritten.
Deleted chunks are not actually deleted on the server to which the file is replicated.
|
|
This file contains metric names for all errors
The naming convention is ErrorSomeThing = "error.some.thing"
|
|
error metrics for filer and store
|
|
add s3api error for copy in file, not directory
|
|
|
|
|
|
fix https://github.com/chrislusf/seaweedfs/issues/2633
|
|
|
|
Otherwise current calls for some methods (i.e. GetObjectAcl) ends up with wrong method selection (i.e. GetObject).
Added generic comment rule of traversing methods
|
|
|
|
|
|
|
|
ubuntu@prod-master-1:~$ aws --endpoint http://10.244.15.66:8333 s3api abort-multipart-upload --bucket prod-cache --key multipart-test --upload-id 5347f936-6adc-43de-8e5c-1fd137c3b2bc
ubuntu@prod-master-1:~$ aws --endpoint http://10.244.15.66:8333 s3api list-parts --bucket prod-cache --key multipart-test --upload-id 5347f936-6adc-43de-8e5c-1fd137c3b2bc
{
"Initiator": null,
"Owner": null,
"StorageClass": "STANDARD"
}
If we abort a multipart upload, it appears that records are left behind. We should get a 404 NoSuchKey error.
|
|
|
|
Protocol
Otherwise any requests to the underlying handlers results in calls to
ListObjects (v1) that may intensively load gateway and volume servers.
Added the following handlers with default responses:
- GetBucketLocation
- GetBucketRequestPayment
Added the following handlers with NotFound and NotImplemented responses:
- PutBucketAcl
- GetBucketPolicy
- PutBucketPolicy
- DeleteBucketPolicy
- GetBucketCors
- PutBucketCors
- DeleteBucketCors
|
|
fix https://github.com/chrislusf/seaweedfs/issues/2626
minuend - subtrahend
|
|
|
|
|
|
|
|
|
|
following https://github.com/grpc/grpc-go/issues/3170#issuecomment-552517779
|