dcd5a265f6f36551e0bc0feac0e08bfdd09301a3
6262 Commits
| Author | SHA1 | Message | Date | |
|---|---|---|---|---|
|
Yan Xiao
|
dcd5a265f6 |
proxy-logging: Add real-time transfer bytes counters
Currently we can get one proxy-logging transfer stat emission over the duration of the upload/download. We want another stat coming out of proxy-logging: something that gets emitted periodically as bytes are actually sent/received so we can get reasonably accurate point-in-time breakdowns of bandwidth usage. Co-Authored-By: Alistair Coles <alistairncoles@gmail.com> Co-Authored-By: Shreeya Deshpande <shreeyad@nvidia.com> Change-Id: Ideecd0aa58ddf091c9f25f15022a9066088f532b Signed-off-by: Yan Xiao <yanxiao@nvidia.com> |
||
|
Yan Xiao
|
9d7e7e27a5 |
Provide some s3 helper methods for other middlewares to use.
get_s3_access_key_id returns the S3 access_key_id user for the request is_s3_req checks whether a request looks like it ought to be an S3 request parse_path returns a wsgi string extract_bucket_and_key extracts bucket and object from the request's PATH_INFO Co-Authored-By: Alistair Coles <alistairncoles@gmail.com> Co-Authored-By: Clay Gerrard <clay.gerrard@gmail.com> Co-Authored-By: Shreeya Deshpande <shreeyad@nvidia.com> Change-Id: Iaf86a07238cca6700dee736f55d4c0672cccf1b1 Signed-off-by: Shreeya Deshpande <shreeyad@nvidia.com> |
||
|
Zuul
|
48a5d5e42f | Merge "test-db-replicator (trivial): just one tmpdir" | ||
|
Zuul
|
4b7543b2e1 | Merge "trivial test_[db_]replicator cleanup" | ||
|
Clay Gerrard
|
fac55ced3c |
test-db-replicator (trivial): just one tmpdir
Change-Id: I1e53d171faff02e2dbcbcc779ad3a47506b26853 Signed-off-by: Clay Gerrard <clay.gerrard@gmail.com> |
||
|
Zuul
|
c161aa168f | Merge "relinker: allow clobber-hardlink-collision" | ||
|
Zuul
|
4fa1237997 | Merge "common.db_replicator: log container, db paths consistently" | ||
|
Alistair Coles
|
c0fefe80b3 |
trivial test_[db_]replicator cleanup
* add a self.temp_dir in setUp and remove in teraDown * be consistent in order of (expected, actual) args * assert the complete exception error line Related-Change: I289d3e9b6fe14159925786732ad748acd0459812 Change-Id: I185c8cd55db6df593bb3304c54c5160c1f662b86 Signed-off-by: Alistair Coles <alistairncoles@gmail.com> |
||
|
Clay Gerrard
|
be62933d00 |
relinker: allow clobber-hardlink-collision
The relinker has already been robust to hardlink collisions on tombstones for some time; this change allows ops to optionally (non-default) enable a similar handling of other files when relinking the old=>new partdir. If your cluster is having a bunch of these kinds of collisions and after spot checking you determine the data is in fact duplicate copies the same data - you'd much rather have the option for the relinker to programatically handle them non-destructively than forcing ops to rm a bunch of files manually just get out of a PPI. Once the PPI is over and you reconstrcutors are running again, after some validation you can probably clean out your quarantine dirs. Drive-by: log unknown relink errors at error level to match expected non-zero return code Closes-Bug: #2127779 Change-Id: Iaae0d9fb7a1949d1aad9aa77b0daeb249fb471b5 Signed-off-by: Clay Gerrard <clay.gerrard@gmail.com> |
||
|
ashnair
|
41bf72a5cc |
common.db_replicator: log container, db paths consistently
Extract helpers (from container.sharder) that formats log context from either a broker (preferring broker.path/db_file) or a plain db_file string. Use it in common.db_replicator and container.replicator so messages are uniform and robust. Update tests to cover both cases. No functional changes to replication behavior; this is logging/robustness and test updates only. Change-Id: I289d3e9b6fe14159925786732ad748acd0459812 Related-Change: I7d2fe064175f002055054a72f348b87dc396772b Signed-off-by: ashnair <ashnair@nvidia.com> |
||
|
Zuul
|
e963d13979 | Merge "s3api: fix test_service with pre-existing buckets" | ||
|
Clay Gerrard
|
3c6e967a58 |
test: fix AccountBroker.path tests
Move tests to base TestCase, currently they are only running against old "broker w/o metadata" - but the tests and behavior should work on all versions of the account schema Drive-by: reword tests to make assertions stronger and behaviors more obvious Related-Change: Ic7c2aa878caf039b29abb900b4f491130be3d8a8 Change-Id: I59abd956ffa01bd41f29959ff3df89a3a20a00d4 Signed-off-by: Clay Gerrard <clay.gerrard@gmail.com> |
||
|
Zuul
|
ac5c783d65 | Merge "Assert metadata of SLO PUT from container sync" | ||
|
Zuul
|
9a45531942 | Merge "Test each method in test_crossdomain_get_only" | ||
|
Shashirekha Gundur
|
a6bde729c5 |
Test each method in test_crossdomain_get_only
iterate through not allowed methods and assert Change-Id: Ia304709fc56d3e81bb1326b56a4b0d64ed698160 Signed-off-by: Tim Burke <tim.burke@gmail.com> |
||
|
Zuul
|
6da1207489 | Merge "test: move import to top of file" | ||
|
Clay Gerrard
|
64bb041398 |
Assert metadata of SLO PUT from container sync
In addition to being in the pure unmolested ondisk format from the source container; the manifest must also include the normally protected X-Static-Large-Object metadata. Change-Id: Ic6638e8258e9dec755f8d9630f0586bd3c9b4420 Related-Change: I8d503419b7996721a671ed6b2795224775a7d8c6 Signed-off-by: Tim Burke <tim.burke@gmail.com> |
||
|
Zuul
|
a9a1ea4137 | Merge "Adds --skip-commits to s-m-s-r" | ||
|
Clay Gerrard
|
b55f13c758 |
test: move import to top of file
Related-Change-Id: I38c11b7aae8c4112bb3d671fa96012ab0c44d5a2 Change-Id: Ibe5d206d2b96e174f849715fb13562ae0d2f5de2 Signed-off-by: Clay Gerrard <clay.gerrard@gmail.com> |
||
|
Zuul
|
e5d44d669a | Merge "proxy: use cooperative tokens to coalesce updating shard range requests into backend" | ||
|
Christian Ohanaja
|
ba1ab9d11c |
Adds --skip-commits to s-m-s-r
This patch replaces --force-commits with a --skip-commits flag in swift-manage-shard-ranges to determine when to commit object updates. Change-Id: I6de041f5c12dca2618d22d1271efe242b2f35258 Signed-off-by: Christian Ohanaja <cohanaja@nvidia.com> |
||
|
Jianjian Huo
|
d9883d0834 |
proxy: use cooperative tokens to coalesce updating shard range requests into backend
The cost of memcache misses could be deadly. For example, when updating shard range cache query miss, PUT requests would have to query the backend to figure out which shard to upload the objects. And when a lot of requests are sending to the backend at the same time, this could easily overload the root containers and cause a lot of 500/503 errors; and when proxy-servers receive responses of all those 200 backend shard range queries, they could in turn try to write the same shard range data into memcached servers at the same time, and cause memcached to return OOM failures too. We have seen cache misses frequently to updating shard range cache in production, due to Memcached out-of-memory and cache evictions. To cope with those kind of situations, a memcached based cooperative token mechanism can be added into proxy-server to coalesce lots of in-flight backend requests into a few: when updating shard range cache misses, only the first few of requests will get global cooperative tokens and then be able to fetch updating shard ranges from backend container servers. And the following cache miss requests will wait for cache filling to finish, instead of all querying the backend container servers. This will prevent a flood of backend requests to overload both container servers and memcached servers. Drive-by fix: when memcache is not available, object controller will only need to retrieve a specific shard range from the container server to send the update request to. Co-Authored-By: Clay Gerrard <clay.gerrard@gmail.com> Co-Authored-By: Tim Burke <tim.burke@gmail.com> Co-Authored-By: Yan Xiao <yanxiao@nvidia.com> Co-Authored-By: Shreeya Deshpande <shreeyad@nvidia.com> Signed-off-by: Jianjian Huo <jhuo@nvidia.com> Change-Id: I38c11b7aae8c4112bb3d671fa96012ab0c44d5a2 |
||
|
Zuul
|
dd23020c30 | Merge "common: add memcached based cooperative token mechanism." | ||
|
ashnair
|
d353f15fac |
account-broker: add resilient path property with lazy cache
Add a path property for AccountBroker and use lazy, resilient _populate_instance_cache(). Use None attrs as flags, avoid broad try/except in path, and retry if cache population fails. Change-Id: Ic7c2aa878caf039b29abb900b4f491130be3d8a8 Signed-off-by: ashnair <ashnair@nvidia.com> |
||
|
Jianjian Huo
|
707a65ab3c |
common: add memcached based cooperative token mechanism.
Memcached based cooperative token is a improved version of ghetto lock, see the description of ghetto lock at here: https://github.com/memcached/memcached/wiki/ProgrammingTricks It's used to avoid the thundering herd situation which many caching users face: given a cache item that is popular and difficult to recreate, in the event of cache misses, users could end up with hundreds (or thousands) of processes slamming the backend database at the same time in an attempt to refill the same cache content. This thundering herd problem not only often leads to unresponsive backend; and also those writes into memcached cause premature cache eviction under memory pressure. With cooperative token, when lots of in-flight callers try to get the cached item specified by key from memcache and get cache misses, only the first few query requests (limited by by ``num_tokens``) will be able get the cooperative tokens by creating or incrementing an internal memcache key, and then those callers with tokens can send backend requests to fetch data from backend servers and be able to set data into memcache; all other cache miss requests without a token should wait for cache filling to finish, instead of all querying the backend servers at the same time. Co-Authored-By: Tim Burke <tim.burke@gmail.com> Co-Authored-By: Clay Gerrard <clay.gerrard@gmail.com> Co-Authored-By: Yan Xiao <yanxiao@nvidia.com> Co-Authored-By: Alistair Coles <alistairncoles@gmail.com> Signed-off-by: Jianjian Huo <jhuo@nvidia.com> Change-Id: I50ff92441c2f2c49b3034644aba59930e8a99589 |
||
|
Zuul
|
92dd03ed77 | Merge "diskfile: Fix UnboundLocalError during part power increase" | ||
|
Clay Gerrard
|
7b05356bd0 |
test: do not create timestamp collision unnecessarily
Change-Id: Ib6bf702e38495e52e3b2f5ca95ed17c519018474 Signed-off-by: Clay Gerrard <clay.gerrard@gmail.com> |
||
|
Clay Gerrard
|
815393dff4 |
test: fix module state pollution
The disable_fallocate function provided in common.utils doesn't really have a way to undo it - it's tested independently in test_utils. It shouldn't be used on test_diskfile or else test_utils fallocate tests will fail afterwards. Change-Id: I6ffa97b39111ba25f85ba7cfde21440d975dc760 Signed-off-by: Clay Gerrard <clay.gerrard@gmail.com> |
||
|
Alistair Coles
|
c26c7b8edd |
tests: simplify TestGlobalSetupObjectReconstructor setUp
Change-Id: I0168ab113fdda60ed858ed0928356699399d4044 Signed-off-by: Alistair Coles <alistairncoles@gmail.com> |
||
|
Tim Burke
|
397f94c73b |
diskfile: Fix UnboundLocalError during part power increase
Closes-Bug: #2122543 Co-Authored-By: Clay Gerrard <clay.gerrard@gmail.com> Signed-off-by: Tim Burke <tim.burke@gmail.com> Change-Id: I8a2a96394734899ee48e1d9264bf3908968c51a8 |
||
|
Clay Gerrard
|
b5e6964a22 |
s3api: fix test_service with pre-existing buckets
The s3api cross-compat tests in test_service weren't sophisticated enough to account for real s3 session credentials that could see actual aws s3 buckets (or a vsaio you actually use) - however valid assertions on the authorization logic doesn't actually require such a strictly clean slate. Drive-by: prefer test config option without double negative, and update ansible that's based on the sample config. Related-Change-Id: I811642fccd916bd9ef71846a8108d50a462740f0 Change-Id: Ifab08cfe72f12d80e2196ad9b9b7876ace5825b4 Signed-off-by: Clay Gerrard <clay.gerrard@gmail.com> |
||
|
Tim Burke
|
e13f4abcd7 |
tests: Skip some tests if crc32c is not available
Signed-off-by: Tim Burke <tim.burke@gmail.com> Change-Id: I2e2a4e2c448319e6531372ae06ab81eb58edc57e |
||
|
Zuul
|
e10c2bafcb | Merge "proxy-logging: create field for access_user_id" | ||
|
Vitaly Bordyug
|
32eaab20b1 |
proxy-logging: create field for access_user_id
Added the new field to be able to log the access key during the s3api calls, while reserving the field to be filled with auth relevant information in case of other middlewares. Added respective code to the tempauth and keystone middlewares. Since s3api creates a copy of the environ dict for the downstream request object when translating the s3req.to_swift_req the environ dict that is seen/modifed in other mw module is not the same instance seen in proxy-logging - using mutable objects get transfered into the swift_req.environ. Change the assert in test_proxy_logging from "the last field" to the index 21 in the interests of maintainability. Also added some regression tests for object, bucket and s3 v4 apis and updated the documentation with the details about the new field. Signed-off-by: Vitaly Bordyug <vbordug@gmail.com> Change-Id: I0ce4e92458e2b05a4848cc7675604c1aa2b64d64 |
||
|
Zuul
|
457af40c9b | Merge "s3api compat tests: stop asserting DisplayName in Owner" | ||
|
Alistair Coles
|
c4cc83c5e7 |
s3api compat tests: stop asserting DisplayName in Owner
S3 stopped returning DisplayNamme in the Owner field of object listings [1], so the tests need to stop asserting that it is present. Further work is needed to drop DisplayName from the Swift s3api responses [2]. [1] https://docs.aws.amazon.com/AmazonS3/latest/API/API_Owner.html [2] https://bugs.launchpad.net/swift/+bug/2120622 Change-Id: Ia915a65313394910c74ae826c912b5549e833a7b Signed-off-by: Alistair Coles <alistairncoles@gmail.com> |
||
|
Tim Burke
|
81df05a9c4 |
checksums: Ignore broken pyeclib installs
The RuntimeError gives more of a hint than the TypeError, but we haven't really solved the issue. If there's a busted pyeclib install, it's safe to assume it won't have ISA-L alongside it; just proceed assuming it's not available. Closes-Bug: #2120591 Related-Change: I64a85eb739fb72efe41f1ee829e463167246b793 Co-Authored-By: Alistair Coles <alistairncoles@gmail.com> Signed-off-by: Tim Burke <tim.burke@gmail.com> Change-Id: I2791566b208327b1fb536fb56a363337ab3f3941 |
||
|
Zuul
|
c1d34baa78 | Merge "Timestamp: fix inconsistent raw and normal values" | ||
|
Zuul
|
fa3a75b73d | Merge "checksum.py: fail gracefully if pyeclib is broken" | ||
|
Alistair Coles
|
5b2e4e00c1 |
ssync unit test: fix intermittent failure
ssync unit tests would sometimes fail when making assertions about the ssync receiver log messages. Test runner output would show that the messages were eventually being logged. However, the assertions could be made before the ssync receiver request thread had completed. A trampoline had been previously been used to workaround this, but that is clearly insufficient. The author found that increasing the trampoline interval would help reduce the rate of failures, but not eliminate them. This patch introduces a custom GreenPool for the unit test wsgi object server so that tests can deterministically wait for the receiver request handling thread to exit before making assertions. Closes-Bug: #212065 Change-Id: I09ad8bb1becae46a78902d1d384a9f27a3d54b38 Signed-off-by: Alistair Coles <alistairncoles@gmail.com> |
||
|
Alistair Coles
|
74274ec8bc |
checksum.py: fail gracefully if pyeclib is broken
If pyeclib dist is missing files then the isal loading would blow up with "TypeError: 'NoneType' object is not iterable". This patch changes that to a RuntimeError with a more useful message. Change-Id: I64a85eb739fb72efe41f1ee829e463167246b793 Closes-Bug: #2120591 Signed-off-by: Alistair Coles <alistairncoles@gmail.com> |
||
|
Alistair Coles
|
93b88540dc |
Timestamp: fix inconsistent raw and normal values
Previously it was possible for a Timestamp to have inconsistent raw
and normal values. For example:
>>> ts1 = Timestamp(1755077566.523385)
>>> (ts1.normal, ts1.raw, (~ts1).normal)
('1755077566.52339', 175507756652338, '8244922433.47661')
This results in the invert function not being reliably reversible:
(~(~ts1)).normal
'1755077566.52338'
The cause is that the normal value is based on Timestamp.timestamp
which preserves the precision of the value given to the constructor,
whereas the invert function uses the limited precision raw value.
This patch forces Timestamp.timestamp to have the limited precision
value of Timestamp.raw.
Change-Id: I4e7fd6078aae3f284628303f20ced66aa702c466
Signed-off-by: Alistair Coles <alistairncoles@gmail.com>
|
||
|
Zuul
|
bdb052b59b | Merge "ssync-receiver: terminate session if subreq read times out" | ||
|
Zuul
|
044ba6c40f | Merge "Ring v2 follow-up" | ||
|
Tim Burke
|
683218c523 |
Ring v2 follow-up
Signed-off-by: Tim Burke <tim.burke@gmail.com> Change-Id: I75bd005a4a3bc79c1bd8f8fa1153a64059970865 |
||
|
Zuul
|
c8d00108a6 | Merge "Fix another way mw may encouter invalid swift paths" | ||
|
Zuul
|
d535833b0b | Merge "crypto: Fix traceback on non-utf8, non-swift paths" | ||
|
Clay Gerrard
|
a770881b59 |
Fix another way mw may encouter invalid swift paths
Related-Change-Id: I8c342c4751ba3ca682efd152e90e396e9f8eb851 Change-Id: I9df32c5aae4e681d488419f36982ffc36589d50a Signed-off-by: Clay Gerrard <clay.gerrard@gmail.com> |
||
|
Zuul
|
ee432f03ac | Merge "ring: Introduce a v2 ring format" | ||
|
Tim Burke
|
1dc3307eaf |
ssync-receiver: terminate session if subreq read times out
If a PUT subrequest body iterator times out while the object server is reading it, the object server will handle the timeout and return a 500 response to the ssync receiver. Previously, the ssync receiver would attempt to drain the remainder of the subrequest body iterator and then resume reading the next subrequest from the SSYNC body. However, the attempt to drain the subrequest iterator would fail (silently) because the timeout had already caused the iterator to exit. The ssync receiver would therefore treat any remaining subrequest body as the preamble to the next subrequest. This remaining subrequest body content was likely to cause the protocol parsing to fail, but in the worst case could be erroneously interpreted as a valid subrequest. (The exact failure mechanism depends on what state the eventlet.wsgi.Input is left in when the timeout fired.) This patch ensures that the ssync receiver will terminate processing an SSYNC request if an exception occurs while reading a subrequest body. Closes-Bug: #2115991 Change-Id: I585e8a916d947c3da8d7c0e8a85cf0a8ab85f7f0 Co-authored-by: Tim Burke <tim.burke@gmail.com> Signed-off-by: Alistair Coles <alistairncoles@gmail.com> |