I am using the AmazonS3Client in an Android app using a getObject request to download an image from my Amazon S3 bucket.
Currently, I am getting this exception:
com.amazonaws.services.s3.model.AmazonS3Exception:
The specified key does not exist.
(Service: Amazon S3; Status Code: 404; Error Code: NoSuchKey;
Even though I am able to see the object with the specified key in my S3 bucket.
This question is related to
android
amazon-web-services
amazon-s3
For me, the object definitely existed and was uploaded correctly, however, its s3 url still threw the same error:
<Code>NoSuchKey</Code>
<Message>The specified key does not exist.</Message>
I found out that the reason was because my filename contained a #
symbol, and I guess certain characters or symbols will also cause this error.
Removing this character and generating the new s3 url resolved my issue.
I also ran into this issue, but in my case I was inadvertently changing the internal state of my source object key when constructing the destination key:
source_objects.each do |item|
key = item.key.sub!(source_prefix, dest_prefix)
item.copy_to(bucket: dest_bucket, key: key)
end
I'm new to Ruby and missed that sub!
has side effects and sub
should have been used instead.
Well this error is actually rather straight forward. it simply means that your file does not exist up within the S3 bucket. Several things could be wrong:
You could be trying to reference the wrong file. Double check the path that you tried to retrieve.
Whenever the file was uploaded it must have failed. Check the logs for your S3Sync process to see if you can find any relevant output
I encountered this issue in a NodeJS Lambda function that was triggered by a file upload to S3.
My mistake was that I was not decoding the object key, which contained a colon. Corrected my code as follows:
let key = decodeURIComponent(event.Records[0].s3.object.key);
In my case the error was appearing because I had uploaded the whole folder, containing the website files, into the container.
I solved it by moving all the files outside the folder, right into the container.
Step 1: Get the latest aws-java-sdk
<!-- https://mvnrepository.com/artifact/org.apache.hadoop/hadoop-aws -->
<dependency>
<groupId>com.amazonaws</groupId>
<artifactId>aws-java-sdk</artifactId>
<version>1.11.660</version>
</dependency>
Step 2: The correct imports
import com.amazonaws.auth.AWSCredentials;
import com.amazonaws.auth.BasicAWSCredentials;
import com.amazonaws.regions.Region;
import com.amazonaws.regions.Regions;
import com.amazonaws.services.s3.AmazonS3;
import com.amazonaws.services.s3.AmazonS3Client;
import com.amazonaws.services.s3.AmazonS3ClientBuilder;
import com.amazonaws.services.s3.model.ListObjectsRequest;
import com.amazonaws.services.s3.model.ObjectListing;
If you are sure the bucket exists, Specified key does not exists error
would mean the bucketname is not spelled correctly ( contains slash or special characters). Refer the documentation for naming convention.
The document quotes:
If the requested object is available in the bucket and users are still getting the 404 NoSuchKey error from Amazon S3, check the following:
Confirm that the request matches the object name exactly, including the capitalization of the object name. Requests for S3 objects are case sensitive. For example, if an object is named myimage.jpg, but Myimage.jpg is requested, then requester receives a 404 NoSuchKey error. Confirm that the requested path matches the path to the object. For example, if the path to an object is awsexamplebucket/Downloads/February/Images/image.jpg, but the requested path is awsexamplebucket/Downloads/February/image.jpg, then the requester receives a 404 NoSuchKey error. If the path to the object contains any spaces, be sure that the request uses the correct syntax to recognize the path. For example, if you're using the AWS CLI to download an object to your Windows machine, you must use quotation marks around the object path, similar to: aws s3 cp "s3://awsexamplebucket/Backup Copy Job 4/3T000000.vbk". Optionally, you can enable server access logging to review request records in further detail for issues that might be causing the 404 error.
AWSCredentials credentials = new BasicAWSCredentials(AWS_ACCESS_KEY_ID, AWS_SECRET_KEY);
AmazonS3 s3Client = AmazonS3ClientBuilder.standard().withRegion(Regions.US_EAST_1).build();
ObjectListing objects = s3Client.listObjects("bigdataanalytics");
System.out.println(objects.getObjectSummaries());
Don't forget buckets are region specific. That might be an issue.
Also try using the S3 console to navigate to the actual object, and then click on Copy Path
, you will get something like:
s3://<bucket-name>/<path>/object.txt
As long as whatever you are passing it to parses that properly I find that is the safest thing to do.
The reason for the issue is wrong or typo in the Bucket/Key name. Do check if the bucket or key name you are providing does exists.
In my case it was because the filename was containing spaces. Solved it thanks to this documentation (which is unrelated to the problem):
from urllib.parse import unquote_plus
key_name = unquote_plus(event['Records'][0]['s3']['object']['key'])
You also need to upload urllib as a layer with corresponding version (if your lambda is Python 3.7 you have to package urllib in a python 3.7 environment).
The reason is that AWS transform ' ' into '+' (why...) which is really problematic...
Note that this may happen even if the file path is correct due to s3's eventual consistency model. Basically, there may be some latency in being able to read an object after it's written. See this documentation for more information.
Source: Stackoverflow.com