Skip to content

MalformedXML: The XML you provided was not well-formed or did not validate against our published schema

I am having this weird issue while working with AWS S3. I am working on application by which I can store the images to AWS bucket. Using Multer as middleware and S3FS library to connect and upload to AWS.

But the following error pops up when I try uploading the content.

“MalformedXML: The XML you provided was not well-formed or did not validate against our publis hed schema”


var express = require('express');
var router = express();
var multer = require('multer');
var fs = require('fs');
var S3FS = require('s3fs');
var upload = multer({
  dest: 'uploads'
var S3fsImpl = new S3FS('bucket-name', {
  region: 'us-east-1',
  accessKeyId: 'XXXXXXXXXXXX',
/* GET home page. */
router.get('/', function (req, res, next) {
  res.render('profile', {
    title: 'Express'
});'/testupload', upload.single('file'), function (req, res) {
  var file = req.file;
  var path = req.file.path;
  var stream = fs.createReadStream(path);
  S3fsImpl.writeFile(, stream).then(function () {
    fs.unlink(file.path, function (err) {
      if (err) {
module.exports = router;

EDIT Output:

{ fieldname: 'file',
  originalname: '441_1.docx',
  encoding: '7bit',
  mimetype: 'application/vnd.openxmlformats-officedocument.wordprocessingml.document',
  destination: 'uploads',
  filename: '662dcbe544804e4f50dfef1f52b40d22',
  path: 'uploads\662dcbe544804e4f50dfef1f52b40d22',
  size: 13938 }
ReadStream {
   ReadableState {
     objectMode: false,
     highWaterMark: 65536,
     buffer: BufferList { head: null, tail: null, length: 0 },
     length: 0,
     pipes: null,
     pipesCount: 0,
     flowing: null,
     ended: false,
     endEmitted: false,
     reading: false,
     sync: true,
     needReadable: false,
     emittedReadable: false,
     readableListening: false,
     resumeScheduled: false,
     defaultEncoding: 'utf8',
     ranOut: false,
     awaitDrain: 0,
     readingMore: false,
     decoder: null,
     encoding: null },
  readable: true,
  domain: null,
  _events: { end: [Function] },
  _eventsCount: 1,
  _maxListeners: undefined,
  path: 'uploads\662dcbe544804e4f50dfef1f52b40d22',
  fd: null,
  flags: 'r',
  mode: 438,
  start: undefined,
  end: undefined,
  autoClose: true,
  pos: undefined,
  bytesRead: 0 }


  "name": "aws-s3-images",
  "version": "1.0.0",
  "private": true,
  "scripts": {
    "start": "node ./bin/www"
  "dependencies": {
    "body-parser": "~1.17.1",
    "connect-multiparty": "^2.0.0",
    "cookie-parser": "~1.4.3",
    "debug": "~2.6.3",
    "express": "~4.15.2",
    "hbs": "~4.0.1",
    "morgan": "~1.8.1",
    "multer": "^1.3.0",
    "s3fs": "^2.5.0",
    "serve-favicon": "~2.4.2"
  "description": "AWS S3 uploading images",
  "main": "app.js",
  "devDependencies": {},
  "keywords": [
  "author": "reeversedev",
  "license": "MIT"


This code should work for you. You need to remember that: 1) use unique bucket name 2) under your file object use ‘originalname’ instead of ‘name’ <– this property does not exist'/testupload', function(req, res){
    var file = req.files[0];
    console.log('FIRST TEST: ' + JSON.stringify(file));
    var stream = fs.createReadStream(file.path);
    S3fsImpl.writeFile(file.originalname, stream).then(function ()
        console.log('File has been sent - OK');
          throw reason;